Name:
2021 Previews Session: New and Noteworthy Product Presentations
Description:
2021 Previews Session: New and Noteworthy Product Presentations
Thumbnail URL:
https://cadmoremediastorage.blob.core.windows.net/746edf96-b490-4f43-b252-ebfddeaff1fe/thumbnails/746edf96-b490-4f43-b252-ebfddeaff1fe.png
Duration:
T00H53M38S
Embed URL:
https://stream.cadmore.media/player/746edf96-b490-4f43-b252-ebfddeaff1fe
Content URL:
https://cadmoreoriginalmedia.blob.core.windows.net/746edf96-b490-4f43-b252-ebfddeaff1fe/2021 Previews Presentations V5_compr (1).mp4?sv=2019-02-02&sr=c&sig=WT9wV7OkWcnBucbTqK6jynp65CqdkcTzpSq6Rjah8Ig%3D&st=2024-11-22T09%3A44%3A38Z&se=2024-11-22T11%3A49%3A38Z&sp=r
Upload Date:
2024-02-02T00:00:00.0000000
Transcript:
Language: EN.
Segment:0 .
DAVID MYERS: Welcome to the 2021 SSP preview session. The preview session, I like to think, is the most widely anticipated session at the SSP annual meeting. Thank you very much for taking the time out of your day, and I'm sure it'll well be worth your time to listen and learn. The goal of the session is to present to the SSP community the new or novel content services or technologies we should all be aware of.
DAVID MYERS: But really it's about innovation and disruption. The most successful organizations are no longer the ones that offer the best deals. They are the ones the champion the most original ideas and do things that other organizations can't or won't do. This is how great organizations go from ordinary to extraordinary. During this time of pandemic, where we're all still forced to stay close to home and not travel or interact as we would want or need, there's been a seismic shift in the world.
DAVID MYERS: And there's been a significant acceleration in the evolution of the publishing industry. Several random highlights include conducting business virtually rather than in person. Can anyone say Zoom meeting? I hope I never have to hear that again. The dramatic increase in usage of content while we had "nothing else to do." The significant decrease in traditional funding sources or revenue streams, finding ways of conducting research and collaborating in many hybrid ways, and of course, the shift to different subscription models-- RNP, PNR, S2O.
DAVID MYERS: But somehow we're all surviving. Driven by exponentially accelerating rate of technical progress, we now literally have supercomputers in our pockets able to access the world's information at our fingertips. Computational power and storage of data is already abundant and nearly costless. We can 3D print whole houses at a tenth of the cost and at 10 times faster than traditional models.
DAVID MYERS: And most importantly for our current state of affairs, what took years or decades to develop a single drug to treat an illness now took less than a year from inception to injection. And all this progress at an ever-increasing pace creates a wealth of new opportunities and disrupts existing markets faster and more forceful than ever. Disruption can come in many forms, from technology to business models.
DAVID MYERS: What you will see from the presentations fall into three forces of disruption-- disruption is never one feature, your existing skills and processes are no longer relevant-- think the iPhone is a small computer, not a phone-- and disruption always reaches a tipping point at the point of sale. Therefore, it is imperative that you must adapt, you must be flexible, you must evolve.
DAVID MYERS: If you are not the disruptor, you will be disrupted. And of course I love to say, if you are not distinct you will be extinct. And for those of you who know, I will not be showing a roadkill slide this year. But at the heart of it all, it's about data. It's about information. It's what we create, it's what we produce, and it's what we sell.
DAVID MYERS: Once an industry becomes information enabled, it moves on to an exponential curve. Thus finding industries which are not yet information enabled-- and there are plenty of them-- and bringing these to the information age is one of the most promising business ideas these days. And I know and see this firsthand from artificial intelligence, machine learning companies, who we license content for, as I see them disrupting their particular industries.
DAVID MYERS: In fact, this is so true that I was compelled to start a new business to address this opportunity, which of course I'm happy to talk about at some other time-- just reach out. But in an attempt to introduce the SSP community innovative technologies, service, and content, these five minute back-to-back video presentations to give you a good idea of these innovative organizations, their challenges, and the opportunities their products or services are hoping to address.
DAVID MYERS: There of course will be no questions, but I am sure the presenters will be happy to answer the questions offline for those who can reach out to them. So without further delay, here's your first presenter.
MELISSA GONZALEZ: Hi. Good morning and good afternoon. My name is Melissa Gonzalez, and I am a marketing manager at AIP publishing for the Americas territory. I just want to thank you first for joining me today for the session. I'll be speaking to you about bridging the gap, publications to advance discovery in global research.
MELISSA GONZALEZ: AIP publishing is committed to helping global research scientists, students, and educators advance their missions. This endeavor led us to deeply understand researchers and librarians needs and to develop tools, services, and publications to advance discovery and shape the future of the physical sciences. In the last couple of years, our team set out to really put forth the researcher and librarian needs at the forefront of our collective efforts at AIP publishing.
MELISSA GONZALEZ: As an organization, we wanted to step back and re-evaluate who we are and how we can rethink the way we engage but also support the STEM community, backing our mission to make research accessible, discoverable, and permanently available to the communities who can and will use the findings to advance science. As we help focus groups, visited different institutions, and worked with researchers and librarians, a few themes emerge from these conversations-- a struggle to find relevant research on specific topics of interest for new and emerging fields in the physical sciences.
MELISSA GONZALEZ: Students and postdocs were rapidly entering the expanding fields of biophysics and chemical physics and would benefit from comprehensive reviews that survey, summarize, and contextualize the current literature on important research topics of practical experimental approaches. Our journal Applied Physics Reviews was already setting a standard for high impact, cutting-edge science for both emerging and experienced scientists.
MELISSA GONZALEZ: So we followed that same editorial model, leading an in-house team of scientific editors to manage the manuscript and workflow process for these two new publications-- Biophysics Reviews and Chemical Physics Reviews. Biophysics focus on covering fundamental issues in biophysics in addition to the application of biophysics ion other branches of science, medicine, and engineering.
MELISSA GONZALEZ: Chemical Physics features articles on important and emerging topics of interest to the chemical physics community. The articles published within this reviews journal focus on experimental and theoretical research of fundamental issues in chemical physics and its application in other branches of science, medicine, and engineering. A second theme emerged as well.
MELISSA GONZALEZ: Researchers and practitioners communicated that they rely heavily on books for foundational knowledge but required better tools and capabilities for tracking highlighted content and citation. But there's a struggle to find relevant books on specific topics of interest, or that specialized in learning in new fields, or even that supported staying abreast of new and evolving science in a specific field.
MELISSA GONZALEZ: So parallel to the review journals, a publishing team was brought on to design an ebook collection that ultimately provided a comprehensive interactive digital resource for researchers and professionals in the physical sciences, spanning an entire spectrum of information, from reference materials and research methodology, through to cutting-edge primary research. Last year, we released our first 40 book collection that set the stage for providing reliable, up-to-date information and data for researchers across the physical sciences.
MELISSA GONZALEZ: This collection will seek to maintain proficiency in their areas of expertise, attain knowledge about new subjects and topics, learn new techniques for data collection and analytics, and feature cutting edge research by experts in their respective fields. These categories were broken up into four individual book types to cast a wide net to suit the needs and demands of academics, researchers, professionals, and students.
MELISSA GONZALEZ: Principals-- comprehensive overview of topic offering, introductory material, and recent developments. Methods-- protocols, and best practices, and instruction. Professional-- providing guidance on training and development for physics educators and professionals. And perspectives, which offers an in-depth analysis of a specialist topic written by experts in the field.
MELISSA GONZALEZ: Lastly, as we come to a conclusion, as the scholarly publishing universe continues to change, our commitment to the scientific community will always remain constant. Our goal at AIP publishing is to live up to our promise of connecting our authors and readers to a living legacy of published science, to peers and institutions across the globe, and to services tools and platforms that expand the impact and reach of their work.
MELISSA GONZALEZ: Thank you for joining me today and for taking the time to explore ways in which AIP publishing is looking to innovate and keep up with the ever-evolving demands and transformation of the world. You can find some resources here if you'd like to learn more about the publications we just discussed today, and you can also find my LinkedIn handle as well. I look forward to connecting.
STEVE PINCHOTTI: Welcome to outcome understanding, finding connections and measuring impact in the research maelstrom. My name is Steve Pinchotti, and I'm the CEO at Altum. Grants are the way research projects are funded around the world, and grant systems have historically been thought of as productivity tools. How do we automate tasks to make things more efficient and more accurate for grants management and executive staff? Over the years, these systems have evolved into innovation platforms.
STEVE PINCHOTTI: The expectations for them have increased to become easier to use, more data driven, more impact focused. And the community stakeholders has increased as well to include chief scientific officers and research directors, board members, donors, and marketing staff looking for success stories out of all of this grant funding. Our mission at Altum is to revolutionize the research ecosystem to benefit humanity.
STEVE PINCHOTTI: And we do this in three ways-- we empower the entire funding process with our industry-leading grant management platform, Proposal Central, we connect the research community with our over 600,000 users, and we extend the impact of this research long after the award, so funders, and institutions, and researchers can see the impact of that funding. Now, we all know we're inundated with information every day.
STEVE PINCHOTTI: I read a statistic that 90% of the world's data was created in the past two years alone. So how do we make sense of all this? How do we connect the dots? How do we find the needle in the haystack? Well, I'm excited to announce that earlier this year we unveiled a product called Altum Insights And this started out of a project with Dr. Fauci's portfolio analysis team at the NIH.
STEVE PINCHOTTI: They were trying to connect their research funding to output, and they had identified 90 products that hit the market, and they wanted to see how their research funding helped get those products to the market. They had four people on their team, and they manually connected all this information. And it took them four years for 90 products. So now with Insights, we can invest hundreds of millions of data points and more easily search and visualize all this information and the outcomes of the research.
STEVE PINCHOTTI: We can see the connections from the grants to these new discoveries, we can more easily communicate the impact of the funding and the awardees, we can track the progress of the new knowledge, and the exciting part is to be able to identify emerging trends visually that might represent new opportunities for funding or analysis. So what this platform does is really help illustrate the path to new knowledge.
STEVE PINCHOTTI: It helps visualize it. It helps make it easier to see. It helps answer questions-- who are the experts in different areas, who's being funded, who's publishing in these areas, who's getting the most patents? It can easily detect searches and search these millions of data points with subsecond search results. And that will help everybody spend less time finding information and more time analyzing it.
STEVE PINCHOTTI: The really exciting part is that as these platforms continue to evolve, artificial intelligence and machine learning algorithms will help expedite all of this data aggregation and analysis. So it will accelerate all the things we're doing in grants management, and in the scholarly publishing space, and research evaluation. It will help facilitate open science and collaboration, because it will be able to analyze all this data in ways that we can't, and in certainly a time frame that we can't, and surface to us people that should talk based on their areas of expertise or projects they're working on, and help facilitate collaboration.
STEVE PINCHOTTI: It'll expedite the time it takes to get new innovations and hopefully in cures treatments get it to market as quickly as we can. So it will fundamentally change every aspect of what we're doing in all of our fields. It's really exciting to be a part of it. If you'd like to find out more information about Altum or Insights or any of our products, please go to altum.com. And Thank you very much for taking the session.
STEVE PINCHOTTI: You can find me at steve.pinchotti@altum.com or on Twitter @StevePinchotti. Thank you. [MUSIC PLAYING]
BARRY BEALER: My name is Barry Bealer, Senior Vice President of Strategic Alliances at codemantra. Today for the SSP previews event, I'd like to talk to you about our accessibility insight platform. Imagine a world such as this, or imagine a world such as this. There are over seven million people in the United States today with some type of visual impairment. There are over 280 million people in the world today with some type of visual impairment.
BARRY BEALER: Our accessibilityInsight platform helps visually impaired folks be able to consume content, because we make documents accessible. What does it mean to be a non-compliant digital document? On the left-hand side we see a PDF. For those of us who can read it, we understand what a title is, we understand what the authors are, we understand the abstract. But for those who have to use assistive technology, the experience can be much more frustrating, because the document is not tagged appropriately, nor is the reading order set up correctly.
BARRY BEALER: Our accessibility insight platform is an AI-based digital document accessibility compliance platform. This platform is hosted on AWS. We leverage our data lake for AI and machine learning. We have a compliance report module that can report against the WCAG and PDF/UA standards, as well as against five different formats, including PDF, Word, Excel, PowerPoint, and EPUB.
BARRY BEALER: We also are able to automate PDF remediation up to about 80% accessibility. There are still other work to be done. But leveraging our AI, we are able to automate about 80% of the process. The last core component of our platform is around the core alt text editing and QC features. accessibilityInsight itself starts with a dashboard, where you can load up all of your documents as well as check the status of the documents for accessibility.
BARRY BEALER: When you want details associated with accessibility, you can go into the accessibility validation report, which will show you on a document basis the thumbnail images, the alt text, and any issues that exist against the WCAG page and PDF/UA standards. As I mentioned before, we leverage AI to tag the PDF documents from the standpoint of the right structure, the heading 1, heading 2 paragraph as example, and the reading order.
BARRY BEALER: All this can be edited through using our online interface if there are changes that need to be made. In addition, we have very robust capability around how to handle math equations. We are able to extract out an equation from a PDF, create a thumbnail, and then create both MathML and syntactically correct alt text. We also are able to handle tables.
BARRY BEALER: Which those of us who have been in the publishing industry for a while understand tables can be very complex. We are able to tag elements within individual rows and columns. If they need to be adjusted, we can redraw the table and recapture the elements within the table. So what is an assistive technology experience after the document has been remediated?
BARRY BEALER: Again, on the left is the PDF. On the right is the assistive technology experience once that document has been remediated. It is in a much more logical reading order, and it provides context to content. Our accessibilityInsight platform leverages AI machine learning to automate up to 80% of the process.
BARRY BEALER: Our compliance reporting allows for the running of over 100 automated tests against WCAG and PDF/UA. We have a very flexible licensing model, and our software is highly configurable, with an intuitive interface, so that we can remediate documents to be 100% compliant. And the good news is we're able to get our customers onboarded within less than a half a day. If you're interested in talking to us about accessibility, about accessibility issues, or how to make your documents compliant, we'd be happy to talk to you.
BARRY BEALER: Thank you for your time. [MUSIC PLAYING]
MARIANNE CALILHANNA: Hi, my name is Marianne Calilhanna, and I'm with DCL. I'm here to speak with you about our solution called Content Clarity. Content Clarity provides a deep analysis of a publisher's entire corpus to identify obstacles and errors in content structure that hinder interoperability and discoverability. We developed this service for a few reasons-- 1, we do this for every platform migration when DCL was involved, although you need not wait for a migration to conduct this type of analysis.
MARIANNE CALILHANNA: 2, errors and issues with content structure are costly. The key business drivers for XML are to improve interoperability and facilitate search and discovery. 3, it provides a clear picture across your entire collection. Publishers that strategize and plan for organizational change to the development, production, and distribution of content must have a clear picture of content structure across the entire collection.
MARIANNE CALILHANNA: And finally, this service aligns with the NISO recommendation for content platform migrations. Lacking this information at the start of the migration effort means these data points will need to be discovered during the normalization process, which can result in needing to adapt and adjust to new information, often with cost and timeline impacts. Here's a list of some of the clarity checks we perform. Some of the clarity checks are informational, like whether or not NISO content platform migration working group recommends, and other clarity checks reveal errors that will impact downstream processes.
MARIANNE CALILHANNA: That includes getting your content into the researchers' hands. In the next few minutes, I'm going to share a few examples from the content clarity report. After analyzing a publisher's content library, the initial dashboard provides a variety of graphs, counts, and other insights. You can drill down to individual reports that help you understand specific files that require updating.
MARIANNE CALILHANNA: This is an example of the report the details which DTDs are referenced and used. If your content is not consistent against one DTD, or up to date with the latest version of JATS, you may be missing out on functionality in your designated platform. I mentioned that some of what content clarity reveals are informational, such as file metrics, and other results are errors in structure.
MARIANNE CALILHANNA: So this slide shows a few examples of those structural errors, such as an incorrect date in your XML, or multiple issue files, or a variance in how your volume, issues, and print publication dates are structured. Looking at some of the informational data might be valuable to other people in your organization. For example, a listing of article types might be something your marketing department could use, or the resulting accounts might indicate you have content misidentified in the XML.
MARIANNE CALILHANNA: The ISSN report captures print and digital ISSNs. You can see from this example that it's very likely this journal had a name change or is referenced differently in different files. If this did not get corrected, it would show up as additional facets in a platform search, even though it's the same journal. As more media are used in journal collections, you'll want to ensure the IDs and paths are referenced and updated.
MARIANNE CALILHANNA: The metadata and subject categories report helps improve content by eliminating redundancies, correcting typos, applying consistency, and updating taxonomies. When I personally started diving into some of these reports, I couldn't believe that some of the content was missing things like the doctype. It didn't compute, and of course it didn't validate, but indeed we do see these strange but true issues all the time.
MARIANNE CALILHANNA: So that's my super-quick preview of DCLs Content Clarity. You can reach out to me any time, and I'd be happy to spend more than five minutes discussing this with you. Thank you so much, and I hope getting back to the SPP annual meeting brings you as much joy as it brings me. Thank you.
NATE JACOBS: Hi, everyone. My name is Nate Jacobs, and I am the found and CEO of flaspub.io. And I'm excited to share with you a little bit about what we've been working on. So who are we and why did we start flaspub.io? So when I was doing my PhD and my dissertation, I just got really, really frustrated with the publishing process and how it was sort of limiting what I could share, and in some ways kind of dictating how we were running our experiments and how we were choosing what to share and what not to share.
NATE JACOBS: So what we wanted to do was create a more expressive and iterative way to formalize and share results. So we wanted to empower researchers to actively lead their communities in real time and build a movement behind the questions that you're the most passionate about. So what problem are we solving? So like a lot of people, we're really focusing on this problem of barriers, inequities, and inefficiencies in the publishing process.
NATE JACOBS: It takes a really, really long time to get enough results to fill out a full-length paper, and it sometimes takes just as long to actually get it published and in press. Obviously there's also financial barriers and lots of other barriers to publishing, and this ends up resulting in as much as 80% of results that are never published anywhere.
NATE JACOBS: So obviously this is a big problem in the space, and there's a lot of people doing a lot of different things in this area. So where does flaspub.io fit in, and how are how we tackling some of these problems? So what we've done is we have designed a smaller, faster scientific article that can be used to rapidly share individual results. So it's about the length of a single figure.
NATE JACOBS: It reads like a section of the results. There's no intro or discussion. And detailed methods, protocols, data sets, supporting and refuting references can all be linked at the bottom. It's [INAUDIBLE],, free to publish, free to read, and each micropublcation is structured around a single finding that is addressed in a structured assertion.
NATE JACOBS: So one issue that you have with micropublcations, these smaller, faster articles, is by themselves they don't tell a very big story on their own. And so the other side of the coin that we're working on is how do you aggregate them together to tell emerging stories and more compelling larger narratives. And the way that we're approaching this is with research campaigns, which essentially allows a researcher to launch a collaborative effort around a specific research aim that motivates and contextualizes a set of micropublcations down a particular line of inquiry.
NATE JACOBS: So one of the powerful aspects of this is that it allows micropublcations, each individual contribution, to be visualized in context, so that you can get a sense for where there's scientific consensus, a lack of scientific consensus, and gaps in the knowledge base, so that you can then solicit and motivate additional contributions to fill in those gaps and to get a more robust, community-wide, cross-lab replicated findings.
NATE JACOBS: So the first micropublishing journal that we've launched is Outbreak for the infectious disease informatics community. [INAUDIBLE],, for example, we've got a [? feed ?] [? view ?] for top and recent micropublcations and campaigns over there to the right. Each campaign, when you click in on it, opens up a full-page dedicated space to host data visualizations and a couple of different interactive features.
NATE JACOBS: So that's pretty much it for us, kind of the quick and dirty version of it. We have a small but mighty team right now, and we are growing every year. And we have a great advisory board. So if you wanted to reach out to me to learn more, again my name is Nate Jacobs. My email's nate@flashpub.io. You can also find me on Twitter @natesjacbos.
NATE JACOBS: And in terms of how to get how to get involved or people that we're looking to work with, if you're a publisher, we'd love to talk about integrating with publishing platforms or preprint servers. We sort of exist between productivity and full-length, regular journal articles. And so there's a lot of opportunities to embed micropubs, and to publish milestones, and final reports that are coming out of the campaigns.
NATE JACOBS: We're open source. So if you're a developer we'd love to work together on some of the software that we're building. And lastly, if you are a researcher in the infectious disease community that does modeling, or scenarios, or logistics, or anything in the informatic space, we'd love to workshop one of your projects to see if the campaigns would work well and if we could work together on featuring it on [INAUDIBLE].
JOHN LENAHAN: Hello, everybody. Thank you very much for joining me today. I'm going to go over a brief review of some of the models we have of making free and open access content available, and take a look at some of the overall global usage that we've seen for that content. JSTOR is part of ITHAKA. It's a not-for-profit organization utilizing digital technologies to preserve and provide access to scholarly content.
JOHN LENAHAN: One of the ways that we've worked on expanding is looking at free and open access content to add to the licensed content that's on JSTOR and provide more access to content to users who can't pay for or have access to licensed resources. So we're at now close to 1 million items that are free or have open access associated to them on the JSTOR platform, including journals, research reports, ebooks, images, and special collections.
JOHN LENAHAN: When we take a look at the usage, in particular on the journals and the book side, we can see that we continue to have year-over-year increases of usage of our open access books and early journal content. That EJC represents Early Journal Content, which is the out of copyright content, 1923 and earlier, and that always increases every year.
JOHN LENAHAN: And we can see some significant usage increases of that open content. We did see a good jump in licensed book usage content. But that was because we released a COVID-19 collection in March of last year, where many of our publishers made their content freely available to institutions who signed in for it, which led to a significant increase in usage there. The other important part when we look at open content and where that content is made available and discoverable is to ensure that we're working in the discovery services, the workflows for libraries, it's available and indexed in Google.
JOHN LENAHAN: Because we can see that content is discovered in different ways. Research reports had 54% of their content discovered direct from JSTOR, as opposed to some of the out-of-copyright journal content or open access journal content, where it was 50% of that content was discovered through Google, not coming direct on the JSTOR platform. So ensuring that we're in the places where researchers find it is really critical.
JOHN LENAHAN: Now, I'm going to look at a few different models where we've actually been working to make additional content available. One is by looking at content that's already licensed on JSTOR, working with publishers, and making articles and chapters available that's part of more theme-related collection building, and creating an open library format. And this area here, in collaboration with Schomburg Center, we worked with the Black Liberation Reading List and made 2,500 articles and chapters available and open that we're supporting and referencing from that reading list.
JOHN LENAHAN: And we can see, if we look here in 2020, and we look at the weeks of the beginning of each year, we can see it was fairly flat over the first eight weeks in 2020, when this content was available in only licensed collections. And as soon as we made most of that content available after the third week of this year, we saw usage go up significantly. And we've been seeing percentage increases of close to 1,000% used by institutions around the world.
JOHN LENAHAN: So as soon as it was made open, we saw significant increases of usage from users everywhere. The same happened with research reports. Research reports through February of last year were available only as part of a licensed thematic collection at JSTOR, putting it behind a paywall. As soon as we removed that and made them freely available on the JSTOR platform, we saw significant increases of usage, where we're seeing well over now a half a million item requests on research reports every year used across the world from every type of institution.
JOHN LENAHAN: Lastly, I'm going to talk about model we've done for ebooks. And we've been supporting open access ebooks now for several years. We've also been working with publisher. And this was a pilot with CLACSO, which is a publisher in Argentina, and in a collaboration with LARPP, which has many academic institutions who had the print subscriptions to CLACSO, redistributed the print funds over to supporting open access publishing, and we've been able to then, over the last two years, publish their current copyright year content directives open access.
JOHN LENAHAN: And we can see, through April of this year, over 600,000 chapter views and downloads from 217 countries and over 4,000 institutions. So making this content available that was just in print or to institutions who may be paying for those individual subscriptions, we can see a broad demand globally for this content. So a couple of examples of what we've been doing, making more content available.
JOHN LENAHAN: And when we do, we can see we get tremendous global usage of that content. Thank you very much for listening.
TIM LLOYD: Hi, everyone. I'm Tim Lloyd from LibLynx. We help content and service providers to generate more value from their online resources for identity, access, and analytics. Today I'm talking about our new Open Access usage Analytics solution or OA Analytics for short. In case you missed it, our industry hit an important tipping point last year. According to data from the open knowledge site lens.org, 51% of journal articles published last year are available through an open access license.
TIM LLOYD: Readers of the Scholarly Kitchen will have see a profusion of articles over the last year exploring the growth in open access publishing across diverse media types and business models. Underlying the move to open access is a seismic shift in funding, from paying to read to paying to publish. And this, combined with the proliferation of open access business models, means that publishers need to communicate the impact of open access content to a much broader range of stakeholders than the traditional library roles.
TIM LLOYD: For example, senior research officers within institutions want to understand how open access publishing supports their research priorities. Administrators are interested in reflecting audience impact metrics alongside traditional metrics in promotion and tenure committees. Publicly funded research needs to demonstrate whether it impacted key audiences, such as indigenous communities or those living in a particular geographical area.
TIM LLOYD: Institutional and community repositories rely on impact to encourage author engagement and justify institutional funding. And authors want to understand the impact of their own research, information that can also influence their future choice of publisher. There's obviously also a huge value in making this information available to your own publishing staff.
TIM LLOYD: For example, development roles that need to understand which organizations are getting value from open access content in order to identify potential future sources of funding, and editorial roles that want to understand the subjects and topics that are engaged in the community. In short, publishers need to communicate a complex message to an increasingly diverse internal and external audience. So how well do traditional analytic solutions address this need?
TIM LLOYD: In short, they don't. Many open access publishers and open repositories rely on simplistic counts of hits or downloads that offer minimal understanding of impact on the community. Counter reports seem an obvious answer to this problem, one that libraries are already very familiar with. But they're engineered for a very specific use case, helping librarians make informed decisions on acquiring content to meet their library's learning and research goals.
TIM LLOYD: There's no standard for how to attribute usage more broadly or to offer alternative methods of analyzing this audience. To address this growing need, LibLynx and PSI Metrics have partnered to provide organizations with real-time, granular usage analytics. Our solution focuses on three key benefits-- first, understanding your audience, with insights into which organizations they're accessing, when, from where, and how do they engage with your content-- compare usage by industry segments, such as government versus corporate or by country and region.
TIM LLOYD: Second, communicating value with dynamic, context-sensitive dashboards delivered to your customers and other external stakeholders, such as institutional partners, authors, and funders. Third, empowering your staff with on-demand analytics that are easy to understand and explore and easy to share, whether that means a PDF, a tabulator report, or online. Implementation is straightforward. Our lightweight tracking script and APIs enable easy browser-based and/or service side capture of raw event logging.
TIM LLOYD: You can embed reports within web applications to create your own data dashboards or access reports via our [? brandable ?] portal. Our analytics incorporate industry-recognized counter metrics to ensure they're consistent, credible, and comparable with existing metrics, and cross-diverse, open access, and pay-walled content. You can query reports via or IPs or export all processed data nightly to your data warehouse.
TIM LLOYD: It's your data. Capture custom metadata and metrics to power additional analytics specific to your content and publishing models. And at LibLynx, as with all our services, pricing is based on usage to ensure costs scale with your needs. To learn more about our open access analytic solution, please contact me afterwards, or email info@liblynx.com, and thanks.
TIM LLOYD: [MUSIC PLAYING]
MAHA AMAMI: Hi, my name is Maha, and I am a data scientist in MyScienceWork. The MyScienceWork repository includes almost 27 millions of scientific publication. To use this huge amount of data in the most effective and efficient way, we have implemented natural language processing application. To improve the database management through the fair principals, this natural language processing application allow the verification of data in several ways.
MAHA AMAMI: They enable the effective use of ontology. They ensure that name and entity are referred across different data sources so they can be used in retrieval system, recommender system, and so on. In this demo, we will present one of these NLP application that we implemented in MyScienceWork, and its name is DCypher. DCypher is a neural entity recognition tools that is based on machine learning and deep learning techniques to annotate biological publication.
MAHA AMAMI: And we will demonstrate how this tool recognized 22 different biological name entities, such as proteins and chemicals. From these articles in it accepts PDF format or row text as an input. So we are going to start by putting the text in the text field. And then there is different types of [INAUDIBLE] that we can choose, but here we are going to select to extract all the named entities and to have all the text annotated.
MAHA AMAMI: You can see here that we have the tooltips there that indicate the confidence score of this named entity from the predicted model. And we are going to wait between 15 seconds and 20 seconds to have the results. So here we have the result. As you can see, there are these text that is annotated. You have, for example, the protein or the gene ATG7 that is colored with the orange and yellow.
MAHA AMAMI: And you can find also mice that is colored with blue to indicate organism or species. And then you have a table that is indicating for each entity what is the label and the frequency of this named entity in the text. You can order also the entities by frequency. And there is a second table that shows us the named entities and their frequency in the text-- the classes of these named entities.
MAHA AMAMI: As I said before, we can also annotate an entire PDF file by selecting it, and then we are going to obtain, the same as before, the named entity identified in the text, as well as the table of frequencies-- the two tables of frequencies. And in addition to that, DCypher is able to extract the metadata of the PDF publication, and you will find it in the Metadata tab.
MAHA AMAMI: So you can see the title, the author and their affiliation, the abstract, the journal, the year of publication, and the link also using the DOE of the publication to its original repository. And then also it's going to show you the references of this publication.
MAHA AMAMI: I hope that you enjoyed the demo, and thanks for your attention. [MUSIC PLAYING]
SARAH TAYLOR: Hi, I'm Sarah Taylor, COO at Research Square Company, where we provide the Research Square preprint platform as well as AJE manuscript preparation services. Today, I'll be talking about an exciting new tool that we've developed to automate language editing of academic manuscripts and how we've use this tool to increase efficiency and reduce costs in our own production workflows.
SARAH TAYLOR: Over the past four years, we've actively invested in AI technologies to improve our operational efficiency, and now we're productizing these tools for authors and publishers. We've developed several automated tools, and we're happy to discuss how any of these may be helpful to you. But today I'm going to focus on our digital editing tool. This tool was built using deep learning technology, with our corpus of over 800,000 edited manuscripts.
SARAH TAYLOR: Here I'm showing an example of what the automated editing tool can do. In this one example, we see corrections for phrasing, articles, tense, singular-plural corrections, missing words, and word choice. And with this tool, we have the ability to choose between American and British English, as well as style requirements tailored to specific journals.
SARAH TAYLOR: Through our internal editing workflows, we provide language editing services for over 2,000 manuscripts per week. We see an average wordcount of just over 5,000 words per paper. And on average, we make about 1,000 edits per paper. With our AI digital editing tool, over half of those edits are now automated. And in this graph, you can see how we're continuously improving this tool over time.
SARAH TAYLOR: We've also developed a tool to assess language quality on a 1 to 10 scale. This is our language assessment tool, and we use this tool in our operational workflows to automatically direct manuscripts toward the appropriate level and time investment for quality control. In this slide, we're using the language assessment tool to track quality through our editing workflow.
SARAH TAYLOR: The workflow starts with a customer submitting a manuscript with scores on the lower end of the scale. After digital editing, two out of the three sample manuscripts already have high language quality. And of course, the human editing step adds further value. With regard to cost savings, we built the digital editing tool in 2017 and began using it in our internal workflows in 2018.
SARAH TAYLOR: In this graph, you can see the number of manuscripts edited per editor on our team. We've nearly doubled our efficiency due to this tool, which for us has increased our overall scalability and saved over $10 million in personnel costs. As I mentioned, we're now productizing our tools, and we're excited to collaborate with you to help improve your language as an author or your production workflows as a publisher.
SARAH TAYLOR: Please reach out to Lindsay Miller, Neil Blair Christensen, or me at Research Square. We look forward to collaborating with you, and thank you so much for your time.
BYRON RUSSELL: Hello, everyone, and thanks for listening to this preview session on blockchain and scholarly publishing. I'm sure that everyone watching will be aware that blockchain has been much in the news recently. Most of all it's associated with cryptocurrencies, which are now making their way into the world of mainstream banking and finance. However, Bitcoin and other such cryptocurrencies are just one use for blockchain technology.
BYRON RUSSELL: What I'd like to do in the next few minutes is to look at what blockchain actually is, and how this technology can be used in the world of publishing, most especially how the technology can protect author rights and intellectual property from fraud and misrepresentation. Almost all scholarly publishing output is currently stored in digital databases.
BYRON RUSSELL: These databases, with a cloud-based on local data store, are held within a publishing house, society, or association and typically have a central administrator. Blockchain, on the other hand, is a distributed ledger or decentralized database. In other words, not one archive but multiple archives, replicated and linked within the network and synchronized via the internet. Anyone within the network can access the records stored in such a distributed ledger if they have the right key.
BYRON RUSSELL: However, although users can access the records so stored, they cannot tamper or mess with them in any way. Blockchain as a storage mechanism has four key attributes-- first, it's immutable. Documents stored on the database cannot be changed in any way. Secondly, it's open, in that anyone with the right key can unlock the ledger and examine the records stored within it.
BYRON RUSSELL: It's also transparent, in that records so stored are potentially accessible to anyone with appropriate access rights. Finally, and most importantly, it's secure. Even if you have a key to the ledger, you cannot remove or replace any article stored in the archive, nor can you change such records. Unlike cloud storage providers, let alone local storage area networks, the blockchain cannot go down.
BYRON RUSSELL: And again, unlike cloud storage, because nobody owns the blockchain network, it's impossible to hack. Plagiarism, falsification of research data, and incorrect content attribution has never been as easy as it is now, nor is it so costly. One estimate of the average cost of a single investigation into research malpractice in the United States was around half a million dollars. By archiving original research data and original content on the blockchain, the supporting data for scholarly articles can be accessible and verifiable by collaborators, peer reviewers, or funders.
BYRON RUSSELL: All content stored is time stamped and immutable. When and by whom an original document or original data set was produced is easy to prove. As publishers, you may well want to license rights to subscription content to third parties. But once content is out in the world, it becomes increasingly difficult to track. Although blockchain cannot help a publisher to track article misuse, blockchain archiving makes it very easy to demonstrate the original source of the work and, perhaps equally importantly, where the funding for such work came from.
BYRON RUSSELL: The publishing industry itself is a changing sea. Learned societies are absorbed or merged with others, while publishers themselves are open to acquisition. The key to value in all such commercial transactions is, of course, the worth to the purchaser of original intellectual property. But dissecting who owns which assets within even a relatively small publishing business can be complex and time consuming. The fact that blockchain provides irrefutable proof of ownership can lead to a streamlining of the merger or acquisition process itself.
BYRON RUSSELL: So if you truly want to protect your assets and ensure that you can maximize the value of your published assets, blockchain storage is a very sensible option. Here at block.co, a technical incubator from the University of Nicosia, we've devised a simple intuitive interface that belies the complexity of the architecture underpinning it. This makes it very easy to upload documents, such as PDFs, directly into a blockchain with minimal training or technical knowledge.
BYRON RUSSELL: If you'd like to know more, I'll be around for questions. Please feel free to send me any questions at the end of the preview session or visit the block.co site for more information. Thank you very much for listening.