Name:
                                If curation is the future, what actually works?
                            
                            
                                Description:
                                If curation is the future, what actually works?
                            
                            
                                Thumbnail URL:
                                https://cadmoremediastorage.blob.core.windows.net/e8a6e42d-5e86-4a71-a1a2-dc58e8fab8f8/videoscrubberimages/Scrubber_10.jpg
                            
                            
                                Duration:
                                T00H54M29S
                            
                            
                                Embed URL:
                                https://stream.cadmore.media/player/e8a6e42d-5e86-4a71-a1a2-dc58e8fab8f8
                            
                            
                                Content URL:
                                https://cadmoreoriginalmedia.blob.core.windows.net/e8a6e42d-5e86-4a71-a1a2-dc58e8fab8f8/Curation_1920x1080.mp4?sv=2019-02-02&sr=c&sig=e7Rd94ZNjxMxMDCexQNgbCpuKycCPOGJR1qFlOC4bmA%3D&st=2025-10-31T20%3A50%3A43Z&se=2025-10-31T22%3A55%3A43Z&sp=r
                            
                            
                                Upload Date:
                                2020-11-18T00:00:00.0000000
                            
                            
                                Transcript:
                                Language: EN. 
Segment:0 . 
  
WILL SCHWEITZER: The Platform Strategies Webinar Series, like our in-person meeting,  explores the boundaries of information in publishing  business and technology, and the strategies  we employ to fulfill our missions and business  objectives.  Today, I'm joined by Claire Moulton, publisher  of The Company of Biologists, Tasha Mellins-Cohen, director  of technology and innovation at the Microbiology Society,  and Bert Carelli, director of partnerships at TrendMD.  And we're here today to discuss curation,  which is a pretty broad topic, and can mean a lot of things.   
WILL SCHWEITZER: But the Wellcome Trust, through a grant award  that some of our panelists are involved in,  provided a helpful definition that we'll use  to ground our discussion today.  That definition is, signaling the significance  of published research outputs in an open  and a transparent manner.  Each of our panelists will present three  or so slides on the topics.   
WILL SCHWEITZER: And in the remarks, we'll touch on kind  of their organization's approach to curation, and, perhaps,  what they think works or doesn't.  We'll reserve the bulk of our time for discussion and Q&A.  You can use the chat function in Zoom  to submit your questions at any point, which  we'll hold until all the end of the panelists' slides.  We'll first hear from Claire, then Tasha, then Burt.  Claire, would you like to take it from here?   
WILL SCHWEITZER:   
CLAIRE MOULTON: I will just share my screen and start up.  So, as Will said, I the publisher at The Company  of Biologists.  And we are a new publisher with Silverchair for 2020 on.  So thank you for inviting me to the webinar.  So I do think that community publishers play  a significant role in curating content for their communities.  Most commonly, that will be around special issues  and subject collections in the content  relevant to the journals.   
CLAIRE MOULTON: But what I'm going to talk about today  is a service that we launched called preLights, which  is about selecting and writing about preprints.  And this is prefixed with the whole  of the biological sciences.  They're not preprints that have been sent to our journals  for publication.  It would cover any preprint.  So preLights is built around a community of early career  researchers, and we call them preLighters.   
CLAIRE MOULTON: And they've been very enthusiastic  about preprints and transparency in communication.  So a really great group to work with.  And what they do is they select interesting or important  preprints in their area of expertise.  As I said, they can select that from the entire preprint  literature.  But we do notice that they predominantly  select preprints from bioRxiv.   
CLAIRE MOULTON: They then digest the preprint.  They will explain key findings, the scientific background,  and, importantly, they comment on  why they personally found it interesting  or why they selected it.  So it's very much a manual curation.  And it's generally a positive selection,  with some constructive questioning around the content.  So things that I think work well is,  I think, that each post has a short tweetable summary.   
CLAIRE MOULTON: And Twitter is the place where I would  love to talk about preprints, so it  provides a good source of browsable content for readers  to see on Twitter before they come through to our site.  Another thing that's worked well is  that we've used the same subject categories as bioRxiv.  And that's helped us create technology links  between archives and preLights, which has been very useful.  So for example, if you're on bioRxiv,  and you're looking at a preprint that we have covered,  you will see a link to preLights and the discussion  of the preprint.   
CLAIRE MOULTON: And then I also think the service provides a bit of value  adds.  So when the preLighter has written their post,  they send it to the original preprint  author with some questions about the research,  whether they can tell them a bit more,  or challenging some of their approaches.  And the majority of the preprint authors  do respond and provide comments back to our site.   
CLAIRE MOULTON: So I think that that's quite interesting.  Things that don't work so well, although the preprint  authors do comment on the site, we don't get very much general  commenting from readers.  This doesn't surprise us.  It's what we see in our other community sites, as well.  But I think that that's something  that we all hoped would develop more in the future.  And I think that the other thing that maybe doesn't work quite  so well is that each preLighter obviously selects preprints  in their area of expertise.   
CLAIRE MOULTON: And so, although we have quite a lot of preLighters,  we have a little bit of patchy coverage.  So I think that ways of navigating the content,  you can obviously search the site.  We add keywords for posts.  You can use those subject categories, which, as I said,  were the same as bioRxiv.  And you could use those to select which preprint  highlights you see on the site.   
CLAIRE MOULTON: And you could also use them to select  which preprint highlights you see in your email content  alerts.  We think that some readers also follow particular preLighters  based on their subject speciality, which  is quite nice to see.  But, as I said, we did get a little bit of patchy coverage.  So one of the things that we introduced  was something that we call preLists.   
CLAIRE MOULTON: So they were actually suggested by the preLighters.  And they wanted to be able to highlight a larger  number of preprints, and just not write quite as  much about them, so that they could just cover more preprints  with a lower amount of work.  So these have actually been quite popular.  We have a set of preLists that were on particular topics,  in this case, antimicrobials.  But we also have preLIsts that focus on the preprints that  are featured in particular scientific meetings,  and those are generally one-off lists.   
CLAIRE MOULTON: But they have been very popular, as well.  So, as it was mentioned, this year has been all about COVID,  so I did want to mention how that has  affected the site this year.  And in fact, it's been very positive for us.  So we had a huge number of preLighters join the team  because people had more spare time during lockdown.  So we now have about 250 active preLighters writing for us.  And they also wrote more for us this year.   
CLAIRE MOULTON: So we are just approaching our 1,000th post.  And we had more of the preprint authors engaging, as well.  And I guess I should also highlight  that we've now got about 5,500 followers on Twitter.  So they're getting a good amount of attention.  You can imagine that we thought it  would be a good idea to launch a preList about COVID preprints.  And that was incredibly popular initially.  But you will not be surprised to hear that we immediately  lost control of the number of preprints being posted.   
CLAIRE MOULTON: So a few of the preLighters actually  moved to produce something separate from our site.  And that is a site called covidpreprints.com.  And we thought this was quite interesting  because it presents the preprint alongside a timeline  of the pandemic.  And crucially, it calls out preprints  that have been debunked, either because they overstated  their claims, were too preliminary,  or were potentially fraudulent.   
CLAIRE MOULTON: So that's actually been a really interesting  separate initiative.  And I thought just to finish, I would  highlight a new service coming out of eLife that they  have just called Sciety.  And this will be a gathering together  of feeds from a number of services that either highlight  or review preprints.  So I think that that's one to keep an eye on.   
CLAIRE MOULTON: So there we are.  That's me finished.  Thank you very much.   
WILL SCHWEITZER:  Thank you, Claire.  Tasha?    
TASHA MELLINS-COHEN: Apologies.  Let me just unmute myself.  So Will referenced that earlier this year, the Wellcome Trust  and Howard Hughes Medical Institute  launched the Learned Society Curation Awards.  And it was very exciting for the Microbiology Society  that we were successful in our application  to convert one of the Society's journals, which  is Access Microbiology, into an open research platform.  And this along the lines of Faculty  of 1,000 or Wellcome Open Research.   
TASHA MELLINS-COHEN: From a values perspective, this fits very well  with the Society.  It offers us an opportunity to bring transparency  to a traditionally very opaque process of peer review,  and embrace open scholarship, which  is very much the direction that we see publishing traveling in.  It also helps us speed up the process of disseminating  research, which maximizes the speed of communication  as well as the potential impact and influence of each article.   
TASHA MELLINS-COHEN: We've just heard from Claire about preLights,  and how valuable that is proving to be as a way to communicate  research at an early stage.  The reason that we are thinking of this as anti-curation,  as it were, is that a traditional journal  was hyper-selective.  They had to confine the published research to the bound  contents of a print journal.  Modern journals have somewhat broken those boundaries.   
TASHA MELLINS-COHEN: And we now have many journals, many articles  which incorporate open data, open methodology,  so on and so forth.  But we see an open research platform  as a new form of product.  It's not just a journal.  It gives us an opportunity, for example,  to incorporate posters from our conferences,  to potentially publish the theses of our members  if they wish to do so.   
TASHA MELLINS-COHEN: There's so many more things that can be incorporated  into an open research platform that it becomes  less about curation and more about  layering and anti-curation.  So as you can see here, this is a very simplified, but very  traditional workflow for a journal,  well, traditional in the sense of preprints  having come into play about 20 years ago.    
TASHA MELLINS-COHEN: Everyone could see a preprint.  We know that, we appreciate that.  But from the point of preprint to the point  of formal publication is very opaque.  A technical check will happen, but if an editor decides  at that point to reject the article,  nobody sees that information.  Nobody sees why that rejection has happened.  Similarly, peer review, nobody really  sees the peer review reports except for the editors  and the authors themselves.   
TASHA MELLINS-COHEN: You don't have that level of transparency.  And only after many rounds of review and revision  and editing, and is a very highly curated process,  will the real world, the world outside of the editors  and the reviewers and the authors,  see the final version of the article.  So most research, most of the process of research  is curated, and very carefully so.  For an open research platform, the workflow changes.   
TASHA MELLINS-COHEN: And it looks a little bit more like this.  So there are many, many points where versions and reviews  of those versions are posted.  So our intention is that the point an article is submitted--  It might well have already gone onto bioRxiv.  That's entirely up to the authors--  we will run machine learning technology peer review.  So I suppose not truly a peer because it's not human.  But we will be looking at things like SciScore  and authenticate to make sure that the research meets  certain basic standards.   
TASHA MELLINS-COHEN: The initial preprint will then be posted to our platform.  And that's our standard microbiologyresearch.org  platform with the AI review reports  made available alongside it.  Transparent peer review will be managed  through our editorial manager system.  We're not making use of any kind of bespoke software here.  It's all the stuff that we already  have in our technology stack.   
TASHA MELLINS-COHEN: And again, the reviews will be posted  once they've been received.  And the revised article will be submitted.  And we'll see a new version of the preprint,  and so on and so forth.  Everything will include relevant persistent identifiers,  whether that's DOI for all of the versions of the article  and for all of the peer review reports, links to open  data, use of funding data, usage of Research Organization  Registry information, use of ORCID, all of that  will be incorporated at every point in the process.   
TASHA MELLINS-COHEN: So in this way, we get a very clear very transparent, very  versioned view of the research process.  Now, this bit is anti-curation.  What comes next is a bit different.  After this, with this formal publication  of the the version that the reviewers and the editors  and the authors agree is acceptable,  the communities around the Microbiology Society  and, indeed, beyond the microbiology society,  will have an opportunity to create,  effectively, overlay journals or reinvented curation,  but reinvented curation over the top of a complete transparent  process.   
TASHA MELLINS-COHEN: I thin k that is it.  Yes.  I will hand over to Bert.   
WILL SCHWEITZER:  Thank you very much.    
BERT CARELLI: Excellent.  Thanks, Tasha.  Thanks, Will.  Let me just.  So TrendMD is a technology company.  So I'm going to focus a bit on technology.  And to provide a bit of oversimplified context here,  we can look at technology as having  played an evolving and disruptive role in the curation  of scholarly content.   
BERT CARELLI: Up until the 1960s, curation was the province  of editors in libraries.  Manuscripts were selected in issues assembled by editors,  in libraries acquired and organized  by journal title, volume, and issue,  discovering new advances in one's own field,  and first acquiring access to the key journals,  and then systematically browsing tables of contents,  or in a few fields, like medicine,  browsing scholarly indexes that could be months or years out  of date.   
BERT CARELLI: These methods for keeping up with new discoveries  were highly inefficient, and often prohibitively expensive.  NASA learned during the early days of the space race  that it was often more expensive to figure out  if an experiment had already been done than it was to simply  do it again, from scratch.  In fact, that realization led to the creation and funding  of one of the first information retrieval databases  at Lockheed Missiles and Space, and the era of digital creation  curation was launched.   
BERT CARELLI: Database era ushered in major changes  in the way scholarly information could be curated.  Instead of pouring through journal by journal,  issue by issue, a researcher, usually  with the help of a librarian skilled in the specialized  language of command line search, could  look at the entire archive of a journal,  or search a citation database using subject taxonomies  like MeSH, and curate lists of appropriate articles  for the researcher to review.   
BERT CARELLI: Flashing forward to the first two  decades of the 21st century, web technology  has given us the ubiquitous search box,  from which virtually all scholarly material can  be accessed, and where curation is effectively  done at the article level.  A keyword search on any subject instantly "curates"--  I'm using air quotes here--  a list.   
BERT CARELLI: The quality of that list can vary widely.  For example, you might see a peer reviewed article  from an established journal on a subject like the flu  vaccine sitting next to a conspiracy monger's post  claiming vaccines can cause autism.  But accepting that the serious researcher can  deal with its obvious flaws, we can see that curation,  the Google search does fulfill another purpose.  Sometimes you're not just looking  for a specific piece of content, which Google's obviously  very good at providing.   
BERT CARELLI: Sometimes a researcher also wants  to see what other related content might  be interesting or useful.  In this case, the search results page  becomes not the end, but a key step in the larger  process of discovery.  In their survey, "How Readers Discover  Content in Scholarly Publications,"  Tracy Gardner and Simon Inger found  that users are also likely to browse other articles that  surface in a search.   
BERT CARELLI: So the need to browse to learn what you did not  think to ask, maybe allowing for a little serendipity,  is still a vital part of the research process.  And this is where this related articles  function comes into play, which is TrendMD's business.  Going back to the study that I just  mentioned, which has been done every three years since 2005,  users were asked what features of published websites they  use most often.   
BERT CARELLI: Over the five iterations of that survey,  the related articles feature showed consistently  increasing popularity, while other website features,  like publisher produced news, site search,  saved search, alerting, declined.  TrendMD's recommendations widget uses behavioral data  and AI technology to create another type of browsing  experience for the reader, suggesting additional articles  based on what they are reading at the moment,  what other users like them have read,  and what they've read in the past.   
BERT CARELLI: Unlike a search-generated browse list,  which contains thousands of items closely related  to the search terms, TrendMD's focus  is to curate a short list of related items  with the highest likelihood of being  of interest to the reader.  It's similar to the way that Amazon works.  If you buy a teapot through Amazon,  the site's not going to show you more teapots.   
BERT CARELLI: You'll see products that other buyers of teapots  have bought, like recipe books, teacups, or tea.  In TrendMD's case, a medical practitioner  reading an article in The BMJ on smoking hazards,  may see recommended articles on lung cancer.  While a public health policy specialist may see articles  on secondhand smoke, and a biochemical researcher  may see articles on tumor formation.  The other aspect of TrendMD's approach  is that these recommendations are selected from the largest  possible collection of content.   
BERT CARELLI: Half of the recommendations are from the same journal  or publisher as the article being read.  And the other half are from a network  of over 5,000 scholarly sites.  Result is that the user sees content from journals  and publishers that they might otherwise not be familiar with.  And it enables authors and publishers  to reach audiences that will truly value their content.  Number of case studies, including  a 12-month randomized control trial  that demonstrated that articles promoted by TrendMD  were 50% more likely to be cited,  have shown the value that this feature  provides for publishers, authors, and researchers.   
BERT CARELLI: Back to Will.   
WILL SCHWEITZER:  Thank you very much.  So just want to remind the audience,  you all can submit questions via the Q&A, which  you'll find at the bottom of your Zoom screen.  And we have just a few questions to kick off our discussion.  And Claire, I thought I'd come back to you first.  And you've already answered two audience questions  around who are the preLighters, and how were they  selected and trained.  But do you know if the preLighters are  using any technology to either identify, or select, or even  summarize preprints they may find?   
WILL SCHWEITZER: Is that something you all are looking into?   
CLAIRE MOULTON: No, it's not actually.  So I think it's a really good question.  But is a very manual human process at the moment.  And I think that we did have in mind  that we might try those sorts of things in the future.  But I think people quite like the real personal touch.  And one of the things that has really  enthused people about this product  is actually the early career community.  And they are really keen on preprints.   
CLAIRE MOULTON: They write preprints with meta-analysis about preprints.  And they do lots of presentations about preprints.  And so it's very much a human endeavor at the moment.   
WILL SCHWEITZER: Thank you.  So Tasha, I'm going to purposefully  ask a pointed question here to maybe generate  some more discussion?  And, actually, I have two questions.  And one is, do you feel this ant-curation approach  exacerbates the glut of information that's out there?   
TASHA MELLINS-COHEN: I think the flood of information  is happening anyway.  The anti-curation approach, by introducing transparency  and providing levels of reporting even  on that very first preprint, which simply doesn't exist  in bioRxiv, helps readers to understand  that what they're seeing might not be the final version,  but has been through some level of analysis,  some level of checking.   
TASHA MELLINS-COHEN: Interestingly, when Claire mentioned early career  researchers, Access Microbiology as a journal  was originally created specifically  in response to Early Career Microbiology Forum  saying that they needed a place to publish  small papers, replication studies, negative results,  and so forth.  And actually, not only was that journal's scope defined  by the early career group, but the editorial board  is made up primarily of early career mentees,  paired with very experienced mentors.   
TASHA MELLINS-COHEN: And that group has proven to be very  receptive to different ways of working,  including this kind of anti-curation approach.  So it was it was them.  It was the early career group who  were really instrumental in us applying for the Wellcome  Grant.  And they're the ones who are saying,  this will make our research easier, better  because we will be able to see what is valid immediately.   
WILL SCHWEITZER:  Oh that's great.  And then I know it's kind of relatively early days  for the journal over the open resource platform.  But are you seeing any user data that  suggest readers are accessing and using  either the peer review commentary, the decision  letters, or the material around the final version record  article?   
TASHA MELLINS-COHEN:  So we aren't live yet.  We're still very much in build.  It's very early days.  But part of the reason that we went ahead with this is that,  actually, if you look F1,000 or Wellcome Open Research,  you can see how much traction the reviewer reports are  getting on those platforms.  So you can see how many times those have been read.  Now with any download statistic, any usage statistic,  you don't know for sure if somebody's actually  read the piece, or if they've just opened it.   
TASHA MELLINS-COHEN: But we know that people are making use of that.  They are visiting the peer review pages,  which gave us confidence that if we were to do this,  we would be able to add value to the publication process  by adding transparency.   
WILL SCHWEITZER:  That's excellent.  I think it's been, I would say, an increasing trend  we've within this Silverchair community  that publishers are bringing new questions to their platform  analytics and their usage data.   
TASHA MELLINS-COHEN: Yeah.   
WILL SCHWEITZER: It seems like five years ago,  we were all taking for granted that folks were coming  in from Google or a large-scale search engine,  finding the PDF that they wanted, and leaving.  And I think as we dig in further,  we're finding there's really valuable use cases  and services we can add on.   
TASHA MELLINS-COHEN: Yeah.   
WILL SCHWEITZER:  Bert, my next question  is for you, which is you provided some great data  around how frequently the TrendMD widget  and recommendations are used on platform,  some potential citation effects for the widget, which is really  great data.  But I'm thinking about editorial stakeholders in the process,  and the efficacy of the recommendations,  and if you have any data, anecdotal or otherwise, that's  compares, say, TrendMD's recommendations  versus editorial or human-curated links  in the process.   
BERT CARELLI: Well, yeah.  Thanks, Will, for that question.  At your former organization, we went through a very lengthy  trial, starting out on some of the newer journals, the smaller  journals.  And the editors weighed in heavily  in the early days about the TrendMD recommendations.  And the conclusion, and why we're now  on all of the AAAS journals, was that they actually  preferred the engagement that was being shown for the TrendMD  recommendations.   
BERT CARELLI: And it's actually not surprising because from my experience back  working with publishers over the years,  I've noticed that publishers are not  so attentive to the use of features on their website.  And one of the things that has been on websites for many years  is some type of related articles functionality.  And often, it was one of the least used functions  even when editorial staff spent a lot of hours trying  to curate that list of, here's additional content  that the user might be interested in.   
BERT CARELLI: So being able to come up with it,  being able to offer an automated solution  was helpful, certainly from a cost perspective  for the print publishers.  But also, the engagement that was provided clearly  indicated that people were using them.  And that's really what TrendMD functions on  is clickthrough rate as the key indicator of quality.  If people are engaging with the widget,  if they're clicking on the articles,  then it's a pretty good indication  that there must be some quality there.   
BERT CARELLI: So it's really leaving it up to the users, the wisdom  of crowds, if you want to think of it that way,  to dictate whether the quality is there.  And so far, it's been consistently high.   
WILL SCHWEITZER: Thank you.  So I have one last question before you  all if you have questions for each other.  And then we can turn to some audience questions, which is  we've brought up kind of COVID.  And we've seen a proliferation of activity  around preprints and COVID.  Chuckling at this point may be a good coping or safety  mechanism.  But earlier today, Kent Anderson in The Geyser  wrote about the potential problems of preprint servers,  and essentially disinformation.   
WILL SCHWEITZER: And I think it points this really interesting tension  of the role of curation, and essentially signaling  significant and valid research.  And if I were to put an editorial comment on preLights,  or define communities that I see,  it seems to be working with a closed community.  So it may be working within a group of several thousand life  sciences researchers, but we don't  seem to have effective curation mechanisms, yet,  at scale, or at the seam of science and the broader public,  or, perhaps, even science in the policymaking community.   
WILL SCHWEITZER: I was wondering if you all have any thoughts about this.  Are you seeing anything in your communities  with your programs that might give possible solutions  or indications for solutions for this problem  about curating research or even early stage  research for the problems facing us today?   
TASHA MELLINS-COHEN: May I?   
WILL SCHWEITZER: Please.   
TASHA MELLINS-COHEN: I would argue that there are also  problems with research that is published in a journal post  peer review.  One of the biggest public health crises we faced--  this would be prior to COVID--  is the damage done by the Wakefield paper, which  was published after peer review in a very  well-respected journal.  And it has created a hideous situation,  where many people are rejecting vaccination,  which is a perfectly safe and sensible public health measure.   
TASHA MELLINS-COHEN: And it's allowing terrible diseases  to spread again in communities that  should not have to face it.  So, yes, I absolutely agree that the proliferation  of non-peer-reviewed research being available to everybody  could create a problem with disinformation  I don't think the peer review is necessarily the only way  to solve it.  But I do think that the application of the sorts  of machine learning tools that we are proposing  to use for preprints could help to flag,  for example, there is a methodology problem here,  or these authors have not declared  their conflicts of interest, and give people  some context and some guidance as  to where there may be flaws in that manuscript.   
TASHA MELLINS-COHEN: But we have for decades said, peer review  means something is valid.  And that is not necessarily the case.  And Kent and I have had a vocal argument  about that, so [LAUGHS] I won't carry on.   
CLAIRE MOULTON: I would agree with that.  I think that posting or publishing,  wherever if you are in the cycle,  it's still just part of the beginning of the discussion  of the science.  People want to discuss it, they want to reproduce it,  and they to challenge it.  I think that one of the interesting things that  is going to come out of this past year  is that I think that the people who were in favor of preprints  will be feel even more strongly about it because they can see  that the rapid fabrication of work  has played a very important role this year around COVID.   
CLAIRE MOULTON: And the people who have had concerns about preprints  will feel more strongly about that  because they will have seen preprints  that could potentially cause harm  or, certainly, diverted resources.  So I suspect that the viewpoints may  have become even more opposite.   
WILL SCHWEITZER: Seemingly with all things these days.  Bert, any thoughts from you?   
BERT CARELLI: Well, I think that there's  a certain philosophical perspective here.  When Cold Spring Harbor first launched bioRxiv,  I was working with them to help them launch the preprint site.  And, to be honest, I had some skepticism  after working with scholarly journals,  with the peer review journals for so many years,  and knowing how editors were so protective of their space,  of their choices, of what to publish.   
BERT CARELLI: But I became a convert, to be honest,  especially seeing the growth and the enthusiasm.  I have a son, who at the time was in graduate school  and working in biochemistry, and I often  used his group, his peers to show me early examples,  even before bioRxiv launched.  This is the type of thing--  This is what is being attempted here.  And the overall enthusiasm of those young people,  of those early career scientists,  who are now know actually in there  the middle of their careers, was overwhelming.   
BERT CARELLI: And I think that there's a sense that if we  can trust the process--   There are going to be these outliers,  there are going to be the Wakefield papers,  and there are going to be some of the COVID papers that  have been really scandalous.  But there's a certain wisdom in crowds  that we're almost committed to that today  anyway because we can't deny that the tides are  rising around us all the time of information.   
BERT CARELLI: So we have to trust that there's an audience there  that has the ability to make some critical choices  about things.  So, I don't know, that that's my sense in why  I feel like a technology solution,  you can't turn the clock on it.  Once the genie is out of the bottle--  to mix many metaphors here.   
WILL SCHWEITZER:  Still a trust or hope  for audience discernment.  So before turning to some audience questions,  we have three in so far, and some of them  are a bit philosophical, I wanted  to see if you all had any questions for each other.   
CLAIRE MOULTON: I do have a question for Tasha, actually.  It's quite a practical one.  But when I was talking about your new workflow with my team,  they were interested in, what happens if a preprint comes  to you, goes through the review, and then the author  chooses not to do the revisions that you've requested?  Can they then continue and take that paper elsewhere?  How does it work?   
TASHA MELLINS-COHEN: Well, effectively, it is a preprint.  So if they really don't want to continue,  then, yes, they can take that elsewhere.  And they can say, this is my preprint,  it's in the open research platform,  I have chosen not to finish the publication process there.   There's no reason that we couldn't let that happen.  From the conversations that we've  had so far, the researchers that we've  spoken to have sort of said, oh, why would I do that?   
TASHA MELLINS-COHEN: If I've gone through all that process,  and I've gone through the reviews with you,  why would I take my article elsewhere and start again?  So I don't foresee it necessarily  happening particularly frequently,  unless and until there is a pretty broad acceptance  of transfer of peer reviews, which isn't  hugely popular at the moment.  And there's a lot of editors that I work with personally  who've over the years said, I wouldn't trust  somebody else's peer reviewers.   
TASHA MELLINS-COHEN: Why?  If the peer reviewer's good, the peer reviewer's good.  So, yeah.  Hypothetically, people could get partway through the process  and pull out at that point.  There's also a possibility that they get partway  through the process and refuse to meet  the requirements of the editor to amend their paper  to make it satisfactory.   
TASHA MELLINS-COHEN: And at that point, it would remain  simply a posted preprint.  We wouldn't go through the production  process, which I think is possibly more likely to happen.    
WILL SCHWEITZER: Any other questions for each other?   
BERT CARELLI: Yeah, actually I have a question  really probably more for Claire, but also Tasha.  I've been talking with a number of companies  that work with automated summarization and sentiment  analysis, and I'm curious whether you've  discussed or considered adding this type of thing  to somehow scale up the process?  Especially, Claire, it's impressive the numbers  that you mentioned of the preLights involvement.   
BERT CARELLI: But still, it's a very small fraction  of the preprints that are there out there.   
CLAIRE MOULTON: As I say, it's a really good point.  It is one that we have discussed.  I think that we're less likely to do it on preLights  because I think that one of the parts of the post is the,  why did I personally choose this preprint?  And what are the questions that I want  to ask to the print author?  And be able to follow up in more detail with those.  I think that where we more interested is--  So often, with research articles,  we write a research highlight, so just a short summary the  basically encourages readers to go and read the article.   
CLAIRE MOULTON: And that's the sort of work that our in-house team currently  write.  And I can see a time saving for using  an automated process for that.  And we would still have I think that certainly  in the early days, would, certainly  then have a person checking that before it went live.  But, yeah, we are interested in the idea.   
TASHA MELLINS-COHEN:  So we are actually  going to be using one of those tools  as part of our initial pass, initial automated pass  through each preprint, so that we can give people  an idea of what's in there.  Because why not?   
WILL SCHWEITZER: Great.  It'll be exciting to see how that gets on.   This next question's from the audience.  And it's a bit of a philosophical one, which  is, is there a distinction between recommendation  and curation?  There's probably a related point here  about how do we label those recommendations  or curated links for end users.   
WILL SCHWEITZER:   
TASHA MELLINS-COHEN: I think I would draw a distinction  because there are articles that I might collect together  for a particular purpose, and I would  consider that to be curation.  But the purpose might not be positive.  It might not be something I would  recommend that somebody read.  I might be collecting it together  to debunk it or to another way challenge it.  I don't know how the others feel about that.   
CLAIRE MOULTON: Yeah, I would also distinguish between them.  I think, yes, curation is more the collecting  together than a recommendation.  But I suspect the most curated collections  will be recommendations.  But I think they are different.   
TASHA MELLINS-COHEN:  Yeah, and a reference list  is a curated list.  But it's often, this is fake, this is negative.  Sorry, Bert.   
BERT CARELLI: Yeah, well TrendMD comes at this issue,  and it's a big question for us, certainly,  that we think about all the time from really  a slightly different perspective.  The problem that we set out to solve  was really one more for publishers and authors  who wanted to have their content exposed to broader audiences.  So the purpose was somewhat different.  And I really appreciate, Tasha, your mentioning curation  for other kinds of uses.   
BERT CARELLI: So we're really talking about really different intent  more than the difference in what the feature actually just  throws out there on a site.  And I think, probably, there could  be more clarity around that from the perspective of labeling  these lists of recommendations as recommendations that  are coming from a particular perspective,  that these are what the author and publisher wants you to see,  which is what you're going to expect  to find on a journal site.   
BERT CARELLI:  These are the issues that have emerged  out of the content, maybe, perhaps,  but from a different perspective,  from the more qualitative analysis of the site.  That's kind of where I was coming  from about the question about automated summarization  and sentiment analysis.  I wonder if there are multiple opportunities  for recommendation technologies that could be applied  in these different ways.   
CLAIRE MOULTON: And I think recommendation doesn't always  mean that you agree with the finding.  It could be that you have back to back papers  argue two sides of an argument.  But you would recommend that people read both of them  to see the two sides of the argument.    
WILL SCHWEITZER: That's a really good point, great.  So this is a related question, and it  is, how do the goals of users and the goals of publishers  come into play in curation?  And do those goals conflict at times?    
CLAIRE MOULTON:  I think that when  you are launching something like this,  you're aiming it at the user.  Of course you want to, perhaps, promote the fact  that you have some strong content in an area,  or something like that.  Or you want to present some content because you want  more submissions in that area.  But you are always thinking about the user.  That's why you produce it.   
CLAIRE MOULTON: If people aren't reading it, then it hasn't succeeded.  I'm not sure that I would say that aims conflict.  I'm sure that publishers get it wrong sometimes,  and produce a collection that the readers aren't  as interested in as we'd hoped.  For example, I think that when they announce the Nobel Prize  winners, lots of publishers suddenly go, oh, yes, we've  got some papers on that!  And we've got a fantastic paper by that particular author.   
CLAIRE MOULTON: And we pool collections together,  and then not always very successful.  We feel that we need to showcase the work.  But we do at least always have the user in mind.    
TASHA MELLINS-COHEN: I think, as well, the idea of "the user"  is really interesting to me because we  have many different communities we are  trying to serve simultaneously.  So if you're a publisher, you need  to serve authors, reviewers, editors, and librarians,  and funders.  If you're a society, you also need  to think about your members, and the different types of members.  So the idea of a user, I think, generally, we're  trying to serve a specific group of users,  not necessarily everybody, all the time.   
TASHA MELLINS-COHEN:  So, yes, I'm sure there are times when a publisher is  launching a new journal, and a library is just going,  god, we don't need another journal.  Please not another journal.  That's just going to cost more money.  But the editors, the people who want to launch the journal,  are saying, there is a niche and nobody  is publishing in this area.   
TASHA MELLINS-COHEN: So, in that instance, we are curating  content as a publisher with conflicting demands  from the two different types of user.   
BERT CARELLI: And users are not necessarily  monolithic in their process.  When you think about, well, again,  coming from the perspective of TrendMD,  where we find the most usefulness, where users really  raise their hands and say, yeah, this was really useful to us,  is when they are looking for that piece of serendipity.  They're looking for that.  One of the other topics that I may have missed,  so it's, if you think of it as an early part of the research  process, maybe it's the PhD student who is fishing around,  like, where am I going to focus the next few years  of my career?   
BERT CARELLI: And it's also, to some degree, the person that  runs a lab is looking, like, what are the experiments that I  should be planning for in the next two years?  So it's more keeping the wide end of the funnel as  open as possible.  Whereas, you know you could really imagine curation  at a later stage, where you really need to be conscious of,  is someone else publishing a paper that's  close to this particular topic?   
BERT CARELLI: Are these compounds being used in another way  that I should be aware of before I publish?  So I think it plays a role in all those different stages.  So it's very much a moving target,  is what the user really is interested in.   
CLAIRE MOULTON: I think you're spot on there, actually.  I think that it's finding the balance  between wide and narrow.  And, certainly, the feedback that we get from researchers  is that whether a list is manually curated or automated,  they worry that it's not going to include the one  thing that they would have filed if they put the list together.  But at the same time, if you always do a broad list,  then they're not going to find all of the pieces quite so  interesting.   
CLAIRE MOULTON: And it is a really crucial part of the puzzle.    
WILL SCHWEITZER: This question is from our friend Richard  Wynne, which is, does curation include establishing  the specific contributions of each author thinking of credit  attribution?   
TASHA MELLINS-COHEN: It could.  It doesn't always.  In practice, in reality, I think it would be a wonderful thing  to be able to say, yes, every author on this paper  has a credit tag.  But it doesn't always happen.    
WILL SCHWEITZER: And I guess that brings up  a slightly broader point, which is,  how should we be valuing or recognizing  the act of curation?  Is that a similar level of importance, say, as authorship?  I think as the credit taxonomy is designed to recognize.  Any thoughts on that?    
BERT CARELLI: Well, it comes back to the value of libraries,  I think, at its core.  The library was always responsible for curation.  And the value of librarians themselves  and their expertise to field.  The advantage that a librarian has is of impartiality.  Their only skin in the game is to serve  the user, the researcher.   
BERT CARELLI: So I think that that's where the value ultimately should be.  That's the attention that we should always be  putting on this whole process.    
TASHA MELLINS-COHEN: There's also a question in there  about recognizing review and editorial contributions  because, at the moment, again, most of that is hidden, mostly  it's not named.   Even if you're providing ORCID feeds  with information about who's contributed to peer review,  It's not really clear.  And that is a part of the curation process.  That editorial selection and peer review  is a part of the curation process.   
TASHA MELLINS-COHEN: And we don't really have a proper taxonomy around that.  So if somebody has done, for example, a statistical review  or a methods review versus a general peer review,  we have no really clear way, within CRediT,  or any other taxonomy, to define that.  And we don't describe our editorial processes,  particularly in a standardized way.  So there is a question there about the role of taxonomy  for the process of curation, as well as  for credit for the authors.   
TASHA MELLINS-COHEN:   
WILL SCHWEITZER: And I think for our last audience question,  Claire, this one is for you.  And I'm trying to synthesize on the fly here.  But it seems like a company biologist  is making it an investment either in human time  or dollars in the summaries.  And then whether you have thought  about repackaging those summaries for audience,  and whether that would have revenue-generating potential.   
CLAIRE MOULTON: So we did decide to invest in this area.  And we don't currently have any commercial revenue-generating  ideas at all.  We've specifically set those aside,  so that we could just develop the project,  and take ideas from the preLighters themselves,  and invest in those, as well.  So we've done quite a few of those sorts of things.  I think it depends on the nature of the business.  We have tried to set ourselves as a community publisher.   
CLAIRE MOULTON: We have a number of community sites.  We have one called the Node, which  is for developmental biologists, so it's associated  with a particular journal.  And we have one which we launched just this year,  called focal plane, which is about imaging and microscopy.  And that relates to all of our journals in some way.  And we've really just invested in those.  They don't cost a huge amount of money.   
CLAIRE MOULTON: And we feel that they do provide a big benefit to the community.  And so we decided it was part of our--  We're not a society, but we are a community publisher.  And we just thought it was part of our community service  to provide those forums.   
WILL SCHWEITZER: And there's a second part  to this question, which is, do you have any comparative data,  say, comparing how these summaries are used  or how preLight is used relative to, say,  your original peer-reviewed research and in your journals?    
CLAIRE MOULTON: Well, I don't think we've really  tried to compare them as such.  Clearly, they have a very big Twitter presence.  All journals have a Twitter presence, as well.  But this is especially hot on Twitter.  And I think that they will be the full piece  will be read fewer times than one of our published articles  because I'm sure that a lot of people  read that the tweetable summary, and they always  feel like they read the article.   
CLAIRE MOULTON: A little bit like when someone spots  an article on a table of contents,  they think they've read it.  But I think that it does create an excitement and a visibility  for the preLighters and what they've done.  A lot of people tell me that they use preLights a lot.  And so you know I think that the warm feeling that people have  is one of the benefits of having produced this service.   
CLAIRE MOULTON: And one of the reasons--  Obviously, there's always a little bit of a selfish reason  for launching a product--  But we did feel that preprints were hot.  We wanted to show that our journals were preprint  friendly, we accept reprints, we do transparent peer review,  we do a lot of things around transparency and openness.  And we thought that this would, in some ways,  reflect on us as an organization,  that we'd embraced preprints and early career researchers.   
CLAIRE MOULTON: And that has definitely paid off.  So if you were looking for sort of a selfish nugget in what  is really a community initiative, that would be it.  I think it made us look cool.   
WILL SCHWEITZER: Thank you.  Well, do you have any last questions for each other?   
TASHA MELLINS-COHEN: Not for me.   
CLAIRE MOULTON: No, good discussion.   
WILL SCHWEITZER: Well, first, a very big thanks to you,  Claire, and Tasha, and Bert, for joining our panel today.  And a really big thanks to our audience  for joining us and participating.  As my colleague, Stephanie, mentioned  at the beginning of the webinar, registrants and attendees  will receive an email with a link to this recording,  and a link to a participant survey.  Your feedback will be really helpful to us  as we plan for 2021.   
WILL SCHWEITZER: So please, take a moment to complete it.  I'd really appreciate it.  So thank you all again.  And we'll see you on a webinar, or perhaps, even  at a live event sometime soon.   
BERT CARELLI: [LAUGHS] Let's hope so.   
TASHA MELLINS-COHEN:  Thanks and goodbye.   
CLAIRE MOULTON: Thank you, bye.   
BERT CARELLI: Thank you both.  Bye, Claire.  Bye, Tasha.