Name:
                                Integrity by Design: Smart Approaches to Pilots & Partners
                            
                            
                                Description:
                                Integrity by Design: Smart Approaches to Pilots & Partners
                            
                            
                                Thumbnail URL:
                                https://cadmoremediastorage.blob.core.windows.net/50e8d373-5fe0-4d31-acd8-83c6ca42ef33/thumbnails/50e8d373-5fe0-4d31-acd8-83c6ca42ef33.png
                            
                            
                                Duration:
                                T00H58M38S
                            
                            
                                Embed URL:
                                https://stream.cadmore.media/player/50e8d373-5fe0-4d31-acd8-83c6ca42ef33
                            
                            
                                Content URL:
                                https://cadmoreoriginalmedia.blob.core.windows.net/50e8d373-5fe0-4d31-acd8-83c6ca42ef33/PSWS25 April - Research Integrity.mp4?sv=2019-02-02&sr=c&sig=OdOjwZgo0BeSW%2F3LdLLqlBrTQE27iutppvfImlWA%2B4M%3D&st=2025-10-31T23%3A13%3A12Z&se=2025-11-01T01%3A18%3A12Z&sp=r
                            
                            
                                Upload Date:
                                2025-04-21T00:00:00.0000000
                            
                            
                                Transcript:
                                Language: EN. 
Segment:0 . 
 STEPHANIE LOVEGROVE 
HANSEN: Hello.   
HANSEN: Welcome.  My name is Stephanie Lovegrove Hansen.  I'm the VP of marketing at Silverchair,  and I want to thank you for joining us today.  This is the second in the 2025 Platform Strategies Webinar  Series, and I'm really looking forward to today's discussion.  But first, I'm just going to go ahead and cover some logistics.  So this webinar series features three virtual events.  And you can find the registration links  and recordings for all of those on the silverchair.com website.   
HANSEN: Also be sure and save the date for our in-person Platform  Strategies event, which will take place on September 25th  in Washington, DC, and registration for that will open  up next month.  So this year's virtual series spotlights the biggest trends  in publishing technology for 2025.  And that includes accessibility, research integrity,  which is what we'll be talking about today,  and leveraging data to drive revenue.   
HANSEN: So we designed these webinars intentionally  to be more like a discussion among the panelists.  And for that reason, we encourage you to participate.  So you can do that via the chat or via the Q&A features  throughout, and we will have some time  at the end for Q&A. This event is being recorded.  A copy of the recording will be made available freely  on our website and be emailed to you afterwards.  And finally, at the end of the event,  you'll see a survey asking you for your feedback  on the event, which just helps us to plan  more of these in the future.   
HANSEN: So thanks for offering your feedback.  And with that, I'm very excited to have this group with us  today.  We've had some amazing discussions leading up to this,  so I'm very much looking forward to today's conversation.  We have with us Julia MacDonald, Director of Journals Product  at Oxford University Press.  Tiago Barros, co-founder of Signals.  Adam Day, CEO and founder of clear skies.   
HANSEN: And Joris van Rossum, product director at STM Integrity Hub.  So before we I pass over to the panelists,  I'm just going to do a little bit of scene setting.  So the research integrity discussion, as we all know,  has become increasingly urgent in recent years.  And this is thanks in large part to the increasingly advanced  methods for fraud and manipulation  that are enabled by AI tools.  The challenges also stem from incentives  baked into the academic infrastructure, which  demand fast and frequent publication from researchers.   
HANSEN: And peer review, as we know, has long  served as the first line of defense  and serves as the hallmark of scholarly publishing.  But as the scale of the research integrity  challenges grow exponentially, we need the value of peer review  to likewise scale.  And so that's where pilots and partnerships come in,  a lot of the things that we'll be discussing today.  So in addition to the critical need for human peer reviewers,  which is not going anywhere, publishers also  need to leverage technology to assist them  in filtering and flagging and technical checks.   
HANSEN: And these kind of tools can be built in-house,  can be bought off the shelf.  But what we're really seeing more of  is these partnerships with best in class service providers.  These kind of partnerships have become  more and more commonplace, but we  wanted to convene today's panel to offer recommendations,  insights from everyone's experience  based on what they've seen in the partnerships  that they've had and the things that they've  tried over the course of the last few years.   
HANSEN: So with that, I am going to let the speakers give  a brief overview of their backgrounds,  and then we'll get into our discussion.  So we're going to start with Adam.   
ADAM DAY: Hi.  Let me just quickly share my screen.   There we go.  Yeah.  So I'm Adam from Clear Skies.  I think that the whole of research integrity  sort of comes down to peer review in a way.  Peer review is the place where science from one person  meets other people and gets looked at.   
ADAM DAY: And Clear Skies was actually founded  based on that principle that we wanted  to build tools that would support the peer review process.   One thing that we did quite early on actually  was journals that were using ScholarOne  and compared with other journals.  One thing that we found was that the alert rate  that we see in publications from those journals is lower.   
ADAM DAY: And I think one possible reason for that,  and we never quite got right down to the bottom of it,  was the existing tools already in ScholarOne,  which are already helping publishers  to deal with problematic papers.  So Clear Skies uses a blend of artificial intelligence  and network analysis to find signs of problematic behavior  in a research manuscripts and also in published papers.   
ADAM DAY: We've recently launched Oversight,  which is our application which tracks and analyzes every single  paper published in the last decade since 1st of January  2016.  Clear Skies has won multiple awards for innovation,  and we have not yet announced an integration with ScholarOne.  However, we have already had publishers sign up.  So I should probably say that there is integration  available as of right now.   
ADAM DAY: And when you use our integration, what you will see  is an alert in ScholarOne and then full details of what we  find in the Oversight platform.  So thank you.  We'll stop sharing now.  STEPHANIE LOVEGROVE 
HANSEN: All right.   
HANSEN: Thanks, Adam.  Joris, over to you.    
JORIS VAN ROSSUM: Yes.  Sorry about that.  Joris van Rossum, product director,  STM Integrity Hub, but also program director  at STM Solutions.  So STM Solutions is the operational arm of STM,  and STM is the International Association  of Scientific, Technical, and Medical Publishers.  So about 3 and 1/2 years ago, publishers realized they needed  to collaborate to address the challenges around research  integrity, and at that time, we launched the Integrity Hub.   
JORIS VAN ROSSUM: And the aim is to equip the scholarly communication  community with data, intelligence, and technology  to protect research integrity.  We do that on three levels.  Knowledge exchange between publishers on paper mills.  We also develop frameworks and editorial policies.  And last but not least, we built an enabling infrastructure,  tools development and content pipelines  to screen incoming manuscripts.   
JORIS VAN ROSSUM: Our guiding principles.  First of all, prevention instead of remediation,  which means concretely that we want  to catch issues at submission.  So prevents those papers to enter the peer review process,  let alone the published corpus.  Collaboration with all stakeholders.  That means publishers and STM members, non-STM members.  But also submission, editorial systems, sleuth, et cetera.   
JORIS VAN ROSSUM: Very important principle, we want  to ensure that critical knowledge and tools are  available to all publishers.  We've seen in the last years that especially large publishers  have built teams and technologies  to screen manuscripts.  You want to make sure that all the publishers of all shapes  and sizes have access to those tools.  Technology-driven was something that  was mentioned at the beginning but always  with the human in the loop.   
JORIS VAN ROSSUM: Technology can support, but it's always the human being  that makes a decision.  And last but not the least, of course,  fully upholding confidentiality and privacy standards.  Something about where we are.  After 3 and 1/2 years, we built two applications.  One is an on-demand screening tool,  allowing integrity specialists to upload manuscripts  and screen them, but also ambient screening application,  which means that submissions at publishers  are pushed automatically at the hub.   
JORIS VAN ROSSUM: For that, we have integrations with seven editorial systems,  including ScholarOne, which means  the content gets pushed to the hub,  but also that we provide the signals in the environment  of the editorial systems.  We have now 15 tools integrated in both applications,  and those include tools we build ourselves,  but also third party tools, including Adam Day's tool.  And we also launched a pilot program,  so that our members can test tools, and once successful,  we can roll them out to more publishers.   
JORIS VAN ROSSUM: We have now 36 publishers using the applications.  I should also mention that the hub is really  the work of a lot of volunteers, from publishers and editorial  systems as well.  At the moment we screen more than 100,000 manuscripts.  And according to our estimates, we  identify dozens of paper mill manuscripts per day already.  Thank you.    
JORIS VAN ROSSUM: STEPHANIE LOVEGROVE 
HANSEN: Great.   
HANSEN: Thanks.  Over to you, Tiago.    
TIAGO BARROS: So I'm Tiago.  I'm one of the co-founders of Signals.  I'm really excited to be here today to tell you  about Signals, what we do, and also  in the context of Silverchair and the partnership  that we have with ScholarOne.  So a very quick overview of Signals, and I  think it's interesting to go back to also what Stephanie said  at the beginning about this issue around research integrity  in the age of disinformation.   
TIAGO BARROS: And really, that is at the core of Signal's mission.  Our mission is around restoring trust in research.  And that is, I would say that even from a personal point  of view, that is very close to my heart  because as a researcher of background  and who has lots of friends and family  that are researchers as well, that's  hearing all of them talking about this issue, that they come  across papers that they no longer know  whether they can trust or not.   
TIAGO BARROS: And that seems like a really fundamental problem  that needs addressing.  And that was also then one of the motivations  to start Signals.  What we do really is about analyzing the world's  publication data and then combining  that with expert knowledge.  And specifically, we are developing this technology  that we call the Signals data graph, in which we are combining  that publication data that you can think of it as  such as the articles and manuscripts themselves  and their content, but also who the authors are,  from which institutions they come from.   
TIAGO BARROS: Also, looking at the citation network.  And even, in many cases, with partnerships  that we have with publishers, but also  including proprietary publisher data that enhances the analysis.  But we know that that may not be sufficient.  And especially, because we know that there's lots of people  out there with expertise in the fields  that can contribute to the analysis  that we can make for published content or manuscripts.   
TIAGO BARROS: That is the second part of this Signals data  graph around expert knowledge.  And that comes in different flavors but three  that I would like to highlight.  One is the expert contributions from researchers themselves.  That could be, for example, research integrity sleuths  who make invaluable contributions to highlighting  problems with papers out there, and we  can get that data to inform our analysis.   
TIAGO BARROS: But it can also be expertise from teams  within the publishers themselves,  whether that is their research integrity team, their editors,  and also benefiting from the best practices  that those publishers have to again inform our analysis.  So taking all of this together, what we want to do  is to provide the transparent and dynamic evaluation  of the research output credibility.  So that's really around linking back  to the issue of research integrity  in the age of disinformation.   
TIAGO BARROS: And so as an example, there's a very quick overview  of what a Signals evaluation of a publication looks like.  And something that is central to us  was this aspect of transparency.  So it's not about just a black box  saying that this paper has a problem  and not exactly telling you why there is a problem with it.  It really is about providing that transparent evaluation,  which you as an editor or research integrity team  member can look at at the scene of evaluation  and very quickly see what is the problem that we identified  with this publication, so that then you  can make an informed decision on what to do with the publication  itself.   
TIAGO BARROS: And so those pieces of evidence that we can find, again,  may relate to who the authors are  and that track record, especially from a research  integrity point of view.  Or also, what is the foundations of this publication  in terms of what other articles it's citing  and whether there's problems already  with those publications that may hint at this  being also a problematic paper.   
TIAGO BARROS: And of course, in the context of this webinar,  I have to refer to the amazing collaboration  that we've been doing with ScholarOne team and Silverchair.  And one of the things that was really important for us  is to go beyond this notion of a one time check for manuscripts  as they come in because we think that it really maximize  the value of what we can do when we look continuously  at the entire editorial process or actually  the entire publishing process of a manuscript.   
TIAGO BARROS: So that every time there is new information that  is available to us, that could, for example,  could be linked to the peer review process  itself as some of the other speakers  were talking about before.  So anytime there is new information,  we can incorporate that information into our analysis  and continuously update what Signals thinks  about this manuscript itself.   
TIAGO BARROS: And also just to say that this process doesn't even  stop when the manuscript is published.  Post-publication, there's continuous information  that we can continue to acquire, again,  for example, through the research integrity sleuth  contributions that will continuously  incorporate into the evaluation of the publication.  And so they're therefore providing  this very wide-ranging and comprehensive overview  of the research integrity landscape.   
TIAGO BARROS: STEPHANIE LOVEGROVE 
HANSEN: Wonderful.   
HANSEN: Thank you.  And last but not least, Julia.   
JULIA MCDONNELL: Well, I'll just get the slides up.  Thank you.  So I'm Julia.  I am director of Journals Product.  I lead the Journals Product team here at Oxford University Press  or OUP, as I'm much more likely to call it from here on out.  And I just wanted to start for a moment and say, a lot of people  ask me, well, what on Earth does that mean?  What does the Journals Product team do?  And I like to summarize it as fundamentally, we  do two main things.   
JULIA MCDONNELL: We try to solve problems that are happening  throughout the whole end to end journals publishing process,  and we look to explore opportunities.  And that involves lots of different things.  But obviously, what we're going to talk about most here today  is really the center option at the bottom  there, where we're talking about pilots and projects.  And in the context of trust and integrity,  so one of the things that my team and I are responsible for  is a dedicated trust and integrity program within OUP  that sits alongside our journal ethics  team, who are the people handling  advice on complex ethics cases that arise.   
JULIA MCDONNELL: And really, the program works very closely with that team.  And together, we're trying to really drive a step change  in terms of the processes and the level of confidence  in the content that's being published.  I think we all know that we move beyond a situation  where you can default to an assumption of trust,  and that takes a number of different forms.  So very much, as Joris said, the more  we can prevent this upstream, the better.   
JULIA MCDONNELL: We don't want to be handling problems certainly  post-publication.  But also honestly, we don't really  want to be handling them even at the point of submission.  It's better if education and best practices and standards  can catch things even earlier than that.  But obviously, lots of problems can happen.  And we need to have the capabilities to detect them,  wherever that might happen in the process.   
JULIA MCDONNELL: And that's obviously where a lot of the emphasis  is in terms of pilots and looking  at the sorts of technologies that  are available in the market.  And then I always like to say, it's not all bad news.  It is also about signaling, which  is really critical to help readers and users of content  assess for themselves why they should trust  this piece of content and why they should trust  the authors of that content.   
JULIA MCDONNELL: So really thinking about what we can do in that space  to help drive that awareness, drive that transparency,  and through that, help support informed trust  when it comes to content.  And I'm going to stop there.  STEPHANIE LOVEGROVE 
HANSEN: Great.   
HANSEN: Well, thank you.  That gives us a lot of good background  to inform the discussion that we'll kick off now.  So as we heard from everyone, many publishers  are already engaged with a lot of these pilots  and partnerships.  There's a lot already going on.  But at the same time, there's many more  that are just now getting started  in pursuing such offerings.   
HANSEN: So what would you recommend from your experience  that organizations should do in terms  of how they approach addressing research  integrity through partnerships?  So for Julia, how does the build buy partner decision play out  at OUP?  And for our partners, what have you  found are the key factors that play into partnership fit?  And I'll let you kick that off, Julia.   
JULIA MCDONNELL: Yeah.  Thank you.  So it's fairly obvious, but I think the first thing to say  is you have to be really clear what you're trying  to achieve in the first place.  And that's critical to working through whatever  your internal processes might look like to make that  build buy partner decision.  As part of that, it's really worth  spending the time to think through what  is your organization expert in?   
JULIA MCDONNELL: What are you best placed to do yourself?  Versus where can you tap into expertise  beyond what you have available?  I think when you get kind of past that point,  you've sort of thought about it and gone,  actually, we know we need to achieve something here,  but we're not the best people to do it.  We need to partner with somebody.  Obviously, you've got to do your research.   
JULIA MCDONNELL: And this is a very busy, active market  because there's a lot changing, a lot  that's evolving all the time.  And it does really need that investment of time  from whatever team is handling this within your organization,  to go out, to talk to vendors, to attend the industry events,  to really understand what is happening  across the industry as a whole, to think through, well,  which of the many different tools  in any given space on integrity would be the best fit for you?   
JULIA MCDONNELL: And that point about fit goes back again  to understanding your goals in the first place.  But it also ties into who are you as an organization?  What is your mission?  In the case of OUP, are you a not for profit?  What are you what are you thinking  about in those contexts?  And how can you identify which potential partners might  be best aligned with your strategic priorities  and your ethos?   
JULIA MCDONNELL: Obviously then, there's also the critical question  of in a busy market, who do you have the greatest confidence in?  Who has the credibility that this  is going to deliver on all the promises that are being made  and all the informational briefs.  And link to that, and this will vary a lot  depending on your organization.  There's a question in there about  what is your appetite for different levels of maturity?   
JULIA MCDONNELL: Because it's moving so quickly, there  are some brand new companies with brand new products  that are very early in their product lifecycle,  and that obviously requires a lot more time and knowledge  from your side to be able to make that work effectively,  versus increasingly, we're getting to a point  where some of the capabilities have been around for a few years  or in extensive use, and that's then  a very different organizational lift to take part.   
JULIA MCDONNELL: So yeah.  Be really clear what you're trying to do,  and that helps inform whether you should build, or buy,  or partner.  But if you are going to partner or buy,  really think about that ethos and those priorities.  STEPHANIE LOVEGROVE 
HANSEN: Yeah.   
HANSEN: That's great.  And I'm curious to hear from others, like you were saying,  Julia, that the appetite for risk  and the appetite for being the first one to try something.  What have what have you all seen from your experience?   
ADAM DAY: It's interesting that you mentioned risk  because I feel like risk is the thing that  comes up again and again in different contexts.   One of the things that, when I was starting this,  there wasn't anyone else doing any kind of paper mill detection  or anything like that.  So we had to figure out what were the right ways  to do things responsibly.  So we had things like in the publishing world.   
ADAM DAY: We've got things like code guidelines,  and we've got working groups like the ones that STM run,  which are really great for working out  what are the kind of ethical norms,  and what are the rights and wrongs?  But the other side was just looking at the legal side.  Things like considering people's interests.  Considering things like data protection, defamation,  and IP copyright.   
ADAM DAY: All of these things are different and intersect  right at this kind of service.  And so when you put these things together,  the sort of confidentiality, the balancing of interests,  and the need for purpose limitation  and being really clear about exactly what is this data for,  you get to need to know, it's really important  to say, well, where does this data have to be?  Who is it for?   
ADAM DAY: Who owns it?  And be very responsible about how that is handled.  But I think that's a really key part of risk.  It's really thinking things through at that level.  And in that way, being prepared to handle  that kind of volume of data.   
JORIS VAN ROSSUM: Yeah.  If I can add to that.  I agree.  I think about the kind of tools you would like to use,  but think first about your own organization.  Think about your challenges.  So we're talking about paper mills.  Some areas are more affected than others.  So think about the challenges you had.  Think about the specific issues you found.   
JORIS VAN ROSSUM: But also think about your organization.  Because the tools that we provide but others as well,  they give a lot of signals at submission.  So you need to make sure you have the teams.  You have the expertise to handle those.  That's what we saw in the first three years.  Certainly, a lot of signals are coming to your organization.  So think about how you want to handle that.  Do you want to internal integrity teams  to deal with those?   
JORIS VAN ROSSUM: Do you want to give those signals directly  to the editorial teams?  Really think about those aspects as well.   
TIAGO BARROS: Yeah.  And if I can add.  There's also a different perspective on risk,  which is the risk of not doing things and what happens.  And so actually, there's multiple ways  to measure the risk.  And I do think that given how much you can cancel  the threat of research integrity issues  that we see here are for many publishers,  I think there's also a risk of being  too paralyzed for whether we should do something or not.   
TIAGO BARROS: But actually, something needs to be done in many cases.  But then it becomes like more of a practical question  of how do we manage that risk of the piloting  and trying new things?  And actually, I really like the fact  that we started this discussion with links to peer review.  Because in some ways, like the Signals team, half of it  was the problems before.  And we went through a similar experience  in which, for example, at the time,  there were these trends around open peer  review and transparent peer review,  and that was also seen by many publishers as a risk,  like making that transition and something that  was desirable but risky at the same time for the organization.   
TIAGO BARROS: And so the kind of lessons that we learned from that experience  that I think we're seeing again as being valuable in looking  more into research integrity.  And those are things like looking  at the portfolio of a publisher and see what  are the opportunities there.  So it's not really the case that things  have to be on and off all the time  and across the entire portfolio of a publisher.   
TIAGO BARROS: Maybe there are different teams within the publisher  that have a different appetite for piloting.  And so maybe leveraging those opportunities to learn something  that then can be expanded across the organization.  So that's practical steps that can mitigate the risk  for the organization and to mitigate the risk of disruption  of the operation.  And by just being targeted, and again, as Julie said,  with very clear, defined goals and objectives,  so that then, with a small pilot,  then can expand over time as the organization gains  confidence in the partnership.   
TIAGO BARROS: STEPHANIE LOVEGROVE 
HANSEN: Yeah.   
HANSEN: And I completely agree with you about it  seems like there's been a tipping point recently  in terms of there's more risk in doing nothing.  There's been enough high profile instances of research integrity  issues at scale, that having things  available like a pilot eliminates a lot of the risk  while allowing you to address the underlying issue.  And so let's say we've done the pilot.  It's great.  You're ready to move forward, expand it more broadly.   
HANSEN: That's where a lot of challenges come in.  How do you scale beyond the pilot?  How does that transition present unique challenges  as you move from a very specific use case to broader adoption?  Curious to see what you've seen.  And Adam, we can start with you.   
ADAM DAY: Yeah.  I mean, one of the things that we did quite early on  was get straight up to industry scale.  So the first time that we managed  to process the whole of OpenAlex, that was April, 2023.  So two years ago now.  And so since then, we've been able to operate  at industry scale.  So we've been able to run processes  at very high rates, which means we can adapt to things very  quickly.   
ADAM DAY: So as soon as we change our methods,  we can just change our ratings on everything very quickly.   And so in terms of scaling up with publishers,  when publisher's data is coming in  and that data goes up and up and up, it's fine.  We're already processing data at that kind of scale.  So the technical side is absolutely no problem.    
ADAM DAY: The other side of it though is about how as more and more  publishers get involved, the.   diversity of what people need and what people are looking for  has changed.  And so that's been quite interesting because when this  started, it was all about let's get the decision right as  early as possible, right to the moment of submission.  We want to get the alerts there, so people are forewarned  as soon as something comes in.   
ADAM DAY: I know I'm sort of echoing what Joris, what  you said at the start.  But it's really about getting things early as possible.  So that decision get made.  And the idea of that is to reduce the need  to do investigations.  It's to reduce the need to spend a lot of time and money  on digging into all these different avenues  that an investigation could go down.   
ADAM DAY: And they can really go down rabbit holes.  They can take a long, long time.  And so really, we just wanted to avoid the whole need for that.  As time has gone on, I think when  I released the first version, the people who got involved  straight away, were like, yes, we absolutely  know what we want to do with that and straight  away got on board.  As it's developed over time and more people have come on board,  it's more like oh, well, we want to use  this for the early warnings.   
ADAM DAY: But we also want to do investigations,  and we want to figure out how to get those investigations as  short as possible and really nip them in the bud.  And so that diversity, that's been the interesting part,  is seeing how as this has grown, the different kinds of use cases  people have and I guess the differing  expectations they have of the tools as they  come aboard as well.    
ADAM DAY: STEPHANIE LOVEGROVE 
HANSEN: Anybody else   
HANSEN: want to jump in on that?   
TIAGO BARROS: Yeah.  The scaling up also brings another challenge,  which I think is quite interesting,  which is there's a lot of unknown unknowns  that you go into as you scale this pilot  and implement us into a larger scale.  We have discovered that we are experiencing already  with some of our partnerships, the reality  is that no two journals are the same,  and there's often quite specific things on the operation, or just  more generally, in the way that the journal operates  that can actually have an impact on how things can  scale from what the pilot was.   
TIAGO BARROS: And I think that's something that we should accept and just  be prepared that as things go up, often,  it's not just a case of doing the same thing over  and over again.  But it's actually becoming slightly more complex  and slightly more customized for specific use cases.  And I actually find that very thrilling,  if I have confessed because it feels like it never  gets boring in some ways.   
TIAGO BARROS: It's this excitement of looking at even sometimes, even  within the same publisher but different, slightly  different subject areas.  They have different operations, slightly different standards  on things that we look into.  And so just seeing that evolution is quite interesting.  So yeah.  I agree with Adam.  It's not really a technical limitation, which some often  think, people think that it is.   
TIAGO BARROS: But again, I always refer to, and I just  think about every time anyone clicks I  like in Facebook, that's a lot more data than what  we deal with.  And so that's not an issue.  It's more about adapting to the specific use  cases of the publisher.   
JULIA MCDONNELL: Yeah.  So I would say kind of from a publisher side of things,  I agree with what you said, but I  would come at it from a little bit of a different perspective.  I think you need to be really clear on what is your target end  state.  So you've run a pilot, and you've gone great.  This is useful for us.  Which context is it useful for you in?  What is the right scale?   
JULIA MCDONNELL: Some of these tools are going to run on every piece of submitted  content.  Others are much more specialist, and you're only  going to run them on certain articles.  Maybe the ones that are focused on image manipulation  or duplication.  You're not going to run that on every paper.  And so one of the things I think that's really important  is when you're designing your pilot in the first place,  is to make sure that pilot design process is going  to give you as much as possible, in the face  of the unknown unknowns, the answer to your questions  about, well, what should this look like for us in a business  as usual situation?   
JULIA MCDONNELL: And there's also your really practical considerations here  of no one's going to argue that we all need to take action here  to reduce the risk, but these tools  do add up to substantial additional costs, which  may or may not be offset.  You've kind of got to be in a BOU to see that.  And particularly if you're maybe a smaller organization,  there's going to be some really practical questions that you  need to ask yourself about what is the balance that you can  strike between tool used to help you mitigate risk and the cost,  whether that's monetary or resource cost of managing  these capabilities?   
JULIA MCDONNELL: Particularly because I think it's  fair to say very few publishers are piloting one thing.  Most of us are piloting lots of things in parallel  because there's so much happening here.  So your end state for any given pilot  is going to depend on the context of all the other things  that you're doing in this space.  And I think just the last thing I'll  say on this, the other thing that's  really important is when you are coming  towards the end of a pilot, make sure you properly  finish that pilot process and do a retrospective.   
JULIA MCDONNELL: Because so many of the questions you're  going to have to work through in terms of practically  for your journals and your processes,  how are you going to scale this up?  Is it as simple as turn it on those journals and you're done?  Well, probably not because even if that's  the technical side of things, there's almost certainly  an education and training element,  and you're going to have questions that you  need to be able to answer.   
JULIA MCDONNELL: And you will have done a lot of that in the initial stage  of the pilot.  But it's how do you refine that in light of everything  you've learned during the pilot to make it as smooth as possible  to use this at a larger scale?  So yeah.  Really make sure that you design your pilots well  would be my main thing to help you actually scale effectively.  STEPHANIE LOVEGROVE 
HANSEN: Yeah.   
HANSEN: And that's a great point about it's so much easier  to identify what some of the challenges  are or problem areas to key areas when it is in the pilot  before you're dealing with it at the large scale.  So that's a really good point to make  sure you've closed that loop before moving on.   Joris.  I saw you unmuted.  Do you have anything to add?   
HANSEN: OK.  All right.  Well, one thing I really love about this panel  today is that it represents a wide variety of backgrounds.  We have publisher, editorial, researcher, service provider,  product backgrounds.  So from your unique perspectives,  what would you say are the biggest roadblocks that you see,  either to research integrity broadly or to partnerships  specifically?   
HANSEN: I know we had a great time in some of the planning calls,  talking about this, and everyone has  such interesting, unique perspectives that we'd  love to hear from you all.  So Joris, can you kick us off?   
JORIS VAN ROSSUM: Yeah.  One of the challenges that we see  is that we're moving from an era of signal scarcity  to one of signal abundance.  So a few years ago, there were simply fewer tools  that publishers could use when screening manuscripts.  And now, they have access to many tools, which  is, of course, a great development,  but it again poses challenges in terms of resources and workflow.  So our focus is on fine tuning and customizing these signals  and targeting to specific people, specific moments  within specific systems in the workflow.   
JORIS VAN ROSSUM: Also, we don't believe there is or there will be a silver bullet  signal.  So it means that the value in integrity screening  lies in aggregating different signals  and bringing those to the right people in the right moment  of the workflow.  Maybe the largest challenge is GenAI  because it threatens to change the way fraud is conducted.  Our current screening system is based on copying or manipulating  existing content.   
JORIS VAN ROSSUM: So think about plagiarism, which is manipulating  something existing.  Image duplication, manipulating or changing something existing.  And the whole screening paradigm is based on that feature.  Of course, GenAI enables people to create text and images  from scratch, which is far more challenging to detect.  For this reason, alongside our efforts in the hub,  STM is going what we call beyond the manuscript.  So we believe in the short to mid-term, the tools  that we talk about today are going to be very important,  and they will remain very important.   
JORIS VAN ROSSUM: But we need also to explore new ways to validate.  For example, the identity of researchers,  that's one thing we're focusing on.  But also verifying the authenticity  of images and other research data,  which is equally important.  I think there was a research a few years ago  that 70% of all integrity breaches related to images.  So that's going to be extremely important.   
JORIS VAN ROSSUM: In terms of opportunity, maybe the last aspect.  So once again, the biggest opportunity is collaboration.  I believe we are all united in our efforts  to protect research integrity.  Institutions, editors, publishers,  companies like those in the panel today,  sleuth, submission systems, of course, and of course, also  researchers themselves, each bringing specific perspectives.  So we already demonstrated that collaboration is not  only possible, but it's also leading to very concrete,  tangible results.   
JORIS VAN ROSSUM:  STEPHANIE LOVEGROVE 
HANSEN: Wonderful.   
HANSEN: Yeah.  Anyone else want to jump in?   Biggest roadblock.   
ADAM DAY: Joris said just about everything  I could have thought of there.   
STEPHANIE LOVEGROVE HANSEN:  No other roadblocks.   
JORIS VAN ROSSUM: I'm sorry.   
JULIA MCDONNELL:  I mean, I'm going  to echo some of what Joris said.  And I completely agree that's kind of one of the biggest  challenges at the moment.  Putting AI to one side, I talked about AI a lot yesterday,  and you've already said it all, Joris.  It's both a challenge and the opportunity, right?  It is the proliferation of services.  And the fact that at the moment, what we're  seeing in that landscape is some true specialist services  that are focused on very specific aspects.   
JULIA MCDONNELL: But then increasingly, a range of services  that are doing more than one thing that partially overlap  with what then another service that's doing more than one thing  is doing.  And so it would be mad to try and run  everything that's on the market on the same manuscript  because you would just get so much noise.  But I think we're sort of reaching hopefully  the point where we're kind of at peak service proliferation,  and we can start actually getting into a bit of a phase,  less about consolidation, but more about integration.   
JULIA MCDONNELL: As the announcements today have said,  it's about having these system capabilities  be part of your normal workflow in a way  that the people who are making decisions and processing  manuscripts.  It's not something extra that they have to go find or do.  It's just there.  And where multiple services are being run, as much as possible,  the outputs from those different results  are being shown in a way that's really easy for you  to interpret and assess across the different areas that  have been looked at with those papers.   
JULIA MCDONNELL: But I think of bigger picture, it  wouldn't be a talk on trust and integrity  if somebody didn't mention the systematic incentives that  result in this content being created in the first place.  And I think because those incentives are not  likely to go away, we're going to continue  to be stuck in this high demand for problematic behavior,  whether that's individuals or whether that's  through organizations and paper mills.   
JULIA MCDONNELL: And they're really savvy.  Paper mills in particular are really technologically savvy.  And so we're in a bit of a Red Queen hypothesis situation,  and it is an arms race.  And everything that we do will then get countered,  and we will have to develop and change the processes.  And I'm sure Adam and Tiago can speak  to what they see in terms of how quickly change is happening  and how quickly we're having to adapt the tools.   
JULIA MCDONNELL: So I think that's another one of the really big challenges  that I don't really see going away,  because that pressure is going to still be there.   
ADAM DAY: Yeah.  It's kind of funny what you say about the information overload  part.  So I used to edit journals and mostly using ScholarOne.  And I went, I think in the time that I was doing that,  it was something like 10,000 articles I went through.  So it was a huge number of these.  And the one thing I remember was that there was always more work  than could be done in the time available.  And so when creating these services,  it was always about how do we minimize the amount of time  that this is going to take up or reduce the amount of time that  has already been taken up by other things?   
ADAM DAY: And that was always a key goal.  But there's something that comes from the data analysis  part of my experience, which is I think it's a little bit  counterintuitive, which is that usually, when you do data  analysis, the more data you have, the better, right?  The more examples you have.  But not the more complicated your examples are.  So if your examples have a lot of data in them,  then what's interesting is that your ability to make predictions  gets worse.   
ADAM DAY: This is called the curse of dimensionality.  And really, all it means is that like you say,  if a human being is presented with a ton of information,  it's very difficult to make a decision.  It's actually no different for AI  or for a piece of machine learning.  The more data that comes in at once, the worse the outcome is.  So when I started this, we had about 60 different ideas  for things that we could look for in an article.   
ADAM DAY: And I went through all of those.  Some of them turned out to be terrible ideas.  Some of them turned out to be things that worked.  Some of them turned out to be things that worked.  But then when you actually put them together,  you realize that they were doing the same thing,  and actually, you didn't need them.  And we ended up with one out of those 60.  Because that one was all that was needed at the start.   
ADAM DAY: That's moved on from there, but it's always moved on  with the same principle.  It's about having the minimal amount of data.  Because it's not just about the workload  on the human who's using it.  It's also actually about the quality of the analytics.  The less diverse the data is that's going in,  the better the outcome.  And so yeah.   
ADAM DAY: We're just about to release another version.  And I think the total now is just up  to about 20 different things.  But it's not 20 different things that are right in your face.  It's here are the important things.  And the other things are there if you  need to do the investigation and dig in.  But ideally, we want to get to the decision  as quickly as possible.   
ADAM DAY: So they're right up top.    
JORIS VAN ROSSUM: Yeah.  In a way, it's the commoditization of signals.  And having so many available also  puts an emphasis on how we synthesize  what those different signals are telling us.  And so on one hand, again, providing the evidence,  so that people can use it either for investigations or for action  but also importantly on education.  So we are aware of what is happening.  And so we can educate the team and also the research community  at large.   
JORIS VAN ROSSUM: But also then the challenge of synthesizing that  into something that is actionable and that people  can make the right decision with the least amount of time  possible, for sure.  The other thing is also that we simplify things  by talking about precision integrity,  but it's actually that many different flavors of that.  And to just think that all of that  can be condensed to a single data point, it's wrong, right?   
JORIS VAN ROSSUM: So this is the case of a data point,  nor a binary thing quite often.  So it's about thresholds.  And it's like for risk and what is  the acceptable norms in a specific context  of the publication.  So yeah.  It is interesting to see how these things are  coming all together.   
JORIS VAN ROSSUM: And I think also, the other thing I want to say  is there is quite often a focus on analyzing the content itself.  But again, I think that leads exactly  to what Julia was saying, that this arms race between we  are getting better at spotting those things that we see  in the content, but then bad actors  are getting better at making our life very hard.  That's definitely true.  And it's also like why, the Signals,  we always thought that it really cannot be just about  the content.   
JORIS VAN ROSSUM: It also has to be about the context of the publication  and the incentives for people doing what  they're doing, when you do it.  And so in a way, if we can highlight the problems that  are probably because of the incentives that  are behind this malpractice, if we can address that  and we can address the fact that maybe even if the content is not  telling you very much, but you see what is the intention?   
JORIS VAN ROSSUM: What is the source of that intention that we see?  That already gives a lot of information  that takes it away from the arms race  or for ever more clever content.   STEPHANIE LOVEGROVE 
HANSEN: And that   
HANSEN: brings me to another opportunity that we talked about in some  of our planning, which is if these incentives exist  that are causing a lot of these challenges,  maybe we can't work that far upstream and address that.  But what are the things we can do that do  incentivize good practices?  What are the ways that we can flag, and tag, and reward  people or research that is coming  through this in a positive way?  What are the things we can do on our platforms and our tools, et  cetera?   
HANSEN: And it's interesting, Tiago, you talk a lot about,  and I think throughout this conversation today,  we're talking about technology partnerships,  but a lot of what we've talked about is actually about people.  The technology solution is only one part of the solution,  but the people are really significant.  So how do organizations define and maintain  best practices for these specific areas?  And that's going to be different obviously  whether you're an OUP or a one journal society.   
HANSEN: So how do you define those?  Adam talked about cope a little bit.  Like what are some of the best practices?  How can we share with each other, learn from each other?  And what role does education broadly  play not only in addressing this as an industry  but in these specific partnerships and pilots?  Tiago, I will let you go first.   
TIAGO BARROS: Yeah.  That's a really good point.  And it also goes back to one of the fundamentals  I was talking about Signals in terms of this expert  knowledge being added to our evaluations, which I think it  really is at the core of this.  So we can do lots of sophisticated technology  that looks into networks and content and images  and all of that.  But there's still a huge opportunity  to leverage the expert knowledge.   
TIAGO BARROS: And again, that could be from research integrity sleuths, who  are doing an amazing job highlighting problems  that they come across.  But it can also be like the combined knowledge  within the publisher itself, that it shouldn't go to waste.  It shouldn't just be focused around a single publication  for which they spend time and apply their expertise  to flex something and then just ends there.  No.   
TIAGO BARROS: It should be leveraged.  It should be really utilized and really maximize  the collective knowledge.  So I think this is really important.  Also the educational piece, these are different ways  to look at it, I'll say.  Again, there's the education in terms  of understanding what are the incentives  and what are the bad practices that we observe  and to share that knowledge.   
TIAGO BARROS: So everyone is aware.  But perhaps it's not even about really education.  It really is more about sharing of experiences and knowledge  that I think if we can really leverage that more fully,  I think it will really come through to the benefit  of everyone.  We all thought that there's a lot of focus on what publishers  are doing around the research integrity issues.  But by no means this is just a publisher's problem  and they are the only stakeholders that  need to do something about it.   
TIAGO BARROS: So we have researchers that need to be involved.  We have institutions that need to be involved.  And so the share of knowledge and expertise  and that combined efforts really is  going to be what hopefully eventually  will solve this problem.  Again, the incentives are also there,  not necessarily because of what publishers do,  but because of the incentives that are leading to paper mills  or beyond what publisher's action.   
TIAGO BARROS: And so yeah.  And the chair of knowledge also from the people, and again,  going specifically into the topic of the pilots,  that share of knowledge from the people that are involved  in the initial pilot and then sharing the knowledge  with the rest of the team within the publisher is really crucial.  And again, so that's the extent, what the best  ways to use the different tools or what  to look for, how to give feedback to the vendors,  so that they can continuously improve the tools.   
TIAGO BARROS: I think that's really, really key.   
JORIS VAN ROSSUM:  Maybe very concretely.  So with the help, we started with sharing knowledge.  So we have various working groups.  We have monthly meetings.  So all the research integrity specialists, they come together.  They share insights and experiences  dealing with paper mills.  How do you recognize, any new trends we see.  We also have an image integrity working group,  and they have created some promotional material.   
JORIS VAN ROSSUM: So if you go to the STM website and look for the image integrity  group.  So we have also guidelines about how to detect image integrity.  So I would say yeah, that sharing of knowledge  amongst publishers, which is really,  that was one of the big problems we had, right?  Siloed nature of our way of working.  And now, we created discussion groups and working groups  to share that knowledge between publishers,  but also encourage the sharing of information publishers  with sleuth.   
JORIS VAN ROSSUM: Every year, we have a Integrity Day in London in December,  where we invite publishers, but also,  for example, representatives from institutions and sleuth.  When we started, we had about 70 people.  Now, we have 250 last year.  So you see also the community is getting larger,  but also there's more interaction and engagement  within that group.    
JULIA MCDONNELL: I think if you only have--  completely agree with everything that's  been said but taking it in a slightly different direction.  If you only have resource to do one thing,  you probably have to spend that on having some detection  capabilities.  But assuming you're big enough to do more than just that core  basic work, I think it is really about making sure  that you are thinking about this holistically  because some of the best protection  is, as Stephanie said, a lot of this is about people.   
JULIA MCDONNELL: So some of the best protection is training your editors,  is auditing your journal workflows  and making sure that they are not--  that they don't have points of vulnerability in them  that you're unaware of.  And this goes to, as we know, paper mills  are offering substantial monetary bribes.  They are incentivized to corrupt the process.  So there's a lot that you can do that isn't about technology.   
JULIA MCDONNELL: That's about people and education and processes  and policies to try and safeguard that process as much  as possible.  So very much, if you have the capability,  make sure that you are not investing all of it  just purely into technological detection  because it's a lot bigger than that.    
ADAM DAY: I think I have this memory from two or three  years ago when people were still just sort  of waking up to paper mills.  It was sort of a period of time when I think a lot of publishers  were aware, but there were some that still weren't.  But especially people outside of publishing,  you would go and talk to people about it  and say, well, there's these people.  They have these factories where they just make fake papers  and sell them to people.   
ADAM DAY: And you tell them that, and they just go, why on Earth  would they do that?  That's really weird.  Because the incentives were not obvious to most people.  And actually, they wouldn't have been obvious to me,  going back maybe like a decade or so ago.  And so I was editing journals, seeing papers all the time.  I would still have found it really weird at that time  that something like that was going on.   
ADAM DAY: And it was going on at that time.   And I think, yeah, the incentives for we're  not going to solve it, are we?  There's so much, it's so complex,  and it's happening at such a different level to where,  that we're never going to solve that.  But I do think there is something  in that education part of it.   
ADAM DAY: Or maybe it's not education.  Maybe it's just talking about stuff  that we value, stuff that we care about.  I think that really matters.  It's like coming through like uni,  I remember once getting a talk from a professor.  And he was saying and explaining to people.  Here's what's going to happen.  If you ever get caught plagiarizing your work,  you're going to get expelled immediately.   
ADAM DAY: And I was like, I had never thought of that until you just  said it.  Because obviously, you only said that because that's  what people do and the other people  sitting in the room with me.  But from my perspective, I think about it,  I think the main things that would put me  off doing something like that would be, number one, my values.  Like I actually care about doing good science,  and I care about doing things well.   
ADAM DAY: I guess the other thing is just that I would find something  like that to be immensely boring.  The idea that you're just going to get a fake paper written  and publish it.  Like wouldn't it be more fun to just actually do some research?   But yeah.  I think that's it.  It's so complex.   
ADAM DAY: We can't figure it out how to solve that problem.  But I do think there's a lot in just talking  about what we care about.  I think that really matters.   
STEPHANIE LOVEGROVE HANSEN:  I love the idea of fun  is the positive incentive that we put for doing good research.  Let's bring back fun.  Love that.  We've just got a couple minutes left.  So I think one thing maybe to close on  is we've talked about the service  providers, the publishers.  But Tiago, you touched on the fact  that there's so many stakeholders  involved in this conversation.   
STEPHANIE LOVEGROVE HANSEN: There's the authors, there's the editors,  there's the peer reviewers.  That's a lot of stakeholder groups to work across.  Any tips that you have for how to approach  working across all of these groups  as you approach these solutions?   
TIAGO BARROS:  Interesting question.  So one thing that became very obvious to us  is that the different stakeholders, although they  have the shared responsibility and the research integrity  problem, they all have very special needs  and very need for specific products that are not going  to be the same for everyone.  So for example, again, I started the webinar  by talking about my friends and family,  the researchers that come to me and say,  I come across this paper.   
TIAGO BARROS: I don't know whether I can trust it or not.  So having ways for researchers themselves to understand that  and to know that whether you should trust or not  to this paper I think becomes very important.  But that's a very specific use case, right?  That maybe the other stakeholders  don't quite have exactly in the same way.  And so I think that providing quite a bit  product-centric view on these matters.   
TIAGO BARROS: So I think that enabling creating products that  meet specific needs from different stakeholders  will, by itself, by their usage, by the sharing of the data that  is used in the products, will bring together the stakeholders,  even if they don't necessarily see that.  So again, as I said, I talked about researchers also being  involved.  And that's why for example, when we bring closer together  researchers to publishers in the sense that we leverage  those comments from a peer or something that's  research integrity has told Signals about,  and we bring that to the eyes of a publisher,  we're bringing them together, and so that we can leveraging  that shared knowledge.   
TIAGO BARROS: And the same thing goes for institutions.  We're starting to see specific use cases where  the same general principles that I  use for detecting fraud or detecting fake submissions  are hugely valuable for institutions when it comes  to hiring, when it comes to promotion,  and different use cases like this where the same principles  can be applied.  And again, that's a shared experience  that we have with publishers, and researchers  are feeding into those products that they're  delivering to institutions.   
TIAGO BARROS:  STEPHANIE LOVEGROVE 
HANSEN: Yeah.   
HANSEN: Any other quick thoughts before we wrap up?   All right.  Well, thank you so much.  This has been a great conversation.  Really appreciate all of you sharing your insights  and excited to keep these conversations going.  I feel like every conference these days  has some research integrity topics from different angles.  And I really enjoy hearing from everyone.   
HANSEN: So thank you so much to our panelists today.  Thank you to everyone who joined.  Like I said, the recording will be made available  online shortly, will be emailed to attendees.  We also have one more event in the virtual series.  I'll drop a quick link to the registration link.  It is May 15th, and it's covering how to leverage  your data across the publishing landscape.  So thank you all again so much for joining today.   
HANSEN: Really appreciate it.  And hope everyone has a great rest of the day.  Thank you.   
ADAM DAY: Thank you, everyone.