ORIGINAL ARTICLES

Who’s Guarding the Gate? The Reach of Prereviewed Emerging Science and Implications for Family Medicine Education

LaKesha N. Anderson, PhD | Christy J.W. Ledford, PhD

Fam Med. 2021;53(8):670-675.

DOI: 10.22454/FamMed.2021.246112

Return to Issue

Abstract

Background and Objectives: Online publication of prereviewed manuscripts disseminates research simultaneously to scientists, clinicians, and patients, enabling the media and public to act as scientific reviewers for studies that are not yet endorsed by the scientific and clinical community. This study describes the reach of prereview literature and frames it within the pursuit to teach evidence-based medicine.

Methods: In this deductive content analysis, the primary unit of analysis was the individual preprint manuscript submitted to the medRxiv preprint server during the emergence of the COVID-19 pandemic. The coding scheme included study design, negative or positive findings, dissemination status (whether it was withdrawn from the server or eventually published), and three levels of reach: user engagement, news media coverage, and social media engagement.

Results: Prereviewed manuscripts describe a variety of study methods. Dissemination status was significantly related to abstract views, manuscript views, news coverage, and social media exposure. Studies with negative findings had higher counts of abstract views, manuscript views, and news coverage, but no significant relationships were detected.

Conclusions: Results demonstrate that not only are scientists publishing negative findings, but that those studies reach a wide audience. Notably, eventually-withdrawn manuscripts, potentially containing incomplete or uncertain science, is reaching the public domain. Increasingly, family physicians will need to critically appraise emerging literature before it is peer reviewed, whether they encounter it in their own searches or when a patient presents information they found before an appointment.


The practice of evidence-based medicine is a critical skill taught in family medicine.1,2 We teach family medicine trainees how to ask focused clinical questions, search for evidence that addresses those questions, appraise that evidence, and produce an evidence-based answer.3 Historically, this process has focused on evidence found in peer-reviewed literature. However, in 2020, as scientists and clinicians worked together to study the emerging threat of COVID-19, prereview literature entered both the public consciousness and medical literature. Preprints became available to clinicians via traditionally credible evidentiary sources such as the National Library of Medicine’s PubMed index, which, in June 2020, launched a preprint pilot program aimed at providing early access to COVID-19 research.4,5 Much of the early discussion on COVID-19 transmission was informed by preprint articles rather than peer-reviewed research.6 Prereview literature was also cited as evidence in decision-making in health care policy and practice. In June 2020, the Food and Drug Administration released new guidelines on the use of chloroquine phosphate and hydroxychloroquine sulfate, citing eight articles that question the effectiveness of these drugs in treating COVID-19.7 Of these eight articles, five were preprints published on the publicly available medRxiv preprint server.

Peer review serves a gatekeeper function,8 ensuring both quality and integrity as academic scholarship enters the evidence base. The review process acts as a sieve, separating reliable, valid, and rigorous science from rushed, possibly less accurate, and potentially harmful science. Peer review is the greatest factor in determining research quality and trustworthiness.9 Peer-reviewed publications have easily recognized gatekeeper roles. Editors and reviewers work with authors to improve a manuscript before eventually endorsing its quality. Without a positive endorsement from these peer reviewers, an article is not viewed as suitable for publication.

A different mechanism of dissemination, the preprint, seeks to increase the visibility of research via shortcut, by bypassing the peer-review process. The term preprint is different than online ahead-of-print or online first articles, which have completed the process of peer review. It is unclear why preprint servers have framed articles this way instead of as prereview, which is a more accurate term. No inherent gatekeeper ensures these articles meet scientific standards before they are disseminated to audiences. Rather, these articles are checked for basic content and organization, and likelihood to do harm. The medRxiv website describes “a basic screening process for offensive and/or nonscientific content and for material that might pose a health risk.”10 The website also recommends that preprints not be reported in the news media as established information, and that individuals reporting preprint results do so with the caveat that these reports were not evaluated by the medical community and may contain errors. Table 1 explains terms used for preprints and ahead-of-print peer-reviewed literature along with benefits and possible consequences of each manuscript type.

The use of preprints has steadily grown from their start in the 1960s when the National Institutes of Health (NIH) mailed preprint manuscripts to biologists.11 Today, more than 60 preprint servers are available around the world,12 including many that focus on science and medicine, such as medRxiv and F1000Research. In 2017, NIH began encouraging researchers to include draft preprints in their grant applications.13,14 Proponents argue that prereviewed research articles advance science, pointing to faster dissemination of scientific findings,15 opportunities to receive rapid feedback, and benefits to younger scholars to make their research more visible.11 As an example of how quickly preprints can be disseminated, preprints studies exploring the Zika and Ebola outbreaks were posted an average of 100 days prior to their peer-reviewed journal publication.16 In both science and medicine, where clinicians comprise the primary readership, quicker access to research findings is an attractive option.15,17

Preprint dissemination may also overcome publication bias, where articles reporting negative findings are less likely to be published. Research finds that manuscripts reporting negative results are less likely to be published than those reporting positive results.18,19 Studies are most likely to be published when the results are statistically significant, confirm a priori beliefs, or are somehow shocking.20 However, publication bias may lead to unreliable evidence available for decision-making,21 which may hinder informed decision-making by clinicians and patients.22

As family medicine trainees and physicians increasingly encounter preprints in PubMed searches and clinical conversations, we need to develop a better understanding of the phenomenon. The current study describes the reach and potential impact of prereview literature, framing the analysis within the context of teaching evidence-based medicine.

Method

This study is a content analysis of preprint manuscripts available on medRxiv that were written during the emergence of the COVID-19 pandemic. Contemporary preprint articles proliferated throughout the COVID-19 pandemic. The two most popular preprint servers for these articles were bioRxiv and
medRxiv.23 Combined, these servers offered more than 14,700 articles related to the pandemic as of February 27, 2021.

This study describes the reach of preprint studies as well as two specific characteristics of these preprint studies: study design and publication bias, referred to as either positive or negative findings type.24-26, 21 Both reach and preprint characteristics are accessible for each article published on the medRxiv preprint server.

The primary unit of analysis was the individual preprint manuscript. Sampling included one quarter-year (a 13-week period), between February 15, 2020 and May 22, 2020. Analyzed manuscripts were submitted to the medRxiv preprint server during the rush to reach clinicians with evidence for COVID-19 treatments.

We limited the search to two potential COVID-19 treatments, hydroxychloroquine and remdesivir, due to their extensive news coverage in the US news media. We conducted a keyword search for manuscripts on “hydroxychloroquine” and “remdesivir” that were added to the medRxiv preprint server under the category of “infectious disease.” The sample included 27 manuscripts describing studies of hydroxychloroquine, four manuscripts describing studies of remdesivir, and eight describing both hydroxychloroquine and remdesivir for a total of 39 papers that met inclusion criteria.

We used a deductive approach to coding, in which codes were formulated based on themes emergent in the literature. We then applied these codes to each article. Using this method, we collected the following information for each article: title, first author, date manuscript was posted, and dissemination status. If a manuscript is later accepted at a peer-reviewed publication, medRxiv notates the publication of the manuscript. Additionally, if authors withdraw manuscripts from medRxiv, the server maintains the articles’ information but removes the manuscript itself. Authors may post the reason for withdrawal. In the event that either of these situations occurred, this was coded as dissemination status, either publication or withdrawal.

Reach, a concept originating in the communication discipline, is the number of people who were exposed to the information.27 In this study, we collected three different levels of reach: user engagement, news media coverage, and social media engagement. Counts are provided in the metrics section of each medRxiv article. User engagement was operationalized as the number of total abstract views and total full manuscript (PDF) views. News media coverage was operationalized as the number of news stories generated about the manuscript and number of news outlets providing coverage of the manuscript. Social media engagement was operationalized as the number of unique tweets about a manuscript, number of Twitter users engaged with a manuscript, and the estimated Twitter audience-reach of a manuscript.

We coded three additional variables specific to the science of the manuscript: research questions, study design, and study findings (ie, did the treatment work?). A copy of the coding instrument can be obtained by contacting the corresponding author (L.N.A.).

A research assistant coded all abstracts in the sample for the objective items. We coded all full manuscripts for the study findings variable. When author coders disagreed on whether the study presented negative or positive findings, the manuscript PDF was reviewed until codes were agreed upon.

Results

The 39 articles reviewed included three types of research questions: tests of drugs as increased susceptibility markers, tests of prophylactic use of drugs, and tests of drug efficacy. Article methodologies included systematic reviews (including reviews of registered ongoing trials), computer or mathematical modelling, retrospective chart reviews (including case reports, case series, and observational cohorts), and efficacy trials. Manuscripts in the data set also included trial protocols that had not been completed. Table 2 presents study methods and findings grouped by treatment studied.

To understand the potential additional reach of prereview manuscripts, we used analyses of variance to test two independent variables onto reach outcomes. First, we tested how dissemination status (published in peer-review literature, published only in medRxiv, or published then withdrawn from medRxiv) was related to the reach of the manuscripts. Dissemination status was significantly related to abstract views, F(2, 36)=3.84, P=.03; to manuscript views, F(2, 36)=3.98, P=.03; to news coverage, F(2, 36)=4.14, P=.02; and to social media exposure, F(2, 36)=7.78, P=.002. Table 3 presents reach counts by category. Manuscripts that were published in peer-reviewed journals had significantly more abstract views, manuscript views, and news coverage than those that remained only on the medRxiv preprint server. However, manuscripts that were eventually withdrawn from the medRxiv server had significantly higher social media exposure that manuscripts that were eventually published or remained on the medRxiv server alone.

In the second set of ANOVAs, we tested findings type (positive or negative findings) by the reach variables. This test excludes studies that were withdrawn since we were unable to code for study findings. The test also excludes one comparative effectiveness study that determined positive findings for remdesivir and negative findings for hydroxychloroquine. Table 3 presents reach counts by category. Studies with negative findings had higher counts of abstract views, manuscript views, and news coverage, but no significant relationships were detected.

Discussion

We historically teach evidence-based medicine as the “critical appraisal of medical research literature.”2 Many of our evidence-based medicine teaching practices rely on the understanding that medical research literature is peer-reviewed. Findings here indicate that learners are likely to encounter prereview medical research, via their own literature searches or via conversations with patients. As family medicine educators, we need to equip learners with the skills to critically appraise research that may not have been peer reviewed. While steps are being taken to teach medical students about COVID-19 misinformation,28 and how to combat misinformation, these efforts often focus heavily on responding to media inaccuracies.

Results demonstrate that, not only are scientists publishing negative findings, but also that those studies reach a wide audience. In fact, readers engaged with preprint studies with negative findings more than with preprint studies reporting positive findings. Two unique circumstances may influence this relationship. First, the credibility of speakers is important when communicating complex science.29-31 The wider conversation about COVID-19 and its potential treatments has included political figures whose positions afford them credibility, from emergency responders to political leaders. Sensational headlines related to these political figures may have increased scientists’ motivations to share not only the “right” treatment options but also the “wrong” ones, or to correct public miscommunication about treatment options. Second, the emergent and collegial nature of scientists searching for COVID-19 treatments may motivate scientists to disseminate findings to their fellow scientists to use in their own experiments. Additionally, scientists’ desire for a treatment may lead to flawed studies making it to the public domain.32

It is interesting to note the number of negative findings on the preprint server. Authors reporting negative findings often face more critical comments about their research design,33 and clinical trials reporting negative findings take longer to publish and have a lower publication rate overall.34 However, as Oldehinkle states, researchers who report negative findings “should be praised for their resistance to the temptation to give their work a positive gloss.”33 The need for rapid research during a pandemic may help explain the dissemination of negative findings on the medRxiv preprint server. It is possible that negative findings were only reported because of scientists’ desire to combat misinformation or conflicting ideas about COVID-19 treatments. This finding presents an opportunity for family medicine educators to teach trainees about the value of reading studies with both positive and negative findings and considering the implications of these studies for their practice.

An unexpected, yet important finding is the large social media reach of withdrawn manuscripts, which indicates that incomplete or uncertain science reached the public domain. We cannot determine if social media conversations about these withdrawn articles focused on a debate of the methods and/or findings or if these conversations promoted the findings. However, research shows that once information is in the public domain, it may be seen as fact by audiences who are unable, or unwilling, to critically evaluate the science. At minimum, weaker research can be overstated in the media.35 Without a traditional gatekeeper, it may be difficult to ensure the accuracy and safety of information focused on human science and/or medical treatments. This finding puts extra pressure on medical educators to teach evidence-based medicine, incorporating higher-order skills to interpret original research papers such as study design and statistical interpretation.36 Increasingly, family physicians will need to critically appraise emerging literature before it is peer reviewed, whether they encounter it in their own searches or when a patient presents information they found before an appointment.

Lastly, it is important to note that of the articles reviewed in this study, five went on to be published in peer-reviewed journals while two were retracted by the authors. It is unclear if preprint servers play a role in making accurate, rigorous science better. What is known is that two retracted studies made their way to the server, which indicates that potentially rushed, unreliable, and harmful science was made available to the public and to science and health professionals. In the traditional model of peer review, these manuscripts would have been submitted as manuscript drafts that a gatekeeper would have flagged for correction or improvement before entering the public domain. This finding is even more striking because our study discovered that manuscripts withdrawn from medRxiv had significantly greater reach than other manuscripts in our sample.

Three limitations inform interpretation of these findings. First, this study was bound in the context of COVID-19, a novel disease and the first pandemic in the social media age.37 The articles reviewed in this study represent only those found in the author-designated subject area of “infectious disease” and focused only on the potential COVID-19 treatments hydroxychloroquine and remdesivir. The decision to limit the treatments reviewed is a result of heavy news coverage of the two medications examined. This search limitation did reduce the studies identified for inclusion in this study. Second, during a pandemic, science moves quickly in a rush to determine effective treatments. This study was conducted as COVID-19 cases were rapidly increasing in the United States, and therefore presents a static picture of a dynamic information environment. Since data were collected, it is likely that abstracts were updated with newer versions, or even retracted, and it is very likely that engagement numbers have increased. Third, the intense public and media interest in COVID-19 may have generated greater interest in scientific studies like the ones analyzed in this study. This likely would not be the case for more routine scientific studies, such as ones focused on treatments for osteoarthritis. However, the findings of this study indicate what may happen when studies explore topics of high public interest.

Although limitations exist, this study indicates a number of opportunities for future research. Research shows that once an idea reaches the public domain, it may be seen as fact by audiences. Those with low science literacy may be more vulnerable to reading unreliable, inaccurate science. However, an inability to correctly understand or decipher information is not relegated to those with low science literacy. Once misinformation is presented, it can take multiple efforts to unlearn a falsehood as misinformation is often resistant to correction.38 Future research should look into the roles of literacy and memory in both interpreting and disseminating research.

More research is needed to develop and understand the theoretical implications of our findings. COVID-19 is the first pandemic to occur in the age of social media,37 and what separates it from past pandemics is the speed at which information about the disease spreads.39 Clinicians and their patients are left to seek information from preprint servers, without a gatekeeper, such as an editor, to help them determine the best science to disseminate to others.

In the rapidly evolving COVID-19 crisis, preprint servers appear to have amplified emerging science, with both positive and negative findings. These same channels enabled science to enter the public domain before the science was ready for implementing in practice. Although the peer-review process is under heavy scrutiny, as concerns about data manipulation and questionable research practices abound, peer review still plays a critical role as gatekeeper to help readers determine and decipher quality science. Thus, family medicine educators need to modify how they teach evidence-based medicine in this evolving evidentiary landscape.

Acknowledgments

The authors thank Amanda Crawford for her critical role in data collection.

References

  1. Sackett DL, Rosenberg WMC, Gray JAM, Haynes RB, Richardson WS. Evidence-based medicine: what it is and what it isn’t. BMJ. 1996;312(7023):71-72. doi:10.1136/bmj.312.7023.71
  2. Epling JW, Heidelbaugh JJ, Woolever D, et al. Examining an evidence-based medicine culture in residency education. Fam Med. 2018;50(10):751-755. doi:10.22454/FamMed.2018.576501
  3. Epling J, Smucny J, Patil A, Tudiver F. Teaching evidence-based medicine skills through a residency-developed guideline. Fam Med. 2002;34(9):646-648.
  4. “medRxiv”[jour]. PubMed Search [Internet]. Bethesda, MD: National Center for Biotechnology Information, US National Library of Medicine. https://pubmed.ncbi.nlm.nih.gov/?term=%22medRxiv%22%5Bjour%5D&sort=pubdate&sort_order=asc&size=50. Accessed June 10, 2021.
  5. NLM announces NIH preprint pilot to provide early access to COVID-19 research. NLM Technical Bulletin. Published June 9, 2020. Accessed June 10, 2021. https://www.nlm.nih.gov/pubs/techbull/mj20/mj20_preprint_pilot.html.
  6. Majumder MS, Mandl KD. Early in the epidemic: impact of preprints on global discourse about COVID-19 transmissibility. Lancet Glob Health. 2020;8(5):e627-e630. doi:10.1016/S2214-109X(20)30113-3
  7. United States Food and Drug Administration. Memorandum explaining basis for revocation of emergency use authorization for emergency use of chloroquine phosphate and hydroxychloroquine sulfate. Silver Spring, MD: FDA; June 2020. https://www.fda.gov/media/138945/download. Accessed June 10, 2021.
  8. Crane D. The gatekeepers of science: some factors affecting the selection of articles for scientific journals. Am Sociol. 1967;2(4):195-201.
  9. Tenopir C, Levine K, Allard S, et al. Trustworthiness and authority of scholarly information in a digital age: results of an international questionnaire. J Assoc Inf Sci Technol. 2016;67(10):2344-2361. doi:10.1002/asi.23598
  10. About medRχiv. medRχiv: The preprint server for health sciences. https://www.medrxiv.org/content/about-medrxiv. Accessed June 10, 2021.
  11. Kaiser J. The preprint dilemma. Science. 2017;357(6358):1344-1349. doi:10.1126/science.357.6358.1344
  12. Johnson R, Chiarelli A. The second wave of preprint servers: how can publishers keep afloat? The Scholarly Kitchen. Published October 2019. Accessed June 10, 2021. https://scholarlykitchen.sspnet.org/2019/10/16/the-second-wave-of-preprint-servers-how-can-publishers-keep-afloat/.
  13. Kaiser J. NIH enables investigators to include draft preprints in grant proposals. Science. Published March 2017. Accessed June 11, 2021. doi: 10.1126/science.aal0961.
  14. Tennant J, Bauin S, James S, Kant J. The evolving preprint landscape: introductory report for the Knowledge Exchange working group on preprint. MetaArXiv Preprints. Published May 17, 2018. Accessed June 10, 2021. doi:10.31222/osf.io/796tu.
  15. Björk BC, Welling P, Laakso M, Majlender P, Hedlund T, Gudnason G. Open access to the scientific journal literature: situation 2009. PLoS One. 2010;5(6):e11273. doi:10.1371/journal.pone.0011273
  16. Johansson MA, Reich NG, Meyers LA, Lipsitch M. Preprints: an underutilized mechanism to accelerate outbreak science. PLoS Med. 2018;15(4):e1002549. doi:10.1371/journal.pmed.1002549
  17. Coloma J, Harris E. Open-access science: a necessity for global public health. PLoS Pathog. 2005;1(2):e21. doi:10.1371/journal.ppat.0010021
  18. Stern JM, Simes RJ. Publication bias: evidence of delayed publication in a cohort study of clinical research projects. BMJ. 1997;315(7109):640-645. doi:10.1136/bmj.315.7109.640
  19. Hoy MB. Rise of the Rxivs: how preprint servers are changing the publishing process. Med Ref Serv Q. 2020;39(1):84-89. doi:10.1080/02763869.2020.1704597
  20. Andrews I, Kasy M. Identification of and correction for publication bias. Am Econ Rev. 2019;109(8):2766-2794. doi:10.1257/aer.20180310
  21. Dwan K, Gamble C, Williamson PR, Kirkham JJ; Reporting Bias Group. Systematic review of the empirical evidence of study publication bias and outcome reporting bias - an updated review. PLoS One. 2013;8(7):e66844. doi:10.1371/journal.pone.0066844
  22. Meerpohl JJ, Schell LK, Bassler D, et al; OPEN project consortium. Evidence-informed recommendations to reduce dissemination bias in clinical research: conclusions from the OPEN (Overcome failure to Publish nEgative fiNdings) project based on an international consensus meeting. BMJ Open. 2015;5(5):e006666. doi:10.1136/bmjopen-2014-006666
  23. Kwon D. How swamped preprint servers are blocking bad coronavirus research. Nature. 2020;581(7807):130-131. doi:10.1038/d41586-020-01394-6
  24. Tenopir C, Allard S, Bates B, Levine KJ, King DW, Birch B, et al. Research publication characteristics and their relative values: A. 2010. School of Information Sciences – Faculty Publications and Other Works. Available at http://trace.tennessee.edu/utk_infosciepubs/20.
  25. Klitzman R, Chin LJ, Rifai-Bishjawish H, Kleinert K, Leu CS. Disclosures of funding sources and conflicts of interest in published HIV/AIDS research conducted in developing countries. J Med Ethics. 2010;36(8):505-510. doi:10.1136/jme.2010.035394
  26. Bekelman JE, Li Y, Gross CP. Scope and impact of financial conflicts of interest in biomedical research: a systematic review. JAMA. 2003;289(4):454-465. doi:10.1001/jama.289.4.454
  27. Green LW, Kreuter MW. Health program planning: An educational and ecological approach. 4th ed. New York: McGraw Hill; 2005.
  28. Quadri NS, Thielen BK, Erayil SE, Gulleen EA, Krohn K. Deploying medical students to combat misinformation during the COVID-19 pandemic. Acad Pediatr. 2020;20(6):762-763. doi:10.1016/j.acap.2020.05.024
  29. Kreps GL, Alibek K, Neuhauser L, Rowan KE, Sparks L. Emergency/risk communication to promote public health and respond to biological threats. In: Haider M, ed. Global public health communication: challenges, perspectives, and strategies. Sudbury, MA: Jones and Bartlett Publishers; 2005:349-362.
  30. Ledford CJW, Anderson LN. Communication strategies for family physicians practicing throughout emerging public health crises. Fam Med. 2020;52(5):48-50. doi:10.22454/FamMed.2020.960734
  31. Rowan KE, Sparks L, Pecchioni L, Villagran MM. The CAUSE model: a research-supported aid for physicians communicating with patients about cancer risk. Health Commun. 2003;15(2):235-248. doi:10.1207/S15327027HC1502_11
  32. Zagury-Orly I, Schwartzstein RM. Covid-19 - A Reminder to Reason. N Engl J Med. 2020 Jul 16;383(3):e12. doi:10.1056/NEJMp2009405. Epub 2020 Apr 28.
  33. Oldehinkel AJT. Editorial: sweet nothings - the value of negative findings for scientific progress. J Child Psychol Psychiatry. 2018;59(8):829-830. doi:10.1111/jcpp.12952
  34. Suñé P, Suñé JM, Montoro JB. Positive outcomes influence the rate and time to publication, but not the impact factor of publications of clinical trial results. PLoS One. 2013;8(1):e54583. doi:10.1371/journal.pone.0054583
  35. Sheldon T. Preprints could promote confusion and distortion. Nature. 2018;559(7715):445. doi:10.1038/d41586-018-05789-4
  36. Rao G. Physician numeracy: essential skills for practicing evidence-based medicine. Fam Med. 2008;40(5):354-358.
  37. Viswanath K; Harvard T.H. Chan School of Public Health. Why is there so much misinformation circulating about COVID-19? [Video]. June 16, 2020. Accessed June 10, 2021. https://youtu.be/rdz_ImgfkqU.
  38. Lewandowsky S, Ecker UK, Seifert CM, Schwarz N, Cook J. Misinformation and its correction: continued influence and successful debiasing. Psychol Sci Public Interest. 2012;13(3):106-131. doi:10.1177/1529100612451018
  39. Gitlin JM. The preprint problem: unvetted science is fueling COVID-19 misinformation. ARSTechnica. Published May 6, 2020. Accessed June 10, 2021. https://arstechnica.com/science/2020/05/a-lot-of-covid-19-papers-havent-been-peer-reviewed-reader-beware/

Lead Author

LaKesha N. Anderson, PhD

Affiliations: National Communication Association, Washington, DC, and the Johns Hopkins University Advanced Academic Program, Washington, DC

Co-Authors

Christy J.W. Ledford, PhD - Department of Family Medicine, Medical College of Georgia, Augusta University, Augusta, GA

Corresponding Author

LaKesha N. Anderson, PhD

Correspondence: Johns Hopkins University Advanced Academic Program, 1765 N Street, NW, Washington, DC 20036. 202-534-1003.

Email: lander73@jh.edu

Fetching other articles...

Loading the comment form...

Submitting your comment...

There are no comments for this article.

Downloads & Info

Share

Related Content