key: cord-0149461-wak4ow9v authors: Rando, Halie M.; Boca, Simina M.; McGowan, Lucy D'Agostino; Himmelstein, Daniel S.; Robson, Michael P.; Rubinetti, Vincent; Velazquez, Ryan; Consortium, COVID-19 Review; Greene, Casey S.; Gitter, Anthony title: An Open-Publishing Response to the COVID-19 Infodemic date: 2021-09-17 journal: nan DOI: nan sha: fede7d2ccdca0c30b0c192e2e441d9dfce394429 doc_id: 149461 cord_uid: wak4ow9v The COVID-19 pandemic catalyzed the rapid dissemination of papers and preprints investigating the disease and its associated virus, SARS-CoV-2. The multifaceted nature of COVID-19 demands a multidisciplinary approach, but the urgency of the crisis combined with the need for social distancing measures present unique challenges to collaborative science. We applied a massive online open publishing approach to this problem using Manubot. Through GitHub, collaborators summarized and critiqued COVID-19 literature, creating a review manuscript. Manubot automatically compiled citation information for referenced preprints, journal publications, websites, and clinical trials. Continuous integration workflows retrieved up-to-date data from online sources nightly, regenerating some of the manuscript's figures and statistics. Manubot rendered the manuscript into PDF, HTML, LaTeX, and DOCX outputs, immediately updating the version available online upon the integration of new content. Through this effort, we organized over 50 scientists from a range of backgrounds who evaluated over 1,500 sources and developed seven literature reviews. While many efforts from the computational community have focused on mining COVID-19 literature, our project illustrates the power of open publishing to organize both technical and non-technical scientists to aggregate and disseminate information in response to an evolving crisis. Coronavirus Disease 2019 (COVID-19) caused a worldwide public health crisis that has reshaped many aspects of society. The scientific community has, in turn, devoted significant attention and resources towards COVID-19 and the associated virus, SARS-CoV-2, resulting in the release of data and publications at a rate and scale never previously seen for a single topic. Over 20,000 articles about COVID-19 were released in the first four months of the pandemic [1] , causing an "infodemic" [1, 2] . The COVID-19 Open Research Dataset (CORD-19) [3] , which was developed in part with the goal of training machine learning algorithms on COVID-19-related text, illustrates the growth of related scholarly literature ( Figure 1 ). This resource was developed by querying several sources for terms related to SARS-CoV-2 and COVID-19, as well as the coronaviruses SARS-CoV-1 and MERS-CoV and their associated diseases [3] . CORD-19 contained 768,929 manuscripts as of September 6, 2021. Additional curation by CoronaCentral [4] has produced, at present, a set of over 180,000 publications particularly relevant to COVID-19 and closely related viruses. Despite many ad-vances in understanding the virus and the disease, there are also downsides to the availability of so much information. "Excessive publication" has been recognized as a concern for over forty years [5] and has been discussed with respect to the COVID-19 literature [6] . Any effort to synthesize, summarize, and contextualize COVID-19 research will face a vast corpus of potentially relevant material. Figure 1 : Growth of the CORD-19 dataset. The number of articles has proliferated, with both traditional and preprint manuscripts in the corpus. The first release (March 16, 2020) contained 28,000 documents [3] . As of September 6, 2021, this had increased to 768,929 articles. Of these, 30,726 are preprints from arXiv, medRxiv, and bioRxiv. Information was released rapidly by both traditional publishers and preprint servers, and many papers faced subsequent scrutiny. The number of COVID-19 papers retracted may be higher, and potentially much higher, than is typical, although a thorough investigation of this question requires more time to elapse [7, 8] . Many preprints and papers are also associated with corrections or expressions of concern 1 [8] . Preprints are released prior to peer review, but some traditional publishing venues have fast-tracked COVID-19 papers through peer review, leading to questions about whether they are held to typical standards [9] . Therefore, evaluating the COVID-19 literature requires not only digesting available information but also monitoring subsequent changes. Because of the fast-moving nature of the topic, many efforts to summarize and synthesize the COVID-19 literature have been undertaken. These efforts include newsletters 2 [10] , web portals 3 [11] or the now-defunct http://covidpreprints.com 4 , comments on preprint servers 5 [12] , and even a journal 6 . However, the explosive rate of publication presents challenges for such efforts, many of which are no longer active. Similarly, many literature reviews have been written on the available COVID-19 literature [13, 14, 15, 16, 17] , but static reviews quickly become outdated as new research is released or existing research is retracted or superseded. One example is a review of topics in COVID-19 research including vaccine development [17] . This review was published on July 10, 2020, four days before Moderna released the surprisingly promising results of their phase 1 trial [18] that changed expectations surrounding vaccines. Therefore, the COVID-19 publishing climate presented a challenge where curation of the literature by a diverse group of experts in a format that could respond quickly to high-volume, high-velocity information was desirable. We therefore sought to develop a platform for scientific discussion and collaboration around COVID-19 by adapting open publishing infrastructure to accommodate the scale of COVID-19 publishing. Recent advances in open publishing have created an infrastructure that facilitates distributed, version-controlled collaboration on manuscripts [19] . Manubot [19] is a collaborative framework developed to adapt open-source software development techniques and version control for manuscript writing. With Manubot, manuscripts are managed and maintained using GitHub, a popular, online version control interface. We selected Manubot because it offers several advantages over comparable collaborative writing platforms such as Authorea, Overleaf, Google Docs, Word Online, or wikis [19] . Citation-by-identifier ensures consistent reference metadata standards that would be difficult to maintain manually in a manuscript with dozens of authors and over 1,500 citations. Manubot's pull requestbased contribution model balances the goals of making the project open to everyone and maintaining scientific accuracy. All contributions are reviewed, discussed, and formally approved on GitHub before text updates appear in the public-facing manuscript 7 . Continuous integration (CI) seamlessly combines author-produced text and figures with automatically generated and updated statistics and figures derived from external data sources and the manuscript's own content. In addition, the authors who initially launched this project included Manubot developers who had prior successes using Manubot for massively open and traditional manuscript, such as a large-scale collaborative efforts such as a review of developments in deep learning [20] and a re-evaluation of the role of authorship in modern collaborations [21] . Collaboration via massively open online papers has been identified as a strategy for promoting inclusion and interdisciplinary thought [22] . However, the Manubot workflow can be intimidating to contributors who are not well-versed in git [22] . The synthesis and discussion of the emerging literature by biomedical scientists and clinicians is imperative to a robust interpretation of COVID-19 research. Such efforts in biology often rely on What You See Is What You Get tools such as Google Docs, despite the significant limitations of these platforms in the face of excessive publication. We recognized that the problem of synthesizing the COVID-19 literature lent itself well to the Manubot platform, but that the potential technical expertise required to work with Manubot presented a barrier to domain experts. Here, we describe the adaptation of Manubot to facilitate collaboration in the extreme case of the COVID-19 infodemic, with the objective of developing a centralized platform for summarizing and synthesizing a massive amount of preprints, news stories, journal publications, and data. Unlike prior collaborations built on Manubot, most contributors to the COVID-19 collaborative literature review came from biology or medicine. The members of the COVID-19 Review Consortium consolidated information about the virus in the context of related viruses and to synthesize rapidly emerging literature. Manubot provided the infrastructure to manage contributions from the community and create a living, scholarly document integrating data from multiple sources. Its back-end allowed biomedical scientists to sort and distill informative content out of the overwhelming flood of information [23] in order to provide a resource that would be useful to the broader scientific community. This case study demonstrates the value of open collaborative writing tools such as Manubot to emerging challenges. Because it is open source software, we were able to adapt and customize Manubot to flexibly meet the needs of COVID-19 review. Recording the evolution of information over time and assembling a resource that auto-updated in response to the evolving crisis revealed the particular value that Manubot holds for managing rapid changes in scientific thought. First, it was necessary to establish Manubot as a platform accessible to researchers with limited experience working with version control, given that this is not typically emphasized in biology and medicine [24, 25, 26] . Contributors were recruited primarily by word of mouth and on Twitter, and we also collaborated with existing efforts to train early-career researchers. We invited potential collaborators to contribute a short introduction on a GitHub issue in order to collect information about participants and provide an introduction to working with GitHub issues. Interested participants were encouraged to contribute in several ways. One option was to catalog articles of interest as issues. We developed a standardized set of questions for contributors to consider when evaluating an article following a framework often used for assessing medical literature. This approach emphasizes examining the methods used, assignment (whether the study was observational or randomized), assessment, results, interpretation, and how well the study extrapolates [27] . Contributors were also invited to contribute or edit text using GitHub's pull request system. These contributions were not strictly defined and could range from minor corrections to punctuation and grammar to large-scale additions of text. Finally, a small number of contributors (the authors of this paper) contributed technical expertise, either through the development of standardized approaches to the evaluation of papers based on the MAARIE Framework [28] , the writing of code to generate manuscript figures, or the addition of features to Manubot. All of these additions were also submitted as pull requests, either to the COVID-19 review repository or to an external repository, as appropriate. Each pull request was reviewed and approved by at least one other contributor before being merged into the main branch. We tagged potential reviewers based on the introductions they had contributed in order to encourage participation. Authorship was determined based on the Contributor Roles Taxonomy 8 . Due to the permeability of ideas among different sections, contributors to a specific manuscript were recognized with masthead authorship, while all contributors to the project were recognized with consortium authorship on all papers. Emphasizing the use of issues and pull requests was designed to encourage authors with and without git experience to discuss papers and provide feedback (both formal and informal) on proposed text additions or changes. We also used the Gitter chat platform 9 to promote informal questions and sharing of information among collaborators. Applying Manubot's existing capabilities allowed us to confront several challenges common in large-scale collaborations, such as maintaining a record of contributions that allowed us to allocate credit appropriately or to contact the original author if questions arose. Additionally, an up-to-date version of the content was available at all times online in HTML 10 or PDF format 11 . This approach also allowed us to minimize the demand on authors to curate and sync bibliographic resources. Manubot provides the functionality to create a bibliography using digital object identifiers (DOIs), website URLs, or other identifiers such as PubMed identifiers and arXiv IDs. The author can insert a citation in-line using a format such as [@doi:10.1371/journal.pcbi.1007128]. Manubot then obtains reference metadata, exports the citations as Citation Style Language JSON Data Items, and renders the bibliographic information needed to generate the references section [19] . This approach allows multiple authors to work on a piece of text without needing to make manual adjustments to the reference lists. Due to the needs of this project, several new features were implemented in Manubot. Because of the ever-evolving nature of the COVID-19 crisis, figures and statistics in the text quickly became outdated. To address this concern, Manubot and GitHub's CI features were used to create figures that integrated online data sources and to dynamically update information, such as the current number of active COVID-19 clinical trials [29] , within the text of the manuscripts ( Figure 2 ). GitHub Actions runs a nightly workflow to update these external data and regenerate the statistics and figures for the manuscript. The workflow uses the GitHub API to detect and save the latest commit of the external data sources that are GitHub repositories 12 . It then downloads versioned data from that snapshot of the external repositories and runs bash and Python scripts to calculate the desired statistics and produce the summary figures using Matplotlib [30] . The statistics are stored in JSON files that are accessed by Manubot to populate the values of placeholder template variables dynamically every time the manuscript is built. For instance, the template variable {{ebm_trials_results}} in the manuscript is replaced by the actual number of clinical trials with results, 98. The template variables also include versioned URLs to the dynamically updated figures. The JSON files and figures are stored in the external-resources branch of the GitHub repository, providing versioned storage. The GitHub Actions workflow automatically adds and commits the new JSON files and figures to the external-resources branch every time it runs, and Manubot uses the latest version of these resources when it builds the manuscript. The GitHub Actions workflow file is available online 13 , as are the scripts 14 . The Python package versions are also available 15 . Another issue identified was the need for standardized citation to clinical trials. Other researchers identified the same need 16 . Trials that are registered with clinicaltrials.gov receive a unique clinical trial identifier, or "NCT ID." Because clinical trials are registered long before results are published, referencing clinical trial identifiers was a priority. Manubot uses the Zotero translation server 17 to extract citation metadata for some types of citations. However, Zotero did not support clinical trial identifiers and could not extract relevant metadata from their URLs. In order to pull clinical trial metadata associated into Manubot, we added Zotero support for these identifiers. To achieve this, we query clinicaltrials.gov to retrieve XML metadata associated with each identifier using JavaScript 18 . This extension enables citing a trial as @clinicaltrials:NCT04280705 instead of the URL. Then, when Manubot requests clinical trial metadata from the Zotero translation server, the response includes the trial sponsors, responsible investigators, title, and summary. Manubot now supports directly citing hundreds of registered Compact Uniform Resource Identifiers 19 , beyond just the clinicaltrials identifier. Because of the large number of citations used in this manuscript and the fast-moving nature of COVID-19 research, keeping track of retractions, corrections, and notices of concern also became a challenge. We implemented a new Manubot plugin to support "smart citations" in the HTML build of manuscripts. The plugin uses the scite [31] service to display a badge below any citation with a DOI. The badge contains a set of icons and numbers that indicate how many times that source has been mentioned, supported, or disputed and whether there have been any important editorial notices. We were thus able to identify references that needed to be reevaluated by an expert. This addition was invaluable given the nature of the project, where we were disseminating rapidly evolving information of great consequence from over 1,500 different sources. The badges also allow readers to ascertain a rough approximation of the reliability of cited sources at a glance. Because most collaborators were writing and editing text through the GitHub website rather than in a local text editor, we also needed to add spell-checking functionalities to Manubot. We integrated an existing Pandoc 20 spell-check extension with AppVeyor CI to automatically post spelling errors as comments in a GitHub pull request. The comment reported both unique misspelled tokens and all locations where the token was detected. Project maintainers managed a custom dictionary to al- low over 1,500 scientific and technical terms that were not common English words. Spell-checking also helped standardize the writing style across dozens of authors by detecting features such as British versus American English spellings. The actual spell-checking was implemented using GNU Aspell 21 and the Pandoc spellcheck filter 22 . The filter enables checking only the manuscript text, ignoring URLs and formatting. Manubot can render a manuscript in several formats that serve different purposes. Prior to this project, Manubot could use Pandoc to convert the markdownformatted manuscript to HTML, PDF, and DOCX formats. We expanded this functionality to export individual sections of the manuscript as separate DOCX files while still rendering the complete manuscript in HTML and PDF formats. This development was necessary because the manuscript grew so large that it needed to be split into seven separate papers for journal submission while still maintaining shared GitHub discussion across topics. When exporting an individual section, Manubot customizes the manuscript title, authors, and author con-21 http://aspell.net 22 https://github.com/pandoc/lua-filters/tree/master/spellcheck tributions to pertain to that specific section. In addition, we expanded the export formats to include partial La-TeX support via Pandoc. Pandoc converts the markdown content for an individual section to TeX and the Citation Style Language JSON, which contains reference metadata generated by Manubot, to BibTeX. We customized a LaTeX template and reformatted the Manubot metadata, such as authors and their affiliations, for the LaTeX template. The exported TeX file requires manual refinement but contains all manuscript content and most of the formatting. Because LaTeX is required for manuscript submission in many fields, automating most of the process of converting markdown to a submission-friendly format expands Manubot's potential user base. Manubot users can write in the simple markdown format, render the manuscript in continuously-updated PDF or interactive HTML formats, and export the manuscript in DOCX or TeX and BibTeX for submission to traditional publishers, taking full advantage of Pandoc's powerful document conversion capabilities and Manubot's automation. Coverage by Nature Toolbox [32] and an associated tweet 23 about the project on April 1, 2020 attracted the interest of the scientific community ( Figure 3 ). Because GitHub issues are similar to other common web commenting systems, authors learned these tools quickly. The Gitter chat also presented a low barrier to entry. The manuscript continued to grow throughout the first year and a half of the project in both word count and the number of references ( Figure 3 ). Though only a fraction of potential contributors contributed to the text included in the manuscripts (Figure 3 ), many contributors remained engaged over the long term ( Figure 4) . Additionally, new contributors continued to join even into the second year of the project. In order to make the project more accessible, we developed resources explaining how to use GitHub's web interface to develop and edit text for Manubot assuming no prior experience with version control. These tutorials explained how to open an issue, open a pull request, and review a pull request 24 . Additionally, the framework for evaluating literature was converted into issue templates to simplify the review of new articles. Articles were classified as diagnostic, therapeutic, or other, with an associated template developed to guide the review of papers and preprints in each category. A total of 285 new paper issues had been opened as of September 13, 2021. The manuscripts produced by the consortium (excluding this one) will be submitted to mSystems as part of a special issue that provides support for continuous updates as more information becomes available. One has been published and two are available as preprints. This approach allows for a version of record to be maintained alongside the most recent version, which is always available through GitHub. These manuscripts cover a wide range of topics including the fundamental biology of SARS-CoV-2 (pathogenesis [33] and evolution), biomedical advances in responding to the virus and COVID-19 (pharmaceuticals [29] , nutraceuticals [34] , vaccines, and diagnostic technologies), and biological and social factors influencing disease transmission and outcomes. To date, 50 authors are associated with the consortium (Figure 3 ). More formal recruitment efforts to integrate with existing projects providing support for undergraduate students during COVID-19 were also successful. We incorporated summaries written by the students, post-docs, and faculty of the Immunology Institute at the Mount Sinai School of Medicine 25 [12] . Additionally, two of the con-sortium authors were undergraduate students recruited through the American Physician Scientist Association's Virtual Summer Research Program. Thus, the consortium was successful in providing a venue for researchers across all career stages to continue investigating and publishing at a time when many biomedical researchers were unable to access their laboratory facilities. We integrated data into the manuscripts from several sources (Figure 2 ). Worldwide cases and deaths were tracked by the COVID-19 Data Repository by the Center for Systems Science and Engineering at Johns Hopkins University 26 . The clinical trials statistics and figure were generated based on data from the University of Oxford Evidence-Based Medicine Data Lab's COVID-19 Trial-sTracker [35] . Information about vaccine distribution was extracted from Our World In Data 27 [36] . Figure 1 integrates data from the CORD-19 dataset [3] . Manubot's bibliographic management capabilities were critical because the amount of relevant literature published far outstripped what we had anticipated at the beginning of the project. As of September 10, 2021, there were 1,676 references ( Figure 3 ). The scite plugin provided a way to visually inspect the reference list to identify possible references of concern. This and the other new features required for the COVID-19 project are now included in Manubot's rootstock, which is the template GitHub repository for creating a new manuscript. Using CI, Manubot now checks that the manuscript was built correctly, runs spell-checking, and cross-references the manuscripts cited in this review. In addition, Manubot now supports citing clinical trial identifiers such as clinicaltrials:NCT04292899 [37] . The current project was based in the GitHub repository greenelab/covid19-review using Manubot [19] to continuously generate the manuscript. The Manubot framework facilitated a massive collaborative review on an urgent topic. We demonstrated the utility of Manubot to a project where many contributors lacked expertise or even experience working with version control. This effort has produced not only seven literature reviews on topics relevant to the COVID-19 pandemic, but has also generated cyberinfrastructure for training novice users in GitHub. We also extended the functionalities of Manubot to provide more of the benefits of What You See Is What You Get platforms such as Google Docs (Table 1) . Open publishing thus allowed us to harness the domain expertise of a large group of non-technical users to respond to the flood of COVID-19 publications. Several existing and new features in Manubot aid in responding to the challenges posed by the infodemic. Manuscripts are written in markdown and can be rendered in several formats providing different advantages to users. For example, beyond building just a PDF, Manubot also renders the manuscript in HTML, DOCX, and now, LaTeX (in a more limited capacity). The interactive HTML manuscript format offers several advantages over a static PDF to harmonize available resources and address specific problems related to COVID-19. The integration of scite into the HTML build makes references more manageable by visually indicating whether their results are contested or whether they have been corrected or retracted. Cross-referencing different pieces of the manuscript, such as cited preprints with reviews stored in an appendix, is another interactive option presented by HTML. The DOCX format was preferred by most non-technical users for reviewing the final version of the manuscript and was useful for creating submissions to a biological journal. Additionally, because of the heavy emphasis on Word processing in biology, Manubot's ability to generate DOCX outputs was expanded to allow users to generate DOCX files containing only a section of the manuscript. In our case, where the full project is nearly 150,000 words, this allows individual pieces to be shared more easily. Finally, the preliminary addition of LaTeX output is useful for researchers from computational fields who submit papers in TeX format and removes the step of reformatting markdown prior to submission. notices of concern Outputs Improved support for Pandoc's LaTeX output Outputs Build complete manuscript alongside individual sections as standalone documents The COVID-19 Review Consortium provided a platform for researchers to engage in scientific investigation early in the pandemic when many biological scientists were unable to access their research spaces. In turn, by seeking to adapt Manubot to allow for broader participation, we made a number of improvements that are expected to increase its appeal to researchers from all backgrounds. Manubot provided a way for contributors from a variety of backgrounds, including early-career researchers, to join a massive collaborative project while demonstrating their individual contributions to the larger work and gaining experience with version control. The licensing and infrastructure also provide the basis for individuals to adapt from this project to create their own snapshots of the COVID-19 literature that derive from, but are not wholly identical to, the primary versions of these reviews. This project suggests that massive online open publishing efforts can indeed advance scholarship through inclusion [22] , including during the extreme challenges presented by the COVID-19 pandemic. Some challenges did arise in efforts to include an academically diverse set of authors. The barriers to entry posed by git and GitHub likely still reduced participation from individuals who might have otherwise been interested. Using pull requests as a tool for writing text is also unfamiliar to many or most scientists, and the review process can be slow, which might cause interested contributors to lose interest. Additionally, the pull request model may limit people from providing general feedback on the manuscript or a section of the manuscript. As a result, some feedback came through email or comments on the DOCX outputs that were then translated into issues or pull requests by the project managers. Given that our approach hinged on these version control tools, it is likely that our group of contributors was biased towards those who were interested in or experienced with computational tools. The trajectory of the pandemic itself also likely influenced participation: engagement waned over the course of the pandemic as labs opened back up and researchers were able to return to their work, and we recruited very few senior clinicians to the project, which is unsurprising given the load on medical professionals during this time. Engagement that waxes and wanes is, however, typical when writing massively open online papers [22] . Adding features such as spell-check did improve usability, and additional features such as automatically checking the formatting of citations could further improve the usability of this tool. In the future, a formal study of participation could allow for quantification of these biases and improved efforts to foster inclusion. Additional limitations are challenges associated with massively open online papers in general. With such a large amount of text, it is not possible to keep all sections of the manuscript up to date at all times. Readers are not able to distinguish when each section was updated. Even GitHub's blame functionality does not distinguish minor changes from substantive updates to the text. While much of the data and statistics update automatically, the text itself required updating by human experts. This asynchronicity could potentially introduce incompatibility between the figures and the surrounding text. Similarly, in line with the collaboration-related challenges of the project, some authors returned to update their text, while others did not. As a result, the lead authors of each paper often spent several weeks prior to journal submission updating the text to reflect new developments in each area. In the future, it may be possible to streamline this process through integration with a tool such as CoronaCentral [4] to automatically identify relevant, high-impact papers that need to be included, although expertise would still be required to incorporate them. Another challenge involves tracking preprints as they are reviewed or critiqued, revised, and potentially published. While updating the content of the manuscript would likely fall to human contributors, automatic detection of published versions of preprints [38] could be integrated in the future. These challenges are exacerbated by the scale of the infodemic, but developing solutions would benefit future projects tracking more typical trends in publication. Similarly, outputting machine readable summaries of key information in the COVID-19 review manuscripts could reduce their contribution to the infodemic. As it stands, the integration of Compact Uniform Resource Identifier does make a step in this direction. Formal identifiers could be used to extract relationships among clinical trials, genes, publications, and other entities. Thus, the experience of using Manubot for a massive project has laid the foundation for future additions to enhance user experience and inclusivity. With the worldwide scientific community uniting during 2020 and 2021 to investigate COVID-19 from a wide range of perspectives, findings from many disciplines are relevant on a rapid timescale to a broad scientific audience. As many other efforts have described, the publishing rate of formal manuscripts and preprints about COVID-19 has been unprecedented [1] , and efforts to review the body of COVID-19 literature are faced with an ever-expanding corpus to evaluate. In the case of the seven manuscripts produced by the COVID-19 Review Consortium, Manubot allows for continuous updating of the manuscripts as the pandemic enters its second year and the landscape shifts with the emergence of promising therapeutics and vaccines [29] . These manuscripts pull data from external sources and update information and visualizations daily using CI. By off-loading some updates to computational pipelines, domain experts can focus on the broader implications of new information as it emerges. Centralizing, summarizing, and critiquing data and literature broadly relevant to COVID-19 can expedite the interdisciplinary scientific process that is currently happening at an advanced pace. As of September 13, 2021, 2,886 commits have been made to the manuscript across 575 merged pull requests. The efforts of the COVID-19 Review Consortium illustrate the value of including open source tools, including those focused on open publishing, in these efforts. By facilitating the versioning of text, such platforms also allow for documentation of the evolution of thought in an evolving area and formal analysis of a collaborative project. This application of version control holds the potential to improve scientific publishing in a range of disciplines, including those outside of traditional computational fields. While Manubot is a technologically complex tool, this project demonstrates that it can be applied to a variety of projects. Future work can address remaining limitations and continue to advance Manubot as an inclusive tool for open publishing projects. Proliferation of Papers and Preprints During the Coronavirus Disease 2019 Pandemic: Progress or Problems With Peer Review? May own stock or stock options How to fight an infodemic CORD-19: The COVID-19 Analyzing the vast coronavirus literature with CoronaCentral The impact of preprint servers and electronic publishing on biomedical research Too Many Papers An alarming retraction rate for scientific publications on Coronavirus Disease 2019 (COVID-19) An "alarming" and "exceptionally high" rate of COVID-19 retractions? Queries on the COVID-19 quick publishing ethics Idle medical students review emerging COVID-19 research Scientists are drowning in COVID-19 papers. Can new tools keep them afloat? Project, Trainees, Faculty, Advancing scientific knowledge in times of pandemics Covid-19: Epidemiology, Evolution, and Cross-Disciplinary Perspectives COVID-19 diagnostics in context COVID-19 Research in Brief: December Pathophysiology, Transmission, Diagnosis, and Treatment of Coronavirus Disease 2019 (COVID-19) An mRNA Vaccine against SARS-CoV-2 -Preliminary Report Open collaborative writing with Manubot Opportunities and obstacles for deep learning in biology and medicine Is authorship sufficient for today's collaborative research? A call for contributor roles Introducing Massively Open Online Papers (MOOPs), KULA: Knowledge Creation, Dissemination, and Preservation Studies How you can help with COVID-19 modelling Advancing Open Science with Version Control and Blockchains Curating Research Assets: A Tutorial on the Git Version Control System, Advances in Methods and Practices in Psychological Science Git can facilitate greater reproducibility and increased transparency in science Using the MAARIE Framework To Read the Research Literature Studying a Study & Testing a Test: Reading Evidence-Based Health Research COVID-19 Review Consortium Matplotlib: A 2D Graphics Environment Scite: A smart citation index that displays the context of citations and classifies their intent using deep learning Synchronized editing: The future of collaborative writing COVID-19 Review Consortium, A. Gitter, C. S. Greene, Pathogenesis, Symptomatology, and Transmission of SARS-CoV-2 through analysis of Viral Genomics and Structure COVID-19 Review Consortium Coronavirus pandemic (COVID-19) A Phase 3 Randomized Study to Evaluate the Safety and Antiviral Activity of Remdesivir (GS-5734 TM ) in Participants With Severe COVID-19 Linguistic Analysis of the bioRxiv Preprint Landscape, bioRxiv We thank Josh Nicholson 29 . key: cord-0751052-j6v3bzhk authors: Rando, Halie M.; Boca, Simina M.; McGowan, Lucy D’Agostino; Himmelstein, Daniel S.; Robson, Michael P.; Rubinetti, Vincent; Velazquez, Ryan; Greene, Casey S.; Gitter, Anthony title: An Open-Publishing Response to the COVID-19 Infodemic date: 2021-09-17 journal: ArXiv DOI: nan sha: fede7d2ccdca0c30b0c192e2e441d9dfce394429 doc_id: 751052 cord_uid: j6v3bzhk The COVID-19 pandemic catalyzed the rapid dissemination of papers and preprints investigating the disease and its associated virus, SARS-CoV-2. The multifaceted nature of COVID-19 demands a multidisciplinary approach, but the urgency of the crisis combined with the need for social distancing measures present unique challenges to collaborative science. We applied a massive online open publishing approach to this problem using Manubot. Through GitHub, collaborators summarized and critiqued COVID-19 literature, creating a review manuscript. Manubot automatically compiled citation information for referenced preprints, journal publications, websites, and clinical trials. Continuous integration workflows retrieved up-to-date data from online sources nightly, regenerating some of the manuscript’s figures and statistics. Manubot rendered the manuscript into PDF, HTML, LaTeX, and DOCX outputs, immediately updating the version available online upon the integration of new content. Through this effort, we organized over 50 scientists from a range of backgrounds who evaluated over 1,500 sources and developed seven literature reviews. While many efforts from the computational community have focused on mining COVID-19 literature, our project illustrates the power of open publishing to organize both technical and non-technical scientists to aggregate and disseminate information in response to an evolving crisis. Coronavirus Disease 2019 (COVID-19) caused a worldwide public health crisis that has reshaped many aspects of society. The scientific community has, in turn, devoted significant attention and resources towards COVID-19 and the associated virus, SARS-CoV-2, resulting in the release of data and publications at a rate and scale never previously seen for a single topic. Over 20,000 articles about COVID-19 were released in the first four months of the pandemic [1] , causing an "infodemic" [1, 2] . The COVID-19 Open Research Dataset (CORD-19) [3] , which was developed in part with the goal of training machine learning algorithms on COVID-19-related text, illustrates the growth of related scholarly literature ( Figure 1 ). This resource was developed by querying several sources for terms related to SARS-CoV-2 and COVID-19, as well as the coronaviruses SARS-CoV-1 and MERS-CoV and their associated diseases [3] . CORD-19 contained 768,929 manuscripts as of September 6, 2021. Additional curation by CoronaCentral [4] has produced, at present, a set of over 180,000 publications particularly relevant to COVID-19 and closely related viruses. Despite many ad-vances in understanding the virus and the disease, there are also downsides to the availability of so much information. "Excessive publication" has been recognized as a concern for over forty years [5] and has been discussed with respect to the COVID-19 literature [6] . Any effort to synthesize, summarize, and contextualize COVID-19 research will face a vast corpus of potentially relevant material. Figure 1 : Growth of the CORD-19 dataset. The number of articles has proliferated, with both traditional and preprint manuscripts in the corpus. The first release (March 16, 2020) contained 28,000 documents [3] . As of September 6, 2021, this had increased to 768,929 articles. Of these, 30,726 are preprints from arXiv, medRxiv, and bioRxiv. Information was released rapidly by both traditional publishers and preprint servers, and many papers faced subsequent scrutiny. The number of COVID-19 papers retracted may be higher, and potentially much higher, than is typical, although a thorough investigation of this question requires more time to elapse [7, 8] . Many preprints and papers are also associated with corrections or expressions of concern 1 [8] . Preprints are released prior to peer review, but some traditional publishing venues have fast-tracked COVID-19 papers through peer review, leading to questions about whether they are held to typical standards [9] . Therefore, evaluating the COVID-19 literature requires not only digesting available information but also monitoring subsequent changes. Because of the fast-moving nature of the topic, many efforts to summarize and synthesize the COVID-19 literature have been undertaken. These efforts include newsletters 2 [10] , web portals 3 [11] or the now-defunct http://covidpreprints.com 4 , comments on preprint servers 5 [12] , and even a journal 6 . However, the explosive rate of publication presents challenges for such efforts, many of which are no longer active. Similarly, many literature reviews have been written on the available COVID-19 literature [13, 14, 15, 16, 17] , but static reviews quickly become outdated as new research is released or existing research is retracted or superseded. One example is a review of topics in COVID-19 research including vaccine development [17] . This review was published on July 10, 2020, four days before Moderna released the surprisingly promising results of their phase 1 trial [18] that changed expectations surrounding vaccines. Therefore, the COVID-19 publishing climate presented a challenge where curation of the literature by a diverse group of experts in a format that could respond quickly to high-volume, high-velocity information was desirable. We therefore sought to develop a platform for scientific discussion and collaboration around COVID-19 by adapting open publishing infrastructure to accommodate the scale of COVID-19 publishing. Recent advances in open publishing have created an infrastructure that facilitates distributed, version-controlled collaboration on manuscripts [19] . Manubot [19] is a collaborative framework developed to adapt open-source software development techniques and version control for manuscript writing. With Manubot, manuscripts are managed and maintained using GitHub, a popular, online version control interface. We selected Manubot because it offers several advantages over comparable collaborative writing platforms such as Authorea, Overleaf, Google Docs, Word Online, or wikis [19] . Citation-by-identifier ensures consistent reference metadata standards that would be difficult to maintain manually in a manuscript with dozens of authors and over 1,500 citations. Manubot's pull requestbased contribution model balances the goals of making the project open to everyone and maintaining scientific accuracy. All contributions are reviewed, discussed, and formally approved on GitHub before text updates appear in the public-facing manuscript 7 . Continuous integration (CI) seamlessly combines author-produced text and figures with automatically generated and updated statistics and figures derived from external data sources and the manuscript's own content. In addition, the authors who initially launched this project included Manubot developers who had prior successes using Manubot for massively open and traditional manuscript, such as a large-scale collaborative efforts such as a review of developments in deep learning [20] and a re-evaluation of the role of authorship in modern collaborations [21] . Collaboration via massively open online papers has been identified as a strategy for promoting inclusion and interdisciplinary thought [22] . However, the Manubot workflow can be intimidating to contributors who are not well-versed in git [22] . The synthesis and discussion of the emerging literature by biomedical scientists and clinicians is imperative to a robust interpretation of COVID-19 research. Such efforts in biology often rely on What You See Is What You Get tools such as Google Docs, despite the significant limitations of these platforms in the face of excessive publication. We recognized that the problem of synthesizing the COVID-19 literature lent itself well to the Manubot platform, but that the potential technical expertise required to work with Manubot presented a barrier to domain experts. Here, we describe the adaptation of Manubot to facilitate collaboration in the extreme case of the COVID-19 infodemic, with the objective of developing a centralized platform for summarizing and synthesizing a massive amount of preprints, news stories, journal publications, and data. Unlike prior collaborations built on Manubot, most contributors to the COVID-19 collaborative literature review came from biology or medicine. The members of the COVID-19 Review Consortium consolidated information about the virus in the context of related viruses and to synthesize rapidly emerging literature. Manubot provided the infrastructure to manage contributions from the community and create a living, scholarly document integrating data from multiple sources. Its back-end allowed biomedical scientists to sort and distill informative content out of the overwhelming flood of information [23] in order to provide a resource that would be useful to the broader scientific community. This case study demonstrates the value of open collaborative writing tools such as Manubot to emerging challenges. Because it is open source software, we were able to adapt and customize Manubot to flexibly meet the needs of COVID-19 review. Recording the evolution of information over time and assembling a resource that auto-updated in response to the evolving crisis revealed the particular value that Manubot holds for managing rapid changes in scientific thought. First, it was necessary to establish Manubot as a platform accessible to researchers with limited experience working with version control, given that this is not typically emphasized in biology and medicine [24, 25, 26] . Contributors were recruited primarily by word of mouth and on Twitter, and we also collaborated with existing efforts to train early-career researchers. We invited potential collaborators to contribute a short introduction on a GitHub issue in order to collect information about participants and provide an introduction to working with GitHub issues. Interested participants were encouraged to contribute in several ways. One option was to catalog articles of interest as issues. We developed a standardized set of questions for contributors to consider when evaluating an article following a framework often used for assessing medical literature. This approach emphasizes examining the methods used, assignment (whether the study was observational or randomized), assessment, results, interpretation, and how well the study extrapolates [27] . Contributors were also invited to contribute or edit text using GitHub's pull request system. These contributions were not strictly defined and could range from minor corrections to punctuation and grammar to large-scale additions of text. Finally, a small number of contributors (the authors of this paper) contributed technical expertise, either through the development of standardized approaches to the evaluation of papers based on the MAARIE Framework [28] , the writing of code to generate manuscript figures, or the addition of features to Manubot. All of these additions were also submitted as pull requests, either to the COVID-19 review repository or to an external repository, as appropriate. Each pull request was reviewed and approved by at least one other contributor before being merged into the main branch. We tagged potential reviewers based on the introductions they had contributed in order to encourage participation. Authorship was determined based on the Contributor Roles Taxonomy 8 . Due to the permeability of ideas among different sections, contributors to a specific manuscript were recognized with masthead authorship, while all contributors to the project were recognized with consortium authorship on all papers. Emphasizing the use of issues and pull requests was designed to encourage authors with and without git experience to discuss papers and provide feedback (both formal and informal) on proposed text additions or changes. We also used the Gitter chat platform 9 to promote informal questions and sharing of information among collaborators. Applying Manubot's existing capabilities allowed us to confront several challenges common in large-scale collaborations, such as maintaining a record of contributions that allowed us to allocate credit appropriately or to contact the original author if questions arose. Additionally, an up-to-date version of the content was available at all times online in HTML 10 or PDF format 11 . This approach also allowed us to minimize the demand on authors to curate and sync bibliographic resources. Manubot provides the functionality to create a bibliography using digital object identifiers (DOIs), website URLs, or other identifiers such as PubMed identifiers and arXiv IDs. The author can insert a citation in-line using a format such as [@doi:10.1371/journal.pcbi.1007128]. Manubot then obtains reference metadata, exports the citations as Citation Style Language JSON Data Items, and renders the bibliographic information needed to generate the references section [19] . This approach allows multiple authors to work on a piece of text without needing to make manual adjustments to the reference lists. Due to the needs of this project, several new features were implemented in Manubot. Because of the ever-evolving nature of the COVID-19 crisis, figures and statistics in the text quickly became outdated. To address this concern, Manubot and GitHub's CI features were used to create figures that integrated online data sources and to dynamically update information, such as the current number of active COVID-19 clinical trials [29] , within the text of the manuscripts ( Figure 2 ). GitHub Actions runs a nightly workflow to update these external data and regenerate the statistics and figures for the manuscript. The workflow uses the GitHub API to detect and save the latest commit of the external data sources that are GitHub repositories 12 . It then downloads versioned data from that snapshot of the external repositories and runs bash and Python scripts to calculate the desired statistics and produce the summary figures using Matplotlib [30] . The statistics are stored in JSON files that are accessed by Manubot to populate the values of placeholder template variables dynamically every time the manuscript is built. For instance, the template variable {{ebm_trials_results}} in the manuscript is replaced by the actual number of clinical trials with results, 98. The template variables also include versioned URLs to the dynamically updated figures. The JSON files and figures are stored in the external-resources branch of the GitHub repository, providing versioned storage. The GitHub Actions workflow automatically adds and commits the new JSON files and figures to the external-resources branch every time it runs, and Manubot uses the latest version of these resources when it builds the manuscript. The GitHub Actions workflow file is available online 13 , as are the scripts 14 . The Python package versions are also available 15 . Another issue identified was the need for standardized citation to clinical trials. Other researchers identified the same need 16 . Trials that are registered with clinicaltrials.gov receive a unique clinical trial identifier, or "NCT ID." Because clinical trials are registered long before results are published, referencing clinical trial identifiers was a priority. Manubot uses the Zotero translation server 17 to extract citation metadata for some types of citations. However, Zotero did not support clinical trial identifiers and could not extract relevant metadata from their URLs. In order to pull clinical trial metadata associated into Manubot, we added Zotero support for these identifiers. To achieve this, we query clinicaltrials.gov to retrieve XML metadata associated with each identifier using JavaScript 18 . This extension enables citing a trial as @clinicaltrials:NCT04280705 instead of the URL. Then, when Manubot requests clinical trial metadata from the Zotero translation server, the response includes the trial sponsors, responsible investigators, title, and summary. Manubot now supports directly citing hundreds of registered Compact Uniform Resource Identifiers 19 , beyond just the clinicaltrials identifier. Because of the large number of citations used in this manuscript and the fast-moving nature of COVID-19 research, keeping track of retractions, corrections, and notices of concern also became a challenge. We implemented a new Manubot plugin to support "smart citations" in the HTML build of manuscripts. The plugin uses the scite [31] service to display a badge below any citation with a DOI. The badge contains a set of icons and numbers that indicate how many times that source has been mentioned, supported, or disputed and whether there have been any important editorial notices. We were thus able to identify references that needed to be reevaluated by an expert. This addition was invaluable given the nature of the project, where we were disseminating rapidly evolving information of great consequence from over 1,500 different sources. The badges also allow readers to ascertain a rough approximation of the reliability of cited sources at a glance. Because most collaborators were writing and editing text through the GitHub website rather than in a local text editor, we also needed to add spell-checking functionalities to Manubot. We integrated an existing Pandoc 20 spell-check extension with AppVeyor CI to automatically post spelling errors as comments in a GitHub pull request. The comment reported both unique misspelled tokens and all locations where the token was detected. Project maintainers managed a custom dictionary to al- low over 1,500 scientific and technical terms that were not common English words. Spell-checking also helped standardize the writing style across dozens of authors by detecting features such as British versus American English spellings. The actual spell-checking was implemented using GNU Aspell 21 and the Pandoc spellcheck filter 22 . The filter enables checking only the manuscript text, ignoring URLs and formatting. Manubot can render a manuscript in several formats that serve different purposes. Prior to this project, Manubot could use Pandoc to convert the markdownformatted manuscript to HTML, PDF, and DOCX formats. We expanded this functionality to export individual sections of the manuscript as separate DOCX files while still rendering the complete manuscript in HTML and PDF formats. This development was necessary because the manuscript grew so large that it needed to be split into seven separate papers for journal submission while still maintaining shared GitHub discussion across topics. When exporting an individual section, Manubot customizes the manuscript title, authors, and author con-21 http://aspell.net 22 https://github.com/pandoc/lua-filters/tree/master/spellcheck tributions to pertain to that specific section. In addition, we expanded the export formats to include partial La-TeX support via Pandoc. Pandoc converts the markdown content for an individual section to TeX and the Citation Style Language JSON, which contains reference metadata generated by Manubot, to BibTeX. We customized a LaTeX template and reformatted the Manubot metadata, such as authors and their affiliations, for the LaTeX template. The exported TeX file requires manual refinement but contains all manuscript content and most of the formatting. Because LaTeX is required for manuscript submission in many fields, automating most of the process of converting markdown to a submission-friendly format expands Manubot's potential user base. Manubot users can write in the simple markdown format, render the manuscript in continuously-updated PDF or interactive HTML formats, and export the manuscript in DOCX or TeX and BibTeX for submission to traditional publishers, taking full advantage of Pandoc's powerful document conversion capabilities and Manubot's automation. Coverage by Nature Toolbox [32] and an associated tweet 23 about the project on April 1, 2020 attracted the interest of the scientific community ( Figure 3 ). Because GitHub issues are similar to other common web commenting systems, authors learned these tools quickly. The Gitter chat also presented a low barrier to entry. The manuscript continued to grow throughout the first year and a half of the project in both word count and the number of references ( Figure 3 ). Though only a fraction of potential contributors contributed to the text included in the manuscripts (Figure 3 ), many contributors remained engaged over the long term ( Figure 4) . Additionally, new contributors continued to join even into the second year of the project. In order to make the project more accessible, we developed resources explaining how to use GitHub's web interface to develop and edit text for Manubot assuming no prior experience with version control. These tutorials explained how to open an issue, open a pull request, and review a pull request 24 . Additionally, the framework for evaluating literature was converted into issue templates to simplify the review of new articles. Articles were classified as diagnostic, therapeutic, or other, with an associated template developed to guide the review of papers and preprints in each category. A total of 285 new paper issues had been opened as of September 13, 2021. The manuscripts produced by the consortium (excluding this one) will be submitted to mSystems as part of a special issue that provides support for continuous updates as more information becomes available. One has been published and two are available as preprints. This approach allows for a version of record to be maintained alongside the most recent version, which is always available through GitHub. These manuscripts cover a wide range of topics including the fundamental biology of SARS-CoV-2 (pathogenesis [33] and evolution), biomedical advances in responding to the virus and COVID-19 (pharmaceuticals [29] , nutraceuticals [34] , vaccines, and diagnostic technologies), and biological and social factors influencing disease transmission and outcomes. To date, 50 authors are associated with the consortium (Figure 3 ). More formal recruitment efforts to integrate with existing projects providing support for undergraduate students during COVID-19 were also successful. We incorporated summaries written by the students, post-docs, and faculty of the Immunology Institute at the Mount Sinai School of Medicine 25 [12] . Additionally, two of the con-sortium authors were undergraduate students recruited through the American Physician Scientist Association's Virtual Summer Research Program. Thus, the consortium was successful in providing a venue for researchers across all career stages to continue investigating and publishing at a time when many biomedical researchers were unable to access their laboratory facilities. We integrated data into the manuscripts from several sources (Figure 2 ). Worldwide cases and deaths were tracked by the COVID-19 Data Repository by the Center for Systems Science and Engineering at Johns Hopkins University 26 . The clinical trials statistics and figure were generated based on data from the University of Oxford Evidence-Based Medicine Data Lab's COVID-19 Trial-sTracker [35] . Information about vaccine distribution was extracted from Our World In Data 27 [36] . Figure 1 integrates data from the CORD-19 dataset [3] . Manubot's bibliographic management capabilities were critical because the amount of relevant literature published far outstripped what we had anticipated at the beginning of the project. As of September 10, 2021, there were 1,676 references ( Figure 3 ). The scite plugin provided a way to visually inspect the reference list to identify possible references of concern. This and the other new features required for the COVID-19 project are now included in Manubot's rootstock, which is the template GitHub repository for creating a new manuscript. Using CI, Manubot now checks that the manuscript was built correctly, runs spell-checking, and cross-references the manuscripts cited in this review. In addition, Manubot now supports citing clinical trial identifiers such as clinicaltrials:NCT04292899 [37] . The current project was based in the GitHub repository greenelab/covid19-review using Manubot [19] to continuously generate the manuscript. The Manubot framework facilitated a massive collaborative review on an urgent topic. We demonstrated the utility of Manubot to a project where many contributors lacked expertise or even experience working with version control. This effort has produced not only seven literature reviews on topics relevant to the COVID-19 pandemic, but has also generated cyberinfrastructure for training novice users in GitHub. We also extended the functionalities of Manubot to provide more of the benefits of What You See Is What You Get platforms such as Google Docs (Table 1) . Open publishing thus allowed us to harness the domain expertise of a large group of non-technical users to respond to the flood of COVID-19 publications. Several existing and new features in Manubot aid in responding to the challenges posed by the infodemic. Manuscripts are written in markdown and can be rendered in several formats providing different advantages to users. For example, beyond building just a PDF, Manubot also renders the manuscript in HTML, DOCX, and now, LaTeX (in a more limited capacity). The interactive HTML manuscript format offers several advantages over a static PDF to harmonize available resources and address specific problems related to COVID-19. The integration of scite into the HTML build makes references more manageable by visually indicating whether their results are contested or whether they have been corrected or retracted. Cross-referencing different pieces of the manuscript, such as cited preprints with reviews stored in an appendix, is another interactive option presented by HTML. The DOCX format was preferred by most non-technical users for reviewing the final version of the manuscript and was useful for creating submissions to a biological journal. Additionally, because of the heavy emphasis on Word processing in biology, Manubot's ability to generate DOCX outputs was expanded to allow users to generate DOCX files containing only a section of the manuscript. In our case, where the full project is nearly 150,000 words, this allows individual pieces to be shared more easily. Finally, the preliminary addition of LaTeX output is useful for researchers from computational fields who submit papers in TeX format and removes the step of reformatting markdown prior to submission. notices of concern Outputs Improved support for Pandoc's LaTeX output Outputs Build complete manuscript alongside individual sections as standalone documents The COVID-19 Review Consortium provided a platform for researchers to engage in scientific investigation early in the pandemic when many biological scientists were unable to access their research spaces. In turn, by seeking to adapt Manubot to allow for broader participation, we made a number of improvements that are expected to increase its appeal to researchers from all backgrounds. Manubot provided a way for contributors from a variety of backgrounds, including early-career researchers, to join a massive collaborative project while demonstrating their individual contributions to the larger work and gaining experience with version control. The licensing and infrastructure also provide the basis for individuals to adapt from this project to create their own snapshots of the COVID-19 literature that derive from, but are not wholly identical to, the primary versions of these reviews. This project suggests that massive online open publishing efforts can indeed advance scholarship through inclusion [22] , including during the extreme challenges presented by the COVID-19 pandemic. Some challenges did arise in efforts to include an academically diverse set of authors. The barriers to entry posed by git and GitHub likely still reduced participation from individuals who might have otherwise been interested. Using pull requests as a tool for writing text is also unfamiliar to many or most scientists, and the review process can be slow, which might cause interested contributors to lose interest. Additionally, the pull request model may limit people from providing general feedback on the manuscript or a section of the manuscript. As a result, some feedback came through email or comments on the DOCX outputs that were then translated into issues or pull requests by the project managers. Given that our approach hinged on these version control tools, it is likely that our group of contributors was biased towards those who were interested in or experienced with computational tools. The trajectory of the pandemic itself also likely influenced participation: engagement waned over the course of the pandemic as labs opened back up and researchers were able to return to their work, and we recruited very few senior clinicians to the project, which is unsurprising given the load on medical professionals during this time. Engagement that waxes and wanes is, however, typical when writing massively open online papers [22] . Adding features such as spell-check did improve usability, and additional features such as automatically checking the formatting of citations could further improve the usability of this tool. In the future, a formal study of participation could allow for quantification of these biases and improved efforts to foster inclusion. Additional limitations are challenges associated with massively open online papers in general. With such a large amount of text, it is not possible to keep all sections of the manuscript up to date at all times. Readers are not able to distinguish when each section was updated. Even GitHub's blame functionality does not distinguish minor changes from substantive updates to the text. While much of the data and statistics update automatically, the text itself required updating by human experts. This asynchronicity could potentially introduce incompatibility between the figures and the surrounding text. Similarly, in line with the collaboration-related challenges of the project, some authors returned to update their text, while others did not. As a result, the lead authors of each paper often spent several weeks prior to journal submission updating the text to reflect new developments in each area. In the future, it may be possible to streamline this process through integration with a tool such as CoronaCentral [4] to automatically identify relevant, high-impact papers that need to be included, although expertise would still be required to incorporate them. Another challenge involves tracking preprints as they are reviewed or critiqued, revised, and potentially published. While updating the content of the manuscript would likely fall to human contributors, automatic detection of published versions of preprints [38] could be integrated in the future. These challenges are exacerbated by the scale of the infodemic, but developing solutions would benefit future projects tracking more typical trends in publication. Similarly, outputting machine readable summaries of key information in the COVID-19 review manuscripts could reduce their contribution to the infodemic. As it stands, the integration of Compact Uniform Resource Identifier does make a step in this direction. Formal identifiers could be used to extract relationships among clinical trials, genes, publications, and other entities. Thus, the experience of using Manubot for a massive project has laid the foundation for future additions to enhance user experience and inclusivity. With the worldwide scientific community uniting during 2020 and 2021 to investigate COVID-19 from a wide range of perspectives, findings from many disciplines are relevant on a rapid timescale to a broad scientific audience. As many other efforts have described, the publishing rate of formal manuscripts and preprints about COVID-19 has been unprecedented [1] , and efforts to review the body of COVID-19 literature are faced with an ever-expanding corpus to evaluate. In the case of the seven manuscripts produced by the COVID-19 Review Consortium, Manubot allows for continuous updating of the manuscripts as the pandemic enters its second year and the landscape shifts with the emergence of promising therapeutics and vaccines [29] . These manuscripts pull data from external sources and update information and visualizations daily using CI. By off-loading some updates to computational pipelines, domain experts can focus on the broader implications of new information as it emerges. Centralizing, summarizing, and critiquing data and literature broadly relevant to COVID-19 can expedite the interdisciplinary scientific process that is currently happening at an advanced pace. As of September 13, 2021, 2,886 commits have been made to the manuscript across 575 merged pull requests. The efforts of the COVID-19 Review Consortium illustrate the value of including open source tools, including those focused on open publishing, in these efforts. By facilitating the versioning of text, such platforms also allow for documentation of the evolution of thought in an evolving area and formal analysis of a collaborative project. This application of version control holds the potential to improve scientific publishing in a range of disciplines, including those outside of traditional computational fields. While Manubot is a technologically complex tool, this project demonstrates that it can be applied to a variety of projects. Future work can address remaining limitations and continue to advance Manubot as an inclusive tool for open publishing projects. Proliferation of Papers and Preprints During the Coronavirus Disease 2019 Pandemic: Progress or Problems With Peer Review? May own stock or stock options How to fight an infodemic CORD-19: The COVID-19 Analyzing the vast coronavirus literature with CoronaCentral The impact of preprint servers and electronic publishing on biomedical research Too Many Papers An alarming retraction rate for scientific publications on Coronavirus Disease 2019 (COVID-19) An "alarming" and "exceptionally high" rate of COVID-19 retractions? Queries on the COVID-19 quick publishing ethics Idle medical students review emerging COVID-19 research Scientists are drowning in COVID-19 papers. Can new tools keep them afloat? Project, Trainees, Faculty, Advancing scientific knowledge in times of pandemics Covid-19: Epidemiology, Evolution, and Cross-Disciplinary Perspectives COVID-19 diagnostics in context COVID-19 Research in Brief: December Pathophysiology, Transmission, Diagnosis, and Treatment of Coronavirus Disease 2019 (COVID-19) An mRNA Vaccine against SARS-CoV-2 -Preliminary Report Open collaborative writing with Manubot Opportunities and obstacles for deep learning in biology and medicine Is authorship sufficient for today's collaborative research? A call for contributor roles Introducing Massively Open Online Papers (MOOPs), KULA: Knowledge Creation, Dissemination, and Preservation Studies How you can help with COVID-19 modelling Advancing Open Science with Version Control and Blockchains Curating Research Assets: A Tutorial on the Git Version Control System, Advances in Methods and Practices in Psychological Science Git can facilitate greater reproducibility and increased transparency in science Using the MAARIE Framework To Read the Research Literature Studying a Study & Testing a Test: Reading Evidence-Based Health Research COVID-19 Review Consortium Matplotlib: A 2D Graphics Environment Scite: A smart citation index that displays the context of citations and classifies their intent using deep learning Synchronized editing: The future of collaborative writing COVID-19 Review Consortium, A. Gitter, C. S. Greene, Pathogenesis, Symptomatology, and Transmission of SARS-CoV-2 through analysis of Viral Genomics and Structure COVID-19 Review Consortium Coronavirus pandemic (COVID-19) A Phase 3 Randomized Study to Evaluate the Safety and Antiviral Activity of Remdesivir (GS-5734 TM ) in Participants With Severe COVID-19 Linguistic Analysis of the bioRxiv Preprint Landscape, bioRxiv We thank Josh Nicholson 29 .