key: cord-1046167-mm100571 authors: Soderberg, Courtney K.; Errington, Timothy M.; Nosek, Brian A. title: Credibility of preprints: an interdisciplinary survey of researchers date: 2020-10-28 journal: R Soc Open Sci DOI: 10.1098/rsos.201520 sha: e5c26a8e486b4efb36880f6ec0eb2b7e9acd2708 doc_id: 1046167 cord_uid: mm100571 Preprints increase accessibility and can speed scholarly communication if researchers view them as credible enough to read and use. Preprint services do not provide the heuristic cues of a journal's reputation, selection, and peer-review processes that, regardless of their flaws, are often used as a guide for deciding what to read. We conducted a survey of 3759 researchers across a wide range of disciplines to determine the importance of different cues for assessing the credibility of individual preprints and preprint services. We found that cues related to information about open science content and independent verification of author claims were rated as highly important for judging preprint credibility, and peer views and author information were rated as less important. As of early 2020, very few preprint services display any of the most important cues. By adding such cues, services may be able to help researchers better assess the credibility of preprints, enabling scholars to more confidently use preprints, thereby accelerating scientific communication and discovery. Scientific outputs have been growing at a rapid rate since the end of World War II; a recent estimate suggested the growth rate has been doubling approximately every 9 years [1] . Researchers are faced with far more available scholarship than they have time to read and evaluate. How do they decide what is credible, or at Reading scholarly works is an important part of how researchers keep up with the emerging evidence in their field, and explore new ideas that might inform their research. In an information-rich environment, researchers have to make decisions about how to invest their time. Effective filters help researchers make decisions about continuing with deeper review or stopping and moving on. For example, on substance, an effective title will provide the reader with cues about whether the topic of the paper is relevant to them. If it is, an effective abstract will help the reader determine whether it is not relevant after all, that the paper is worth reading in full, or that the gist from the abstract is enough. Assessing substantive relevance is easier than assessing whether the paper meets quality standards that make it worth reading and using. Journal reputation and peer-review are methods of signalling that others independently assessed the work and deemed it worthy enough to publish. Even though previous research has shown that the peer-review process is unreliable (e.g. low test-retest reliability of article acceptance [9] , reviewers catch a small proportion of major mistakes in submissions [10] , strong confirmation biases of reviewers [11] , and poor inter-rater reliability between reviewers and between editors and reviewers [12] ), the signalling function is highly attractive to researchers that need some cues to help filter an overwhelming literature with their limited time. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 Without journal reputation and peer-review, are there any useful cues that can be provided to help researchers filter preprints based on quality or credibility? To date, no work that we know of has investigated credibility cues on preprints specifically. There are, however, models that propose heuristics for how people judge the credibility of online information broadly [13] [14] [15] [16] . Previous work on credibility judgements of scholarly work mostly assessed cues that were already present, rather than investigating potential new cues. For the present work, we sampled cues that could be helpful for assessing credibility, whether they currently existed or not. Transparency, openness, and reproducibility are seen as important features of high-quality research [17, 18] and many scientists view these as disciplinary norms and values [19] . However, information about the openness of research data, materials, code or pre-registrations, either in preprints or published articles, is often not made clear. Previous work by the Center for Open Science (COS) has led to over 60 journals adopting badges on published articles to indicate open data, materials, and/or pre-registrations, and such signals could also be adopted by preprint services [20, 21] . Such information could match up well with the more systematic, thorough processes researchers go through when assessing the credibility of articles. Previous work [5] found that many researchers rate determining whether the 'data presented in the paper are credible' as very important for use/reading decisions. A recent Pew Research survey [22] found that a majority of U.S. adults trust scientific findings more if the data is made publically available. Additionally, work by Piwowar & Vision [23] suggested that an increase in citations for articles which shared data could be due at least in part to data sharing signalling credibility of the work. To the extent that communities value certain research behaviours (e.g. data sharing, code sharing, pre-registration), providing cues that indicate a preprint engages in these behaviours could signal the credibility of the preprint. Work on credibility judgements of internet content identified a number of heuristics that rely on cues about the opinions of others: bandwagon/consensus heuristic (if many others [known or anonymous] think it is good, so should I; [14, 24] ), endorsement (trust sites and sources that are recommended by others [16] ), and liking/agreement heuristic (tending to agree with the opinions of others I like [24] ). Peer-review may function as this type of heuristic cue, and so displaying information such as download or view counts, endorsements, or community comments on preprints could serve a similar purpose. Previous work found that displaying download count information about papers alongside abstracts can alter download behaviours [25] . Though downloads are not a direct measure of credibility, they are moderately correlated with citations [26, 27] and past research has used them as an indication of user trust and satisfaction [28, 29] . Authority [14] and source reputational heuristics [16] suggest a number of cues that might be added to preprints. Preprints do not have journal reputation to rely on for credibility cues, but adding information about authors (e.g. author institutions, verified identities markers like ORCID digital identifiers, links to Google Scholar pages) could increase the extent to which author reputational cues are used [30, 31] . Cues about conflicts of interest of authors or funders, which may carry their own reputational clout, could also tap into these heuristics. Consistency heuristics (a piece of information should be judged as credible when it is found to agree with information from other independent sources [16] ) suggest other types of useful cues. If preprint services could signal the extent to which others could verify author claims (e.g. that data is actually available, that results are robust to other analysis choices), these cues could affect the credibility judgements of individual preprints. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 1.2.5. Variability in cue use Previous research indicates that the cues mentioned above are used generally for making credibility judgements, but there is variability in the relative importance placed on different types of cues. For example, when judging the credibility of scholarly work, all research disciplines strongly endorsed the trustworthiness of the peer-review process. But, life-scientists endorsed peer-review as an indicator of credibility more strongly than respondents from other disciplines, and social scientists more strongly endorsed recommendations by colleagues [5] . There were also differences by the age [32] and Human Development Index (HDI) of the countries researchers were in [33] . This suggests that different types of researchers might find different heuristic cues more relevant or salient than others when assessing credibility. In the spring of 2019, we conducted a survey of active researchers. We assessed the extent to which cues are considered important for credibility judgements about preprints, and how this varied across disciplines and career stages. In total, 4325 researchers consented to take the survey. Of those, 13.09% answered no questions after consent leaving us with a sample of 3759 respondents. Of those, 13.25% dropped out part-way through the survey. All questions in the survey were optional, so even those who completed the survey did not necessarily have complete data. We retained respondents who only partially completed the survey. We cannot calculate an overall response rate to the survey because we cannot determine the total number of people contacted with the diverse outreach methods. All analyses reported below are exploratory, as specified in our pre-registration (https://osf.io/wbpxy) [34] , and the data underlying these analyses can be found at https://osf.io/j7u6z/ [35] . Overall the sample was quite familiar with preprints, with 52.83% of the sample saying they were very or extremely familiar with preprints, and only 5.43% reporting that they were not at all familiar with preprints. The sample was largely balanced in terms of academic career stage, with 33.97% of the sample being graduate students or post docs, 33.28% being an assistant, associate, or full professor and 32.75% either not answering the question or not falling into one of the previously listed career stages. In terms of discipline, the four largest categories, based on the bepress taxonomy, were the social sciences (35.20%), life sciences (24.10%), physical science and mathematics (10.90%), and medical and health sciences (7.95%). Within social scientists, 67.07% self-identified as psychologists. Of respondents, 15.30% either did not report this discipline or did not report something that could be clearly categorized. Finally, the majority of the sample was from North America (31.7%) or Europe (35.3%), with US researchers specifically making up 26.96% of the total sample. Of respondents, 15.8% did not list their country. The vast majority of respondents, 71.70%, also came from countries with very high HDI scores. We had hoped to analyse geographical differences, but unfortunately, we felt that our sample was far too skewed towards very high/high HDI and United States and Western European samples for us to draw any meaningful conclusions based on differences we might see in our data. The sample was mostly favourable towards preprints; 69.73% of the sample felt slightly to strongly favourable towards preprints, while only 15.16% felt opposed to preprints and 14.95% felt neutral. On average, all disciplines and career stages felt favourable towards preprints, though there were slight differences in the extent of this favourability. Psychology and other social science disciplines showed the highest level of favourability, with engineering, biology, and physical and mathematical sciences showing only slightly lower favourability (figure 1). Only one discipline, medicine, had fewer than 60% of respondents favour the use of preprints, and even there, the majority of respondents (51%) royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 reported feeling favourable. Among careers stages, graduate students (80%) and postdocs (78%) showed the highest level of favourability while full professors (61%) showed the lowest levels (figure 2). Of respondents, 72.92% had either viewed/downloaded or submitted preprints at least a few times. However, repeated usage was associated with preprint viewers/downloaders more than by preprint submissions. Of respondents, 70.63% had viewed/downloaded preprints either a few or many times, while only 29.85% had submitted a preprint a few or many times. Among disciplines, medicine had by far the lowest levels of viewing/downloading of any discipline (figure 3a). The life sciences and psychology showed slightly higher percentages of respondents who had either never viewed/downloaded a preprint or only done it a few times as opposed to many times. This may reflect the fact that dedicated preprint services in these disciplines are newer than in other disciplines, and so they have had less time to develop norms around such behaviour. Preprint submission was much lower overall than rates of reading/citing; across all disciplines, the most common response category was that respondents had never submitted a preprint (figure 3b). Even in the physical sciences and mathematics, which have the longest history with preprint services, rates of submission were low. The use of preprints varied by academic career stage, though not in a clear linear pattern. Though postdocs showed the highest levels of viewing/downloading, and full professors the lowest, graduate students and associated professors were not markedly different (figure 4a). As with discipline, the rate of preprint submissions was quite low (figure 4b). Graduate students and postdocs showed the lowest levels of submission, perhaps because they have had fewer opportunities to post/have their work posted as preprints. Professors of all levels were more likely to have submitted a preprint. Of the 19 questions on what information would be important when judging the credibility of a preprint, eight items were rated as either very or extremely important by a majority of respondents (figure 5 Figure 1 . Favourability towards preprints by discipline. Respondents favourability towards the use of preprints in their discipline, broken out by the six most common disciplines in our sample. Numbers to the left of the bars indicate the percentage of respondents who responded with 'very unfavorable', 'somewhat unfavorable' or 'slightly unfavorable', the numbers in the centre of the bars indicate the percentage who responded, 'neither unfavorable nor favorable', and the numbers to the right of the bars indicate the percentage who responded, 'very favorable', 'somewhat favorable' or 'slightly favorable'. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 the exception of COIs, these eight cues are related to indicators of transparency/openness of research content and process (e.g. links to data, links to pre-analysis plans) or to verification of author claims by independent groups (e.g. computational reproducibility, accessing linked information). Only two cues were rated as not at all or slightly important by a majority of respondents (i.e. simplified endorsements and anonymous comments). In general, information related to field perceptions/usage of preprints was rated as less important by respondents, and information related to preprint authors (e.g. institutional information, previous work) received more muted support. To increase the credibility of preprints, it may be particularly useful to know what items were rated as important for judging credibility by respondents who use preprints less or view them less favourably. To investigate this, we ran correlations between participants' ratings of the 19 potential cues and the extent to which respondents, or their co-authors, had submitted preprints; the extent to which they view/download preprints (both converted into ordinal variables, with 1 representing 'never', and 4 coded as 'yes, many times'); and the extent to which they favoured the use of preprints. The results can be seen in figure 6 . We used Spearman correlations for the view/download and submit correlations due to the ordinal nature of the data and a Pearson correlation for favourability. Figure 6 shows that most of the correlations are small (median absolute value r = 0.064). The largest correlations involved information about a preprint having been submitted to a journal, with those that feel less favourability towards preprints (r = −0.26) and use them less (r = −0.20) rating this as more important for making credibility judgements. In general, it appears that more 'traditional' metrics (e.g. author information, peer-review information) are negatively correlated with preprint views/ downloads and favourability, indicating that these items were seen as more important by those lower on these scales, while the open science and independent verification indicators tended to be rated as more important by those who tended to favour and use preprints. In general, the preprint submission variable did not correlate as strongly with the open science or independent verification indicators as either preprint viewing/downloading behaviour or favourability towards preprints. We investigated the extent to which the importance of cues varied across discipline and academic career stage [5, 32] . Figure 7 shows the breakdown of the importance of cues by discipline 1 . Though there are small differences between disciplines (e.g. psychology rates author's previous work and institutions lower than other disciplines; biology rates comments higher than other disciplines), there is consistency across disciplines on the rank ordering of average importance ratings. Differences between cues are bigger than differences between disciplines. To quantify the difference in variance in importance ratings explained by discipline and cue question, we ran a mixed-model, with discipline, cue, and their interactions as fixed effects and a random effect for participants and calculated the R 2 for each fixed effect [36] . In the model, cues explained 10.13% of the overall variation in importance ratings while discipline explained only 1.29% of the variation. In terms of the magnitude of the differences among cues and disciplines, the largest mean differences (on a 1-to-5 point Likert scale) between disciplines on any given question was 0.67, while the largest difference between questions for any given discipline was 1.82. Similarly, the mean differences in item importance between career stages (e.g. graduate students and postdocs rating author information more negatively than full professors) are quite small (figure 8). We again ran a mixed-model with career stage, cue, and their interaction as fixed effects and participant as a random effect. We found that cues explain 18.90% of the variation in important ratings, and career stage explained only 1.36% of the variation. The maximum difference between the mean for any given question across career stages was 0.51, while the maximum mean difference (on a 1-to-5 Likert scale) between questions in any given career stage was 1.71. Thus, across both discipline and career stage, cues appear to be more important for driving importance ratings than participants' discipline or career stage. If certain cues are judged similarly by respondents, this could reflect classes of indicators, with specific indicators within each class perhaps being psychologically interchangeable. Space on preprint landing 35 Figure 3 . Preprint use by discipline. Whether respondents had ever 'viewed/downloaded a preprint' (a) and whether they or a coauthors had 'submitted a preprint' (b), broken up by discipline. Respondents who did not answer the question or who answered 'not sure', are not included in the graphs. pages is not infinite, and so understanding how researchers cluster various indicators could help services prioritize which information to show to most efficiently use presentation space. Additionally, for services that can display more indicators, understanding which cues provide similar types of information would allow services to group psychologically related cues. Grouping related cues can increase the salience of that class of information thus making it more likely to be used in making judgements [37] . We used exploratory factor analysis (EFA) to investigate potential factor structures of the 19 items. A parallel analysis suggested that the data contained six factors. We extracted a six-factor solution using maximum likelihood (ML) and an oblimin rotation. Overall, the fit of the model was adequate, with a TLI = 0.965 and an RMSEA of 0.041, 90% CI [0.038, 0.044]. Though four of the factors are small, with eigenvalues less than 1 and explaining less than 10% of the variance, the fit of the model rapidly grew worse with simpler factor structures. For example, a four-factor solution resulted in a TLI = 0.779 and an RMSEA = 0.103, 90% CI [0.100, 0.105]. Because of fit indices, the parallel analysis, and interpretability, we retained the six-factor solution. A diagram of the structure can be seen in figure 9 . The first two factors conceptually appear to map well onto an 'openness/transparency' and an 'independent verification' concept. The third and fourth factors contain questions related to 'peers views' and 'external support', respectively. Finally, the fifth and sixth factor contains most of the questions related to 'usage metrics' and 'author information'. 2 Based on the correlation structure of the factors, the 'openness/transparency' and 'independent verification' factors are highly related as are the 'peers views', 'usage metrics', and 'author information' factors. This may indicate that more traditional metrics are more closely associated than newer metrics. We explored the extent to which the various cues we investigated are implemented on existing preprint services. In September 2019, we coded eight services, representing a number of disciplines and companies 32 Figure 4 . Preprint use by academic career stage. Whether respondents had ever 'viewed/downloaded a preprint' (a) and whether they or a co-author had 'submitted a preprint' (b), broken up by career stage. Respondents who did not answer the question or who answered 'not sure' are not included in the graphs. 2 The factor structure appears to be broadly used by our respondents, as we found strong measurement invariance of the structure across discipline and career stage using the [38] criterions based on a combination of CFI and RMSEA changes. This indicates that different disciplines and career stages interpret the factors in conceptually similar ways, and the same items are grouped together across disciplines and career stages. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 in the preprints space. For companies that host multiple different preprint services (e.g. Figshare and COS), we chose the first preprint service offered by each company. We coded whether each service displayed each cue on their preprint pages. To be marked 'yes', the preprint service had to meet two criteria. First, the cue had to be displayed on the page that the service controls, rather than in the uploaded preprint document itself. This avoided scoring services based on idiosyncratic features of the uploaded preprints that happened to be examined during the coding process. Second, the service needed to specifically identify the cue content. For example, as of September 2019, OSF Preprints enabled uploading of supplementary material including data and code, but the preprint pages only indicated a link to 'supplementary material' rather than specifically identifying the cue content (e.g. open data). So, OSF Preprints were coded as 'no' for data sharing cues. The results are shown in figure 10 . None of the 19 items were cued on all preprint services, and only the 'usage metrics' cue was displayed by a majority of the services. The preprint service that displayed the most cues was PeerJ (7 of the 19 cues, 36.84%). At the time of coding, this service was still active, but it stopped accepting new preprints in December 2019. Across services, many of the few cues that were present were not prominently placed, often appearing far down the page or on a separate tab. This means that even the few cues that are being displayed may not have their intended impact because users are never exposed to them or they are not very salient for users. We observed especially poor coverage of the eight cues rated as the most important by researchers. Though many preprint services have some way of uploading 'supplementary material,' most do not differentiate between different types of information or cue particular types of content and few services enable preprint authors to provide links if relevant material (e.g. open data, pre-registrations) are stored on another service. As such, while it is possible for preprint authors to engage in the open practices that readers report as credibility enhancing, the preprint services are not yet effective at providing visual cues that those behaviours are occurring. When assessing the credibility of a preprint how important would it be to have each of the following pieces of information? We surveyed 3759 researchers about their perceptions of the importance of different cues for assessing the credibility of preprints. We observed that cues related to openness and independent verification of author assertions were rated more highly than cues related to author identities and peer-review and usage indicators. However, researchers more skeptical of preprints tended to rate author-and peerreview-related cues somewhat more highly than researchers who supported preprints. Nevertheless, there was broad agreement that transparency of the underlying research content (i.e. data, materials, code, pre-registration) and evidence of independent verification of content and research claims were the most important factors for assessing the credibility of preprints. We observed small differences in cue ratings for both academic career stage and discipline, but the differences between career stages and disciplines were much smaller than the differences between cues. The factor structure of items was consistent across the researcher type, indicating that the cues tend to be treated similarly across discipline and career stage. In all subgroups, openness and independent verification cues had higher importance ratings than author identity and peer-review and usage cues. There may be cues that could show large disciplinary or career-stage differences-we only asked participants about a subset of all possible cues-but we did not identify them in this study. Instead, the present evidence suggests that there are a shared set of cues that can be applied across scholarly preprint communities to improve assessment of research credibility. We also observed that existing preprint services display few of the cues that we investigated, particularly those rated as most important: openness and independent verification. This suggests that preprint services could improve support of preprint readers' assessment of research credibility by implementing some of these cues prominently with each preprint. Openness cues should be relatively royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 easy to implement, but cues for independent verification are more difficult to implement for wide application because of conceptual issues in understanding verification practices, implementation challenges in creating widely applicable verification workflows, and resource challenges in conducting verification processes. Given the high value placed on these cues by the community, services and funders should work together to build capabilities to gather and display evidence of verification. This could include creating tools for the community to provide information about independent verification attempts or tools for computationally reproducible manuscripts to be uploaded to services. Though we did not directly ask respondents about how they would rate the importance for these cues for journal articles, it is reasonable to infer journals could also benefit readers by providing cues related to openness and independent verification. Evidence from a recent Pew Research Center survey [22] found that openly available data increased trust in scientific research findings, indicating that there may be broad support for certain types of cues on papers of all types, preprints or peer-reviewed publications. Our sample is large, but it is a convenience sample recruited by marketing to a variety of research communities. As a consequence, our survey does not provide population estimates of researcher opinions, and the estimates could be biased by factors influencing self-selection to participate. Notably, our sample had decent coverage of career stage and research discipline, however, the vast majority of our sample was from the U.S. and/or Western Europe and was quite favourable towards preprints. We observed that favourability towards preprints was correlated with the ratings of certain cue items. So, to the extent that the population of researchers has lower favourability towards preprints than our sample, the overall importance placed on different cues may look different in the Figure 7 . Responses to preprint credibility questions by disciplines. Mean and standard deviation of response to preprint credibility questions by disciplines. Respondents who either skipped the question or could not be categorized into a bepress tier 1 taxonomy are not included in the table. Additionally, participants who listed their discipline as Business, Law, Education, Engineering or Arts and Humanities were also excluded because there were too few respondents in these categories. The response scale is 1-not at all important, 2-slightly important, 3-moderately important, 4-very important, 5-extremely important. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 broader research population. The variation by career stage and discipline was rather weak, suggesting that the top-line conclusions of our findings may be generalizable. However, the significant underrepresentation of non-Western researchers is a significant limitation for making any inferences to those research communities. Researchers were our main population of interest for this study, but they are not the only potential consumers of preprints; funders, journalists, policy makers, and the general public could all potentially benefit from preprints. For example, during the COVID-19 pandemic, major news sources used preprints regularly to keep the engaged public informed about the latest evidence. The Pew survey [22] described above indicates that there may be at least some consistency between what the public and researchers might value, and a recent preprint has called for more open science behaviours in COVID-19 preprints to increase rigor and the ability to assess rigor [39] . But more work is needed to understand if researchers and others generally favour similar or different cues on preprints. We investigated researchers' self-reported opinions about the importance of various cues for assessing research credibility, but not their behaviour towards preprints. There is substantial evidence that self-reported intentions and attitudes are not always aligned with their behaviours [40, 41] . Participants may have reported the beliefs about which cues should be important for assessing credibility, or what they think they would use when assessing credibility, but it is not clear whether these would be the most predictive of researchers' actual likelihood to read, believe, and cite papers. Therefore, it is important to directly measure the relative impact of these cues on behaviours. For example, would the presence of transparency and independent verification cues overwhelm the impact of author identity and usage cues on readers' likelihood to engage with a preprint? royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 researchers' ideals for credibility cues, this insight is important for informing the design and implementation of services to help researchers behave according to their ideals and to reduce the impact of unwanted biases. We observed broad agreement among researchers from different disciplines and career stages that cues on preprints related to open content and independent verification of author claims would be very important for judging the credibility of preprints. It will be useful to replicate and extend our findings to other research communities and other potential cues for assessing research credibility. Open content and independent verification cues are rarely displayed by preprint services, so there is a big opportunity for services to add cues valued by researchers and potentially increase the credibility of preprints. We drafted items starting with information that is commonly displayed on published work, initiatives related to best practices in research reporting (e.g. badges for open practices, Transparency and Figure 9 . Factor structure of information items. The six-factor structure resulting from the EFA. Two items, 'authors general level of open scholarship' and 'preprint submitted to a journal', did not load onto any factor particularly strongly, and were not retained in further analyses. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 Openness Promotion [TOP] guidelines), and cues that had been identified during literature review as important for trust/credibility judgements. Then, we invited members of the OSF preprint services community, and other preprint stakeholders to revise the survey and assess clarity and pertinence for researchers from a broad variety of fields. Following this round of refinement, we pilot tested with researchers from a few disciplines to gather additional feedback on the clarity of the questions and the adequacy of the response options. The final survey included questions in four categories: engagement information, importance of cues for credibility, credibility of service characteristics, and demographics (see https://osf.io/4qs68/ for the full version of the questionnaire [42] ). Four items asked participants about their familiarity with preprints, the extent to which they favour the use of preprints in their discipline, and how often (if ever) they viewed/downloaded preprints, and how often (if ever) they or one of their co-authors has posted a preprint. Because the word 'preprint' is not universal in all disciplines (e.g. 'working papers' are often posted before publication in economics) 1 service only shows this information for author who uploaded preprint 2 service shows which pre prints have been accepted to journals, but not which submitted 3 service has a general place to link/upload other files, but types of files are not clearly identified 4 service has a specific location for data/code information, but does not differentiate between the two 5 service require commenters to have a public username, but username doesn't have to be a real name Figure 10 . Information cues presented by preprint services. Coding by preprint service for whether they cue the information discussed in each preprint credibility question. royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 and some disciplines use 'preprint' broadly to include both preprints and postprints, we defined 'preprint', 'postprint', and 'preprint service' for our participants. We asked participants to rate 19 different types of information for how important it would be to have them when assessing the credibility of a preprint. Some types of information (e.g. author[s] institutional information, links to available data) are presently listed on at least some preprint services ( figure 10 ). However, other types of information (e.g. simplified endorsements of preprints by other researchers, information about whether independent groups could reproduce the findings in the preprint) were not displayed by any services at the time of the survey. The response scale was a 5-point Likert scale from 1 (not at all important) to 5 (extremely important). We asked participants the extent to which 18 characteristics would increase or decrease the credibility of the service as a whole. Some features related to specific functionality that could be provided by a preprint service (e.g. services screen for spam), while others related to behaviours users could engage in on the service (e.g. service allows the posting of new versions of preprints). The response scale was a 7-point Likert scale from −3 (decrease a lot) to 3 (increase a lot). Analysis of this data can be found in the electronic supplementary materials. We asked participants five questions to gather demographic information. These included the country they currently reside in, their job title/position, age, discipline, sub-discipline, and how they heard about the survey. We collected survey responses from late March 2019 to the end of June 2019. To investigate potential differences between discipline, career stage, and geographical location, we collected as diverse a sample as possible. We partnered with groups from different research communities, preprint services, publishers, scholarly societies, and individuals to diversify our outreach. Outreach included social media posts, emails to various listservs, emails to journals or platform lists, pop-up windows on journal pages, and some direct emails to departments. Following consent, we defined 'preprints', 'preprint service', and 'postprint' for all respondents, and then asked them the preprint engagement questions. This was followed by the importance of cues for credibility section. During this section, we defined 'data', 'materials', 'pre-registration', 'replication', and 'reproduction' for respondents, as these terms are used differently in different disciplines and may have been unfamiliar to some respondents. They then responded to the service credibility questions, followed by the demographics. For discipline, respondents could self-select one of 29 disciplines that we knew to have preprint services or could select 'other'. They were then asked to enter their sub-discipline or discipline as free response text, and we used this information to code them into the three tiers of the bepress taxonomy of disciplines [43] . If respondents put a discipline that could not be cleanly coded into a single discipline at a given level of the bepress taxonomy, we treated the data as missing at that level. Ethics. The protocol for the survey was approved by the IRB at the University of Virginia (Protocol Number: 2192). Informed consent was obtained from all participants in the first page of the survey. Data accessibility. The data, materials, and code underlying this paper can be found on OSF: https://osf.io/6kz2j/. Growth rates of modern science: a bibliometric analysis based on the number of publications and cited references Accelerating scientific publication in biology Why do people use preprints? What is their value? Why Preprint. Retrieved Trustworthiness and authority of scholarly information in a digital age: results of an international questionnaire Editorial: preprints, citations, and nucleic acids research 2020 The effects of bioRxiv preprints on citations and altmetrics Trust and authority in scholarly communication in the light of the digital transition: setting the scene for a major study Peer-review practices in psychological journals: the fate of published articles, submitted again What errors do peer reviewers detect, and does training improve their ability to detect them? Publication prejudices: an experimental study of confirmatory bias in the peer review system Editorial peer reviewers' recommendations at a general medical journal: are they reliable and do editors care? Making sense of credibility on the web: models for evaluating online information and recommendations for future research The MAIN model: a heuristic approach to understanding technology effects on credibility Believe it or not: factors influencing credibility on the web Social and heuristic approaches to credibility evaluation online Journals unite for reproducibility Promoting transparency in social science research Normative dissonance in science: results from a national survey of U.S. scientists Badges to acknowledge open practices: A simple, low-cost, effective method for increasing transparency Trust and mistrust in Americans' views of scientific experts Data reuse and the open data citation advantage The heuristic model of persuasion Scientists as subjects: How researcher behaviors influence psychological knowledge The spread of scientific information: Insights from the web usage statistics in PLoS article-level metrics Altmetrics in the wild: Using social media to explore scholarly impact Viewing and reading behavior in a virtual environment: The full-text download and what can be read into it To boldly go beyond downloads: how are journal articles shared and used? Cues filtered out, cues filtered in: Computer-mediated communication and relationships Signals in social supernets Do younger researchers assess trustworthiness differently when deciding what to read and cite and where to publish? How scholars implement trust in their reading citing and publishing activities: geographic difference Credibility of Preprints Survey: OSF-Standard Pre-Data Collection Registration An R2 statistic for fixed effects in the linear mixed model Heuristic versus systematic processing and the use of source versus message cues in persuasion Sensitivity of goodness of fit indexes to lack of measurement invariance Open science saves lives: lessons from the COVID-19 pandemic. bioarxiv Telling more than we can know: verbal reports on mental processes Implicit social cognition bepress three-tiered list of academic disciplines Acknowledgements. We had many individuals and groups assist us with survey drafting and/or recruitment. We would like to thank Oya Y. Rieger from Ithaka S+R (formerly from arXiv), Peter Binfield from PeerJ, Jessica Polka from royalsocietypublishing.org/journal/rsos R. Soc. Open Sci. 7: 201520 15 ASAPbio, Naomi Penfold from eLife (formerly from ASAPbio), Darla Henderson from ACS/ChemRxiv, Jon Tennant, Katherine Hoeberling from BITSS, Philip Cohen from SocArXiv, Wendy Hasenkamp from the Mind and Life Institute, Tom Narock from EarthArXiv, and Vicky Steeves from LISSA for helping to draft and refine survey items. We would also like to thank ASAPbio, PLOS, BITSS, FABSS, EACR, AMSPC, The Electrochemical Society, Lindau Alumni Network, F1000Research, Springer Nature, BMJ, SREE, Elsevier, Cambridge University Press, Wiley, eLife, iTHRIV, preprints.org, HRA, preLight, APA, APS, the Psychonomic Society, SocArXiv, PaleoRxiv, LISSA, EarthRxiv and Inarxiv for their help with survey recruitment. Competing interests. C.S., T.E. and B.N. are paid employees of the non-profit Center for Open Science that has a mission to increase openness, integrity and reproducibility of research including offering services supporting preprints.Funding. This research was funded by a grant from the Alfred P. Sloan Foundation (Grant Number: G-2018-11108).