key: cord-0622170-mr4whsvv authors: Wadden, David; Lo, Kyle; Wang, Lucy Lu; Cohan, Arman; Beltagy, Iz; Hajishirzi, Hannaneh title: MultiVerS: Improving scientific claim verification with weak supervision and full-document context date: 2021-12-02 journal: nan DOI: nan sha: 2743e66939b30c43affb3c9e31f20cfac2109045 doc_id: 622170 cord_uid: mr4whsvv The scientific claim verification task requires an NLP system to label scientific documents which Support or Refute an input claim, and to select evidentiary sentences (or rationales) justifying each predicted label. In this work, we present MultiVerS, which predicts a fact-checking label and identifies rationales in a multitask fashion based on a shared encoding of the claim and full document context. This approach accomplishes two key modeling goals. First, it ensures that all relevant contextual information is incorporated into each labeling decision. Second, it enables the model to learn from instances annotated with a document-level fact-checking label, but lacking sentence-level rationales. This allows MultiVerS to perform weakly-supervised domain adaptation by training on scientific documents labeled using high-precision heuristics. Our approach outperforms two competitive baselines on three scientific claim verification datasets, with particularly strong performance in zero / few-shot domain adaptation experiments. Our code and data are available at https://github.com/dwadden/multivers. The proliferation of scientific mis-and disinformation on the web has motivated the release of a number of new datasets (Saakyan et al., 2021; Sarrouti et al., 2021; Wadden et al., 2020; Kotonya and Toni, 2020) and the development of modeling approaches (Pradeep et al., 2021; Zhang et al., 2021) for the task of scientific claim verification. The goal of the task is to verify a given scientific claim by labeling scientific research abstracts which SUPPORT or REFUTE the claim, and to select evidentiary sentences (or rationales) reporting the findings which justify each label. A common approach to this task is to first extract rationales from the larger document context, and then make label predictions conditioned on the Ibuprofen worsens COVID-19 symptoms and avoiding Ibuprofen. Figure 1 : A claim from the HealthVer dataset, refuted by a research abstract. The sentence in red is a rationale reporting a finding that REFUTES the claim. However, this finding cannot be interpreted properly without the context in blue, which specifies that the finding applies to Ibuprofen as a treatment for COVID. MUL-TIVERS incorporates the full context of the evidencecontaining abstract when predicting fact-checking labels. selected rationales. This "extract-then-label" approach has two important drawbacks, which we aim to address in this work. First, the rationales may lack information required to make a prediction when taken out-of-context; for instance, they may contain acronyms or unresolved coreferences, or lack qualifiers that specify the scope of a reported finding (Figure 1 provides an example). Second, the "extract-then-label" approach requires training data annotated with both sentence-level rationales and abstract-level labels. While sentence-level rationale annotations are costly and require trained experts, abstract-level labels can be created cheaply using high-precision heuristics, e.g., the titles of research papers sometimes make claims that are supported by their abstracts. Motivated by these challenges, we introduce MULTIVERS (Multitask Verification for Science): Given a claim and evidence-containing scientific abstract, MULTIVERS creates a shared encoding of the entire claim / abstract context, using the Long-former encoder (Beltagy et al., 2020) to accommodate long sequences. Then, it predicts an abstractlevel fact-checking label and sentence-level rationales in a multitask fashion, enforcing consistency between the outputs of the two tasks during decoding. This modeling approach ensures that label predictions are made based on all available context, and enables training on instances derived via weak supervision for which abstract-level labels are available, but sentence-level rationales are not. In experiments on three scientific claim verification datasets, we find that MULTIVERS outperforms two state-of-the-art baselines, one of which has more than 10x the parameters of our system. In addition, we show that training MULTIVERS on weakly-labeled in-domain data substantially improves performance in the zero / few-shot domain adaptation settings. The ability to achieve reasonable performance given limited labeled data is especially valuable in specialized domains, due to the high cost of collecting expert annotations. In summary, our contributions are as follows: 1. We introduce MULTIVERS, a multitask system for full-context scientific claim verification. MULTIVERS improves fully-supervised factverification performance by an average of 11% on three datasets over two state-of-the-art baselines, with improvements of 14% and 26% in the few-shot and zero-shot settings. 2. We present weak supervision heuristics to assign fact-checking labels to two large scientific datasets, and show that training on these annotations more than doubles zero-shot domain adaptation performance. 3. Through ablations and analysis, we demonstrate that our multitask modeling approach achieves our goals of incorporating full-document context into label predictions, and facilitating zero / few-shot domain adaptation. We use the definition of scientific claim verification from the SCIFACT task (Wadden et al., 2020) , and provide a brief overview of the task here. Other works have cast scientific claim verification as a sentence-level natural language inference (NLI) task; in §4.1, we describe how we process these datasets to be compatible with the task as considered in this work. Task definition Given a claim c and a collection of candidate abstracts which may contain evidence relevant to c, the scientific claim verification task requires a system to predict a label y(c, a) ∈ {SUPPORTS, REFUTES, NEI 1 }, which indicates the relationship between c and a for each candidate a. For all abstracts labeled SUPPORTS or REFUTES, the system must also identify rationales R(c, a) = {r 1 (c, a), . . . , r n (c, a)}, where each r i (c, a) is a sentence from a that either entails or contradicts the label y(c, a). 2 The rationales may not be self-contained, and may require additional context from elsewhere in the abstract to resolve coreferential expressions or acronyms, or to determine qualifiers specifying experimental context or study population. 3 Examples of these situations are provided in Figure 1 and Appendix A.3. Evaluation The SCIFACT task reports four evaluation metrics. We have found that two of these metrics are sufficient to convey the important findings for our experiments: (1) abstract-level label-only evaluation computes the model's F1 score in identifying abstracts that SUPPORT or REFUTE each claim. Predicting the correct label y(c, a) is sufficient; models do not need to provide rationales. (2) Sentence-level selection+label evaluation computes the point-wise product of the model's F1 score in identifying the rationales R(c, a), with the model's abstract-level label y(c, a); this metric rewards precision in identifying exactly which sentences contain the evidence justifying the label. In this work, we refer to these two metrics as "abstract" and "sentence" evaluation respectively. Retrieval settings For open scientific claim verification, the system must retrieve candidate abstracts from a corpus of documents. In the abstractprovided setting, candidate abstracts for each claim are given as input. We describe the retrieval settings for all datasets in §4.1. We consider three supervision settings. In the zero-shot domain adaptation setting, models may not train on any in-domain factchecking data, though they may train on generaldomain fact-checking data and other available scientific datasets. In the few-shot domain adaptation setting, models may train on 45 claims from the target dataset. In the fully-supervised setting, models may train on all claims from the target dataset. While most existing work on scientific factchecking has focused on the fully-supervised setting, some recent work has examined the zero-shot setting. Lee et al. (2021) use language model perplexity as a measure of claim veracity. Wright et al. (2022) generate claims based on citation sentences, and verify each generated claim against the abstracts mentioned in the claim's source citation. Given the high potential impact of fact verification systems for specialized domains, combined with the substantial cost of creating these datasets, we believe that the development of techniques for zero / few-shot domain adaptation represents an important area for continued research. Several scientific claim verification datasets have been released in the past few years. COVIDFact (Saakyan et al., 2021) and HealthVer (Sarrouti et al., 2021) verify COVID-19 claims against scientific literature. PUBHEALTH (Kotonya and Toni, 2020) verifies public health claims against news and web sources. SCIFACT (Wadden et al., 2020) verifies claims made in citations in scientific papers. CLIMATE-FEVER (Diggelmann et al., 2020) verifies claims about climate change against Wikipedia. In this work, our focus is verifying claims against scientific literature. We therefore perform experiments on the COVIDFact, HealthVer, and SCIFACT datasets. Preprocessing details and summary statistics for these datasets are included in §4.1. Motivated in part by the SCIVER shared task (Wadden and Lo, 2021) and leaderboard, a number of models have been developed for SCIFACT (the focus of the shared task). The two strongest systems on the shared task were VERT5ERINI (Pradeep et al., 2021) and PARAGRAPHJOINT , which we adopt as baselines. More recently, ARSJOINT (Zhang et al., 2021) achieved performance competitive with these two systems. 4 Given a claim c and candidate abstract a, these models make predictions in two steps. First, they predict rationales R(c, a) = { r 1 (c, a), . . . , r n (c, a)} likely to contain evidence. Then, they make a label prediction y(c, f R ( R(c, a)) based on the claim and predicted rationales, where 4 Recent progress can be found on the SciFact leaderboard. f R is a function which creates a representation of the predicted rationales. While existing models share this general approach, they use different functions f R to construct rationale representations. For VERT5ERINI, rationale selection and label prediction are performed by two separate T5-3B models, and f R concatenates the text of the selected rationales. As a result, the label predictor may not have access to all context needed to make a correct label prediction. PARA-GRAPHJOINT and ARSJOINT attempt to address this issue by encoding the claim and full abstract (truncating to 512 tokens), and using these representations as the basis for both rationale selection and label prediction. The function f R consists of selfattention layers over the (globally-contextualized) token representations of the predicted rationales. Thus, PARAGRAPHJOINT and ARSJOINT can incorporate abstract-level context into label decisions. However, the mechanism by which this occurs is more complex than for our proposed system and requires rationale supervision for all training instances. We propose the MULTIVERS model for fullcontext claim verification. In §3.1, we describe our modeling approach. Rather than predicting rationales R(c, a) followed by the overall fact-checking label y(c, f R ( R(c, a))), we predict y(c, a) directly based on an encoding of the entire claim and abstract, and enforce consistency of R(c, a) with y(c, a) during decoding. A similar idea has been shown to be effective on sentiment analysis and propaganda detection with token-level rationales (Pruthi et al., 2020) . In §3.2, we explain how our approach facilitates few-shot domain adaptation using weakly-labeled scientific documents. Long-document encoding Given a claim c and candidate abstract a consisting of title t and sentences s 1 , . . . , s n , we concatenate the inputs separated by tokens. The token following each sentence s i is notated as i : c t s 1 1 . . . s n n The model input sometimes exceeds the 512-token limit common to transformer-based language models like BERT (Devlin et al., 2019) and RoBERTa ; see Table 1 for details on how frequently this occurs. Therefore, we use the Longformer model (Beltagy et al., 2020) as our encoder. We assign global attention to the token, as well as all tokens in c and all tokens. Multitask rationale selection and label prediction Given the full-context Longformer encoding, we predict whether sentence s i is a rationale via a binary classification head, consisting of two feedforward layers followed by a two-way softmax, on top of the globally-contextualized token i . Similarly, we predict the overall fact-checking label y(c, a) by adding a three-way classification head over the encoding of the token. Since the token is trained with global attention, the model makes predictions based on a representation of the entire claim and abstract. During training, we compute the cross-entropy losses for the label and rationale predictions, and train to minimize the multitask loss: where λ rationale is tuned on the dev set. At inference time, we first predict y(c, a) to be the label with the highest softmax score. If the predicted label is NEI, we predict no rationales. If the predicted label is either SUPPORTS or RE-FUTES, then we predict rationales as all sentences with an assigned softmax score of greater than 0.5. If no sentences have a rationale softmax over 0.5, then we predict the highest-scoring sentence as the sole rationale. In §6.2, we show that this ability to condition the rationale predictions on the label prediction (as opposed to conditioning the label on the predicted rationales) leads to substantial improvement in the zero-shot domain adaptation setting. Candidate abstract retrieval For datasets that require retrieval of candidate abstracts, we rely on the VERT5ERINI (Pradeep et al., 2021 ) retrieval system, which achieved state-of-the-art performance on the SCIVER shared task (SCIVER used the SCIFACT dataset for evaluation). This model first retrieves abstracts using BM25 (Robertson and Zaragoza, 2009), then refines the predictions using a neural re-ranker based on Nogueira et al. (2020) , which is trained on the MS MARCO passage dataset (Campos et al., 2016) . Three types of data are available to train scientific claim verification systems. (1) In-domain fact-checking annotations are the "gold standard", but they are expensive to create and require expert annotators. (2) General-domain fact-checking datasets like FEVER (Thorne et al., 2018) are abundantly available, but generalize poorly to scientific claims (see §6.1). (3) Scientific documents -either unlabeled or labeled for different tasks -are abundant, and high precision heuristics (described in §4.2) can be used to generate document-level fact-checking labels y(c, a) for these data. We train MULTIVERS as follows: we first pretrain on a combination of general-domain factchecking annotations, combined with weaklylabeled in-domain data. 5 Then, we finetune on the target scientific fact-checking dataset. The multitask architecture of MULTIVERS is well-suited to this strategy, since the model can be trained on data with or without rationale annotations. When no rationales are available, we set λ rationale = 0 in the loss function and train as usual. By contrast, training an "extract-then-label" model on weaklysupervised data requires creating rationale annotations R(c, a), which is quite noisy (see §4.2). We experiment with three scientific claim verification datasets. Table 1 provides a summary of important dataset characteristics. Preprocessing steps and additional statistics can be found in Appendix A. HealthVer and COVIDFact were originally released in an NLI format, pairing claims with (out-of-context) evidentiary sentences. We convert to our task format by identifying the abstracts in the CORD-19 corpus containing these sentences. We use the following terminology: an atomic claim makes an assertion about a single property of a single entity, while a complex claim may make assertions about multiple properties or entities. SCIFACT Claims in SCIFACT (Wadden et al., 2020) were created by re-writing citation sentences occurring in biomedical literature into atomic claims, which were verified against the abstracts of the cited documents. REFUTED claims were created by manually negating the original claims. Abstracts that were cited but which annotators judged not to contain evidence were labeled NEI. SCI-FACT requires retrieval of candidate abstracts. Table 1 : Summary of datasets used in experiments. The top group of datasets are scientific claim verification datasets, and the bottom group are for pretraining. Datasets with a for "Open" require that candidate abstracts be retrieved from a corpus; those with a provide candidate abstracts as input. Datasets with a for "Has NEI" require three-way (SUPPORTS / REFUTES / NEI) label prediction, while those with an are (SUPPORTS / REFUTES) only. The "> 512 tokens" column indicates the percentage of claim / abstract contexts that exceed 512 tokens. HealthVer (Sarrouti et al., 2021) consists of COVID-related claims obtained by extracting snippets from articles retrieved to answer questions from TREC-COVID (Voorhees et al., 2020) , verified against abstracts from the CORD-19 corpus . Claims in HealthVer may be complex. REFUTED claims occur naturally in the article snippets. HealthVer provides candidate abstracts for each claim, but some of these candidates do not contain sufficient information to justify a SUPPORTS / REFUTES verdict and are labeled NEI. COVIDFact (Saakyan et al., 2021 ) collects claims about COVID-19 scraped from a COVID-19 subreddit, and verifies them against linked scientific papers, as well as documents retrieved via Google search. Claims in COVIDFact may be complex, and candidate abstracts for each claim are provided. All candidates either SUPPORT or REFUTE the claim. Claim negations were created automatically by replacing salient words in the original claims, and as a result the labels y(c, a) are somewhat noisy (see Appendix A). We briefly describe our pretraining datasets and the weak supervision heuristics used to construct them. Detailed descriptions of these heuristics can be found in Appendix A.1. FEVER (Thorne et al., 2018) consists of claims created by re-writing Wikipedia sentences into atomic claims, verified against Wikipedia articles. EVIDENCEINFERENCE (Lehman et al., 2019; was released to facilitate understanding of clinical trial reports, which examine the effect of an intervention on an outcome, relative to a comparator ("ICO" elements). The dataset contains ICO prompts paired with (1) labels indicating whether the outcome increased or decreased due to the intervention, and (2) rationales justifying each label. We use rule-based heuristics to convert these prompts into claims -for instance "[intervention] increases [outcome] relative to [comparator]". PUBMEDQA (Jin et al., 2019) was released to facilitate question-answering over biomedical research abstracts. We use the PQA-A subset, which is a large collection of abstracts with "claim-like" titles -for instance, "Vitamin B6 supplementation increases immune responses in critically ill patients." We treat the paper titles as claims and the matching abstracts as the evidence sources. To train models requiring rationale supervision, we create weakly-supervised rationales by selecting the sentences with highest similarity to the claim as measured by cosine similarity of Sentence-BERT embeddings (Reimers and Gurevych, 2019) . These annotations are not used to train MUL-TIVERS. To estimate the precision of our rationale labeling heuristic, we predict rationales in the same fashion for our supervised datasets and compute the Precision@1 with which this method identifies gold rationales. The scores are relatively low: 49.4, 48.8, and 43.4 for SCIFACT, COVIDFact, and HealthVer respectively. We describe our model training procedure, the systems against we compare MULTIVERS, and our ablation experiments. Our complete training procedure consists of pretraining on the three datasets from §4.2, followed by finetuning on one of the target datasets from §4.1. We conduct experiments with three different levels of supervision. For zero-shot experiments, we perform pretraining only. For few-shot experiments, we pretrain followed by finetuning on 45 target examples. For fully-supervised experiments, we pretrain and then train on all target data. Following , we found that negative sampling was important to achieve good precision on SCIFACT, which requires document retrieval. We train with 20 negative samples per claim and retrieve 10 abstracts per claim at inference time. Appendix C.3 shows results without negative sampling. For the other datasets, no negative sampling was used. Additional details including batch sizes, learning rates, number of epochs, etc. can be found in Appendix B. During model development, we experimented with training on all three target datasets combined before predicting on each one, but found that this did not improve performance; see Appendix C.4. We use PARAGRAPHJOINT and VERT5ERINI as baselines. VERT5ERINI is the largest model, with 5.6B parameters. MULTIVERS and PARA-GRAPHJOINT are comparably-sized, with 440M and 360M parameters, respectively. In the fully-supervised setting, we compare against both baselines. For prediction on SCIFACT, we use publicly available model checkpoints asis. For training on HealthVer and COVIDFact, we use the code provided by the authors, starting from the available checkpoints trained on SCI-FACT. Model hyperparameters (learning rate, batch size, epoch number, etc.) for all systems including MULTIVERS were tuned based solely on SCIFACT and not adjusted further. Additional details can be found in Appendix B.4. Evaluation in the few-shot and zero-shot settings requires pretraining and finetuning as described in §5.1. Due to the expense of pretraining T5-3B, we do not perform these experiments for VERT5ERINI, and compare only against PARA-GRAPHJOINT (which shows comparable performance in the fully-supervised setting). We pretrain PARAGRAPHJOINT on the data described in §4.2. Since PARAGRAPHJOINT and VERT5ERINI differ from MULTIVERS along a number of important dimensions (e.g. model architecture, number of parameters, and base encoder), we conduct ablations to characterize the performance contributions of three key components of MULTIVERS. Pretraining data We compare the results of three different pretraining strategies. For FEVER-SCI, we pretrain on all available data as described in §5.1. For FEVER, we pretrain on FEVER only. For No-Pretrain, we perform no pretraining. Base encoder We compare the performance achieved using LongFormer as the encoder for MULTIVERS, compared to the results when we swap in RoBERTa but keep other settings identical. We use Longformer-large and RoBERTa-large. Modeling approach We compare three modeling approaches: (1) The Multitask approach is the method used by MULTIVERS as described in §3.1. (2) The Pipeline approach consists of two separate Longformer modules. The first selects rationales as described in §3.1, but with L label removed from Eq. 1, and the second module predicts a label given the text of the rationales selected by the first module as input. When pretraining on PUBMEDQA, we train on the rationales chosen by Sentence-BERT as described in §4.2. (3) The Multitask train / Pipeline inference (MT / PI) approach takes the model trained using the Multitask approach, and performs inference using the Pipeline approach. Specifically, MT / PI is trained to make label predictions based on full abstracts, but must make testtime label predictions based on predicted rationales only. By contrast, the Pipeline model makes label predictions based on gold and predicted rationales at train and test time, respectively. We compare MULTIVERS performance relative to our baseline systems, and present ablation results. Table 2 : Performance of MULTIVERS and baselines. In the fully-supervised setting, we compare to PARA-GRAPHJOINT and VERT5ERINI, which exhibit comparable performance. In the zero and few-shot settings, we compare to PARAGRAPHJOINT only due to the high cost of pretraining VERT5ERINI. We report performance using abstract-level and sentence-level evaluation as defined in §2.1. that this is due to the automatically-generated nature of COVIDFact negations. Third, we observe that HealthVer appears to be the most challenging dataset of the three. Few-shot abstract-level F1 scores for COVIDFact and SCIFACT are generally within 10 F1 of their fully-supervised values, while the gap is roughly 20 F1 for HealthVer. This may be due to the high complexity of HealthVer claims. The results of all ablations are shown in Table 3 . We report abstract and sentence-level F1 scores in the main text; full results can be found in Table 9 in Appendix C. In-domain pretraining substantially improves zero / few-shot performance In Table 3a , we compare the performance of models pretrained on FEVERSCI, FEVER, and No-Pretrain. In the zeroshot setting, removing scientific data during pretraining results in a relative performance decrease of 65%, averaged over the three datasets and two evaluation methods. The decrease is driven primarily by very low recall (see Table 9a ). In the few-shot setting, FEVER pretraining scores within 4% of FEVERSCI, while No-Pretrain results in a 39% decrease relative to FEVERSCI. This suggests that training on a handful of target examples is sufficient to recalibrate a model trained for a different domain, but not to learn the task from scratch. In the fully-supervised setting, FEVER pretraining is only slightly worse than FEVERSCI, while No-Pretrain lags by roughly 9%. Overall, the results indicate that pretraining always helps, and pretraining on weakly-labeled in-domain data helps especially when target data are scarce. Table 3b compares the performance of MULTIVERS when Longformer and RoBERTa are used as the base encoder. Using Longformer consistently helps on SCIFACT, but does not help on the other two datasets. This is unsurprising, since 27% of SCIFACT instances exceed the RoBERTa token limit, compared to less than 15% for the other two datasets (Table 1) . Multitask modeling improves zero / few-shot performance Results comparing our three different modeling approaches are shown in Table 3c . In the zero-shot setting, we find that Multitask performs best, with both MT / PI and Pipeline exhibiting performance drops greater than 50%. The Multitask approach of predicting rationales conditioned on the predicted label leads to improved recall (see Table 9c ). Similarly, in the few-shot setting, both Pipeline and MT / PI perform roughly 10% worse than Multitask. Collectively, the results indicate that Multitask makes the best use of the available data when target annotations are limited. We also find that MT / PI outperforms Pipeline in the zero-shot setting. This supports our intuition from §3.2 that, while training on weaklysupervised document-level labels improves zeroshot performance, training on weakly-supervised sentence-level rationales (as for Pipeline) leads to worse performance than not training on these rationales (as for MT / PI). In the fully-supervised setting, Multitask performs best on SCIFACT, while Pipeline slightly outperforms Multitask on HealthVer and COVID-Fact. MT / PI performs substantially worse than the other approaches on all datasets. We investigate (a) Effect of pretraining data. In-domain pretraining is very effective in the zero-and few-shot settings. In the zero-shot setting, "No-Pretrain" metrics are not shown since this would correspond to no training at all. HealthVer COVIDFact SCIFACT Table 4 : Performance of the Multitask, Pipeline, and MT / PI modeling approaches on SCIFACT instances with rationales that are self-contained (can be interpreted in isolation) or context-dependent (must be interpreted in the context of the abstract). Evaluation is performed in the abstract-provided setting. We report abstract-level metrics; sentence-level results are similar. The %∆ indicates the drop in F1 score on contextdependent instances relative to self-contained instances. Multitask suffers the smallest performance loss, while MT / PI suffers the largest. stance, an annotator indicated whether the annotated rationales were "self-contained" -i.e. sufficient to justify the fact-checking label when taken in isolation, or "context-dependent" -i.e. only sufficient when taken in the context of the abstract. Figure 1 and Table 4 compares the performance of the three modeling approaches on instances with selfcontained vs. context-dependent evidence. We find that all approaches have lower performance on context-dependent instances relative to selfcontained instances, but the size of the performance drop varies widely. The Multitask approach performs 14.0% worse on context-dependent instances, while the Pipeline approach performs 22.8% worse. Most interestingly, MT / PI performs 66.4% worse, driven predominantly by low recall. The MT / PI model frequently (and correctly) predicts that context-dependent rationales are not sufficient to justify a SUPPORTS / REFUTES decision. These findings suggest that (1) the Multitask approach is, as expected, best at verifying claims with context-dependent evidence, and (2) the Pipeline approach has, in effect, over-fit to context-dependent rationales and learned to make predictions based on insufficient evidence. 6 Unlike Choi et al. (2021) , we do not include the presence of acronyms as "context-dependent," since an acronym can be matched with its expansion based on surface-level textual features. See Appendix C.2 for further analysis of acronyms. Table 5 : Performance on SCIFACT in the "abstractprovided" setting. Models exceed human agreement as measured by sentence-level F1, but not abstract-level. To determine an "upper bound" on the achievable performance of scientific fact-checking models, we assign 151 claim-evidence pairs from SCIFACT for independent annotation by two different annotators. We estimate human-level performance by treating the first annotator's results as "gold," and the second annotator's results as predictions. For comparison, we make predictions using MULTIVERS and our two baseline models, with candidate abstracts provided as input. The results are shown in Table 5 . Existing systems already exceed human agreement for sentence-level evaluation, but not abstract-level, indicating that experts tend to agree on the overall relationship between claim and abstract, but may disagree about exactly which sentences contain the best evidence. This constitutes another reason not to rely solely on selected rationales when predicting a fact-checking label: the choice of rationales is itself somewhat subjective. In addition, these results suggest that one key subtask of scientific claim verification -namely, predicting whether an evidence-containing abstract SUPPORTS or REFUTES a claim -may be nearly "solved" in the setting where (1) the claims are atomic and (2) roughly 1,000 in-domain labeled claims are available for training. Background on scientific claim verification is covered in §2; we discuss other relevant work here. have previously observed that document-level context is often required to properly interpret scientific findings. use an "extract-then-label" pipeline for the original EVIDENCEINFERENCE task. Multitask label prediction and rationale selection was proposed by Pruthi et al. (2020) and applied to sentiment analysis and propaganda detection. As in this work, the authors condition on the predicted label when predicting rationales. Another alternative to supervised rationale selection is to treat rationales as latent variables (Lei et al., 2016; Paranjape et al., 2020) . Long-document encodings for fact verification have been explored by Stammbach (2021) , who use Big Bird (Zaheer et al., 2020) for full-document evidence extraction from FEVER. Domain adaptation for scientific text has been studied in a number of works, including Gururangan et al. (2020); Beltagy et al. (2019) ; Lee et al. (2020) ; Gu et al. (2021) . In those works, the primary focus is on language model pretraining. Here, we focus on training on the target task using out-of-domain and weaklylabeled data. This work points to a number of promising future directions for scientific claim verification. These include applying the approach presented here to develop scientific claim verification models for new scientific sub-domains or other specialized fields given a handful of labeled examples, and extending the task definition to verify claims against longer contexts (e.g. full scientific papers) or larger corpora. Our task formulation also offers an opportunity to study the effects of rationale decontextualization (Choi et al., 2021) , especially in cases where models may be making predictions based on insufficient evidence. In presenting the MULTIVERS system, we addressed two challenges associated with scientific claim verification: incorporating relevant information beyond rationale boundaries by modeling fulldocument context, and facilitating zero / few-shot domain adaptation through weak supervision enabled by a multitask modeling approach. Our experiments show that MULTIVERS outperforms existing systems across several scientific claim verification datasets. We hope that the task, data, and modeling resources presented in this paper will encourage further work and progress towards the broader goals of identifying and addressing scientific mis-and disinformation. One long-term goal of research on scientific claim verification is to build systems that can automatically identify mis-and dis-information, which we believe would be socially beneficial given the current prevalence of mis-and dis-information online. In the shorter term, this work presents two potential risks. First, automated systems for scientific fact-checking are not mature enough to inform realworld medical decisions. We will include a disclaimer with released software to this effect. Second, bad actors could potentially use this work to develop disinformation generators trained to "fool" automated fact-checkers. While this risk cannot be ruled out, we believe that the benefits of publishing this work and making our models available to the community to facilitate further research outweigh the risks that this work will be misused by malicious actors. A.1 Data preprocessing SCIFACT We use SCIFACT in its original form, as it was released by the paper authors (Wadden et al., 2020) . HealthVer The HealthVer (Sarrouti et al., 2021) data release available at https://github.com/ sarrouti/HealthVer appears in NLI format, pairing claims with evidence-containing sentences; the documents from which the sentences were extracted are not provided. We match evidencecontaining sentences to their abstracts in the CORD-19 corpus using a simple substring search, after normalizing for capitalization and whitespace differences. Evidence for which no match was found in the corpus is discarded. We then segment the abstracts into sentences. Any sentence in the abstract with a string overlap of > 50% with the evidence provided in the original data is marked as a rationale. A small number of claims in HealthVer had both supporting and refuting evidence in the same abstract; we remove these claims as well to conform to our task definition. Modeling conflicting evidence is a promising direction for future work. COVIDFact The COVIDFact data available at https://github.com/asaakyan/covidfact is released in a similar format to HealthVer. Like HealthVer, we perform string search over CORD-19 to identify the abstracts containing evidence, and use the same procedure for assigning rationale labels to sentences from the abstract. COVIDFact also includes evidence from sources scraped from the web that are not contained in CORD-19, such as news articles. These sources are not provided with the data release; we discard evidence from non-CORD-19 sources 7 . Refuted claims in COVIDFact are generated automatically by replacing words in the original claim. Based on a manual inspection, we found this process to generate a truly refuted claim roughly a third of the time; in most other cases, it generated a claim that was either ungrammatical or for which the provided evidence was irrelevant. A few cases are provided in Table 6 . FEVER We use the FEVER dataset as-is. EVIDENCEINFERENCE The EVIDENCEINFER-ENCE dataset consists of "ICO" (intervention / comparator / outcome) prompts, paired with labels indicating whether the intervention leads to an increase, decrease, or no change in the outcome with respect to the comparator. The dataset is available at https://evidence-inference.ebm-nlp. com/. We use templates to convert these prompts to claims. See Figure 2 for an example. Rationale annotations are provided for this dataset. Additional examples of templates are below; the full list will be included in the code release. Refuted claims are generated by swapping "increase" and "decrease" templates. Figure 2 : An example showing how an evidence inference prompt (top) can be converted into a claim (bottom) using templates. A refuted claim could be generated by substituting "increases" for "decreases" in the prompt text. PUBMEDQA We use the PQA-A subset released at https://pubmedqa.github.io/, which is filtered for "claim-like" titles. We generate negations by identifying titles with the phrases "does not", "do not", "are not", "is not". "Does not" and "do not" are removed and the relevant verbs are modified to have the correct inflection; for instance "smoking does not cause cancer" is converted to "smoking causes cancer". Similarly, "are not" and "is not" are replaced by "are" and "is". To generate rationales needed to train pipeline models on PUBMEDQA, we employ the following procedure. First, we encode the claim and all abstract sentences using the all-MiniLM-L6-v2 model from the Sentence-Transformers package https://www.sbert.net/. Then, we rank abstract sentences by cosine similarity with the claim and label the top-k sentences as rationales, where k is randomly sampled from {1, 2, 3} with a 4:2:1 frequency ratio (this matches the distribution of k in SCIFACT). Table 7 provides counts showing the number of claim / evidence pairs with each label (SUPPORTS, REFUTES, NEI), in each of our target datasets. Note that a given claim may be (and often is) paired with more than one abstract containing evidence. HealthVer is the largest dataset. COVIDFact is the smallest, in part due to the aggressive evidence filtering described in §A.1. Table 8 provides an example of a contextdependent rationale (as defined in §7.1), as well as an example of a rationale with an undefined acronym. The latter occurs when an acronym appears in a rationale but its full expansion does not; an analysis of undefined acronyms is included in Sars-cov-2 reactive t cells . . . are likely expanded by beta-coronaviruses Sars-cov-2 reactive t cells . . . are not expanded by beta-coronaviruses Successful negation Regn-cov2 antibody cocktail prevents and treats sars-cov-2 . . . On-cov2 antibody cocktail prevents and treats sars-cov-2 infection . . . Ungrammatical; "On-cov2" isn't a scientific entity. . . . immunity is maintained at 6 months following primary infection . . . immunity is maintained at 6 weeks following primary infection Not refuted; The original claim entails the negation. Immunity at 6 months implies immunity at 6 weeks. Appendix C.2. The code and data release will contain full annotations indicating which of the 128 human-annotated examples described in §7.1 are context-dependent, and which contain undefined acronyms. In §7, we report an analysis based on annotations performed on the SCIFACT dataset. These annotations were performed by students and / or professional annotators associated with the authors' research institutions. Annotators were paid between $15 and $20 / hour. We implement MULTIVERS using PyTorch Lightning (https://www.pytorchlightning.ai/), which relies on PyTorch (https://pytorch. org/). Pretraining For pretraining, we train for 3 epochs on FEVER, EVIDENCEINFERENCE, and PUBMEDQA, with the data randomly shuffled. We train on 4 negative samples (i.e. abstracts containing no evidence) per claim, which we find improves precision. We train on 8 NVIDIA RTX 6000 GPUs with a batch size of 1 / gpu (effective batch size of 8), using a learning rate of 1e − 5, using the PyTorch Lightning implementation of the AdamW optimizer with default settings. We initialize from a Longformer-large checkpoint pretrained on the S2ORC corpus . Finetuning For finetuning, we train for 20 epochs on the target dataset (SCIFACT, HealthVer, or COVIDFact) . For SCIFACT, we train on 20 negative samples / claim. To create "hard" negatives -i.e. abstracts that have high lexical overlap with the claim -we create a search index from 500K abstracts randomly selected from the biomedical subset of the S2ORC corpus. For each claim, we obtain negative abstracts by using the VERT5ERINI retrieval system from §3.1 to retrieve the top-1000 most-similar abstracts from this index, removing abstracts that are annotated as containing evidence, and randomly sampling 20 abstracts to be used as negatives during training. Since HealthVer and COVIDFact do not have a retrieval step, they do not require negative sampling, and we train on the original datasets as-is. Retrieval For SCIFACT, we performed dev set experiments retrieving 10, 20, or 50 abstracts / claim, and found that 10 was the best. We use that in our final experiments. No organized hyperparameter search was performed. We consulted with the authors of the Longformer paper for suggestions about good model parameters, and generally followed their suggestions. The loss function in Section 3.1 requires a weight λ rationale . This is set to 15 for all final experiments. We informally experimented with values of 1, 5, and 15; no organized hyperparameter search was performed. We selected the learning rate from the values [9e − 5, 5e − 5, 1e − 5]. We performed all experiments with the same random seed, 76, used by invoking the Category Example Contextdependent Claim: Errors in peripheral IV drug administration are most common during bolus administration Context: OBJECTIVES: To determine the incidence of errors in the administration of intravenous drugs . . . Evidence: . . . Most errors occurred when giving bolus doses Explanation: The evidentiary sentence reporting the finding does not specify the type of error. Claim: Hematopoietic stem cells segregate their chromosomes randomly. we tested these hypotheses in hematopoietic stem cells (HSCs). . . Evidence: . . . indicated that all HSCs segregate their chromosomes randomly. Explanation: HSCs is an acronym for Hematopoietic stem cells. COVIDFact and HealthVer, we follow the instructions in that repository to convert the data to the required format, and train using the available training code as-is, beginning from the available SCIFACT checkpoint. We used Google Cloud TPU for training and inference. PARAGRAPHJOINT We use the code available at https://github.com/jacklxc/ ParagraphJointModel. For predictions on SCIFACT, we make predictions using the publicly available checkpoint. For the other two target datasets, we use the training code in the repo without modification. We used PARAGRAPHJOINT as our baseline for zero / few-shot learning experiments, and hence also performed pretraining on PARAGRAPHJOINT. The repository provides code to train on the FEVER dataset, which we used for pretraining with EVI-DENCEINFERENCE and PUBMEDQA added to the data. In Table 3 , we presented F1 scores for ablations comparing pretraining data, model architecture, and encoder used. Table 9 presents the full results, including precision and recall. In §7.1, we examined the difference in performance on instances with self-contained vs. contextdependent evidence. Here, we show the results of evaluation on instances containing an undefined acronym vs. cases without one. We find that undefined acronyms do not pose a challenge for Multitask and Pipeline, but do cause a small performance drop on MT / PI. In §5.1 we described how, for SCIFACT, we trained on 20 negative abstracts per claim. The effect of training on these additional negative samples is shown in Figure 11 . In the abstract-provided setting, negative sampling is not very beneficial. However, when the model must select evidence from retrieved abstracts, precision drops off dramatically without negative sampling. This is worth noting since it suggests that performance reported when models are provided with "gold" candidate abstracts may not offer an accurate estimate of the accuracy these systems would achieve when deployed in a real-world setting, which could require systems to verify claims over hundreds of thousands of documents. In §5, we discussed how the available scientific factchecking datasets differ in a number of important respects. Here, we explore whether models trained on one system are able to generalize to another despite these differences. We train MULTIVERS on each of our three datasets and then evaluate its performance on the other two. We also train a version of MULTIVERS on all three datasets together, and evaluate on each one. Since COVIDFact has no NEI instances, during evaluation we remove all NEI instances from the other two datasets, and evaluate in the abstract-provided setting. The results are shown in Table 12 . The sentencelevel evaluation results (Table 12b) indicate that none of the datasets generalize well to each other in their ability to identify rationales. The situation is better for abstract labeling (Table 12a) . SCIFACT and HealthVer each generalize reasonably well to each other, but not to COVIDFact. COVIDFact generalizes well to SCIFACT, but not to HealthVer. In general, SCIFACT appears the "easiest" dataset to generalize to; this could be explained by the fact that SCIFACT claims were written to be atomic and therefore simple to verify. Finally, a model trained on all datasets combined manages to achieve reasonable performance across all three datasets, though falling short of the performance of models trained specifically for each individual dataset. SciB-ERT: A Pretrained Language Model for Scientific Text Longformer: The Long-Document Transformer MS MARCO: A Human Generated MAchine Reading COmprehension Dataset Dipanjan Das, and Michael Collins. 2021. Decontextualization: Making Sentences Stand-Alone. TACL BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding Evidence Inference 2.0: More Data, Better Models CLIMATE-FEVER: A Dataset for Verification of Real-World Climate Claims Tristan Naumann, Jianfeng Gao, and Hoifung Poon. 2021. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing 2020. Don't Stop Pretraining: Adapt Language Models to Domains and Tasks PubMedQA: A Dataset for Biomedical Research Question Answering Explainable Automated Fact-Checking for Public Health Claims BioBERT: a pre-trained biomedical language representation model for biomedical text mining Towards Few-shot Fact-Checking via Perplexity Inferring Which Medical Treatments Work from Reports of Clinical Trials Rationalizing Neural Predictions A Paragraph-level Multi-task Learning Model for Scientific Fact-Verification RoBERTa: A Robustly Optimized BERT Pretraining Approach 2020. S2ORC: The Semantic Scholar Open Research Corpus Document Ranking with a Pretrained Sequence-to-Sequence Model Understanding Clinical Trial Reports: Extracting Medical Entities and Their Relations An Information Bottleneck Approach for Controlling Conciseness in Rationale Extraction Scientific Claim Verification with VerT5erini Weakly-and Semisupervised Evidence Extraction Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks The Probabilistic Relevance Framework: BM25 and Beyond. Foundations and Trends in Information Retrieval COVID-Fact: Fact Extraction and Verification of Real-World Claims on COVID-19 Pandemic Yassine Mrabet, and Dina Demner-Fushman. 2021. Evidencebased Fact-Checking of Health-related Claims Evidence Selection as a Token-Level Prediction Task FEVER: a large-scale dataset for Fact Extraction and VERification TREC-COVID: Constructing a Pandemic Information Retrieval Test Collection Overview and Insights from the SciVer Shared Task on Scientific Claim Verification Fact or Fiction: Verifying Scientific Claims CORD-19: The COVID-19 open research dataset Generating Scientific Claims for Zero-Shot Scientific Fact Checking Big Bird: Transformers for Longer Sequences Abstract, Rationale, Stance: A Joint Model for Scientific Claim Verification