key: cord-0595401-iz6yu8t6 authors: Rawat, Mrinal; Kanojia, Diptesh title: Automated Evidence Collection for Fake News Detection date: 2021-12-13 journal: nan DOI: nan sha: c72de5eeab81e87dbbbed6a55da4c9b161571947 doc_id: 595401 cord_uid: iz6yu8t6 Fake news, misinformation, and unverifiable facts on social media platforms propagate disharmony and affect society, especially when dealing with an epidemic like COVID-19. The task of Fake News Detection aims to tackle the effects of such misinformation by classifying news items as fake or real. In this paper, we propose a novel approach that improves over the current automatic fake news detection approaches by automatically gathering evidence for each claim. Our approach extracts supporting evidence from the web articles and then selects appropriate text to be treated as evidence sets. We use a pre-trained summarizer on these evidence sets and then use the extracted summary as supporting evidence to aid the classification task. Our experiments, using both machine learning and deep learning-based methods, help perform an extensive evaluation of our approach. The results show that our approach outperforms the state-of-the-art methods in fake news detection to achieve an F1-score of 99.25 over the dataset provided for the CONSTRAINT-2021 Shared Task. We also release the augmented dataset, our code and models for any further research. The ability to consume readily available information from the internet is alarming for both individuals and organizations. The quality of content on social media platforms has been significantly affected due to the spread of fake news, misinformation and unverifiable facts. The current tally of internet users stands at 4.66 billion 2 (Kemp, 2015) ; and many of these users generate, post and consume content without any regulation, in a large number of countries 3 . Due to the unrestricted nature of online platforms, there is a significant increase in the amount of misinformation on social media (Allen et al., 2020) , especially in developing nations (Badrinathan, 2020; Wasserman and Madrid-Morales, 2019) . Studies show that events such as the presidential election of the United States in 2016 were affected due to moderated fake news campaigns (Tavernise, 2016) . Shu et al.(2017) (Shu et al., 2017) propose that fake news is intentionally written, verifiably false, and is created in a way that makes it look authentic. Manual efforts by other online platforms such as Poynter 4 , FactCheck 5 , AltNews 6 etc. to detect fake news, requires a lot of human effort and can prove to be cumbersome. Such manual efforts can be time-consuming, challenging, and at times, can also be ineffective as fake news can spread faster than verified claims over social media platforms. Automatic Fake News Detection is a task that aims to mitigate the problem of misinformation with the help of evidence supported by various sources. Most of the approaches in this recently devised task aim to use the classical machine learning-based methods or the recent deep learning-based methods to help classify news items as fake or as real. Initially proposed methods for the task applied machine learning-based techniques but cited insufficient data as a major concern (Vlachos and Riedel, 2014) . Recent deep learning and ensemble approaches (Malon, 2018; Roy et al., 2018) were proposed on the FEVER (Thorne et al., 2018a) and LIAR (Wang, 2017) datasets, and have been shown to perform very well. Studies have proposed a combination of evidence detection with textual entailment concerning the claim (Vijjali et al., 2020) . FEVER Shared Tasks (Thorne et al., 2018b (Thorne et al., , 2019 have helped the automatic fact verification task gather attention towards the problem and helped generate approaches to mitigate the issues with previously proposed solutions. This study shows that our novel approach improvises over state-of-the-art approaches and helps detect fake news related to COVID-19. Our approach performs web-search for evidence collection and uses BERT-score similarity to match the unverified claim with the top-k searches. Further, we propose the use of summarization to mitigate problems with the evidence collection. We summarize the top-n selected lines from these articles and use them as evidence to support or reject the news item claim. Our experiments perform an extensive evaluation of the approach over the datasets released as a part of the CONSTRAINT-2021 Shared Task (Patwa et al., 2020) shared task. Our summarized contributions with this paper are: • We propose a novel approach to help automate the evidence collection for any fake news detection dataset. • Additionally, we incorporate a summarization component that helps outperform the state-ofthe-art approaches for automatic fact verification on the CONSTRAINT-2021 dataset. Automatic detection and classification of fake news, especially in epidemic situations like COVID-19, is a significant issue for society. Most of the recent works have identified that fake news is written intentionally and factually false (Shu et al., 2017) . Several datasets have been released for the AI community in the field of fake news detection, such as LIAR (Wang, 2017), Fake News Challenge-1, and FEVER (Thorne et al., 2018a) . Some recent techniques extract the evidence from Wikipedia to classify a claim as SUPPORTED, REFUTED or NOTENOUGHINFO (Thorne et al., 2018a) . They formulate the problem as a three-step process (i) first the top-k documents are identified based on the TF-IDF based approaches (ii) then top-k sentences are identified from the documents, and (iii) finally the textual entailment based approaches (Parikh et al., 2016) are used to classify the claim. Team Papelo (Malon, 2018) used the Transformer-based approach for the textual entailment and selected the evidence-based on tf-idf and entities present in the title. Hanselowski et al. (2018) selects the documents and sentence using the entity mentions and recognizes textual entailment using Enhanced Sequential Inference Model (ESIM) (Chen et al., 2017) . Despite the several attempts, fake news detection is a challenging problem and countering fake news is a typical issue that requires continuous studies. Recently, some researchers released the datasets related to COVID-19 fake news detection. Shahi and Nandini (2020) (Shahi and Nandini, 2020) proposed the first multi-lingual cross-domain dataset for COVID-19 that consists of 5182 factchecked news articles from Jan-2020 to May-2020. They collected data from 92 different websites and manually classified them into 23 classes. Kar et al. (Kar et al., 2020) also released a multi-indiclingual dataset besides English to detect the fake news in social media tweets. (Chen et al., 2021) trained the model with additional words such as covid-19, coronavirus, pandemic, indiafightscorona since the BERT tokenizer will split these words into separate tokens. Some works leveraged the fine-tuned models like COVID-Twitter-BERT (CT-BERT) (Müller et al., 2020) and demonstrated a boost in performance (Li et al., 2021; Glazkova et al., 2020; Wani et al., 2021) . The fake news detection methods described above mainly uses the claim for the classification. Our method focuses on extracting and summarizing the evidence from the external source and uses it to classify the claim on the COVID-19 fake news detection dataset (Li et al., 2021; Patwa et al., 2020) . Training 3360 3060 6420 Validation 1120 1020 2140 Test 1120 1020 2140 Total 5600 5100 10700 For our work, we use the pre-released COVID-19 fake news dataset as a part of the CONSTRAINT-2021 shared task (Patwa et al., 2020) . This goldstandard manually annotated dataset comprises social media posts and articles which are related to COVID-19. Each post or tweet contains content in the English language and is classified in either of the two categories-(1) Real: where tweets or articles which are factually correct and verified from authentic sources, for example, "Wearing mask can protect you from the virus. (Twitter)"; or (2) Fake: where tweets or posts related to COVID-19 which are factually incorrect and verified as false, for example, "If you take Crocin thrice a day you are safe. (Facebook)". The authors collect fake news from two different sources-social media platforms and public factchecking platforms. The social media posts include text from Facebook posts, Instagram posts, and Twitter posts, whereas the fact-checking websites such as PolitiFact, Snopes, and Boomlive are used to collect fact-checked news items. To further collect real news, they sample tweets from official government channels, news channels, and medical institutes. Overall, a total of 14 such sources were used to prepare this dataset. The dataset comprises 10700 manually annotated samples and is split into (60%) train, (20%) validation and (20%) test sets. We provide the exact numbers for each split/class in Table 1 for clarity. The dataset is class-balanced as it contains 52.3% samples of real posts and 47.7% samples of fake posts. As an analysis on it, we obtained a word-cloud illustration for both real and fake samples and observed a high lexical overlap between both the classes, where words like 'coronavirus', 'covid19', 'people', 'cases', 'number', 'test', etc. are repeatedly used in both the sets. We do not show the word cloud due to space constraints. We create and present a wordcloud for the dataset in Figure 1 . In this section, we provide details of our novel approach to augment the dataset with evidence from web search and the use of this evidence to complement the task of fact verification. The algorithm for our approach can be seen in Algorithm 1. As discussed above, we collect this evidence and prune to top-k related news items based on semantic similarity via BERTScore (Devlin et al., 2019) . We also select top-n lines from each article for further building an evidence repository, as detailed below in further subsections. In the original dataset of COVID-19, evidence is not released along with the claim. We hypothesize that evidence is equally relevant to classify the claim as proposed by Thorne et al. (2018a) . As per our approach, given a claim or post text, we first select K relevant articles using a BERT-based sentence similarity score as detailed here; and can be seen as an architecture component in Figure 2 . For each claim c, we search the claim as a query using a publicly available search API. The response returned by this API consists of (heading, text) pairs. We use the spacy (Honnibal et al., 2020) library to get the similarity score of response text with respect to the input claim. Based on this similarity score, we select top K results that have the similarity score greater than 0.7 7 . While selecting documents, we prune for webpages in other languages and pages which are direct links to PDF or other such non-text files. As an immediate next step, we scrape the selected web pages to obtain the matching N sentences concerning the claim as detailed here. Algorithm 1: Algorithm to collect the evidence from the input claim Input: Claim c, Blocked URLs u Output: Evidence e 1 Function ArticleRet(c, k = 3): 16 return e; In the previous step, we extract the relevant article URLs U = (u 1 , u 2 , u 3 ). We employ a similar method to find the sentences within each article. For every url u, we first scrape the webpage and extract the text from < h > and < p > tags. Further, we use the same similarity score to select the top N sentences with respect to the claim. We obtain a similarity threshold of 0.5 after performing a similar empirical evaluation as mentioned in the footnote. Eventually, we concatenate the selected sentences from these articles, which act as our evidence for the claim. We would like to note that increasing the threshold significantly higher returned an empty set in some case and hence we choose a relatively lower threshold (0.5). An example of evidence collected via our approach is shown in Table 2 where the column titled "Evidence" shows the output after these steps. To map claims with evidence, we pre-process both the dataset and the evidence collected from external sources. Following are the details of the preprocessing steps: (1) URL Mapping: We observe that some posts contain URLs in a masked form, e.g., https://t.co/z5kk XpqkYb. Our approach extracts these URLs using a regular expression-based match and maps them to the original URL using the python 'requests' library. Any additional information from the URL is removed, and only appropriate URLs remain in the text. For example, https://t.co/z5kkXpqkYb − → https://www.cdc.gov/; (2) Special symbols: We removed extra whitespaces, special symbols and brackets like, (, ), Summarization-1 (S1) Summarization-2 (S2) There is no evidence that children have died because of a COVID-19 vaccine. No vaccine currently in development has been approved for widespread public use. https://t.co/9ecvMR8SAf Currently there is no coronavirus vaccine that has been approved for the American public. And there is no evidence that children have died because they received one of the COVID-19 vaccines being developed. PolitiFact found no evidence that anyone has died from complications related to a trial COVID-19 vaccination. There is no evidence that children have died because of a COVID-19 vaccine. There is no evidence that children have died because they received a COVID-19 vaccine. No evidence that anyone has died from complications related to a trial COVID-19. There is no evidence that children have died because they received one of the COVID-19 vaccines being developed. PolitiFact found no evidence that anyone has died from complications related to a trial COVID-19 vaccination. Similarly, we also replace mentions "@" with "MENTION:" token. At the end, we convert emojis to their text form using the 'demoji' library 8 ; (4) Lowercasing: Eventually, we lowercase the claim and the evidence text to obtain the input data used for the next Summarization step. Our pre-processed evidence for claims in many cases were multiple paragraphs resulting in performance degradation. Therefore, we propose the addition of a summarization component to our pipeline which utilizes state-of-the-art Text-to-Text Transfer Transformer (T5) language model (Raffel et al., 2019) for the inherent summarization task 9 . Due to the nature of the usual summarization task input, a large body of text (full documents), we believe that our comparatively short paragraphs would be better summarized. This language model is fine-tuned for the task of summarization helps us obtain a summarized text for each piece of evidence resulting in what we call Summarization-1 or S1. An output obtained is shown in Table 2 . As an additional experimental step, we further finetune the Text-to-Text Transfer Transformer (T5) model using the original FEVER dataset (Thorne et al., 2018a) . The original T5 summarization model is trained on the CNN/Daily Mail (Hermann et al., 2015) data where the input is the news article text, and the objective is to highlight summarized text as the output. The T5 is an encoder-8 GitHub: Demoji 9 This language model can perform the summarization task with the help of a prefix "summarize" to the input text provided. decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format. This allows for the use of the same model, loss function, hyperparameters, etc. across our diverse set of tasks. T5 works well on a variety of tasks out-of-the-box by prepending a different prefix to the input corresponding to each task, e.g., for the task of translation− → translate English to German: , for the task of summarization− → summarize: . For our experiments, the aim is to summarize the pre-processed evidence while including the claim. Thus, we hypothesize that fine-tuning on an auxiliary dataset will improve the quality of the generated summary. For fine-tuning, we use the same hyperparameters as described in their paper to generate another model. We perform another iteration of the summarization step using this fine-tuned model to generate a parallel set of evidence and label the output as Summarization-2 or S2 as shown in Table 2 . Further, we provide the details of as the classification task, which uses either S1 or S2 as evidence input to classify the claims as real or fake (Figure 2 ). In this section, we discuss the experiment setup in detail. We perform the task of fake news detection as a binary classification task in a supervised setting. We choose to perform our experiments with both conventional machine learningand deep learning-based classifiers. From the machine learning-based approaches, we choose Logistic Regression (LR) and Support Vector Machines (SVM) with the GridSearch implementation for best results over multiple hyperparameters (values of c, different kernels, etc.) We also utilize LSTMs with various contextual language models from the deep learning methods. From the deep learning-based approaches, we use a simple LSTM Table 3 : Results obtained after the fake news classification task where the values for previous approaches are from the latest shared task results and the results for each iteration of our approach are shown [P (Precision), R (Recall), and F (F-Score)]. (-) − → No Evidence, S1 − → Summarization-1 as Evidence, S2 − → Summarization-2 as Evidence. Our Approach w/ various Deep Learning Classification methods BERT base RoBERTa base XLNet base -S1 S2 -S1 S2 -S1 S2 P 0. Table 4 : Results obtained after the fake news classification task where the results for each iteration of our approach with various deep learning classification methods are shown [P (Precision), R (Recall), and F (F-Score)]. (-) − → No Evidence, S1 − → Summarization-1 as Evidence, S2 − → Summarization-2 as Evidence. implementation with pre-trained GloVE 10 vectors, BERT base , RoBERTa base , and XLNET base -based classifiers. Our LSTM implementation uses Adam optimizer with a learning rate of 0.001, and 256 as the batch size. For classifiers based on BERT base , RoBERTa base , and XLNET base , we use the Hug-gingFace implementations with a batch size of 32, L2 regularization and cross-entropy loss. The regularization parameter λ was set to 0.1. Each classification method is iterated (1) without evidence (-), (2) with augmented summarized evidences from S1, (3) and then with S2, thus giving us three sets of results for each method; as shown in Table 4 . As an input to the classifier, we use the claim asis from the dataset as described above. We have a dataset D = (x n , y n ) N n=1 comprising of N training samples. Here x n = (c n , e n ), where c n represents the claim, and e n represents evidence gathered using our approach. X ∈ X is defined on input space, and Y ∈ Y = {0, 1} are the corresponding labels. Thus, given a claim c and evidence e, the aim of this task is to train a classifier such that the claim c is predicted as fake news or not, i.e where F (c, e) is the function our model aims to learn over each iteration or epoch. 10 https://nlp.stanford.edu/projects/glove/ The results for our classification task are shown in Table 4 . Using our approach, we are able to marginally outperform (+0.24, F-Score) the previous state-of-the-art (SoTA) approaches for the task of fake news detection as shown in the last column (XLNet base , S2). Even the RoBERT base model is able to outperform the SoTA approaches by a small margin. We present the values of our top two best models in boldface in Table 4 . Although the improvement margin is small, we would like to note that the previous SoTA approaches are already performing at almost a 0.99 F-score. We executed our model run multiple times to ensure that our improvement margin is indeed truly obtained. We also observe that RoBERTA base , and XLNet base outperform the SoTA approaches (Chen et. al. / Li et. al.) even with S1 summarization component. Classical machine learning-based approaches are also shown to perform very well for this task as the scores of 0.96 can be considered to be a good performance for any classification method. However, this is not the only key takeaway from these results. We observe that by using our novel approach, a consistent improvement is seen in the task results. The efficacy of our approach can be seen from Table 4 , as either S1 or S2 consistently outperforms all the base models (-) [no evidence] in the table. Moreover, using our approach, we are able to gather key evidence for such a dataset where, to begin with, only claims were present with manually annotated labels. Table 5 illustrates the success and failure cases from XLNet, Logistic Regression and Support Vector Machines. We observe that first two cases were incorrectly predicted by the XLNet but were predicted correctly by LR and SVM. Last case was predicted incorrectly by all of the models. We believe that the absence of source in the text could be a potential reason for this failure. Our approach can gather the evidence using a fully automated method with summarization component(s) in the pipeline. The importance of this component can be gathered from manual observations of examples in the augmented dataset. We observe that summarized evidences shorten the length of the evidences, which helps the Transformer architecture-based classifiers like BERT base , RoBERTa base , XLNet base perform better. These pre-trained models have a token length limitation of 512 tokens which is easily able to capture our summarized evidence. We also manually observe that the summarization component helps reduce redundancy in the generated sentences and removes duplicates. Hence, improving the quality of evidence used as additional input helps reduce the training time. The performance of our models with the fine-tuned summarization component (S2) seems to perform better than S1, and the model without any evidence, as can be seen in Table 4 . We acknowledge that the CONSTRAINT dataset is saturated in terms of possible improvements. However, with this paper, our aim is to show the efficacy of our summarization technique which can help the evidence detection for news. We chose this dataset at an early stage of our work, and our experiments do show that improvements can, in fact, still be shown on this dataset. Our best-performing system surpasses the state-of-the-art by 0.23% points. In this paper, we present an automated method to collect evidence for the fake news detection task. We use our novel approach to augment the dataset, released in the CONSTRAINT-2021 Shared Task, with evidence sets collected from the web. Our method helps process these evidence sets, clean them and use them to generate summarized evidence based on two different methodologies. We use either of the summarized evidence as an additional input to the fake news classification task and perform an evaluation of our approach. We discuss the results of the classification task and conclude that our approach helps outperform the previous SoTA approaches by a small margin, however, helping generate evidence for a crucial dataset. We show that a summarization module can help collect evidence more effectively. We augment this dataset with the summarized evidence and release it along with the code and generated models for further research. We would also like to conclude that our method is generalizable; since it uses pre-trained metrics (BERTScore) and models (T5), it can be used to gather evidence for other datasets. The overall pipeline is also not very time-consuming (2 seconds per sample) once fine-tuned models are included in it. We hope our method and the resources are helpful to the NLP community. In future, we would like to use our method to gather evidence for other fact detection/verification datasets as well. Our initial aim is to reproduce this study with other datasets and ensure that our method performs well in a real-world scenario. We would also like to apply this method and gather further evidence for existing fake news datasets, and perform our experiments to evaluate this approach over multiple exisiting datasets, including existing multilingual datasets. Evaluating the fake news problem at the scale of the information ecosystem Educative interventions to combat misinformation: Evidence from a field experiment in india Transformer-based language model fine-tuning methods for covid-19 fake news detection Enhanced LSTM for natural language inference BERT: Pre-training of deep bidirectional transformers for language understanding g2tmn at constraint@aaai2021: Exploiting CT-BERT and ensembling learning for COVID-19 fake news detection UKP-athene: Multi-sentence textual entailment for claim verification Teaching machines to read and comprehend Sofie Van Landeghem, and Adriane Boyd. 2020. spaCy: Industrial-strength Natural Language Processing in Python No rumours please! A multi-indic-lingual approach for COVID faketweet detection Global digital & social media stats: 2015. Social Media Today Exploring text-transformers in aaai 2021 shared task: Covid-19 fake news detection in english Team papelo: Transformer networks at FEVER and Per Egil Kummervold. 2020. Covid-twitter-bert: A natural language processing model to analyse COVID-19 content on twitter A decomposable attention model for natural language inference Amitava Das, and Tanmoy Chakraborty. 2020. Fighting an infodemic: Covid-19 fake news dataset Exploring the limits of transfer learning with a unified text-to-text transformer Vikram Keswani, and Vasudeva Varma. 2021. Identifying COVID-19 fake news in social media A deep ensemble framework for fake news detection and classification Fakecovid -A multilingual cross-domain fact check news dataset for COVID-19. CoRR, abs Saiful Islam. 2021. A transformer based approach for fighting COVID-19 fake news Fake news detection on socialwang media: A data mining perspective TUDublin team at Constraint@AAAI2021 -COVID19 Fake News Detection. arXiv e-prints As fake news spreads lies, more readers shrug at the truth. The New York Times FEVER: a large-scale dataset for fact extraction and VERification Christos Christodoulopoulos, and Arpit Mittal. 2018b. The fact extraction and verification (fever) shared task Christos Christodoulopoulos, and Arpit Mittal. 2019. The FEVER2.0 shared task Two stage transformer model for COVID-19 fake news detection and fact checking Fact checking: Task definition and dataset construction liar, liar pants on fire": A new benchmark dataset for fake news detection Evaluating deep learning approaches for covid19 fake news detection An exploratory study of "fake news" and media trust in kenya, nigeria and south africa