key: cord-0627859-9khqpxa2 authors: Jachim, Peter; Sharevski, Filipo; Pieroni, Emma title: "TL;DR:"Out-of-Context Adversarial Text Summarization and Hashtag Recommendation date: 2021-04-01 journal: nan DOI: nan sha: 4408a1ceb906652c0263f2c394d8e681c66137c6 doc_id: 627859 cord_uid: 9khqpxa2 This paper presents Out-of-Context Summarizer, a tool that takes arbitrary public news articles out of context by summarizing them to coherently fit either a liberal- or conservative-leaning agenda. The Out-of-Context Summarizer also suggests hashtag keywords to bolster the polarization of the summary, in case one is inclined to take it to Twitter, Parler or other platforms for trolling. Out-of-Context Summarizer achieved 79% precision and 99% recall when summarizing COVID-19 articles, 93% precision and 93% recall when summarizing politically-centered articles, and 87% precision and 88% recall when taking liberally-biased articles out of context. Summarizing valid sources instead of synthesizing fake text, the Out-of-Context Summarizer could fairly pass the"adversarial disclosure"test, but we didn't take this easy route in our paper. Instead, we used the Out-of-Context Summarizer to push the debate of potential misuse of automated text generation beyond the boilerplate text of responsible disclosure of adversarial language models. format. The simplicity associated with text summarization is preferable to lengthy readings, and therefore leaves the technique ripe for adversarial misuse. Social media has become an outlet for news, and users on sites like Twitter or Parler rely on short pieces of text in order to understand polarizing topics or narratives. Text summaries of news articles are continuous feeds of fresh content that can be used by adversaries in a variety of different ways, including as newsletters, or continuous Twitter or Parler content. If an adversary were to upload a summary to Twitter, it would need to fall within the character limit or use a conversational thread with few consecutive tweets [57] . Similarly, Parler deploys a "Read More..." functionality which deters the use of messages longer than 1000 characters [40] . Studies suggest that when users are scrolling social media, particularly looking for political content, they are conditioned to respond well to relatively short messages or threats that evoke a feeling of being informed [2, 18] . Even more, people tend to take content from trusted sources like legitimate news articles at face value as a heuristic that saves time [64] . Text summarization is able to satiate the need for speedy news consumption, by giving users the "tl;dr" critical information from an article, so that a user can keep scrolling. Summaries are usually accompanied with embedded features like links or hashtags in order to facilitate content diffusion to reach a broad audience in a crowded social media landscape [35] . Understanding that this increases audiences' exposure, adversaries were quick to diffuse alternative narratives on social media including polarizing hashtags and links to posts on less regulated platforms like 4chan [25, 57] . The addition of links is to avoid hard moderation by mainstream platforms for openly disseminating trolling content. The particular addition of hashtags, which help to organize content and track discussions based on keywords, is not to avoid moderation but to yield inferences about the summaries that reinforce the polarizing positions [64] . In anticipation of an adversary utilizing the approach to bolster a position, we have also designed Out-of-Context Summarizer to predict potential hashtag keywords to accompany the suggested out-of-context summary. Often, in trying to understand trolling online or detect it, researchers utilize complex automated language models [4, 14] . While these models certainly have the potential for adversarial misuse, the complexity associated provides little incentive for trolls. Yet the text summarization method used in Out-Of Context Summarizer provides a relatively quick and efficient way to generate adversarial narratives from already existing pieces of public discourse like news articles. Using machine learning, Out-of-Context Summarizer "borrows" the experience of journalists who understand their audiences in order to weaponize that relationship, by swaying the summarization to seem either left or right-leaning, without raising alarm from users who trust the new sources they read. After generating alternative narratives under the guise of legitimate journalism, trolls can upload summaries quickly to social media sites like Twitter or Parler, to change the context of the narrative. By manipulating the summarization to favor either conservative or liberal sentences, the perspective of the polarizing topic is manipulated, but remains undetectable, as the selected sentences are actually present in publicly available news articles. Traditional trolling may be more easily detectable, as the alternative narratives are often known and also widely dispelled, but by advancing adversarial agendas through legitimate pieces of public discourse, trolling becomes much harder to identify and poses a security risk worth disclosing. Automatically condensing large texts to contain few important sentences has been, and remains, an imperative of many language models in a sincere effort to maximize the entropy of information delivered to a user for consumption in minimum required time [17, 26] . The sheer volume of online text content in particular makes it hard for users to sift through textual data that is often redundant, unverified, and incorrectly formatted. Robust automatic text summarization systems, therefore, received a widespread attention from the natural language processing and artificial intelligence communities. These systems usually are classified based on the language, input size, summarization approach, nature of the output summary, summarization algorithm, summary content, summarization domain, language, and type of summary [11] . The Out-of-Context Summarizer manipulates a single neutral input article, based on a machine learning model trained on a corpus of polarized texts, to create a polarized article summary, as described in more detail in Section 3. The modular design of the system allows the adversary to quickly change the contexts of the system to target different articles in different contexts, or even different languages. The summarization approach is extractive-our Out-of-Context Summarizer selects sentences from the input article and then concatenates them to form the summary. In Section 7 we expand on the future adaptation for generating abstractive summaries, by using sentences that are different than the original sentences. This adaptation is possible given that approaches for automated word manipulation in various forms of online text such as email [47] , Twitter posts [25] , Wikipedia articles [46] , or text converted to speech by an voice assistant [48] exist. The nature of the the Out-of-Context Summarizer output is generic, but the hashtag generator can be repurposed to be used for querying particular keywords for appending the generic output, as shown in Section 4. The summary output of Out-of-Context Summarizer algorithm is both indicative, alerting the user about the source content, and specifically informative, outlining the main contents of the original text (though with an original twist, which makes the Out-of-Context Summarizer appealing for adversarial text summarization). Out-of-Context Summarizer can produce various lengths of summaries, including headlines, sentence-level, highlights, or a full summary [9] . Although we focus on full-text summaries, in Section 7 we provide examples of more abbreviated outputs. We also focus on a specific domain of temporary events like elections or the COVID-19 pandemic, but Out-of-Context Summarizer could be customized to produce summaries from different domains, for example incorporating elements from medical documents or journal papers to supplement the COVID-19 summaries. We also focus on creating alternative narratives out of context from news articles, but one might use Out-of-Context Summarizer for example to summarize a legislative bill for the same purposes. The scope of our adversarial text summarization mechanism includes: the gathering and structuring of data, data modelling, and text summarization aided by analysis of the fitted model. Figure 1 shows the high-level overview of the language model behind the Out-of-Context Summarizer. The out-of-context summarization starts when a website of interest publishes an article. The text from the article is then scraped and formatted into a cohesive, labelled dataset. The labelled dataset is passed to a script that fits a machine learning pipeline to the labelled dataset, and saves the fitted pipeline. Finally, the text summarization process takes the fitted model, analyzes the prediction probability to decide the out-ofcontext weights, and creates text summaries that can be passed along to the delivery mechanism. The sources used can be adjusted, and new scrapers can be created to enable the adversary to choose from a wide range of online news outlets. For the purposes of this paper, we trained the classification model used in the Out-of-Context Summarizer on articles from Vox and Breitbart initially scraped in February of 2021. We selected these two websites to establish liberal and conservative baselines, respectively [58] , but future adaptations could select text from any number of news sources. To demonstrate the Out-of-Context Summarizer capabilities, we generated three adversarial text summaries in three different contexts: first, we summarized the articles included in The Hill as a politically-centered news outlet based on the learned data from Vox and Breitbart; next, we repeated the experiment on March 10 ℎ of 2021 but with a limited scope of the articles of interest to the COVID-19 pandemic; and lastly, we tried creating text summaries from BuzzFeed, while Out-Of-Context Summarizer was again trained on text scraped from Vox and Breitbart. The choice of the articles was to provide for a balanced editorial input with Breitbart on the political right in the U.S., Vox and Buzzfeed on the political left, and The Hill at the center [15] . This was a deliberate choice but for adversarial purposes, the articles can be selected from different sources -not just on the political spectrum but on any polarizing editorial stance. We deliberately avoided using mainstream outlets like Fox News, New York Times, or MSNBC to demonstrate the versatility of the out-of-context language modeling, but also to showcase summarizes that could not be immediately associated with the tune of a mainstream outlet. For the language modeling, we must note, an adversary might not be using news articles per se; they can train the Out-of-Context Summarizer on the textual portion of available Twitter information operations datasets [57] , for example on a particular topic like COVID-19 [25] . To obtain the data, we scraped four sites on the political spectrum: The Hill, Buzzfeed, Vox, and Breitbart (shown in Table 1 ). To scrape the data, in our baseline we used a python script that leveraged the requests and bs4 (Beautiful Soup 4) libraries, to extract and parse the HTML for each site. To find the links, we found high level pages that linked to each of the articles we were interested in, collected the links, and used those links to figure out which text we needed to collect. Due to the context in which we built the Out-of-Context Summarizer, we decided that it was easier to manually delete long passages of JavaScript, and it would allow us to avoid repeatedly sending traffic to the Breitbart and Vox pages to scrape the data. An adversary could use regular expressions to avoid downloading the JavaScript in the first place with additional experimentation with the scraper. We decided to demonstrate the Out-of-Context Summarizer by altering the context based on the U.S. political spectrum. In order to make the sentences appear right or left leaning, we needed an automated way to tell whether a piece of text is left or right-leaning. To accomplish this and meet the requirements we anticipate that an adversary would encounter in a practical scenario, we wanted a pipeline that could be re-trained for different contexts, performed reasonably well on imbalanced datasets with minimal tuning, and we needed it to provide us with a probability of a sentence being conservative or liberal. We expected that someone trying to use this method would not necessarily have the technical skills required to fully understand how the language model works, so it needed to perform well out of the box. To meet these requirements, our final model pipeline was as follows: TF-IDF is a text vectorizer that weights each word in a sentence so less common words have a bigger impact on the value used to represent each word. We did customize this step a little bit, to include bigrams and trigrams, as well as to provide the standard list of English stopwords that Scikit-Learn provides. To balance the classes, we used the Synthetic Minority Oversampling TEchnique, or SMOTE. SMOTE generates new vectors to represent sentences in the less dominant class, to make it so that the model has the same number of conservative and liberal sentence vectors. Finally, we used a Random Forest Classifier to actually make the predictions. We selected the random forest classifier [3] because it works relatively reliably with minimal tuning (in fact we got reasonable performance without tuning the model at all), it is relatively fast, and the Scikit-Learn implementation of the RandomForestClassifier has a built in .predict_proba() method that returns the probability for each prediction [36, 51] , which is required to weight sentence scores for the text summarization. The model performance metrics are outlined in Table 2 . Overall, these classifiers performed reasonably well without any special hyperparameter tuning, and all of these models were trained on identical pipelines. Each of the classification models that we set up had a class imbalance, with the most dramatic being the class imbalance with the results for COVID specific articles. While our scraper sampled 23,089 sentences in Vox articles discussing COVID-19, it only found 693 Breitbart sentences discussing the pandemic. Despite the relative imbalance, the SMOTE oversampling improved the recall of the Breitbart sentence identification so all Breitbart sentences were identified, though the precision lacked, and only 58% of those samples were confirmed to be Breitbart sentences. For our application, the weak precision, means text summaries for COVID-19 related articles in The Hill are more likely to incorrectly include sentences that are liberal in conservative summaries. The success of these models without hyperparameter tuning, means that generating summaries for new sources, or based on new types of data is as simple as retraining the model on different data. We acknowledge that in addition to the systematic differences between Vox and Breitbart articles which we attempt to model, the classifiers identify other patterns. These patterns include differences between the style guides that the different news sources use, as well as random noise in the word choices used by different authors or editorial approach. To look inside the model we used the LIME representations [42] , which we calculated using the eli5 python library [28] , then we tested the model using a few sentences that we wrote (i.e. they did not appear in the dataset). One of this sentences was: "Alexandria Ocasio-Cortez's terrible green new deal will destroy our way of life. " Between Breitbart and Vox, the classifier determined that the sentence had an 87.8% chance of being in Breitbart, based on the permutation importances in Table 3 . It looks like the language that makes the text more likely to appear in Breitbart is the words based aound the "way of life," and the name "Alexandria Ocasio-Cortez. " Additionally, some adjectives like "terrible" seem to register as being more typical among the Breitbart sentences. This could reflect a difference between the two style guides, where Vox may not give writers as much freedom to use adjectives as writers in Breitbart, or it could be that the Vox editors are more likely to find alternatives for those specific adjectives. It is also possible that the data we scraped happened to only include the word terrible in the text scraped from Breitbart, and overall, if we had scraped more data, then the word "terrible" would appear as frequently (or even more so) in Vox than in Breitbart. Our text summarization model is a modification of an extractive text summarization model [12] to select the sentences that the classifier determines to be the most likely to be from Vox or Breitbart. To make the summaries seem conservative or liberal, we use the probability that a sentence is in a class. The base score that we started out with, and built up from, is a sum of the total appearances of each word in the article for each word in a sentence. To calculate the score for a given sentence , we are adding up the total number of appearances of each word in the full article: On its own, the probability that the model thinks that a sentence is liberal or conservative doesn't weigh the sentences enough. To overcome this, for 1 the probability that a sentence is in class 1, we are taking it to the power : An additional benefit of this parameter is that it allows the adversary to consciously throttle how much of an impact they would like for the machine learning model to have on the sentence summaries, allowing the adversary to use a boiled frog approach and gradually increase how much influence the prediction probabilities might have on the final output. In real articles, some sentences are longer than others. We want to make sure that the sentences that are very long are not considered more important than others, because they are able to include more words. For the length of a sentence (calculated by the number of words in the sentence), we used a simple binomial expression , with upper and lower cutoffs: We wrote this function like this to allow for rapid experimentation with different first and second order functions. So far we have set as −1 2 11 , as 0, and as 1.1. This could also be adjusted to even further reduce the length of sentences, or the values could be reduced to allow longer sentences, or dropped entirely. In practice, one benefit of the weight function is to help correct errors with the sentence tokenizer function separating articles into sentences, which occasionally misses a sentence break. In the event that adversaries might utilize out-of-context summaries on social media platforms like Twitter or Parler, they may seek to manufacture hashtags to accompany the text and potentially amplify the reach of the content. To this end, Out-of-Context Summarizer can also extrapolate potential hashtags regarding a specific article. While a hashtag is too reductive for summarization, it nonetheless can grab a user's attention enough to read a summary accompanying it. Such an approach for crafting politically polarizing Twitter hashtags was used by the Russian troll farms during the US elections in 2016 [1] , for example. To start out with, we calculated the expected number of times that each word would appear in the article and we created a contingency table where each cell represented the number of times a single word being used in Breitbart or Vox, with representing the total number of times in the dataset a word was used, and representing the number of words in a specific news source. We used to represent the total number of occurrences of all words in Vox and Breitbart. The equation below provides a table of the expected values: A short list of 15 hashtags is initially complied, based on which words were more likely to make a post seem politically polarizing. To identify words more closely associated as being interpreted as either liberal or conservative, we identified a hashtag score ℎ for a word that we multiplied the number of times a word appears in an article , by the feature importance of the feature, by the difference between the number of times a word appeared in the Breitbart or Vox dataset and the expected number of times each word would appear: The list of individual words gives an opportunity to leverage fifteen keywords that the adversary can choose between to create hashtags to supplement the text summaries. It is relatively easy to extend the Out-of-Context Summarizer to combine the keywords into bigrams or trigrams for more realistic hashtags, but we opted out not to do in order to provide the opportunity for a user to give the "final touch" of combining reinforcing hashtags. Another reason is to avoid difficult cognition of the Out-of-Context Summarizer output because concatenated text strings in a hashtags are harder to decipher without spaces between words [35] . With that, we wanted to allow for users to have an opportunity for ownership in the adversarial text summarization, or more so, to present our language model as a possible trolling content feeder. We are aware that this could be an inhibiting factor for defenses, because adversaries could use one keyword and then add words outside of the recommendations in an unpredictable manner, and we discuss the human agency in weaponizing adversarial text summarization in our extensive ethical treatise in Section 6. In this section we provide example output of the Out-of-Context Summarizer from the datasets mentioned above. We chose to show verbose summaries that balance our the lower bound of 280 characters posts imposed on Twitter and the upper bound of 1000 character limit or parleys imposed by Parler. We chose this because one can easily create a conversational model by splitting the verbose summary into two or three consecutive tweets in a thread. Table 4 provides an example liberal-leaning summary of The Hill's article on a proposed way ahead after a couple of unusually polarizing and dramatic election cycles [30] . The liberal summary discusses the "heroic efforts" of election workers, and it discusses the roles that businesses play. It also mentions wealthy individuals, which makes parallels the main premise of the article that funding an election defaults to wealthy individuals if Congress is not willing to step in. The recommended hashtag terms provided a couple fruitful combinations, which appear left-leaning, at face value, in the position against "big money": #AmericanCrisis, #NeedAction, #WorkersSupport. In the context of social media trolling, the summary is 697 characters long and could be split in three posts to create more of a conversational model of trolling [21] for Twitter and be taken verbatim to Parler. Why Congress should provide sustained funding in elections In addition to the heroic efforts of election workers, private philanthropy and businesses stepped up to fill some of the gaps after Congress deadlocked on a bill that would have provided increased emergency funding for states and localities to administer elections. While community leaders have important roles to play in fostering civic engagement, funding elections should not be the responsibility of businesses, nonprofits, or wealthy individuals. Lawmakers from both parties must uphold their Constitutional oaths to support a democracy that guarantees free and fair elections -and they must recognize that free and fair elections depend on both preparation and sufficient, steady funding. Hashtag Keywords Recommendations like, new, support, way, need, help, year, community, important, workers, start, crisis, American, address, action The conservative summary of the same article shown in Table 5 builds on the roles Congress could have in ensuring safe and secure elections without directly undermining the option for private election funding. The summary ends with a sentence that discusses the "critical role" that "private donors" play, emphasizing a position of approval for the support coming from wealthy individuals. Though both instances summarize the article fairly effectively, both of the summaries play up one side of the discussion a little bit more heavily. Because all of the sentences were written by a human writer from The Hill, neither of the summarizations are conspicuously liberal or conservative. If the summary raised the reader's suspicion, they would find each of the offending sentences in the Why Congress should provide sustained funding in elections Members of Congress, however, take such an oath, and it should be their responsibility to ensure that American elections are both safe and sufficiently funded. The challenges to holding safe and secure elections in 2020 were imminently predictable, and the need for robust federal assistance was clear since last spring -for items like personal protective equipment, poll worker recruitment, and additional equipment to process the record numbers of absentee ballots. This year, private donors played a critical role in ensuring our nation could hold our elections. Hashtag Keywords Recommendations 2020, election, states, federal, members, private, run, safe, general, congress, responsibility, nation, funding, director original article. The recommended conservative-leaning hashtags give some promising ideas, particularly #ElectionResponsibility, #PrivateFunding, both of which might help to reinforce a threepiece conversational Tweet or a one-piece Parler post with the 566 characters of content. As trolls usually seek to target a specific polarizing topic or event [25] , we also tested Out-of-Context Summarizer on a topic-specific dataset, consisting of articles related to the COVID-19 pandemic. In order to create conservative and liberal summaries of COVID-related articles on The Hill, Vox, and Breitbart, we utilized each site's respective search functionality (additionally, Vox had a dedicated "Coronavirus" tab to pull articles from). Tables 6 and 7 demonstrate the ability of Out-of-Context Summarizer to produce the "tl;dr" of articles related to COVID-19 that could be taken to verbatim or in conversational format to Parler or Twitter. Progress regarding the passing of COVID-19 relief legislation While the Senate made modest changes to the legislation, some of those changes undermined parts of the bill I do support, and others were insufficient to address my concerns with the overall size and scope of the bill, Golden said. But Senate Democrats replaced the GOP amendment hours later with a deal of their own so that the weekly unemployment insurance payments are at $300 and last through Sept. 6. Another major provision of the bill provides $1,400 stimulus checks for individuals making $75,000 or less with phased-out partial payments for those earning up to $80,000. Hashtag Keywords Recommendations pandemic, trump, said, vaccine, going, stimulus, help, testing, republicans, far, democrats, local, unemployment, schools, billion The example article discusses the latest COVID-19 Relief Bill and its intended provisions [29] . Each summary highlights a different part of the article and larger narrative surrounding COVID-19 legislation, in order to take it out of context. In 582 characters, the liberal summary expounds on the specific bill amendments, emphasizing the parts that will directly benefit some constituents, since part of the summary is spoken in the first person tense, it may be more effective for readers. Some of the suggested hashtag terms provided a couple potential combinations that may have the ability to amplify the chosen out-of-context liberal-leaning narrative if posted to social media: #DemocratsHelp and #VaccinesHelp might be used to convey a positive message regarding the impact of Democrat-backed COVID-19 relief legislation while conversely, #RepublicansDontHelp or a hashtag like #TrumpSaid might be used to mock the opposition party and amplify negative sentiment associated with the GOP (used extensively to ridicule and troll on statements about injecting disinfectants, for example). With only 424 characters, the conservative-leaning summary developed by Out-Of-Context Summarizer seems to highlight the lack of bipartisan cooperation for the bill's passing, in stark contrast to COVID-19 relief measures passed in 2020. The recommended hashtag keywords may not have successfully captured the negative sentiment within the Republican Party associated with the passing of COVID-19 relief legislation but an adversary could identify the negative context and alter hashtags while still using recommended keywords like: #CorruptDemocrats, #SenateCronies, or #BidenAd-ministrationFailure. A conversational model of tweets carrying alternative narratives was characteristic for the tweeting during the pandemic and both summaries could pass muster in a series of three consecutive Tweets [7] or one "influencer" Parler post [40] . In the final iteration of Out-of-Context Summarizer, we selected a fourth news source to analyze, BuzzFeed News, since it regularly uploads a lot of text content. We summarized a collection of BuzzFeed articles, including the article utilized in Tables 8 and 9 , discussing the status of former President Trump's post-election court proceedings to challenge the election outcome [56] . The liberal-leaning summary seems to contextualize the failure by the Trump campaign in court after the election, regarding "meritless" claims of interference and even, unsympathetic Trumpappointed judges. In 616 characters, the summary also manages to touch on a major point of polarization, suggesting that the Trump campaign used its post-election court proceedings to "promote lies about widespread voter fraud. " If posting the summary to social media, some of the recommended keywords could be used to construct hashtags like #FakeEvidence, #RepublicansLost, or #BidenWon to potentially amplify the post and its out-of-context narrative. Status of the Trump campaign's post-election litigation They petitioned the Supreme Court to hear a small number of these cases, and the justices either rejected them right away or didn't take any action before Biden was sworn in on Jan. 20, a clear sign that they wouldn't interfere. Trump and his allies had used postelection legal challenges to promote lies about widespread voter fraud, and they denounced the judicial system, including the Supreme Court, as biased when they repeatedly lost. Judges at every level -including some who were nominated by Trump -concluded that these cases were either procedurally deficient or, after reviewing the evidence, meritless. Hashtag Keywords Recommendations filed, right, court, didn, clear, change, republican, evidence, federal, day, small, wasn, won, case, action The lengthier of the two summaries, with 723 characters, the conservative summary is much less harsh regarding the losses, and targeted objective sentences with little opinion and more factual content. Out-Of-Context Summarizer seemingly chose less-emotional sentences which were used in the original article solely to convey the facts without offering much commentary since the outcome was unfavorable to the Republican Party. Some of the recommended keywords were already prevalent in hashtags that went viral at the height of the election challenges including #NotMyPresident, #IllegitimateElection, or #WeLovePresidentTrump, suggesting that Out-Of-Context Summarizer could be used to anticipate narrative trends on social media in the future. For each iteration of Out-of-Context Summarizer, we used the Linguistic Inquiry and Word Count (LIWC) tool to perform a quantitative content analysis of the three summarization datasets (i.e., all compiled original, conservative, and liberal summaries) [37] . Original summaries were not manipulated to appear right or leftleaning, but Out-Of-Context Summarizer was trained on text from Vox and Breitbart, to select liberal and conservative-leaning summaries. LIWC has been extensively used to study the politically polarizing rhetoric on wide range of topics with datasets similar to ours containing news articles [19, 54, 59] . We decided to compare the summarization datasets on the four main categories: (1) analytical thinking; (2) clout; (3) authenticity; and (4) emotional tone. We didn't analyzed each article's three summaries individually, as LIWC requires a larger-scale database to perform effectively. We first performed a linguistic analysis on the initial model of Out-Of-Context Summarizer in which summaries were developed from articles published by The Hill. Table 10 illustrates the results of the LIWC analysis on The Hill dataset. The high "analytical thinking" score for all three summary sets indicate that overall, they follow logical and thoughtful patterns, although this is not highly surprising considering that all summarizations include unaltered sentences from legitimate news articles published by The Hill. Potentially notable is the decreased analytical score of liberal summaries compared to both the unaltered summaries and conservative groups which are fairly similar. All summaries are rated similarly in the"clout" category, for perhaps a similar reason: this metric measures confidence level, and journalists may want to exude a high level of confidence in their writing, in hopes of adding a perception of legitimacy to their work. In the "authenticity" category, the liberal summaries contain the most unique content, but all three groups fall under 50, indicating that the summaries are prone to regurgitate known phrases and narratives. And finally, the "emotional tone metric indicates that while liberal summaries picked slightly more positive sentences than the original group, the conservative summaries chosen were more negative. After the initial iteration of Out-Of-Context Summarizer, we scraped additional text from The Hill, Vox, and Breitbart related to the ongoing COVID-19 pandemic and performed a subsequent linguistic analysis of accumulated summary groups (i.e. unaltered original summaries, liberal-leaning summaries, conservative summaries). Consistent with the LIWC results of the previous The Hill iteration, each group of COVID-19 summaries ranks fairly high in the category of "analytical thinking", with liberal summaries ranking lowest of the three summary groups in both iterations. As shown in Table 11 , all summary groups score similarly in the "clout" category, indicating that each maintains a fairly confident writing style, which is also unsurprising considering the source of the text. In the "authenticity" category, conservative summaries include seemingly more unique content than the original and liberal summaries; conversely, the liberal summaries contain even less unique content than the original group. Finally, each group of summaries has an "emotional tone" score under 50, which indicates a negative and impolite style of writing, with the conservative group of summaries falling even lower in this category. Liberal summaries selected more positive sentences, increasing the "tone" score. This suggests that while both groups of summaries still selected negative sentences, conservative sentences preserved negative content from the original articles, potentially increasing the likelihood of polarization of a topic like COVID-19 if viewed out of context. In our final iteration of Out-Of-Context Summarizer, we targeted an additional news source, BuzzFeed, and scraped news articles, then summarized the article three different times (i.e. unaltered original, left-leaning liberal, right-leaning conservative) and performed a linguistic analysis on each compiled set of summaries, shown in Table 12 . The trend established in the previous two iterations regarding the "analytic thinking" category was present in this dataset as well, with original and conservative summaries seemingly following more logical thought-patterns than the liberal group of summaries. On average, the BuzzFeed dataset had the highest "clout" scores of the three iterations, suggesting the high confidence of writers at BuzzFeed. Of the three groups of summaries, the liberal set scored the highest in "authenticity" suggesting those summaries contained the most unique content. Additionally, in regard to "authenticity" of summary contents from BuzzFeed articles, on average, it contained the least original, unique content of the three datasets we analyzed using LIWC. Potentially most striking is the "emotional tone" category, in which the original summaries scored significantly higher, suggesting a negative tone, but not nearly as negative as either the liberal or conservative summaries, which both seemingly targeted more gloomy sentences from BuzzFeed articles. The tendency of Out-Of-Context Summarizer to select negative or impolite sentences may lead to increased polarization and by extension, a more favorable outcome for potential adversaries. Our implementation of Out-of-Context Summarizer worked easily because we were able to very quickly find a high-quality dataset, which fit our purposes. We targeted content that was easily obtainable, and was homogeneous relative to our other sources, both in terms of its lexicon and the contexts in which the three sources were written, that allowed the machine learning models to effectively generalize to the neutral news source. We selected the news sources because they were especially convenient to access. None of the news sources used were behind a paywall, nor did they require a login, and we did not encounter any defenses to prevent us from using the same IP address to systematically access the websites directly in the order they appear in main pages. Additionally, the language is relatively easy to parse using the bs4 [43] python library, which enabled rapid prototyping and scraping of data. We selected these news sources because all of them are written in a relatively similar context. One news-source that we initially considered for our centrist news source was the BBC. We decided against the BBC articles because they are written using British spelling, so the sentence classification would not be as simple. Additionally, articles published in the BBC are written for an international audience. They do not make the same assumptions or provide details without explanations to the same extent that newsletters tailored toward American audiences might. Protecting against an adversarial implementation of a tool like Out-of-Context Summarizer may not be simple, as it weaponizes legitimate news sources in order to change the context of content, and not the content itself. Potentially one of the best defenses to text summarization being taken out of context is read a piece in its entirety to alleviate the constraints of any tool's inadvertent, or intentional, biases. Yet this is easier said than done. Recently, Twitter tested a means of soft moderation, in hopes of confronting this same issue we now face. Select Twitter users were prompted with a message when trying to retweet a post with an attached article link: "Headlines don't tell the fully story. You can read the article on Twitter before retweeting" [52] . More research must be done into the effectiveness of soft moderation tactics, and warnings targeted at specific content may increase the likelihood of the "backlash" effect when users double-down of previously-held beliefs because of the warning [38] . Researchers in a preliminary study found that some means of soft moderation on social media sites like Twitter are more promising than others, including warnings which fully obscure content like Twitter's "Read First" warning [45] . Levying soft moderation techniques like these may encourage users to read the full context of news content, as well as serve as a reminder of the dangers of out-of-context content in perpetuating the spread of alt-narratives online. For sites like Parler, which prides its brand on the lack of moderation, containing the spread of alt-narratives propagated by humans or machines may deem to be more difficult, as the site is designed to help alt-narratives flourish [40] . The misuse of adversarial AI research, particularly automated language models, appears to be a complex problem that cannot be solved with the boilerplate ethical argument for responsible disclosure of security vulnerabilities [50] . The "security value" of disclosing a vulnerability outweighs the risk of an adversary exploiting it, the argument goes, because the patched system and the improved risk management will make for a harder target in the future. Disclosing a fake text language model might not outweigh the risks of an adversary exploiting it, say for trolling or spreading misleading text summarizations. First, there is no obvious patch because we live in a post-truth society where fake text can serve as truth as long as is sufficiently contrarian to positions that one dislikes [60] . A defender could prevent buffer overflow, but could hardly prevent an overflow of automated text summaries because the summary could be used as a seed for humans to generate and evolve new alternative narratives [55] . Second, managing the risk of trolling or spreading alternative narratives is a thorny issue inhibited with the constitutional right for free speech. Take for example the effort for soft moderation on social media platforms to curb misinformation [8] . Adding warning labels on seemingly fake text, even if human generated, not just alienated users to abandon mainstream platforms like Twitter and switch to alternatives like Parler [62] , but inspired some users' further belief in the alternative narratives promoted with the fake text [33] . The intention to publish automated language models despite the adversarial output, after all, is to increase defenders' understanding of the threat from text generation, and contribute to building tools that could guard against the harms [27, 41, 65] . The Kerckhoffs' doctrine that "security through obscurity doesn't work" is thus thrown into the mix as an argumentation for publishing fake text generators because an adversary would figure out the language model themselves, if they haven't already, and proceed with exploitation (akin to zero-day vulnerabilities) [47] . Adversaries, in the traditional security sense, pose extensive knowledge of targeted systems that help them work through the obscurity or find a zeroday vulnerability. There is no doubt that they could craft a fake text generator, especially after the the revelation of massive human troll farms that generated fake news during the 2020 U.S. Elections [57] . Individual users with virtually no knowledge of adversarial language modeling could become "adversaries", by propagating alternative narratives by leveraging existing tools like GPT-2 or Grover. Adversaries could also use rudimentary text modeling, using tools like the TrollHunter[Evader] that uses a Markov chain to subversively replace target words with replacements in trolling Tweets to evade automated trolling detection [25] . A user needs not to replicate TrollHunter[Evader] to employ the evasion idea when generating their own content. The "vulnerability" revealed by publishing automated language models, then, has to be guarded on two fronts, not just one -the system patching -as in the traditional system security sense. Defenders, initially, need automated counter models to detect fake text generated by automated language models, but they also need assistance when dealing with fake text generated by humans. On the first account, the battle of the machines is underway, and work already exists to detect nefarious text generation for adversarial purposes [65] . On the second account, the machine-human battle is a bit more complicated. Back to the case of soft moderation on Twitter, evidence shows that automated soft moderation for COVID-19 misinformation tweets is prone to mislabeling simply because targets COVID-19 rumor words/hashtags like "oxygen" and "frequency" [61] . Defenders also have to work on the human front to help humans distinguish between fake text generated by a machine and fake text generated by another human. This guard might be even superseded by simply guarding the truth regardless of the text origins, but building such a tool seems quite a Sisyphean task. The fake text as a form of deception is conductive to formation of echo chambers [22] , conspiracy theories [53] , and even entire social media platforms like Parler [40] , which are hard to be eradicated or dispelled in the era of post-truth with any constructive argumentation [38] . One could likely expect a similar effect from fake text generated from automated language models. The "adversary knows the system" might upend the argument for publication of adversarial automated language models but the (dis)balance of offense-defense argumentation, in our view, should be upended by the notion that humans are both the adversaries and the direct victims of fake text. Even before fake generators existed, fake or ill-generated text by humans caused great harm. Take for example the translation of intercepted North Vietnamese messages in the Gulf of Tonkin incident, which replaced the words "comrades" with "boats" to create an impression of an attack that was sufficient to justify a declaration of war by the Johnson administration [20] . No fake text generators, or translators existed in this case, but an adverse effect nonetheless was produced by humans and targeted at humans. Certainly, a publication of a tool for fake text generation or translation could be misused to cause similar harms, but for that to happen, an adversarial intention by a human should probably exist in the first place. and Benefits of Disclosure 6.2.1 Attacking. With the argumentation in mind, we apply the framework proposed in [50] for weighing security costs and benefits of disclosure for our tool Out-of-Context Summarizer. First, we consider factors that affect the adversaries' capacity to cause harm. The first factor is counterfactual possession or the possibility that the would-be adversary would either independently discover the language modeling behind the Out-of-Context Summarizer or learn about it from other adversaries. We believe that it is relatively easy for an adversary to independently discover the language modeling; first, the design described in section Section 2 is based on wellknown language modeling techniques like TD-IDF and SMOTE. Even if we did not publish this paper or the adversary doesn't look into academic papers, they have the option to skim numerous trending blogs of language modeling and find out how these techniques work. The politically polarizing summarization and generation of hashtags is just the flavor we chose to adapt these models into, emphasizing the long-played editorial practice of taking statements out-of-context (also known as "contextomy") [31] . This practice of reporting emerged way before any fake text generators surfaced in academic journals or conference proceedings and resourceful adversaries used it to coordinate humans to generate fake text. Take for example the conservative politicians' quotation of Rev. Martin Luther King in their campaigns to eliminate affirmative action programs in the US [23] . Or the infamous Russian Trolls that took a citation to Hillary Clinton's "mentally ill" out of context to suggest that she incites violence on Trump rallies through the Tennessee GOP Twitter account [57] . The Out-of-Context Summarizer does a distantly related adversarial text summarization in a rather benign fashion because it only borrows from, but it does not make fake replacements in already published articles. The second factor is absorption and application capacity of the adversary or the extent to which they are able to grasp and utilize the full potential of Out-of-Context Summarizer. The barrier for absorption is low because the adversary needs only to "get the idea" of Out-of-Context Summarizer and perhaps even implement it with other language modeling techniques without the need to replicate our work [27, 41, 65] . There is little cost in any path of adaptation. The application capacity could be a bit problematic for Out-of-Context Summarizer, but that's the case for all adversarial or fake text generators. An adversary could select biased news articles or even fake text from GTP-2 or Grover for the training phase and summarize texts with hashtags to assign a meaningful context to otherwise coherent-only content. An adversary could pair the Out-of-Context Summarizer output with the techniques from the TrollHunter[Evader] to combine a hashtag and trolling content that can evade both soft and hard moderation on Twitter on any topic of their choice, not just COVID-19 [25] . Take for example the out-of-context summarization and hashtags from the Centers from Disease Control (CDC) about a potential allergic reaction to the first dose of the COVID-19 vaccine, shown in Table 13 [13]: Getting A COVID-19 Vaccine If you had a severe allergic reaction -also known as anaphylaxisafter getting the first shot of a COVID-19 vaccine, CDC recommends that you not get a second shot of that vaccine. If it is not feasible to adhere to the recommended interval and a delay in vaccination is unavoidable, the second dose of Pfizer-BioNTech and Moderna COVID-19 vaccines may be administered up to 6 weeks (42 days) after the first dose. Hashtag Recommendations #delayvaccination, #analylaxis, #COVID-19 Using the TrollHunter[Evader], an adversary can manipulate the out-of-context summarization to read more coherently by replacing "interval" with "abstination" or replace the #COVID-19 with #COVIDIOT and use the resulting output as an anti-vaccination alternative narrative for the COVID-19 vaccine. Per the "Goldilock zone" reasoning provided in [50] , the Out-of-Context Summarizer fits into the "script kidde" case, given that it requires fairly little capacity to absorb the inner workings of the language modelling, but the application range is quite large (as mentioned in the argumentation section above, an adversary could be anybody who's intent is to communicate fake or manipulated text). 6.2.2 Defending. Next, we consider factors that affect the the defenders' ability to mitigate the potential harms. We extend the framework to consider both automated means of defense by also the human ability to discern fake generated text or text generated for adversarial purpose, because we argued that the humans, or the text users, are both the adversaries and the defenders. For the first factor, counterfactual possession, we provide a defensive discussion on potential ways to detect an output of the Out-of-Context Summarizer with automatic means. We also believe the detection mechanism provided in [65] could be adapted for defensive purposes of the summarized text. A correspondence analysis as proposed in [24] could possibly help in detecting the automatically generated polarizing hashtags. On the human front, we don't believe that the Out-of-Context Summarizer sounds the alarm on adversarial text summarization, given the debate over the fake text generation surrounding GPT-2 [50] , the well known existence of "contextomy" [31] , or the existence of "'belief echoes" [39] . The preliminary effort of "defeneding" from fake or adversarial text with both soft and hard moderation on social media, for example, is out to a rocky start, given that the warnings labels can backfire [8] . But, Twitter recently changed the defense strategy to include strikes [44] before banning someone's account for sharing adversarial text and experimented with crowd "fact checkers" in their Birdwatch program [34] . How these mechanisms will affect the human defense remains to be explored and we believe they will provide the similar aid against the Out-of-Context Summarizer output. For the second factor, absorption and application capacity, we believe that the barrier for absorption is low because the defenders, too, need only to "get the idea" of Out-of-Context Summarizer and perhaps implement a baseline context checks to detect any drift in the context or the summarized text with both numerical or linguistic analysis as we propose in our defense section. Perhaps same for humans, that could possibly manually go through the input articles and compare the Out-of-Context Summarizer summarization, provided they are ready to engage in such a tedious task. Anticipating the applications of Out-of-Context Summarizer might seem like a complex task for both the automated and human defenders, given that the adversarial creativity usually stays one step beyond. Not to be discouraged, because this creativity feeds on the contemporary polarizing events -alternative narratives were generated about the Boston Marathon bombings, the Pope endorsement of the Donald Trump, but also for the human manufacturing of the COVID-19 pandemic and the adverse effects of the vaccine [53? ]. In both cases, the resources for solution finding will be considerable, given the perpetuating adversarial value of adversarial language models for any current or future topic of content. Even if the solutions are available, or could be made available relatively fast, the solutions' effectiveness and the solutions' adoption, especially by the human defenders, could range widely based on the wide options for application of the Out-of-Context Summarizer. We applied the framework to determine the security value of disclosure of our Out-of-Context Summarizer language model. We decided to disclose the idea of adversarial out-of-context text summarization and polarizing hashtag generator with the tool we developed to test such automation. We are aware that the Out-of-Context Summarizer might have a slight "offensive bias" particularly in combining the generated summaries and polarizing hashtags to perpetuate a contrarian position to the mainstream stance on COVID-19, elections, or any highly contested topic. The security value of disclosure, we believe, lays in appending the argumentation for "patching social vulnerabilites" as pointed out in [50] , but also in highlighting the ease with which an adversary could automate and weaponize the human need for "contextomy" to either up the ante on polarization or blackbox probing of defensive mechanisms. On social media platforms with a imbalanced number of "influencers" and "followers" like Parler, an "influencer" could use the Out-of-Context Summarizer to further reinforce a negative sentiment as shown for the case of the notorious conspiracy theory QAnon in [40] . An "influencer" could only use the language modeling as an aid, a seed idea to craft an alternative narrative in a way to avoid soft moderation, birdwatching, or strikes on mainstreams platforms like Twitter too. The "offensive bias" might not come from a single post, but more so as a blackbox probing of Twitter's algorithms or policies by an adversary for which they can use the Out-of-Context Summarizer to generate a vast amount of adversarial text samples accompanied by polarizing hashtags. This comes in handy to append the practice of sharing links by buffering the access to the full content of the articles while capitalizing on their credibility or assumed position on a polarizing issue. The disclosure of our language model, in these regards, serves as an information about a new threat of not just another fake text, but original text automatically taken out of context and possibly appended with polarizing hashtags. It helps the defense, automated or at least the human "birdwatchers, " to conceive an automated text summarization beyond the one-dimensional distributing of fakes and originals, but includes a dimension where original text could be taken out of context by automatic means. The techniques that we used for text-summarization in this paper, with the exception of the use of the classifier, are covered in detail in a paper by Ferreira et al. [12] . Text summarization has also been applied in adversarial contexts, with two examples being the 2008 application of extractive text-summarization as a steganography approach to disguise messages [10] . This approach is similar to ours in that it uses text-summarization as a tool for deceiving people into thinking a summarization is sharing one message, but this approach is not intended to actively persuade the target to mis-perceive the manipulated text. Another paper highlighted a possible application of text summarization to help summarize lots of text data about a rival company to build a concise competitor profile [5] . This paper uses text summarization to quickly perform a lot of research on a target than might otherwise be easily available. Adversarial language manipulation is an active area of research, with a variety of recent developments. One recent example is the python library TextAttack, which provides adversaries with the ability to reliably and precisely change the meaning of text based on a variety of different models [32] . Our Out-of-Context Summarizer does not need to manipulate words within sentences, because it uses selective summarization to show one perspective within an article, and presents subsets of the article to show a biased perspective of the article. Unlike TextAttack, which manipulates the language of a piece of text, meaning that a targeted reader could possibly see differences between the original and the manipulated text with a closer inspection of the text out. There are many optimizations available that could be built into our text summarizations to improve the substitutions' quality. Our implementation used a minimally viable implementation based on word frequency with optional sentence length optimizations. Possible enhancements could include the use of the TF-IDF statistic, lexical similarity of terms, text case, part of speech for word scores, cue phrases, sentence position, sentence centrality, enhancement using numerical data, bushy path of the node, and aggregate similarity [12] . Our text summarization mechanism has a lot of different parameters to determine. The ideal parameters might be different for different people, choice of article, and tuning for the application of taking narratives out-of-context. By using a machinelearning driven approach, such as a multi-armed bandit, we might be able to define the text summary parameters based on how a user reacts to the text summaries. Another option would be to automate even more of the process. While an adversary might extend a couple of modules that we built for the Out-of-Context Summarizer summaries to post on social media platforms without any human intervention, with ethical considerations, we could develop automatically-generated simulated tweets or parleys for experimentation to gauge user reactions in a lab setting, similar to user studies investigating the effects of soft moderation on Twitter [45] . In addition to the default extracting summaries, Out-of-Context Summarizer could be adapted to generate abstractive summaries with little effort. We already provided a hint to an adversarial abstractive adaptation when discussing the manipulation of the summary shown in Table 13 . The first thing an adversary could do is replace words or hashtags, like we exemplified with the replacement of "interval" with "abstination" or #COVID-19 with #COVIDIOT. The next thing what the adversary could do is borrow sentences from other CDC articles and abstract the severity of the said allergic reaction to the COVID-19 vaccine as shown in Table 14 [? ]. Compared to the summary in 13, the abstractive summary adds an additional dimension to the out-of-context summarization for full realization of alternative narratives dissimination by taking out-of-context an entire set of articles on a topic, not just one. Though not in an automated fashion, a similar attempt of out-ofcontext trolling was used by Robert F. Kennedy Jr. in his attempt to discredit the COVID-19 vaccination (which earned him a ban from Instagram) [6] . In this context, the misperception-inducing approach of word manipulation from [47] , an adversary could also go further, for example, to append the first summary sentence with ..., which resulted in death in several cases in Florida in order to hint an extreme outcome of the vaccination. Certainly this moves the summary into the "fake news" category, and we provided this example to highlight the dangers of abstractive re-purposing of Out-of-Context Summarizer. We condemn such use, of course, but insinuations and implausible interpretations of a set of individual facts was, and still is, the essence of trolling on Twitter and especially Parler [40] . Getting A COVID-19 Vaccine A severe allergic reaction -also known as anaphylaxis -happens within 4 hours after getting vaccinated with the the first shot of a COVID-19 vaccine and could include symptoms such as hives, swelling, and wheezing (respiratory distress). CDC recommends that you not get a second shot of that vaccine. You need to be treated with epinephrine or EpiPen or go to the hospital. Hashtag Recommendations #severereaction, #analylaxis, #COVID-19 Both the expressive and abstractive summarizations could also be appended or utilized to generate headlines, a one-sentence summary, or an extended full summary. It's trivial to limit the summary and select one sentence to be a headline, for example, selecting the first sentence in 13 verbatim or modifying it to read CDC recommends that you not get a second shot of that vaccine, if you had a severe reaction after getting the first shot of a COVID-19 vaccine. An adversary could use the approach for modifying emails from [47] to simply make an email appear with a subject line CDC recommends that you not get a second shot of that vaccine. Similarly, one could use the approach from [48] and create a third-party COVID-19 skill that will deliver headlines to Amazon Alexa users by summarizing articles from the CDC website. A similar adversarial twist with Alexa was successful in reducing the perceived accuracy of information as to who gets the vaccine first, vaccine testing, and the side effects of the vaccine, as shown in [49] . A full expressive summary could also be generated for the purpose of updating or creating Wikipedia articles with the latest developments of the pandemic, for example. Again, using the approach for evading detection of Wikipedia vandalism [46] , an adversary could simply decide to minimize mentions of a target vaccine maker, for example, Sinopharm, if the adversary wants to implicitly promote other vaccines. Such a trolling attempt already surfaced on Twitter, promoting homegrown Russian vaccines and undercutting rivals [16] . Out-of-Context Summarizer, like every automated text summarization mechanism, comes with a set of limitations. First, the out-ofcontext summaries are generated using a relatively simple design of − → → . Other algorithms for text summarization might yield out-of-context summaries different than the ones generated by Out-of-Context Summarizer. We also allow for simplistic yet flexible calculation of the weights and sentence tockenization, which are couple of segments of the language model that could also be modified to generate different output. One could potentially choose another approach for word count and calculating the hashtag score to produce a nonoverlapping set of hashtag keywords with the set of keywords generated by Out-of-Context Summarizer. Second, we deliberately selected four news outlets other than the main agenda setters or any alternative outlets on the far ends of the spectrum. One could try to train Out-of-Context Summarizer on other outlets than ours, and, with a high probability, generate a separate liberal or conservative context. A similar conclusion holds for the selection of the articles reported in the time frame we conducted the study -COVID-19, elections, or stimulus as polarizing topics might morph and be incorporated in the trolling agenda in unpredictable ways. Finally, we used a modest infrastructure to implement and test the Out-of-Context Summarizer with satisfactory results of generating a summary and recommended hashtag keywords in less than few seconds. Summarizing larger datasets could affect this performance. We used the adversarial language model proposed in this paper to generate two versions of the "tl;dr" conclusion. We hope that the conclusion demonstration together with the argument put forward in the paper can serve the security community to better deal with adversarial language modeling in future. The summary output of Out-of-Context Summarizer algorithm is both indicative, alerting the user about the source content, and specifically informative, outlining the main contents of the original text (though with an original twist, which makes the Out-of-Context Summarizer appealing for adversarial text summarization). For the language modeling, we must note, an adversary might not be using news articles per se; they can train the Out-of-Context Summarizer on the textual portion of available Twitter information operations datasets [57] , for example on a particular topic like COVID-19 [25] . In addition to the default extracting summaries, Out-of-Context Summarizer could be adapted to generate abstractive summaries with little effort. Out-Of-Context Summarizer seemingly chose less-emotional sentences which were used in the original article solely to convey the facts without offering much commentary since the outcome was unfavorable to the Republican Party. For each iteration of Outof-Context Summarizer, we used the Linguistic Inquiry and Word Count (LIWC) tool to perform a quantitative content analysis of the three summarization datasets (i.e., all compiled original, conservative, and liberal summaries) [37] . Original summaries were not manipulated to appear right or left-leaning, but Out-Of-Context Summarizer was trained on text from Vox and Breitbart, to select liberal and conservative-leaning summaries. Analyzing the Digital Traces of Political Manipulation: The 2016 Russian Interference Twitter Campaign IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM) Social network sites and acquiring current affairs knowledge: The impact of Twitter and Facebook usage on learning about the news Random Forests SALSA: Detection of Cybertrolls Using Sentiment, Aggression, Lexical and Syntactic Analysis of Tweets Multi-document Text Summarization for Competitor Intelligence: A Methodology For Spreading Vaccine Misinformation Tracking Social Media Discourse About the COVID-19 Pandemic: Development of a Public Coronavirus Twitter Data Set Real solutions for fake news? Measuring the effectiveness of general warnings and fact-check tags in reducing belief in false stories on social media A repository of corpora for summarization Auto-Summarization-Based Steganography Automatic text summarization: A comprehensive survey A Context Based Text Summarization System Centers for Disease Control and Prevention. 2021. COVID-19 Vaccines and Allergic Reactions A holistic system for troll detection on Twitter False equivalencies: Online activism from left to right Russian Campaign Promotes Homegrown Vaccine and Undercuts Rivals Recent automatic text summarization techniques: a survey Fake News on Facebook and Twitter: Investigating How People (Don't) Investigate Liberals and conservatives rely on different sets of moral foundations Skunks, Bogies, Silent Hounds, and the Flying Fish: The Gulf of Tonkin Mystery Building a conversational model from two-tweets Valence-based homophily on Twitter: Network Analysis of Emotions and Political Talk in the 2012 Presidential Election Affirmative Reaction: Kennedy, Nixon, King, and the Evolution of Color-Blind Rhetoric TrollHunter2020: Realtime Detection of Trolling Narratives on Twitter During the 2020 US Elections TrollHunter [Evader]: Automated Detection [Evasion] of Twitter Trolls During the COVID-19 Pandemic Automatic summarising: The state of the art CTRL: A Conditional Transformer Language Model for Controllable Generation No Republicans back $1.9T COVID-19 Relief Bill COVID-19 Vaccines and Allergic Reactions Contextomy: the art of quoting out of context. Media Di Jin, and Yanjun Qi. 2020. TextAttack: A Framework for Adversarial Attacks, Data Augmentation, and Adversarial Training in NLP When corrections fail: The persistence of political misperceptions Twitter launches crowd-sourced fact-checking project The popularity and virality of political social media: hashtags, mentions, and links predict likes and retweets of 2016 U.S. presidential nominees' tweets Scikit-learn: Machine Learning in Python Linguistic inquiry and word count: LIWC The implied truth effect: Attaching warnings to a subset of fake news headlines increases perceived accuracy of headlines without warnings Prior exposure increases perceived accuracy of fake news Parlermonium: A Data-Driven UX Design Evaluation of the Parler Platform Language models are unsupervised multitask learners Why Should I Trust You Updates to our work on COVID-19 vaccine misinformation COVID-19 Misinformation Warning Labels: Twitter's Soft Moderation Effects on Belief Echoes about the COVID-19 Vaccine WikipediaBot: Machine Learning Assisted Adversarial Manipulation of Wikipedia Articles My Boss is Really Cool: Malware-Induced Misperception in Workplace Communication Through Covert Linguistic Manipulation of Emails Meet Malexa, Alexa's malicious twin: Malwareinduced misperception through intelligent voice assistants Hey Alexa, What do You Know About the COVID-19 Vaccine?" -(Mis)perceptions of Mass Immunization Among Voice Assistant Users The Offense-Defense Balance of Scientific Knowledge: Does Publishing AI Research Reduce Misuse RandomForestClassifier -scikit-learn 0.24.1 documentation_2021 Fact check: Twitter's 'read before you retweet' warnings do not just target conservative articles Examining the alternative media ecosystem through the production of alternative narratives of mass shooting events on Twitter Political psycholinguistics: A comprehensive analysis of the language habits of liberal and conservative social media users Examining trolls and polarization with a retweet network Trump Has Now Officially Lost All Of His Postelection Challenges In The Supreme Court Information Operations Exploring the Ideological Nature of Journalists Conservatives report, but liberals display, greater happiness The elusive backfire effect: Mass attitudes' steadfast factual adherence I Won the Election!":An Empirical Analysis of Soft Moderation Interventions on Twitter. arXiv 2101.07183v1 The Web Centipede: Understanding How Web Communities Influence Each Other through the Lens of Mainstream and Alternative News Sources The Web of False Information: Rumors, Fake News, Hoaxes, Clickbait, and Various Other Shenanigans Searchable Talk: Hashtags and Social Media Metadiscourse Franziska Roesner, and Yejin Choi. 2020. Defending Against Neural Fake News