key: cord-0623313-h00xb3fl authors: Mather, Brodie; Dorr, Bonnie J; Dalton, Adam; Beaumont, William de; Rambow, Owen; Schmer-Galunder, Sonja M. title: From Stance to Concern: Adaptation of Propositional Analysis to New Tasks and Domains date: 2022-03-20 journal: nan DOI: nan sha: 54eee4b10f67f1eb8db62bd00a76e66d116e3e6f doc_id: 623313 cord_uid: h00xb3fl We present a generalized paradigm for adaptation of propositional analysis (predicate-argument pairs) to new tasks and domains. We leverage an analogy between stances (belief-driven sentiment) and concerns (topical issues with moral dimensions/endorsements) to produce an explanatory representation. A key contribution is the combination of semi-automatic resource building for extraction of domain-dependent concern types (with 2-4 hours of human labor per domain) and an entirely automatic procedure for extraction of domain-independent moral dimensions and endorsement values. Prudent (automatic) selection of terms from propositional structures for lexical expansion (via semantic similarity) produces new moral dimension lexicons at three levels of granularity beyond a strong baseline lexicon. We develop a ground truth (GT) based on expert annotators and compare our concern detection output to GT, to yield 231% improvement in recall over baseline, with only a 10% loss in precision. F1 yields 66% improvement over baseline and 97.8% of human performance. Our lexically based approach yields large savings over approaches that employ costly human labor and model building. We provide to the community a newly expanded moral dimension/value lexicon, annotation guidelines, and GT. This paper presents a generalized paradigm for adaptation of tasks involving predicate-argument pairs, i.e., combinations of actions and their participants, to new tasks and domains. Predicateargument analysis has been a longstanding area of research for many tasks: event detection (Du and Cardie, 2020; Zhang et al., 2020) , opinion extraction (Yang and Cardie, 2013) , textual entailment (Stern and Dagan, 2014) , and coreference (Shibata and Kurohashi, 2018) . We refer to the induction of such representations as propositional analysis. We induce a proposition PREDICATE(x 1 ,x 2 ,...) to represent sentences such as John wears a mask: wear (John,mask) . We pivot off this explanatory representation to answer questions such as What is John's stance towards mask wearing? or What concerns does John have about mask wearing? Stance detection has recently (re-)emerged as a very active research area, yet many approaches generally equate stance to raw (bag-of-word) sentiment and often employ machine-learning based models requiring large amounts of (labeled or unlabeled) training data. Within such approaches, the notion of stance varies, but generally falls into one of a handful of "sentiment-like" categories for stance holder X regarding topic Y, i.e., X agrees/disagrees with Y (Umer et al., 2020) , X favors/disfavors Y (Krejzl et al., 2017) , X is pro/anti Y (Samih and Darwish, 2021 ), X has a positive/negative opinion about Y (AlDayel and Magdy, 2021), or X is in favor/against/neither Y ( Küçük and Can, 2020) . We adopt the stance definition of , originally formulated for Covid-19. There, a stance is a belief-driven sentiment, derived via propositional analysis (i.e., I believe masks do not help [and if that belief were true, I would be antimask]), instead of a bag-of-words lexical matching or embedding approach that produces a basic pro/anti label. This variant of stance detection uses a proposition to identify a domain-relevant belief in the Covid-19 domain; the belief is leveraged to compute a belief-driven sentiment and attitude toward a topic in that domain (e.g., masks). For example, I believe masks do not protect me is rendered as a belief type PROTECT with an underlying proposition protect(masks) and a negative overall stance toward the propositional content: masks. We implement and evaluate an analogous propositional framework for a new task, concern detection. Table 1 shows stance and concern detection output on a tweet from an English subset of a Kaggle Twitter dataset (1.9M tweets) for a new domain, the 2017 French Elections (Daignan, 2017) . The proposition common to both is ruin(jean-luc melenchon, economy). For stance, a belief type, DE-STROY, is coupled with values: (1) Belief strength ranges from certainty that the belief is not true (-3) to certainty that the belief is true (+3), with 0 as "uncommitted"; and (2) Belief-driven sentiment strength ranges from extremely negative (-1) to extremely positive (+1), with 0 as "neutral". We define a concern to be a topical type (e.g., ECONOMIC) coupled with a set of values: (1) moral dimensions from Moral Foundations Theory (MFT) (Haidt and Joseph, 2004; Graham et al., 2009 Graham et al., , 2011 , represented as vice/virtue pairings (authority/subversion, care/harm, fairness/cheating, loyalty/betrayal, and purity/degradation); and (2) corresponding endorsement values, where "vice" is between 1 to 5 and "virtue" is >5 to 9 (1 is strong vice and 9 is strong virtue). Dimensions and endorsements shown in Table 1 are derived from a state-of-the-art baseline (Moral 1), see §3. Propositional representation is the centerpiece of both stance detection and concern detection, distinguishing our lexical-based approach from modelbased approaches trained on word-level annotations. Predicate-argument structure captures relationships between multi-word constituents that need not be contiguous, thus inducing explainability. That is, Jean-Luc Melenchon is not adjacent to economy, yet these terms are crucially related via the intermediate term ruin. This enables answers to questions such as What does the author believe Jean-Luc Melenchon did to the economy? Moreover, domain adaptation is streamlined through predicate-argument annotation, reducing effort needed for human annotation. Annotation at the level of predicate-argument pairs factors out commonalities, reducing redundancy in the resource building process. During resource building each verb is visited only once rather than the multiple times required for word-level corpora annotation (see §3). For example, annotation of the word lead is done all in one shot with a handful of automatically presented de-duplicated cases, whereas corpus-based annotation requires repeated annotations of lead, substantially increasing human labor. We demonstrate that the key to reduced adaptation time is the coupling of semi-automatic resource building for concern types with automated expansion of domain-independent concern values using semantic similarity. We develop a ground truth (GT) based on expert annotators and compare concern detection output to GT. We also demonstrate that, with each lexicon expansion, the performance of concern detection improves significantly over a state-of-the-art baseline moral dimension lexicon. We obtain a 231% improvement in recall over a strong baseline for our best performing system, with only a 10% loss in precision. This section provides background and motivation for task and domain adaptability applied to concern detection in the 2017 French Election domain ( §2.1, §2.2), including concern values induced from Moral Foundations Theory (MFT) ( §2.3). Task adaptability and domain adaptability are two supporting areas of research for this work. Prior domain adaptation approaches, surveyed by Ramponi and Plank (2020), have been applied to tasks such as sentiment analysis (Ben-David et al., 2020; Ghosal et al., 2020) , stance detection (Xu et al., 2019) , and event trigger identification (Naik and Rose, 2020) . Task adaptation approaches (Gururangan et al., 2020; Garg et al., 2020; Ziser and Reichart, 2019) have been applied to tasks such as answer selection for question answering. To date, both types of adaptation rely heavily on machine learning (ML) techniques, many of which require a large amount (e.g., 1M+, Gururangan et al. (2020) ) of training data (whether labelled or not). Some approaches employ smaller datasets, e.g., 10K+ Amazon reviews, fake news articles (Ben-David et al., 2020; Xu et al., 2019) . Additionally, while explainability has recently been brought to the fore in deep learning approaches, as surveyed by Xie et al. (2020) , such systems have not focused on task and domain adaptability. We develop a general framework for resource building techniques that features task adaptability and retains the ability to adapt quickly to new domains. Our dataset requirements are much more minimal than prior approaches (2500 tweets per domain), there is no human labeling of corpora, and no model training is required. Moreover, explainability is achieved by virtue of inclusion of propositional information (who did what to whom) that serves as a window into the process of detecting concern types and moral dimensions. Pirolli et al. (2021) apply a belief-based formulation of stance in the Covid-19 domain, with topics such as mask wearing and social distancing. For example, a stance assigned to Wear a mask! includes a PROTECT belief type, where the predicate wear is considered a "trigger" and a mask is considered the "content" of the belief. The values associated with this stance include a belief strength of +3 and a sentiment strength of +1. The final stance is thus a belief-oriented sentiment with this interpretation: the person posting the tweet is positive toward "masks," assuming the belief that masks are protective is true. In the 2017 French Election domain, a stance representation (e.g., for the example in Table 1 ) would be: . While this prior framework lays the groundwork for domain adaptability, it has not been shown to be generalizable to new tasks (within or across domains), which is the focus of this paper. We leverage the propositional underpinnings of the framework of to enable a straightforward adaptation from stance detection to a new task, concern detection, while also retaining domain adaptability. This task involves extraction of a concern type (e.g., immigration, taxation) associated with a given domain (e.g., 2017 French Elections), analogous to the extraction of a belief type for a given stance detection domain. In this paper, we demonstrate that it is straightforward to port belief-targeted stance both to a new domain (French elections) and, through an analogous proposition-based extraction, to a new task: Concern detection. An example of a formal Concern representation in the 2017 French Election domain for the example in Table 1 would be: . The approach described herein focuses on lexicon expansion obtained automatically through semantic similarity to map key terms in propositional statements to moral foundation lexicon words, using WordNet (Fellbaum, 1998) . 1 Three different variants of lexicon expansion (described in §3.3) improve on results obtained using the current stateof-the-art moral lexicon of Araque et al. (2020) , which we take to be a strong baseline (henceforth referred to as 'Moral 1'). The advance beyond this prior work lies in the prudent (automatic) selection of terms designated for expansion, based on propositional structure, and the combination of moral dimensions with concern types. We focus on concern detection because identifying critical issues discussed online within a particular domain is important and useful, as is identifying the moral justifications or deliberate appeals to moral identity in these discussions. We use the Moral Foundations Theory (MFT) framework (Haidt and Joseph, 2004; Graham et al., 2009 Graham et al., , 2011 to encode the moral dimensions of social media contributions. These moral dimensions may serve as potential indicators of influence attempts, as in When it comes to immigration it's not about children, it's about damaging our country!, where the Concern type is IMMIGRATION_REFUGEE and there is an appeal to the vice side (harm) of the care/harm moral dimension. An emphasis on highly controversial and/or polarizing topics in online posts/messages may be indicative of an attempt to sway others. More importantly, if these posts/messages are interwoven with language that reflects (and speaks to) the moral values of the target audience it can increase in-group cohesion, and that may further contribute to polarization. Additionally, deliberate use of morality to justify harmful intentions towards others may foster online outrage disguised as ethical conduct (Bandura et al., 1996; Friedman et al., 2021) . Several studies show that social groups provide a framework in which moral values are endorsed, and when these values are threatened by e.g., opposing political ideology, existing beliefs of the group are strengthened (Van Bavel and Pereira, 2018). Thus, when presented with information that is incongruent with our identity and in-group, we tend to override accuracy motives in favor of social identity goals (partisan bias). When accuracy and identity goals are in conflict, moral values determine which belief to endorse and thus how to engage with information. This makes moral values an ideal breeding ground for influence campaigns, but also very useful for our stance and concern detection tasks. We note that most studies of cross-cultural values, beliefs, and morality have been conducted by WEIRD (western, educated, industrialized, rich, and democratic) countries (Goodwin et al., 2020; Henrich et al., 2010) . Inglehart's model of World Values (Inglehart and Welzel, 2010 ) has surveyed 60 countries over the last 40 years, taking into account that many nations are more concerned with economic and physical security (e.g. survival), while self-expression values are more reflective of Western countries. For example, in Pakistan or Nigeria 90% of the population say that God is extremely important in their lives, while in Japan only 6% take this position. Similarly, Schwartz' Theory of Basic Values (Schwartz, 2012) uses a different set of organizing principles, e.g., values that relate to anxiety (e.g., tradition, security, control of threat) which may lead to an increased belief in misinformation (Jost et al., 2003) . Thus, moral dimensions combined with concern types are a potential indicator of a common actor (possibly an outside influencer) if several individuals or accounts (potentially purporting to be individuals) invoke the same moral dimensions across their messages. Adaptation of stance detection to concern detection gives rise to a new framework for rapid development of a task-adapted system that retains domain adaptability and uses relatively low amounts of data, with only 2-4 hours of human categorization. We generalize to a new task while retaining domain adaptability by leveraging the stance-concern analogy, through: (a) semi-automatic domaindependent extraction of types from propositional analysis, i.e., moving from belief types for stances to concern types for concerns; and (b) fully automatic domain-independent induction of associated values from a combination of propositional arguments and lexical and semantic resources, i.e., belief/sentiment strengths (cf. (Baker et al., 2012) ) for stances and moral dimensions and endorsements (cf. ) for concerns. Resource building for domain-dependent stance types involves propositional analysis using semantic role labeling (SRL) (Gardner et al., 2018) 2 to detect positions with the most highly relevant content terms, e.g., masks. The work of indicates that these positions are ARG0 and ARG1. To port this approach over to the induction of domain-specific concern types, we conducted a similar analysis and found that the same positions (ARG0 and ARG1) contain the most highly relevant terms for concerns, e.g., economy. Semiautomatic induction of concern types thus leverages these positions, as described in §3.2. Just as stance resource building induces domainindependent stance values (belief / sentiment strengths), concern resource building induces domain-independent concern values (moral dimensions / endorsements). A deeper propositional analysis reveals that additional SRL positions have a high likelihood of association with moral dimension terms, e.g., ruin: V, ARG2, ARGM-ADV, ARGM-MNR, ARGM-PRD. This discovery further generalizes the original stance resourcebuilding approach and enables rapid task adaptation to concerns through entirely automatic means. We leverage these additional SRL positions to extract candidate terms for expansion of moral dimensions. Associated endorsements are then inherited from semantically similar terms from baseline Moral 1. Further details about the expansion of moral dimensions are provided in §3.3. Domain adaptability is retained-on analogy with stance detection-by separating and independently addressing two aspects of concern detection: (a) induction of domain-specific concern types; (b) induction of domain-independent moral dimensions. Lexicon expansion using this approach can thus be applied to domains beyond the French Elections presented herein. We adopt a generalized semi-automatic process for lexically based concern type induction that retains domain adaptability (later referred to as 'Concern 1'). A small set (approximately 15) of domainrelevant key terms (e.g., health, taxation, immigration), is provided by a domain expert as input to a semi-automatic resource building tool. These terms are used to filter the domain-relevant dataset. The filtered tweet subset (214k) is then divided into training (2500), 3 and development (211,500) subsets, and propositional analysis is applied to the training set in a 3-step process. First, the top 25 most frequent terms are extracted (e.g., economy), 4 ignoring functional elements such as stop words. Second, the verbs whose relevant SRL positions (defined in §3.1) contain any of these top 25 terms are extracted, and the top 40 most frequent verbs e.g., ruin, restrict are selected for further processing. Lastly, the top 10 most frequent terms associated with relevant SRL positions (for each of the 40 verbs) are extracted automatically from domain-relevant data, e.g., economy and business, yielding 400 propositions, e.g., ruin(economy), restrict(business). 5 These, coupled with terms from the original domain-relevant key terms, are presented to the domain expert who constructs a small set of concern types-10 in the case of French Election: textscimmigration_refugee, ELEC- CRIMINAL_JUSTICE. Terms left uncategorized by the expert are dropped. This semi-automated concern-type induction takes 2-4 hours owing to the automatic extraction of high frequency domain relevant propositions. Concern values leverage MFT (Haidt and Joseph, 2004; Graham et al., 2009 Graham et al., , 2011 and particularly the moralstrength library (Araque et al., 2020) , which serves as a strong baseline (referred to as 'Moral 1'). 6 This baseline lexicon includes manually developed moral dimensions (e.g. care/harm, loyalty/betrayal) and endorsement values (1-9). 3 Training data are strictly for one-time semi-automatic resource building, not for model training. 4 spaCy 3.1.0 with model en_core_web_sm (Honnibal et al., 2020) is used for sentence splitting and POS tagging. 5 The thresholds of 25, 40, and 10 are selected empirically in preliminary experiments (not reported here) ascertaining a balance between adequate coverage of the data and time spent manual categorization by the expert. 6 https://github.com/oaraque/moral-foundations. Our approach transcends this earlier paradigm in its application of propositional analysis with semanticrole labeling (SRL Gardner et al. (2018) ), coupled with a more in-depth WordNet (WN Fellbaum (1998) ) expansion to enrich the lexicon. This results in higher recall while retaining linguistically relevant constraints to achieve acceptable precision. Expansion of moral dimensions relies on propositional analysis, SRL, and WordNet expansion. We select candidates for moral dimension expansion through extraction from propositional statements, and then induce three lexicon variants (in addition to the Moral 1 baseline, and a Moral 0 random chooser described in §6) using a progression of finer-grained WordNet-based semanticsimilarity functions. This expansion supports the goal of task adaptability, as required in the transition from stance detection to concern detection. The end result is a general approach to induction of values for specific tasks, rendered in the form of domain-independent lexicons. That is, analogous to belief/sentiment terms of stance detection (might, probably, hate, love), we induce vice/virtue moral dimensions and their corresponding endorsement values for concern detection. This expansion results in three different system variations for moral dimension/values (referred to as Moral 2, Moral 3, Moral 4) beyond the Moral 1 baseline. We note that moral dimensions are assigned automatically to each lexical entry via semantic similarity to Moral 1 terms; the endorsement values are then inherited from the most semantically similar term from the Moral 1 lexicon. An excerpt of lexicon output is shown below, from the best performing lexicon (Moral 4): hypocrite -dim: betrayal; endorse: 1.0 (strong) appreciation -dim: care; endorse: 8.57 (strong) Lexicon Expansion Details: Moral 2-4 rely on automatic propositional analysis for prudent selection of words from the training data (via SRL, see §3.1) to be considered candidates for lexicon expansion. The highest similarity match is recorded to inherit the corresponding moral endorsement value from Moral 1. A brief description of all lexicon expansions used for induction of moral dimensions and values for concern detection is provided below. (See detailed description in Appendix A and links to Moral 2-4 lexicons in Appendix B.) Note: The term "lemma-matched" below refers to a match between a word in a training tweet and a synset's first lemma. Moral 1: This initial moral dimension lexicon developed by Araque et al. (2020) serves as a strong baseline, with 2800+ terms across five moral dimensions. This was derived by expanding an initial crowd sourced lexicon of about 480+ terms, annotated for moral dimensions and endorsements. Expansion to 2800+ terms was via WordNet synset matching, without regard to propositional analysis. Moral 2: (Added 214 terms, for total of 3064) This lexicon expansion yields a set of terms whose lemma-matched synsets are semantically similar (above a threshold) to lemma-matched synsets of the words in the strong baseline (Moral 1) lexicon. Moral 3: (Added 995 terms, for total of 3845) This lexicon expansion yields a set of terms whose lemma-matched synsets and their descendents are semantically similar (above a threshold) to lemmamatched synsets and their descendents of the words in the strong baseline (Moral 1) lexicon. Moral 4: (Added 5623 terms, for total of 8473) This lexicon expansion yields a set of terms drawn from all synsets and their descendents that are semantically similar (above a threshold) to all synsets and their descendents of words in the strong baseline (Moral 1) lexicon, without lemma matching. We conduct an annotation task to develop Ground Truth (GT), against which to compare our concern detection system variants, based on 50 heldout tweets from the held-out development portion of the English subset of the Kaggle Twitter dataset on 2017 French elections (Daignan, 2017) . Ground truth was produced for concern types and vice/virtue pairs for any of five moral dimensions, in accordance with guidelines in Appendix C. Annotation was completed by two non-algorithm developers (one with expertise in linguistics, the other with expertise in psychosocial moral indicators). For concern types the inter-rater reliability (IRR) is calculated through macroaveraging of kappa scores (Carletta, 1996) which produces a 66% agreement, considered Strong according to (McHugh, 2012) . By contrast, the macroaveraged IRR for moral dimensions is low (16%), which is considered Weak. Given the high annotator reliability for concern types, system output is compared against the union of both annotators, yielding the concern type scores in Table 3 . However, the lower IRR for moral dimensions is an indication that research in this realm is still in nascent stages and significant training for the task is required to achieve a reliable GT. We note that "ground truth" is inherently problematic with moral indicators given the complex way in which dimensions vary with socio-cultural factors. Thus, for moral dimensions, system output is compared only against the single annotator with expertise in psychosocial moral indicators. Appendix D presents the Ground Truth resulting from these annotations. Kaggle data (Daignan, 2017) are open and publicly available, intended for research on text analytics. The data carry no privacy or copyright restrictions. Furthermore, IRB designates our work as non human subject research. Annotators spent two hours apiece on the task. We have implemented/validated a system to detect concerns based on induced lexical resources, using spaCy and SRL. We derive a proposition coupled with a concern type and values (moral dimensions/endorsements). Systems produce pairs, with an average runtime of 0.6s per tweet on Mac, with no GPUs required. Representative outputs on a 1K-tweet held-out portion of the development dataset from Kag-gle 2017 French Elections are shown in Table 2 . These are taken from our best performing variant described in §6. In Example 1, IMMIGRA-TION_REFUGEE (triggered by refugees) is coupled with moral dimension/endorsement values, Harm (wall, entrance, guard), Authority (wall, guard), and Degradation (entrance). In Example 2, CRIM-INAL_JUSTICE (triggered by justice) is coupled with moral dimension/endorsement values: Fairness (justice). Both are reasonable outputs. In example 3, ELECTORAL_PROCESS_ VOTING_LAWS (triggered by voters) is coupled with moral dimension/endorsement values, Care (care), Loyalty (country), Authority (care, country), and Degradation (care, voters). This example overgenerates, assigning Degradation based on the terms care and voters. A further lexicon enhancement is needed (as alluded to in §7) for elimination of spurious lexical entries that lead to false positives (i.e., a reduction in Precision). We explore the performance of each of our four moral dimension lexicons by comparing outputs against GT for each lexicon. Concerns types are evaluated for an exact match against the GT concern (e.g., CRIMI-NAL_JUSTICE) and moral values are evaluated for an exact match against the binary choice in the GT (vice or virtue). Concern 1 represents the lexiconbased chooser described in §3.2. Our (strong) baseline for moral values is "Moral 1" (Araque et al., 2020) . Subsequent variants (Moral 2-4) are the expanded moral dimension lexicons using the techniques described in §3.3. We also compare against random choosers for Concern Types (Concern 0) and Values (Moral 0). Table 3 shows the results of each system output compared to GT. System performance is measured by weighted macro-averaged precision (P), recall (R), and F1 scores. Domain-dependent Concern Type is not affected by moral lexicon variants and independently has its own P/R/F1 scores. Concern Values (i.e., moral Dimensions) are impacted by lexicon variants and therefore have a row corresponding to each variant. We note the importance of applying a weighted macro average to these scores due an imbalance in the distribution of classes (Delgado and Tibau, 2019), where the probability of one class can be substantially higher or lower from others. For example, we observed that the num- ber of ELECTORAL_PROCESS_VOTING_LAWS annotations is 2.3 times higher than the number of INTERNATIONAL_TRADE annotations. One might expect random choosers (Concern 0 and Moral 0) to have a decent likelihood of getting many hits, with an expected rate of about 50% that each Concern type and Moral value will be selected. If so, this would result in an expected 250 positive selections for concern types (whereas the two annotators together only made 60 positive selections) and an expected 250 positive selections for moral values (whereas the expert annotator only made 66 positive selections). However, the results in Table 3 indicate that, while the increased number of hits leads to a reasonably high recall, the number of false positives (233) swamps out the hit rate, leading to a low F1 score (20.54). Accordingly, Concern 1 easily beats the random choice baseline by a healthy margin, with an F1 score of 77.17. In contrast, Moral 0 achieves 79.65% of the performance of the Moral 1 baseline, with an F1 score of 19.65-not too far off from the 24.67 baseline F1 score. Moreover, the F1 scores for Moral 2-4 surpass this baseline, with statistically significant improvements indicated between all system pairs at the 3.5% level or better, according to the McNemar statistical test (McNemar, 1947) . 7 That is, all lexicon expansion improvements are statistically significant. Notably, a 231% improvement in Recall is achieved for Moral 4 over the strong baseline (Moral 1): 65.15 vs. 19.69. This is achieved with only a 10% loss of precision, ultimately yielding an F1 score of 40.85 which is a 66% improvement. We observe that the precision-recall gap increases as the notion of similarity is loosened: (1) lead is similar to strip in Moral 4, but not in Moral 3 (reducing Moral 4 precision); (2) price is similar to value in Moral 4 but not in Moral 3 (increasing Moral 4 recall). We further assess system performance by comparing Concern 1 performance to human performance. Average F1 score for the two annotators is 78.88, and Concern 1 performance is 77.17 F1. Concern detection thus yields 97.8% of human performance on concern type detection. Error analysis of FP/FN's for concern types reveals that concern detection fails to assign any concern type (FN) to Infighting among left-wing could hand Front National VICTORY, which is annotated as ELECTORAL_PROCESS_VOTING_LAWS, because terms like Front and left-wing are not present in the concern type lexicon. For concern values, the annotator does not assign a moral dimension to Macron is center right, yet Moral 4 inaccurately detects (FP) Care, Authority, and Betrayal due to the existence of the word center and also detects Purity from the word right. For this same sentence Moral 1 also inaccurately detects Fairness from the word right. Many cases similar to these impact precision values for each Moral 1-4, potentially requiring lexicon tuning (see §7). We conduct further analysis to determine whether performance is impacted by potential overfitting of concern value detection to the domain of interest during development. We compare Araque's original concern value detection tools (Araque, 2021) , a unigram model trained on Hurricane Sandy data from the Moral Foundations Twitter Corpus (MFTC) (Hoover et al., 2020) , to our proposition-based concern value detection that uses semantically expanded lexicons based on the Kaggle French elections data. We level the playing field by applying both approaches to both datasets, measuring each against their respective GTs. Araque (2021)'s Unigram Model performs best on Hurricane Sandy test data (Table 4) , with weighted macroaverage F1, using both Moral 1 and Moral 4 lexicons (60.55), but does not perform as well on Kaggle test data (30.38 and 22.43) . This is a potential indicator of overfitting to the Hurricane Sandy data during training. Similarly, propositionbased concern value detection ( §3.3) performs better on Kaggle test data (40.85) than on Hurricane Sandy Data (29.99)-a sign of overfitting in the opposite direction during lexicon expansion. Testing our moral lexicon expansion for stability retention across domains, we level the playing field by relaxing the constraint that only a propositional sub-portion of the text is considered as input. We apply our algorithm to the entire text (full tweet) to induce a fairer comparison with Araque et al's Unigram model. Table 5 shows comparable or better macroaveraged F1 scores than those shown for "Proposition" in Table 4 . More importantly, our expanded Moral 4 lexicon demonstrates stability across domains: 42.61 for 2017 French Elections and 42.03 for Hurricane Sandy (vs. 26.14 and 34.13, respectively, for Araque et al's Moral 1 lexicon). This illustrates the potential for proposition-based domain adaptability but highlights the need for hybridized detection to ensure both task adaptability and explainability (discussed further below). elections). Our primary contribution is the provision of a framework for rapid adaptability, based on: (a) semi-automatic domain-dependent extraction of types from propositional analysis; and (b) fully automatic domain-independent induction of values from propositional arguments and semantic resources. We demonstrate that the coupling of (a) and (b) leads to rapid ramp-up resource construction for a new task and domain (2-4 hours). We demonstrate that a deeper propositional analysis is key to generalizing domain-adaptable resource-building for new tasks. We develop an automatic procedure for expanding moral dimensions that incorporates propositional analysis, semanticrole labeling, and in-depth WordNet (WN) expansion, to produce three increasingly expanded moral dimension/endorsement lexicons. We develop a ground truth (GT) based on expert annotators and compare our concern detection output to GT, to yield 231% improvement in recall over baseline, with only a 10% loss in precision. F1 yields 66% improvement over baseline and 97.8% of human performance. Our approach yields large savings over those that employ costly human labor and model building. Work produced herein provides the community with a newly expanded moral dimension/value lexicon, annotation guidelines, and GT for 50 tweets, intended for research purposes. We show that our proposition-based moral lexicon expansion provides stability across domains. However, the results of concern value detection using the full text of an input (a tweet) highlight the importance of adopting a hybrid approach that captures fine-grained distinctions and produces an explainable representation that is not otherwise available (e.g., in the unigram language model of Araque (2021)). An avenue of future research is to explore a hybridized approach that makes fine-grained concern value distinctions for domain adaptability, while leveraging an explainable propositional representation for task adaptability. As a first step, we will apply our propositional analysis iteratively on full-text inputs to detect answers to questions not otherwise extractable from raw textual strings (see related discussion in §1). Our results do not require any tuning of the lexicons to remove terms that result in a high number of false positives (FP) and false negatives (FN). Future work will explore fine tuning of the lexicons to address cases seen in §6 with an eye toward improving the precision without a large drop in recall, to yield an even higher F1-score. A current limitation (for future study) is the omission of additional moral dimensions, e.g., liberty/oppression (Haidt, 2012) . This reflects political equality related to dislike of oppression and concern for victims, not a desire for reciprocity. In political discourse, this is apparent in antiauthoritarianism and anti-government anger, which makes it an important dimension for the topic of the French election and widens the range of potentially relevant information that could indicate influence attempts. Future work also includes exploration of other cultural frameworks, in addition to or in place of Moral Foundations Theory, e.g., Inglehart's Cultural Map model (Inglehart and Welzel, 2010) and Schwartz Value Theory (Schwartz, 2012) . Cultural models that allow more room for non-Western values (e.g. survival needs) are important for reducing bias, and a feasible avenue for improving the performance and applicability of concern detection. Another limitation is that GT development for a given cultural context is difficult, especially with diverse annotators. Prabhakaran et al. (2021) show that systematic disagreements between annotators with differing socio-cultural backgrounds are obfuscated through aggregating crowd-sourced annotations. Future work will explore vector-based approaches to assign weights based on their representativeness in a given culture. Cultural values are also reflected in language (e.g., gendered vs. non-gendered languages, culturally intrinsic concepts). Accordingly, our future work involves processing concerns for other languages through adaption of SRL (Gardner et al., 2018) to multilingual input, starting with French, employing multilingual preprocessing via spaCy (Honnibal et al., 2020) and EuroWordNet and related multilingual WordNets (Bond and Paik, 2012; Bond and Foster, 2013) , to expand moral dimensions to other languages. Annotation was completed by two non-algorithm developers (one with expertise in linguistics, the other with expertise in psychosocial moral indicators), compensated appropriately for their work. A two-level institutional review board (at both the sponsored site and at the sponsoring site) deemed this work as "research not involving human subjects," as it does not involve a living individual about whom an investigator conducting research obtains information through intervention or interaction with the individual, or obtains, uses, studies, analyzes, or generates identifiable private information. The data used for the software development is provided from Kaggle's existing public research Twitter dataset, focusing on an English subset for the 2017 French Elections of 1.9 million tweets (Daignan, 2017) at https://www.kaggle.com/jeanmidef/ french-presidential-election. Kaggle data contain user names, but the dataset is open and publicly available. Potential risks may emerge from language biases in standard resources on which some of our work is built. For example, cultural idioms like fluctuat nec mergitur may be translated from Latin into the correct literal meaning of tossed [by the waves], but not sunk, but the culturally distinct values (Paris' coat of arms and motto with a deep affective history) will get lost in translations into English, and with it, its cultural meaning. Similarly, while English WordNet provides one of the most comprehensive semantic ontology of words in English, embedded biases are still present today, e.g., offensive, racist, and misogynistic slurs (Crawford and Paglen, 2021) . These issues need to be addressed within the language resource community. Risk of misuse of this technology is mitigated by the transparent nature of concern detection, owing to the propositional representations that underlie and inform algorithmic decisions. In contrast to ML approaches, misuse within the technology would be easily discoverable. The technology further serves as a framework within which cultural distinctions may be studied and better understood, thus mitigating the potential for cross-culturally undetected misuse. Stance detection on social media: State of the art and trends Oscar Araque. 2021. moral-foundations Moralstrength: Exploiting a moral lexicon and embedding similarity for moral foundations prediction. Knowledge-Based Systems Use of modality and negation in SIMT. CL Mechanisms of moral disengagement in the exercise of moral agency PERL: Pivot-based domain adaptation for pre-trained deep contextualized embedding models Linking and extending an open multilingual Wordnet A survey of wordnets and their licenses Assessing agreement on classification tasks: The kappa statistic Excavating ai: The politics of images in machine learning training sets French presidential election: Extract from twitter about the french election, kaggle data set Why cohen's kappa should be avoided as performance measure in classification Event extraction by answering (almost) natural questions WordNet: An Electronic Lexical Database. Language, Speech, and Communication Toward transformer-based nlp for extracting psychosocial indicators of moral disengagement AllenNLP: A deep semantic natural language processing platform Tanda: Transfer and adapt pre-trained transformer models for answer sentence selection KinGDOM: Knowledge-Guided DOMain Adaptation for Sentiment Analysis Cross-cultural values: a meta-analysis of major quantitative studies in the last decade Moral foundations theory: The pragmatic validity of moral pluralism Liberals and conservatives rely on different sets of moral foundations Mapping the moral domain Don't stop pretraining: Adapt language models to domains and tasks The righteous mind: Why good people are divided by politics and religion Intuitive ethics: how innately prepared intuitions generate culturally variable virtues Most people are not weird Sofie Van Landeghem, and Adriane Boyd. 2020. spaCy: Industrial-strength NLP in Python Jun Yen Leung, Arineh Mirinjian, and Morteza Dehghani. 2020. Moral foundations twitter corpus: A collection of 35k tweets annotated for moral sentiment The wvs cultural map of the world Understanding libertarian morality: The psychological dispositions of selfidentified libertarians Political conservatism as motivated social cognition Stance detection in online discussions Stance detection: A survey A general framework for domain-specialization of stance detection A covid-19 response use case Interrater reliability: The kappa statistic Note on the sampling error of the difference between correlated proportions or percentages Towards open domain event trigger identification using adversarial domain adaptation Mining Online Social Media to Drive Psychologically Valid Agent Models of Regional Covid-19 Mask Wearing On releasing annotator-level labels and information in datasets Neural unsupervised domain adaptation in NLP-A survey A few topical tweets are enough for effective user stance detection An overview of the schwartz theory of basic values Entitycentric joint modeling of Japanese coreference resolution and predicate argument structure analysis Recognizing implied predicate-argument relationships in textual inference Fake news stance detection using deep learning architecture (cnn-lstm) The partisan brain: An identity-based model of political belief Explainable deep learning: A field guide for the uninitiated Adversarial domain adaptation for stance detection Joint inference for fine-grained opinion extraction A two-step approach for implicit event argument detection Task refinement learning for improved accuracy and stability of unsupervised domain adaptation Below is the detailed description of each expansion algorithm. Basic definitions: w = word in training tweet (e.g., 'concerned') m = moral foundations word (e.g., 'concern') wn = wordnet package S w = wn.synsets(w) (set of synsets for word w) s w,1 = wn.synsets(w)[0] (1st synset for word w) L i = lemmas associated with S i l i,k = s.lemmas() (lemmas for synset s i ) l i,1 = s.lemmas() [0] .name() (1st lemma synset s i ) Note: Use of the term "lemma-matched" below refers to a match between a word in a training tweet and a synset's first lemma. Lexicon Expansion: All lexicon expansions (Moral 2-4 below) beyond a strong baseline (Moral 1) rely on propositional guidance to select words from the training data to be considered candidates for lexicon expansion. The highest similarity match is recorded for future selection of the moral endorsement value. All moral lexicon expansions (Moral 2-4 below) beyond a strong baseline (Moral 1) apply to each pair (w, m) for each word of interest w and each moral foundations word m (w × m iterations):Moral 1: This initial moral dimension lexicon developed by Araque et al. (2020) 8 serves as a strong baseline, with 2800+ terms across five moral dimensions. This was derived by expanding an initial crowd sourced lexicon of about 480+ terms, annotated for moral dimensions and endorsements. Expansion to 2800+ terms was via WordNet synset matching, without regard to propositional analysis. Moral 2: (Added 214 terms) This lexicon expansion yields a set of terms whose lemma-matched synsets are semantically similar (above a threshold) to lemma-matched synsets of the words in the strong baseline (Moral 1) lexicon, using the following steps: (a) Extract all synsets s w,i in S w (for word w) whose first lemma l i,1 exactly matches w, producing S x .(b) Extract all s m,j in S m (for word m) whose first lemma l j,1 exactly matches m, producing S y . (c) Return all lemmas l k,1 of any s x,k in S x that matches any s y,h in S y using wordnet wup_similarity w/ threshold 0.9 (if any). Example: concerned → concerned.a.01.Moral 3: (Added 995 terms) This lexicon expansion yields a set of terms whose lemma-matched synsets and their descendents are semantically similar (above a threshold) to lemma-matched synsets and their descendents of the words in the strong baseline (Moral 1) lexicon, using the following steps: (a) Extract all synsets s w,i , in S w (for word w) whose first lemma l i,1 matches exactly w, producing S x . Extract all synsets s m,j in S m (for word m) whose first lemma l j,1 matches exactly m, producing S y . (First part of 2 up to this point.) (b) Expand lemmas from both sets: (i) collect all lemmas for all synsets in S x , producing L x ; (ii) collect all lemmas for all synsets in S y , producing L y . (c) Extract synsets for these lemma expansions: (i) collect all synsets for all lemmas in L x , producing S a ; (ii) collect all synsets for all lemmas in Steps below refer to column labels in the Blank Annotation Sheet found here.1. For each tweet excerpt in the "Text" column A, apply steps 2-8 below.2. Columns B through K are the potential Concern types. Put a "1" in a single column corresponding to the applicable Concern type. Leave empty if none appears to apply.3. Columns L through W are moral dimensions. Put a "1" in any column that has an applicable moral dimension, in either the vice subcolumn or the virtue subcolumn. Leave the vice/virtue cells empty for any moral dimension that is not applicable. Refer to the moral dimensions described in the Graham and Haidt tables for this task (see links in 4a and 4b below).4. Provide annotations only for explicitly represented material. Do not infer context that is not stated and do not apply any subject-matter background knowledge. The only background knowledge to be used for this task is the moral dimensions description in tables below from:(a) (see page 16) (b) [Graham et al., 2013] (see page 68) 5. Do not try to force the Text into a particular moral dimension. If a moral dimension appears to be applicable, but it is unclear whether the vice or virtue is active, put "DK" in either the vice cell or virtue cell, instead of leaving it blank.6. Consider only the content of the Text without regard to grammaticality or punctuation.7. Assume no sarcasm is present. Annotate the literal sense.8. Use the last column (optionally) for any annotation notes, eg, the reasoning behind the chosen annotations. We conduct an annotation task to develop Ground Truth (GT), against which to compare our concern detection system variants, based on 50 held-out tweets from the held-out development portion of the Kaggle English Twitter dataset on 2017 French elections (Daignan, 2017) . Annotation was completed by two non-algorithm developers (one with expertise in linguistics, the other with expertise in psychosocial moral indicators). Annotators were provided guidelines (Appendix C) for both concern types and moral dimensions. For concern types the interrater reliability (IRR) is calculated through macroaveraging of kappa scores (Carletta, 1996) which produced a 66% agreement, considered Strong according to (McHugh, 2012) . By contrast, the macroaveraged IRR for moral dimensions is considered none to slight (16%), which is deemed Weak. [Link to GT]