| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T14:39:06.053559Z" |
| }, |
| "title": "COVIDLIES: Detecting COVID-19 Misinformation on Social Media", |
| "authors": [ |
| { |
| "first": "Tamanna", |
| "middle": [], |
| "last": "Hossain", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Califonia", |
| "location": { |
| "settlement": "Irvine" |
| } |
| }, |
| "email": "tthossai@uci.edu" |
| }, |
| { |
| "first": "Robert", |
| "middle": [ |
| "L" |
| ], |
| "last": "Logan", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Califonia", |
| "location": { |
| "settlement": "Irvine" |
| } |
| }, |
| "email": "rlogan@uci.edu" |
| }, |
| { |
| "first": "Arjuna", |
| "middle": [], |
| "last": "Ugarte", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Califonia", |
| "location": { |
| "settlement": "Irvine" |
| } |
| }, |
| "email": "dugarte@uci.edu" |
| }, |
| { |
| "first": "Yoshitomo", |
| "middle": [], |
| "last": "Matsubara", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Califonia", |
| "location": { |
| "settlement": "Irvine" |
| } |
| }, |
| "email": "yoshitom@uci.edu" |
| }, |
| { |
| "first": "Sean", |
| "middle": [], |
| "last": "Young", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Califonia", |
| "location": { |
| "settlement": "Irvine" |
| } |
| }, |
| "email": "syoung5@uci.edu" |
| }, |
| { |
| "first": "Sameer", |
| "middle": [], |
| "last": "Singh", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Califonia", |
| "location": { |
| "settlement": "Irvine" |
| } |
| }, |
| "email": "sameer@uci.edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "The ongoing pandemic has heightened the need for developing tools to flag COVID-19related misinformation on the internet, specifically on social media such as Twitter. However, due to novel language and the rapid change of information, existing misinformation detection datasets are not effective for evaluating systems designed to detect misinformation on this topic. Misinformation detection can be divided into two sub-tasks: (i) retrieval of misconceptions relevant to posts being checked for veracity, and (ii) stance detection to identify whether the posts Agree, Disagree, or express No Stance towards the retrieved misconceptions. To facilitate research on this task, we release COVIDLIES 1 , a dataset of 6761 expert-annotated tweets to evaluate the performance of misinformation detection systems on 86 different pieces of COVID-19 related misinformation. We evaluate existing NLP systems on this dataset, providing initial benchmarks and identifying key challenges for future models to improve upon.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "The ongoing pandemic has heightened the need for developing tools to flag COVID-19related misinformation on the internet, specifically on social media such as Twitter. However, due to novel language and the rapid change of information, existing misinformation detection datasets are not effective for evaluating systems designed to detect misinformation on this topic. Misinformation detection can be divided into two sub-tasks: (i) retrieval of misconceptions relevant to posts being checked for veracity, and (ii) stance detection to identify whether the posts Agree, Disagree, or express No Stance towards the retrieved misconceptions. To facilitate research on this task, we release COVIDLIES 1 , a dataset of 6761 expert-annotated tweets to evaluate the performance of misinformation detection systems on 86 different pieces of COVID-19 related misinformation. We evaluate existing NLP systems on this dataset, providing initial benchmarks and identifying key challenges for future models to improve upon.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Detecting spread of misinformation such as, rumors, hoaxes, fake news, propaganda, spear phishing, and conspiracy theories, is an important task for natural language processing (Thorne et al., 2017; Shu et al., 2017; . Online social media networks provide particularly fertile ground for the spread of misinformationthey lack gate-keeping and regulations, users publish content without having to go through an editor, peer review, verification of qualification, or providing sources, and social networks tend to create \"echo chambers\" or closed networks of communication insulated from disagreements.", |
| "cite_spans": [ |
| { |
| "start": 177, |
| "end": 198, |
| "text": "(Thorne et al., 2017;", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 199, |
| "end": 216, |
| "text": "Shu et al., 2017;", |
| "ref_id": "BIBREF30" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Tweet: \"Coronavirus CV19 was a top secret biological warfare experiment. That is why it is only affecting the poor.\" Misconception: \"Coronavirus is genetically engineered.\" Label: Agree Tweet: \"It looks like we are all going to have to wait much longer for a #COVID19 vaccine.\" Misconception: \"We're very close to a vaccine.\" Label: Disagree Tweet: \"CDC: Coronavirus spreads rapidly in dense populations with public transit and regular social gatherings.\" Misconception: \"Coronavirus cannot live in warm and tropical temperatures.\" Label: No Stance Figure 1 : COVIDLIES Dataset. Given a tweet, we annotate whether any of the known misconceptions are expressed in the tweet, in particular, if the tweet spreads the misconception (e.g., they Agree), combats the spread of the misconception (e.g., they Disagree), or takes No Stance towards the misconception.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 549, |
| "end": 557, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The COVID-19 pandemic has created a pressing need for tools to combat the spread of misinformation. Since the pandemic affects the global community, there is a wide audience seeking information about the topic, whose safety is threatened by adversarial agents invested in spreading misinformation for political and economic reasons. Furthermore, due to the complexity of medical and public health issues, it is also difficult to be completely accurate and factual, leading to disagreements that get exacerbated with misinformation. This difficulty is compounded by the rapid evolution of knowledge regarding the disease. As researchers learn more about the virus, statements that seemed true may turn out to be false, and vice versa. Detecting this spread of pandemic-related misinformation, thus, has become a critical problem, receiving significant attention from government and public health organizations (WHO, 2020), social media platforms (TechCrunch, 2020) , and news agencies (BBC, 2020; CNN, 2020; New York Times, 2020) . Figure 2 : Misconception Detection Pipeline consisting of two sub-tasks, (a) Misconception Retrieval that identifies the known misconceptions that are relevant to the given tweet, and (b) Stance Detection that identifies whether the tweet agrees, disagrees, or expresses no stance, for each of the relevant misconceptions.", |
| "cite_spans": [ |
| { |
| "start": 945, |
| "end": 963, |
| "text": "(TechCrunch, 2020)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 984, |
| "end": 995, |
| "text": "(BBC, 2020;", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 996, |
| "end": 1006, |
| "text": "CNN, 2020;", |
| "ref_id": null |
| }, |
| { |
| "start": 1007, |
| "end": 1028, |
| "text": "New York Times, 2020)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1031, |
| "end": 1039, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we introduce the COVIDLIES dataset for misconception detection on Twitter. COVIDLIES comprises of 86 common misconceptions about COVID-19 along with 6761 related tweets, identified and annotated by researchers from the UCI School of Medicine. Given a tweet, we annotate whether any of the known misconceptions, curated by the researchers, are expressed by the tweet. If they are not, then they are considered No Stance. If they are, we further identify whether the tweet propagates the misconception (Agree) or is informative by contradicting it (Disagree). Example misconception-tweet pairs for each label are illustrated in Figure 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 641, |
| "end": 649, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We provide benchmark results for existing NLP models for this task. First, we evaluate text similarity models on their ability to detect whether a tweet is relevant to a given misconception (a.k.a misconception retrieval). Following prior work on fact verification and fake news detection (Yang et al., 2019) , we evaluate NLI models on misinformation (a.k.a. stance detection), by equating the class labels Agree, Disagree, and No Stance to Entailment, Contradiction, and Neutral, respectively. Our results show that existing models struggle at both tasks (38.7 Hits@1 for retrieval and 32.5 macro F1 on stance detection), however improve considerably after domain adaptation (Gururangan et al. (2020) ; 61.3 Hits@1 for retrieval and 50.2 macro F1 on stance detection).", |
| "cite_spans": [ |
| { |
| "start": 289, |
| "end": 308, |
| "text": "(Yang et al., 2019)", |
| "ref_id": "BIBREF46" |
| }, |
| { |
| "start": 677, |
| "end": 702, |
| "text": "(Gururangan et al. (2020)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "While our initial results using domain adaptation are encouraging, they leave much room for improvement. There is still much work that needs to be done before NLP systems can be seriously considered for combating COVID-19-related misinformation, and we hope COVIDLIES will be useful to help researchers understand when such systems are ready to be deployed.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We assume access to a collection of positively phrased known misconceptions M = {m 1 , . . . , m |M | }, e.g., \"Wearing masks does not prevent spread of COVID-19.\" is a misconception. As we describe later, the set of misconceptions in this work are vetted, curated, and maintained by medical researchers. Given a collection of tweets, T = {t 1 , . . . , t |T | }, the task is to determine, for each input t, whether there exists a misconception m \u2208 M that is being discussed, and if so, whether the discussion propagates the misconception (i.e., identifies m as true, and thus is spreading the misconception) or refutes the misconception (i.e., identifies m as false). This task is naturally separated into the following steps (shown in Figure 2 ):", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 737, |
| "end": 745, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "1. Misconception Retrieval: Given t return a subset M t \u2286 M of relevant misconceptions. 2. Stance Detection: For each (m, t) pair (m \u2208 M t )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": ", predict whether the m and t Agree, Disagree, or t takes No Stance with respect to m.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Due to limited availability of labeled data specific to this problem, we expect that models will need to be supervised on other, related tasks. For misconception retrieval, for example, relevant misconceptions can be ranked by measuring the semantic similarity between the tweet and each misconception, e.g., using cosine similarity between average word embeddings or more recent transformerbased methods such as BERTSCORE (Zhang et al., 2019) . For the stance detection sub-task, the problem can be recast as natural language inference (NLI), mapping the tweet t to the premise, the misconception m to the hypothesis, and the Agree, Disagree, and No Stance labels to Entailment, Contradiction, and Neutral, respectively. Due to novel language used to describe the disease and its associated misconceptions, existing misinformation detection dataset are unlikely to be effective for evaluating systems designed to detect COVID-19-related misinformation on social media. To facilitate research on this problem, we collect an evaluation dataset, COVIDLIES; the collection process is described below.", |
| "cite_spans": [ |
| { |
| "start": 423, |
| "end": 443, |
| "text": "(Zhang et al., 2019)", |
| "ref_id": "BIBREF47" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Misconceptions We extract misconceptions from a Wikipedia article about misinformation related to the COVID-19 pandemic (Wikipedia, 2020) . The extracted statements are manually examined, and statements that are not misinformation are removed. We manually rephrase the misinformation statements to a positive expression of that misinformation, e.g. \"Some conspiracy theorists also alleged that the coronavirus outbreak was cover-up for a 5G-related illness\" is shortened to \"Coronavirus is caused by 5G\".", |
| "cite_spans": [ |
| { |
| "start": 120, |
| "end": 137, |
| "text": "(Wikipedia, 2020)", |
| "ref_id": "BIBREF43" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Tweets Our source of tweets is the collection of COVID-19-related tweets identified by Chen et al. (2020) . We only use tweets from March and April 2020, and filter out non-English tweets.", |
| "cite_spans": [ |
| { |
| "start": 87, |
| "end": 105, |
| "text": "Chen et al. (2020)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Annotation Process To help identify tweets related to our list of misconceptions, we use BERTSCORE (Zhang et al., 2019) to compute a similarity metric on tweet-misconception pairs. For each given misconception, the 100 most similar tweets are selected for annotation. Each of these tweet-misconception pairs is manually labeled by researchers in the UCI School of Medicine as either: Agree (tweet is a positive expression of the misconception), Disagree (tweet contradicts/disagrees with the misconception), or No Stance (tweet is neutral or not relevant to the misconception).", |
| "cite_spans": [ |
| { |
| "start": 99, |
| "end": 119, |
| "text": "(Zhang et al., 2019)", |
| "ref_id": "BIBREF47" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Annotation Quality To evaluate inter-rater reliability, we randomly chose a subset of 200 tweetmisconception pairs and had four researchers manually label the subset. Percent agreement between researchers was 79%. Fleiss Kappa score was 0.69 which indicates substantial agreement between researchers (0.61-0.8). Disagreements were discussed and resolved before continuing to label the remaining tweet-misconception pairs in the dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Most disagreements came down to labeler interpretation. For example, given the misconception, \"Drinking large amounts of water will protect against coronavirus\", and a tweet of \"It's a good thing everyone is stocking up on water to survive the Coronavirus because the 128 OZ of Diet Coke, the double cheeseburgers, and radiation from our phones definitely won't kill us first\", one researcher labeled the pair as No Stance because it does not address any protective benefits but another researcher labeled the pair as Agree because people were stocking up on water to survive Coronavirus, which, to that researcher, implied water was protective. After discussions among the researchers, we concluded the pair was No Stance as it did not implicitly address the benefits of water and the statement was stated in a sarcastic tone. Other labeling challenges included deciding whether or not links or images in the tweet should be taken into account, as these could potentially change context of the tweet. We concluded that we would only evaluate the text as is since the various models would not be able to take images and links into account.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Problem Setup", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The current dataset contains 86 misconceptions, along with 6761 annotated tweet-misconception pairs. Statistics about the distribution of labels are provided in Table 2 . The distribution is heavily skewed, containing mostly No Stance tweets, and a higher proportion of Agree tweets than Disagree. The heavy skew towards No Stance tweets could be a due to the dataset construction methodology, specifically using BERTSCORE without fine-tuning to retrieve tweets per misconception. As we show in 4.2, domain adaptation significantly improves misconception matching. Further, presence of more Agree than Disagree tweets could be due to a bias in BERTSCORE towards scoring agreement higher.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 161, |
| "end": 168, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dataset Statistics", |
| "sec_num": null |
| }, |
| { |
| "text": "Top misconceptions for each class are shown in Table 1 . We only consider misconceptions with more than 80 annotated tweets, and rank the misconceptions for each class by the proportion of tweets that are annotated as that class. We present the top three misconceptions for each class with their corresponding percentage. There are misconceptions for which 100% of the paired annotated tweets express No Stance, which we do not see for the other two classes. We also notice that there are misconceptions with greater than 60% of paired tweets labeled as Agree; however, the highest proportion of Disagree labeled tweets found for any misconception in the Disagree class was 51%.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 47, |
| "end": 54, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dataset Statistics", |
| "sec_num": null |
| }, |
| { |
| "text": "COVIDLIES, however, is an evolving dataset; annotation is not yet complete for all 86 Wikipedia misconceptions matched to 100 tweets using BERTSCORE, and we are continually identifying additional misconceptions, as well as collecting more recent tweets for annotation. Further, we will gather more relevant tweets by using domainadapted retrieval models, which, as we will see in the next section, considerably outperform the current approach to retrieval, BERTSCORE.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dataset Statistics", |
| "sec_num": null |
| }, |
| { |
| "text": "Supervised classifiers have been used extensively for detecting misinformation (Wang, 2017; Karimi et al., 2018; Shu et al., 2017 Shu et al., , 2019 . However, existing tasks involve static or slowly evolving domains, and topics that do not require specific expertise to annotate. Gathering an annotated dataset large enough to be used for training a COVID-19 misinformation detector is difficult: the way misconceptions are expressed rapidly evolves, and identifying whether or not something is a misconception requires expertise in public health and medicine. Further, even the misconceptions themselves change over time as we learn more about the disease and the pandemic. Thus, it is desirable that COVID-19 misinformation detection systems are: (i) data efficient, e.g., trained with little to no supervision, and (ii) flexible, e.g., allow the addition, removal, or modification of the known misconceptions.", |
| "cite_spans": [ |
| { |
| "start": 92, |
| "end": 112, |
| "text": "Karimi et al., 2018;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 113, |
| "end": 129, |
| "text": "Shu et al., 2017", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 130, |
| "end": 148, |
| "text": "Shu et al., , 2019", |
| "ref_id": "BIBREF31" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Performance of Benchmark Models", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In this section, we investigate whether models trained for related tasks in natural language processing can be adapted to misinformation detection on the COVIDLIES dataset without additional training. We specifically focus on models that can be used to score two input sequences, i.e., tweet-misconception pairs. Because these models come pretrained on different tasks, they are naturally data efficient, and furthermore, due to their pairwise nature, are also flexible as modification of supported misconceptions is performed at the input level. Our code, dataset, and a demo of our best performing system are all available at https://ucinlp.github.io/covid19.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Performance of Benchmark Models", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In the misconception retrieval sub-task, for a each tweet, t, the goal is to retrieve all the misconceptions that the tweet refers to (i.e. may be labeled Agree or Disagree by the annotators). Note: for clearer description, we introduce a new \"pseudolabel\", Relevant, to refer to misconceptions that either Agree or Disagree with a given tweet. We treat this as a ranking task, where for each tweet, t, the system ranks the list of misconceptions, M , in decreasing order of relevancy. We evaluate this ranking using the standard information retrieval metrics Hits@k and Mean Reciprocal Rank (MRR) for each Relevant misconception m \u2208 M * t . The stance detection sub-task is a standard classification problem with three classes (Agree, Disagree, and No Stance). As such we perform evaluation by measuring the precision, recall, and F1score of the predicted classes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Relevant (Agree or Disagree)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Agree", |
| "sec_num": null |
| }, |
| { |
| "text": "H@1 H@5 H@10 MRR H@1 H@5 H@10 MRR Table 4 : Stance Detection Performance. We present evaluation for classification of tweet-misconception pairs into Agree, Disagree, and, No Stance classes. Precision (P), Recall (R), and F1-Score (F1) are presented for each class as well as macro averaged values. DA indicates domain-adaptive pretraining on COVID-19 tweets.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 34, |
| "end": 41, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Agree", |
| "sec_num": null |
| }, |
| { |
| "text": "We evaluate a number of information retrieval and semantic similarity approaches for the misconception retrieval sub-task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Misconception Retrieval", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Information Retrieval We use two information retrieval approaches. The first approach uses TF-IDF vectorization of tweets and misconceptions. Cosine similarity is used to score each tweetmisconception pair. Misconceptions are retrieved for each tweet in decreasing order of this score. NLTK is used for tokenization and vectorization.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Misconception Retrieval", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The second approach uses the BM25 algorithm, a bag-of-words retrieval technique which retrieves documents in decreasing probability of relevance of the query term. IDF and document lengths are used to determine probability of relevance. We use the pyserini implementation of BM25 to retrieve misconceptions for each tweet.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Misconception Retrieval", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We obtain vectorized representations of tweets and misconceptions using word embeddings. We then use two approaches for computing the semantic similarity between them: (i) cosine similarity computed between average to-ken embeddings, and (ii) BERTSCORE (Zhang et al., 2019) , which involves computation over BERT token embeddings of the tweet and misconception to obtain an F1-score-like measurement that we use as a similarity score.", |
| "cite_spans": [ |
| { |
| "start": 253, |
| "end": 273, |
| "text": "(Zhang et al., 2019)", |
| "ref_id": "BIBREF47" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic Similarity", |
| "sec_num": null |
| }, |
| { |
| "text": "For the cosine similarity approach, we experiment with both non-contextualized and contextualized word embeddings. For non-contexualized word embeddings we use 300D GloVe trained on 2014- Wikipedia and Gigaword embeddings (Pennington et al., 2014) . For contexualized embeddings we use a pretrained BERT-LARGE (Devlin et al., 2018) model. However, Since BERT is not trained on COVID-19-related text we also use COVID-Twitter-BERT 2 (M\u00fcller et al., 2020) which uses domain-adaptive pretraining (Gururangan et al., 2020) on 160M tweets about COVID-19. For sake of brevity, we will append the suffix (DA) to models that use COVID-Twitter-BERT instead of spelling out the full model name.", |
| "cite_spans": [ |
| { |
| "start": 188, |
| "end": 247, |
| "text": "Wikipedia and Gigaword embeddings (Pennington et al., 2014)", |
| "ref_id": null |
| }, |
| { |
| "start": 299, |
| "end": 331, |
| "text": "BERT-LARGE (Devlin et al., 2018)", |
| "ref_id": null |
| }, |
| { |
| "start": 432, |
| "end": 453, |
| "text": "(M\u00fcller et al., 2020)", |
| "ref_id": "BIBREF21" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic Similarity", |
| "sec_num": null |
| }, |
| { |
| "text": "We present the performance of similarity models in Table 3 . Average embedding, both with GloVe and BERT embeddings, perform the worst (and are fairly similar to each other). Although information retrieval based approaches, TF-IDF and BM25, considerably outperform the average embedding techniques, BERTSCORE captures the similarity as accurately as well. Domain adaptation, however, further improves the embedding-based similarity techniques, improving average BERT embeddings to be as good as others, while making BERTSCORE much more accurate than all other techniques. Thus we see that using domain adaptation and BERTSCORE are both important for performing accurate misconception retrieval.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 51, |
| "end": 58, |
| "text": "Table 3", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": null |
| }, |
| { |
| "text": "We illustrate the differences in the similarity models using example predictions in Table 5 . The first example provides a challenging case of retrieval that requires taking both COVID-19 knowledge and contextual information (e.g. multiple sentences, 'testing' vs 'tests') into account, and thus only the BERTSCORE (DA) model is able to retrieve the correct misconception. The second example primarily requires domain knowledge that 'coronavirus' and 'Sars-cov-2' are very similar, and only domain-adapted models are able to score the correct misconception highest. The last example shows when contextual embeddings (BERT) outperform non-contextual embedding (GloVe).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 84, |
| "end": 91, |
| "text": "Table 5", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": null |
| }, |
| { |
| "text": "Due to the lack of adequately large datasets for stance detection with pairs of sentences (Mohammad et al., 2016; Ferreira and Vlachos, 2016; Gorrell et al., 2018) , we cannot use existing datasets to train models for our setup. However, since classes in misinformation detection correspond to those in natural language inference (NLI), a task with much larger training datasets, we instead experiment with adapting NLI models on this task.", |
| "cite_spans": [ |
| { |
| "start": 90, |
| "end": 113, |
| "text": "(Mohammad et al., 2016;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 114, |
| "end": 141, |
| "text": "Ferreira and Vlachos, 2016;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 142, |
| "end": 163, |
| "text": "Gorrell et al., 2018)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Stance Detection, using NLI Models", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "We train linear classifiers on three common NLI datasets-SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) , and MedNLI (Shivade, 2019) . These classifiers use the following features, respectively: (i) concatenated unigram and bigram TF-IDF vectors for each input, (ii) concatenated average GloVe embeddings for each input, (iii) Bidirectional LSTM encoding, and (iv) the Sentence-BERT (SBERT) (Reimers and Gurevych, 2019) representation that uses siamese and triplet networks to obtain semantically meaningful sentence embeddings. Note that for (iii) and (iv), the transformer architectures (BiLSTM and SBERT) are jointly trained with the linear classifier. BERTSCORE (DA) + NLI Since BERTSCORE with domain adaptation performs best at retrieval for relevant classes, we use it to improve stance detection. We combine BERTSCORE (DA) with NLI models, initially classifying tweet-misconception pairs with high BERTSCORE scores (>0.4) as Relevant, subsequently using the NLI model to determine whether the pair Agree or Disagree. We denote such \"combined models\" by inserting a plus sign between the retrieval model and the NLI model, e.g., BERTSCORE (DA) + BiLSTM denotes a model that uses BERTSCORE (DA) to determine retrieve relevant misconceptions and a BiL-STM NLI model for classifying the stance. Table 4 show that, generally, most models do not perform well on the Agree and Disagree classes, which are minority classes in our dataset. On the other hand, performance on No Stance is high; quite a few models achieve an F1-score of 89% or higher. BERTSCORE (DA) + SBERT (DA) (on MultiNLI) achieves the highest F1 (41.2) for the Agree class, while also obtaining the highest macro averaged Precision (55.9%) and F1 (50.2). The combined BERTSCORE (DA) + NLI approach, in general, improves F1 across all classes for all models.", |
| "cite_spans": [ |
| { |
| "start": 94, |
| "end": 117, |
| "text": "(Williams et al., 2018)", |
| "ref_id": "BIBREF45" |
| }, |
| { |
| "start": 131, |
| "end": 146, |
| "text": "(Shivade, 2019)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 405, |
| "end": 433, |
| "text": "(Reimers and Gurevych, 2019)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1312, |
| "end": 1319, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Stance Detection, using NLI Models", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Tweet:In order for accurate information about the #coronavirus to be obtained, you have to be able to do widespread testing. The U.S. is behind many other countries for a variety of reasons. #COVID19 #COVID19US https://t.co/muzxHjY0XP Misconception: Anybody in the U.S. who wants a COVID-19 test can get a test. Tweet: There is evidence that coronaviruses can live on inanimate surfaces for up to nine days, but its not yet clear how likely humans are to be infected by touching these surfaces. https://t.co/DJ99AAISWw Misconception: Sars-cov 2 can survive for prolonged periods of time on surfaces. Tweet: Covid-19 is about 43 times more deadly if you get it, but China's number of cases is leveling off at around 80K which is much less than the number of US flu cases. If the number of cases is kept small then Covid-19 will be minor compared to the flu. Misconception: COVID-19 is only as deadly as the seasonal flu. We present examples to demonstrate the difference in performance between some of the semantic similarity models: (I) BERTSCORE (DA), (II) Avg. BERT (DA), (III) Vanilla BERTScore, and (IV) Avg. GloVe. = The model retrieved the relevant misconception for a tweet with rank 1; = The model did not score the relevant misconception for the tweet with rank 1.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "I II III IV Example", |
| "sec_num": null |
| }, |
| { |
| "text": "SBERT Disagree Tweet: @IVANISTHEMAN @OriginalDWoods @jboog3000 It didnt come from an animal chinese spies smuggled a form of the virus to china from a canadian lab and was then \"leaked\". SBERT (DA) Agree Misconception: Coronavirus was taken from a Canadian lab or is the result of bioweapons defense research in China.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Models Labels Example", |
| "sec_num": null |
| }, |
| { |
| "text": "Tweet: @Acyn The corona virus can live on a surface for up to 9 days. Just saying.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "No Stance", |
| "sec_num": null |
| }, |
| { |
| "text": "Agree Misconception: Sars-cov 2 can survive for prolonged periods of time on surfaces.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "BERTSCORE (DA) + BiLSTM", |
| "sec_num": null |
| }, |
| { |
| "text": "No Stance Tweet: @alexsalvinews Alex. Check out Dean Koontz, The Eyes of Darkness. 1981. He predicts the Wuhan-400 virus. He said in \"around\" 2020, a pneumonia-like virus will be spread worldwide. BERTSCORE (DA) + SBERT (DA) Agree Misconception: Dean Koontz predicted the pandemic in his 1981 novel The Eyes of Darkness. Examples of stance predictions in Table 6 illustrate the differences between these models. The first example demonstrates that knowledge about the domain vocabulary helps domain adapted models in predicting the correct stance, as it did for retrieval. The remaining two examples both show the advantage of the combined BERTSCORE (DA) + NLI approach, in particular, demonstrating that retrieval models are effective at identifying relevant misconceptions, which the NLI models are then able to correctly classify the stance of.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 355, |
| "end": 362, |
| "text": "Table 6", |
| "ref_id": "TABREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "SBERT (DA)", |
| "sec_num": null |
| }, |
| { |
| "text": "COVID-19: In the social sciences, there have been recent efforts to quantify COVID-19 misinformation on social media (Brennen et al., 2020; Kouzy et al., 2020) , as well experimental efforts to prevent propagation of misinformation (Pennycook et al., 2020) . At the same time, members of the NLP community have been working on developing tools for the automatic detection of COVID-19related misinformation online. Serrano et al. (2020) detect YouTube videos spreading conspiracy theories using features of user comments, and Dharawat et al. (2020) classify tweets by the severity of health risks associated with them. McQuillan et al. (2020) study the behaviour of COVID-19 misinformation networks on Twitter using mapping, topic modeling, bridging centrality, and divergence. Penn Medicine launched a chatbot to provide patients with accurate information about the virus (Volpp-Kevin et al., 2020) , and a crowdsourced chatbot, Jennifer, is also available to answer questions about the pandemic (Li et al., 2020) . We are the first to frame COVID-19 misinformation detection as a two-stage task of misconception retrieval and pair- wise classification of stance, and add to this body of work by providing a dataset and benchmark models for automated identification of misinformation.", |
| "cite_spans": [ |
| { |
| "start": 117, |
| "end": 139, |
| "text": "(Brennen et al., 2020;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 140, |
| "end": 159, |
| "text": "Kouzy et al., 2020)", |
| "ref_id": null |
| }, |
| { |
| "start": 232, |
| "end": 256, |
| "text": "(Pennycook et al., 2020)", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 414, |
| "end": 435, |
| "text": "Serrano et al. (2020)", |
| "ref_id": null |
| }, |
| { |
| "start": 872, |
| "end": 898, |
| "text": "(Volpp-Kevin et al., 2020)", |
| "ref_id": null |
| }, |
| { |
| "start": 996, |
| "end": 1013, |
| "text": "(Li et al., 2020)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Misinformation Detection: There are several datasets for misinformation detection with binary veracity labels , for example, FakeNewsNet (Shu et al., 2017 (Shu et al., , 2019 (Shu et al., , 2020 consisting of news articles, Some Like It Hoax (Tacchini et al., 2017) consisting of Facebook posts, and PHEME containing twitter threads. Misinformation detection is also closely related to fact-checking since both tasks aim to assess the veracity of claims. FEVER (Thorne et al., , 2019 ) is a dataset of claims and evidence pairs with Supported, Refuted or NotEnoughInfo labels to facilitate research in automated fact checking. This is similar to Emergent (Ferreira and Vlachos, 2016) , a stance classification dataset consisting of rumored claims and associated news articles with labels of For, Against, or Observing the claim. Stance detection is also the focus of the Fake News Challenge (FNC-1) 3 consisting of pairs of news article headlines and body texts with Agrees, Disagrees, Discusses, and Unrelated labels.", |
| "cite_spans": [ |
| { |
| "start": 137, |
| "end": 154, |
| "text": "(Shu et al., 2017", |
| "ref_id": "BIBREF30" |
| }, |
| { |
| "start": 155, |
| "end": 174, |
| "text": "(Shu et al., , 2019", |
| "ref_id": "BIBREF31" |
| }, |
| { |
| "start": 175, |
| "end": 194, |
| "text": "(Shu et al., , 2020", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 242, |
| "end": 265, |
| "text": "(Tacchini et al., 2017)", |
| "ref_id": "BIBREF32" |
| }, |
| { |
| "start": 461, |
| "end": 483, |
| "text": "(Thorne et al., , 2019", |
| "ref_id": null |
| }, |
| { |
| "start": 655, |
| "end": 683, |
| "text": "(Ferreira and Vlachos, 2016)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Our proposed models for detecting misinformation by using classifiers fall within the framework of detecting misinformation using content features (Volkova et al., 2017; Wei and Wan, 2017) . Other approaches include using crowd behaviour (Tschiatschek et al., 2018; Mendoza et al., 2010) , reliability of the source (Lumezanu et al., 2012; Li et al., 2015) , knowledge graphs (Ciampaglia et al., 2015) , or a combination of these approaches 6 Conclusions and Future Work", |
| "cite_spans": [ |
| { |
| "start": 147, |
| "end": 169, |
| "text": "(Volkova et al., 2017;", |
| "ref_id": "BIBREF39" |
| }, |
| { |
| "start": 170, |
| "end": 188, |
| "text": "Wei and Wan, 2017)", |
| "ref_id": "BIBREF42" |
| }, |
| { |
| "start": 238, |
| "end": 265, |
| "text": "(Tschiatschek et al., 2018;", |
| "ref_id": "BIBREF38" |
| }, |
| { |
| "start": 266, |
| "end": 287, |
| "text": "Mendoza et al., 2010)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 316, |
| "end": 339, |
| "text": "(Lumezanu et al., 2012;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 340, |
| "end": 356, |
| "text": "Li et al., 2015)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 376, |
| "end": 401, |
| "text": "(Ciampaglia et al., 2015)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The ongoing COVID-19 pandemic has been accompanied by a corresponding 'infodemic' of misinformation about the virus. It is important to develop tools to automatically detect misinformation online, especially on social media sites where the volume and speed of the spread are high. However, rapidly evolving information and novel language make existing misinformation detection datasets and models ineffective for detecting COVID-19 misinformation. In this paper, to initiate research on this important and timely topic, we introduced COVIDLIES, a benchmark dataset containing known COVID-19 misconceptions accompanied with tweets that Agree, Disagree, or express No Stance for each misconception, annotated by experts. Our code, dataset, and a demo of our best performing system are publicly available at https://ucinlp.github.io/covid19.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Given a tweet, we formulate the task of detecting misinformation as retrieving relevant misconceptions, and classifying whether the tweet supports or refutes it. We evaluate a number of approaches for this task, including common semantic similarity models for retrieval, accurate models trained on a variety of NLI datasets, and domain adaptation by pretraining language models on a corpus of COVID-19 tweets. We demonstrate domain adaptation significantly improves results for both subtasks of misinformation detection. We also show that it is feasible to detect the stance of tweets towards misconceptions using models trained on existing NLI datasets. However, the performance has considerable scope for improvement since existing NLI datasets do not contain texts on COVID-19 and are linguistically different from tweets. Future work will involve using models trained on more domain specific and linguistically similar text. We plan to continually expand our annotated dataset by including posts from other domains such as news articles and Reddit, and misconceptions from sources beyond Wikipedia, such as Poynter (2020). We invite researchers to build COVID-19 misinformation detection systems and evaluate their performance using the presented dataset.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "5" |
| }, |
| { |
| "text": "https://huggingface.co/digitalepidemiologylab/ covid-twitter-bert", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "http://www.fakenewschallenge.org/(Castillo et al., 2011;Kumar et al., 2016). Adapting these techniques to COVID-19 misinformation is a promising direction for future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We thank the anonymous reviewers of EMNLP 2020 NLP COVID-19 workshop for their comments and the authors of related work for publishing their code and data. We would like to acknowledge Nicole Woodruff, an undergraduate at UCLA, Aileen Guillen, a medical student at UCI, Sadhika Yamasani, and Victoria Rong, undergraduates at UCI, for volunteering to help annotate data for this project. We would also like to thank Lidia Flores, staff research associate in Dr. Young's Lab, for her help in compiling tweets. Finally, we would also like to thank Elena Kochkina and Maria Liakata of the Alan Turing Institute for helpful discussions and feedback. This material is based upon work sponsored in part by NSF award #IIS-1817183 and in part by the DARPA MCS program under Contract No. N660011924033 with the United States Office Of Naval Research.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Hundreds dead' because of Covid-19 misinformation", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bbc", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "BBC. 2020. 'Hundreds dead' because of Covid-19 mis- information. https://www.bbc.com/news/world- 53755067 [Accessed on June 30, 2020].", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "A large annotated corpus for learning natural language inference", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Samuel", |
| "suffix": "" |
| }, |
| { |
| "first": "Gabor", |
| "middle": [], |
| "last": "Bowman", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Angeli", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Potts", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large anno- tated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Types, sources, and claims of COVID-19 misinformation", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Scott Brennen", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Felix", |
| "suffix": "" |
| }, |
| { |
| "first": "Philip", |
| "middle": [ |
| "N" |
| ], |
| "last": "Simon", |
| "suffix": "" |
| }, |
| { |
| "first": "Rasmus Kleis", |
| "middle": [], |
| "last": "Howard", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Nielsen", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J Scott Brennen, Felix M Simon, Philip N Howard, and Rasmus Kleis Nielsen. 2020. Types, sources, and claims of COVID-19 misinformation. Reuters Insti- tute.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Information credibility on twitter", |
| "authors": [ |
| { |
| "first": "Carlos", |
| "middle": [], |
| "last": "Castillo", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcelo", |
| "middle": [], |
| "last": "Mendoza", |
| "suffix": "" |
| }, |
| { |
| "first": "Barbara", |
| "middle": [], |
| "last": "Poblete", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the 20th international conference on World wide web", |
| "volume": "", |
| "issue": "", |
| "pages": "675--684", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Carlos Castillo, Marcelo Mendoza, and Barbara Poblete. 2011. Information credibility on twitter. In Proceedings of the 20th international conference on World wide web, pages 675-684.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Tracking Social Media Discourse About the COVID-19 Pandemic: Development of a Public Coronavirus Twitter Data Set", |
| "authors": [ |
| { |
| "first": "Emily", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Lerman", |
| "suffix": "" |
| }, |
| { |
| "first": "Emilio", |
| "middle": [], |
| "last": "Ferrara", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "JMIR Public Health and Surveillance", |
| "volume": "6", |
| "issue": "2", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emily Chen, Kristina Lerman, and Emilio Ferrara. 2020. Tracking Social Media Discourse About the COVID-19 Pandemic: Development of a Public Coronavirus Twitter Data Set. JMIR Public Health and Surveillance, 6(2):e19273.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Computational fact checking from knowledge networks", |
| "authors": [ |
| { |
| "first": "Giovanni", |
| "middle": [], |
| "last": "Luca Ciampaglia", |
| "suffix": "" |
| }, |
| { |
| "first": "Prashant", |
| "middle": [], |
| "last": "Shiralkar", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Luis", |
| "suffix": "" |
| }, |
| { |
| "first": "Johan", |
| "middle": [], |
| "last": "Rocha", |
| "suffix": "" |
| }, |
| { |
| "first": "Filippo", |
| "middle": [], |
| "last": "Bollen", |
| "suffix": "" |
| }, |
| { |
| "first": "Alessandro", |
| "middle": [], |
| "last": "Menczer", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Flammini", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "PloS one", |
| "volume": "10", |
| "issue": "6", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Giovanni Luca Ciampaglia, Prashant Shiralkar, Luis M Rocha, Johan Bollen, Filippo Menczer, and Alessandro Flammini. 2015. Computational fact checking from knowledge networks. PloS one, 10(6):e0128193.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Covid-19 misinformation is spreading in 25 different languages, leading to deaths and injuries, study finds", |
| "authors": [], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "CNN. 2020. Covid-19 misinformation is spreading in 25 different languages, lead- ing to deaths and injuries, study finds. https://www.cnn.com/2020/08/10/health/covid- misinformation-spread-study-wellness/index.html [Accessed on June 30, 2020].", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Bert: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1810.04805" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understand- ing. arXiv preprint arXiv:1810.04805.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Drink bleach or do what now? Covid-HeRA: A dataset for riskinformed health decision making in the presence of COVID19 misinformation", |
| "authors": [ |
| { |
| "first": "Ismini", |
| "middle": [], |
| "last": "Arkin R Dharawat", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Lourentzou", |
| "suffix": "" |
| }, |
| { |
| "first": "Chengxiang", |
| "middle": [], |
| "last": "Morales", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Zhai", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arkin R Dharawat, Ismini Lourentzou, Alex Morales, and ChengXiang Zhai. 2020. Drink bleach or do what now? Covid-HeRA: A dataset for risk- informed health decision making in the presence of COVID19 misinformation. https://openreview. net/forum?id=PmY1SNmJlEC [Accessed on July 1, 2020].", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Emergent: a novel data-set for stance classification", |
| "authors": [ |
| { |
| "first": "William", |
| "middle": [], |
| "last": "Ferreira", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Vlachos", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: Human language technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "1163--1168", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "William Ferreira and Andreas Vlachos. 2016. Emer- gent: a novel data-set for stance classification. In Proceedings of the 2016 conference of the North American chapter of the association for computa- tional linguistics: Human language technologies, pages 1163-1168.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Rumoureval 2019: Determining rumour veracity and support for rumours", |
| "authors": [ |
| { |
| "first": "Genevieve", |
| "middle": [], |
| "last": "Gorrell", |
| "suffix": "" |
| }, |
| { |
| "first": "Kalina", |
| "middle": [], |
| "last": "Bontcheva", |
| "suffix": "" |
| }, |
| { |
| "first": "Leon", |
| "middle": [], |
| "last": "Derczynski", |
| "suffix": "" |
| }, |
| { |
| "first": "Elena", |
| "middle": [], |
| "last": "Kochkina", |
| "suffix": "" |
| }, |
| { |
| "first": "Maria", |
| "middle": [], |
| "last": "Liakata", |
| "suffix": "" |
| }, |
| { |
| "first": "Arkaitz", |
| "middle": [], |
| "last": "Zubiaga", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1809.06683" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Genevieve Gorrell, Kalina Bontcheva, Leon Derczyn- ski, Elena Kochkina, Maria Liakata, and Arkaitz Zubiaga. 2018. Rumoureval 2019: Determining rumour veracity and support for rumours. arXiv preprint arXiv:1809.06683.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Don't stop pretraining: Adapt language models to domains and tasks", |
| "authors": [ |
| { |
| "first": "Ana", |
| "middle": [], |
| "last": "Suchin Gururangan", |
| "suffix": "" |
| }, |
| { |
| "first": "Swabha", |
| "middle": [], |
| "last": "Marasovi", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyle", |
| "middle": [], |
| "last": "Swayamdipta", |
| "suffix": "" |
| }, |
| { |
| "first": "Iz", |
| "middle": [], |
| "last": "Lo", |
| "suffix": "" |
| }, |
| { |
| "first": "Doug", |
| "middle": [], |
| "last": "Beltagy", |
| "suffix": "" |
| }, |
| { |
| "first": "Noah", |
| "middle": [ |
| "A" |
| ], |
| "last": "Downey", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Suchin Gururangan, Ana Marasovi, Swabha Swayamdipta, Kyle Lo, Iz Beltagy, Doug Downey, and Noah A. Smith. 2020. Don't stop pretraining: Adapt language models to domains and tasks. In Proceedings of ACL.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Multi-source multi-class fake news detection", |
| "authors": [ |
| { |
| "first": "Hamid", |
| "middle": [], |
| "last": "Karimi", |
| "suffix": "" |
| }, |
| { |
| "first": "Proteek", |
| "middle": [], |
| "last": "Roy", |
| "suffix": "" |
| }, |
| { |
| "first": "Sari", |
| "middle": [], |
| "last": "Saba-Sadiya", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiliang", |
| "middle": [], |
| "last": "Tang", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 27th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "1546--1557", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hamid Karimi, Proteek Roy, Sari Saba-Sadiya, and Jiliang Tang. 2018. Multi-source multi-class fake news detection. In Proceedings of the 27th Inter- national Conference on Computational Linguistics, pages 1546-1557.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Elie W Akl, and Khalil Baddour. 2020. Coronavirus goes viral: quantifying the COVID-19 misinformation epidemic on Twitter", |
| "authors": [ |
| { |
| "first": "Ramez", |
| "middle": [], |
| "last": "Kouzy", |
| "suffix": "" |
| }, |
| { |
| "first": "Joseph", |
| "middle": [], |
| "last": "Abi Jaoude", |
| "suffix": "" |
| }, |
| { |
| "first": "Afif", |
| "middle": [], |
| "last": "Kraitem", |
| "suffix": "" |
| }, |
| { |
| "first": "Molly", |
| "middle": [ |
| "B" |
| ], |
| "last": "El Alam", |
| "suffix": "" |
| }, |
| { |
| "first": "Basil", |
| "middle": [], |
| "last": "Karam", |
| "suffix": "" |
| }, |
| { |
| "first": "Elio", |
| "middle": [], |
| "last": "Adib", |
| "suffix": "" |
| }, |
| { |
| "first": "Jabra", |
| "middle": [], |
| "last": "Zarka", |
| "suffix": "" |
| }, |
| { |
| "first": "Cindy", |
| "middle": [], |
| "last": "Traboulsi", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ramez Kouzy, Joseph Abi Jaoude, Afif Kraitem, Molly B El Alam, Basil Karam, Elio Adib, Jabra Zarka, Cindy Traboulsi, Elie W Akl, and Khalil Bad- dour. 2020. Coronavirus goes viral: quantifying the COVID-19 misinformation epidemic on Twitter. Cureus, 12(3).", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Disinformation on the Web: Impact, Characteristics, and Detection of Wikipedia Hoaxes", |
| "authors": [ |
| { |
| "first": "Srijan", |
| "middle": [], |
| "last": "Kumar", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "West", |
| "suffix": "" |
| }, |
| { |
| "first": "Jure", |
| "middle": [], |
| "last": "Leskovec", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 25th international conference on World Wide Web", |
| "volume": "", |
| "issue": "", |
| "pages": "591--602", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Srijan Kumar, Robert West, and Jure Leskovec. 2016. Disinformation on the Web: Impact, Characteristics, and Detection of Wikipedia Hoaxes. In Proceedings of the 25th international conference on World Wide Web, pages 591-602.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "On the discovery of evolving truth", |
| "authors": [ |
| { |
| "first": "Yaliang", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Qi", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Jing", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| }, |
| { |
| "first": "Lu", |
| "middle": [], |
| "last": "Su", |
| "suffix": "" |
| }, |
| { |
| "first": "Bo", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Fan", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiawei", |
| "middle": [], |
| "last": "Han", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "675--684", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yaliang Li, Qi Li, Jing Gao, Lu Su, Bo Zhao, Wei Fan, and Jiawei Han. 2015. On the discovery of evolv- ing truth. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 675-684.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Jennifer for COVID-19: An NLP-Powered Chatbot Built for the People and by the People to Combat Misinformation", |
| "authors": [ |
| { |
| "first": "Yunyao", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Tyrone", |
| "middle": [], |
| "last": "Grandison", |
| "suffix": "" |
| }, |
| { |
| "first": "Patricia", |
| "middle": [], |
| "last": "Silveyra", |
| "suffix": "" |
| }, |
| { |
| "first": "Ali", |
| "middle": [], |
| "last": "Douraghy", |
| "suffix": "" |
| }, |
| { |
| "first": "Xinyu", |
| "middle": [], |
| "last": "Guan", |
| "suffix": "" |
| }, |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Kieselbach", |
| "suffix": "" |
| }, |
| { |
| "first": "Chengkai", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Haiqi", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "ACL", |
| "volume": "2020", |
| "issue": "", |
| "pages": "Workshop NLP--COVID", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yunyao Li, Tyrone Grandison, Patricia Silveyra, Ali Douraghy, Xinyu Guan, Thomas Kieselbach, Chengkai Li, and Haiqi Zhang. 2020. Jennifer for COVID-19: An NLP-Powered Chatbot Built for the People and by the People to Combat Misinformation. ACL 2020 Workshop NLP-COVID.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "# bias: Measuring the tweeting behavior of propagandists", |
| "authors": [ |
| { |
| "first": "Cristian", |
| "middle": [], |
| "last": "Lumezanu", |
| "suffix": "" |
| }, |
| { |
| "first": "Nick", |
| "middle": [], |
| "last": "Feamster", |
| "suffix": "" |
| }, |
| { |
| "first": "Hans", |
| "middle": [], |
| "last": "Klein", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Sixth International AAAI Conference on Weblogs and Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cristian Lumezanu, Nick Feamster, and Hans Klein. 2012. # bias: Measuring the tweeting behavior of propagandists. In Sixth International AAAI Confer- ence on Weblogs and Social Media.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Cultural convergence: Insights into the behavior of misinformation networks on twitter", |
| "authors": [ |
| { |
| "first": "Liz", |
| "middle": [], |
| "last": "Mcquillan", |
| "suffix": "" |
| }, |
| { |
| "first": "Erin", |
| "middle": [], |
| "last": "Mcaweeney", |
| "suffix": "" |
| }, |
| { |
| "first": "Alicia", |
| "middle": [], |
| "last": "Bargar", |
| "suffix": "" |
| }, |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Ruch", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:2007.03443" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liz McQuillan, Erin McAweeney, Alicia Bargar, and Alex Ruch. 2020. Cultural convergence: Insights into the behavior of misinformation networks on twitter. arXiv preprint arXiv:2007.03443.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Twitter under crisis: Can we trust what we RT?", |
| "authors": [ |
| { |
| "first": "Marcelo", |
| "middle": [], |
| "last": "Mendoza", |
| "suffix": "" |
| }, |
| { |
| "first": "Barbara", |
| "middle": [], |
| "last": "Poblete", |
| "suffix": "" |
| }, |
| { |
| "first": "Carlos", |
| "middle": [], |
| "last": "Castillo", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the first workshop on social media analytics", |
| "volume": "", |
| "issue": "", |
| "pages": "71--79", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marcelo Mendoza, Barbara Poblete, and Carlos Castillo. 2010. Twitter under crisis: Can we trust what we RT? In Proceedings of the first workshop on social media analytics, pages 71-79.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "SemEval-2016 Task 6: Detecting Stance in Tweets", |
| "authors": [ |
| { |
| "first": "Saif", |
| "middle": [], |
| "last": "Mohammad", |
| "suffix": "" |
| }, |
| { |
| "first": "Svetlana", |
| "middle": [], |
| "last": "Kiritchenko", |
| "suffix": "" |
| }, |
| { |
| "first": "Parinaz", |
| "middle": [], |
| "last": "Sobhani", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaodan", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| }, |
| { |
| "first": "Colin", |
| "middle": [], |
| "last": "Cherry", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016)", |
| "volume": "", |
| "issue": "", |
| "pages": "31--41", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Saif Mohammad, Svetlana Kiritchenko, Parinaz Sob- hani, Xiaodan Zhu, and Colin Cherry. 2016. SemEval-2016 Task 6: Detecting Stance in Tweets. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), pages 31- 41.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter", |
| "authors": [ |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "M\u00fcller", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcel", |
| "middle": [], |
| "last": "Salath\u00e9", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Per", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Kummervold", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:2005.07503" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Martin M\u00fcller, Marcel Salath\u00e9, and Per E Kummervold. 2020. COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter. arXiv preprint arXiv:2005.07503.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Coronavirus doctors battle another scourge: Misinformation", |
| "authors": [], |
| "year": 2020, |
| "venue": "New", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "New York Times. 2020. Coronavirus doctors bat- tle another scourge: Misinformation. https: //www.nytimes.com/2020/08/17/technology/ coronavirus-disinformation-doctors.html [Ac- cessed on June 30, 2020].", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "GloVe: Global Vectors for Word Representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Empirical Methods in Natural Language Processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D. Manning. 2014. GloVe: Global Vectors for Word Representation. In Empirical Methods in Natural Language Processing (EMNLP), pages 1532-1543.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Fighting COVID-19 misinformation on social media: Experimental evidence for a scalable accuracy nudge intervention", |
| "authors": [ |
| { |
| "first": "Gordon", |
| "middle": [], |
| "last": "Pennycook", |
| "suffix": "" |
| }, |
| { |
| "first": "Jonathon", |
| "middle": [], |
| "last": "Mcphetres", |
| "suffix": "" |
| }, |
| { |
| "first": "Yunhao", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Rand", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gordon Pennycook, Jonathon McPhetres, Yunhao Zhang, and David Rand. 2020. Fighting COVID- 19 misinformation on social media: Experimental evidence for a scalable accuracy nudge intervention. PsyArXiv Preprints, 10.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "The coronavirusfacts/datoscoronavirus alliance database", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Poynter", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Poynter. 2020. The coronavirusfacts/datoscoronavirus alliance database. https://www.poynter.org/ifcn- covid-19-misinformation/ [Accessed on June 30, 2020].", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Sentence-BERT: Sentence Embeddings Using Siamese BERT-Networks", |
| "authors": [ |
| { |
| "first": "Nils", |
| "middle": [], |
| "last": "Reimers", |
| "suffix": "" |
| }, |
| { |
| "first": "Iryna", |
| "middle": [], |
| "last": "Gurevych", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1908.10084" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nils Reimers and Iryna Gurevych. 2019. Sentence- BERT: Sentence Embeddings Using Siamese BERT- Networks. arXiv preprint arXiv:1908.10084.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Orestis Papakyriakopoulos, and Simon Hegelich. 2020. NLP-based Feature Extraction for the Detection of COVID-19 Misinformation Videos on YouTube", |
| "authors": [ |
| { |
| "first": "Juan Carlos Medina", |
| "middle": [], |
| "last": "Serrano", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "ACL 2020 Workshop NLP-COVID", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Juan Carlos Medina Serrano, Orestis Papakyriakopou- los, and Simon Hegelich. 2020. NLP-based Feature Extraction for the Detection of COVID-19 Misinfor- mation Videos on YouTube. ACL 2020 Workshop NLP-COVID.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "MedNLI -A Natural Language Inference Dataset For The Clinical Domain", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Shivade", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| ".org/10.13026/C2RS98" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shivade. 2019. MedNLI -A Natural Language Inference Dataset For The Clinical Domain. doi.org/10.13026/C2RS98.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "FakeNewsNet: A Data Repository with News Content, Social Context, and Spatiotemporal Information for", |
| "authors": [ |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Shu", |
| "suffix": "" |
| }, |
| { |
| "first": "Deepak", |
| "middle": [], |
| "last": "Mahudeswaran", |
| "suffix": "" |
| }, |
| { |
| "first": "Suhang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Dongwon", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Huan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Studying Fake News on Social Media. Big Data", |
| "volume": "8", |
| "issue": "3", |
| "pages": "171--188", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kai Shu, Deepak Mahudeswaran, Suhang Wang, Dong- won Lee, and Huan Liu. 2020. FakeNewsNet: A Data Repository with News Content, Social Context, and Spatiotemporal Information for Studying Fake News on Social Media. Big Data, 8(3):171-188.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Fake news detection on social media: A data mining perspective", |
| "authors": [ |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Shu", |
| "suffix": "" |
| }, |
| { |
| "first": "Amy", |
| "middle": [], |
| "last": "Sliva", |
| "suffix": "" |
| }, |
| { |
| "first": "Suhang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiliang", |
| "middle": [], |
| "last": "Tang", |
| "suffix": "" |
| }, |
| { |
| "first": "Huan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "ACM SIGKDD explorations newsletter", |
| "volume": "19", |
| "issue": "1", |
| "pages": "22--36", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kai Shu, Amy Sliva, Suhang Wang, Jiliang Tang, and Huan Liu. 2017. Fake news detection on social me- dia: A data mining perspective. ACM SIGKDD ex- plorations newsletter, 19(1):22-36.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Beyond news contents: The role of social context for fake news detection", |
| "authors": [ |
| { |
| "first": "Kai", |
| "middle": [], |
| "last": "Shu", |
| "suffix": "" |
| }, |
| { |
| "first": "Suhang", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Huan", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining", |
| "volume": "", |
| "issue": "", |
| "pages": "312--320", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kai Shu, Suhang Wang, and Huan Liu. 2019. Beyond news contents: The role of social context for fake news detection. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, pages 312-320.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Some like it hoax: Automated fake news detection in social networks", |
| "authors": [ |
| { |
| "first": "Eugenio", |
| "middle": [], |
| "last": "Tacchini", |
| "suffix": "" |
| }, |
| { |
| "first": "Gabriele", |
| "middle": [], |
| "last": "Ballarin", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco L Della", |
| "middle": [], |
| "last": "Vedova", |
| "suffix": "" |
| }, |
| { |
| "first": "Stefano", |
| "middle": [], |
| "last": "Moret", |
| "suffix": "" |
| }, |
| { |
| "first": "Luca", |
| "middle": [], |
| "last": "De Alfaro", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1704.07506" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Eugenio Tacchini, Gabriele Ballarin, Marco L Della Ve- dova, Stefano Moret, and Luca de Alfaro. 2017. Some like it hoax: Automated fake news detection in social networks. arXiv preprint arXiv:1704.07506.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "Facebook, reddit, google, linkedin, microsoft, twitter and youtube issue joint statement on misinformation", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Techcrunch", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "TechCrunch. 2020. Facebook, reddit, google, linkedin, microsoft, twitter and youtube issue joint state- ment on misinformation. TechCrunch. https: //techcrunch.com/2020/03/16/facebook-reddit- google-linkedin-microsoft-twitter-and-youtube- issue-joint-statement-on-misinformation/ [Ac- cessed on June 30, 2020].", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "Fake news stance detection using stacked ensemble of classifiers", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Thorne", |
| "suffix": "" |
| }, |
| { |
| "first": "Mingjie", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Giorgos", |
| "middle": [], |
| "last": "Myrianthous", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiashu", |
| "middle": [], |
| "last": "Pu", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaoxuan", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Vlachos", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 2017 EMNLP Workshop: Natural Language Processing meets Journalism", |
| "volume": "", |
| "issue": "", |
| "pages": "80--83", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/W17-4214" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Thorne, Mingjie Chen, Giorgos Myrianthous, Jiashu Pu, Xiaoxuan Wang, and Andreas Vlachos. 2017. Fake news stance detection using stacked en- semble of classifiers. In Proceedings of the 2017 EMNLP Workshop: Natural Language Processing meets Journalism, pages 80-83, Copenhagen, Den- mark. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Automated fact checking: Task formulations, methods and future directions", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Thorne", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Vlachos", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 27th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "3346--3359", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Thorne and Andreas Vlachos. 2018. Automated fact checking: Task formulations, methods and fu- ture directions. In Proceedings of the 27th Inter- national Conference on Computational Linguistics, pages 3346-3359, Santa Fe, New Mexico, USA. As- sociation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF36": { |
| "ref_id": "b36", |
| "title": "The fact extraction and verification (fever) shared task", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Thorne", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Vlachos", |
| "suffix": "" |
| }, |
| { |
| "first": "Oana", |
| "middle": [], |
| "last": "Cocarascu", |
| "suffix": "" |
| }, |
| { |
| "first": "Christos", |
| "middle": [], |
| "last": "Christodoulopoulos", |
| "suffix": "" |
| }, |
| { |
| "first": "Arpit", |
| "middle": [], |
| "last": "Mittal", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1811.10971" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Thorne, Andreas Vlachos, Oana Cocarascu, Christos Christodoulopoulos, and Arpit Mittal. 2018. The fact extraction and verification (fever) shared task. arXiv preprint arXiv:1811.10971.", |
| "links": null |
| }, |
| "BIBREF37": { |
| "ref_id": "b37", |
| "title": "Christos Christodoulopoulos, and Arpit Mittal. 2019. The FEVER2.0 Shared Task", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Thorne", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Vlachos", |
| "suffix": "" |
| }, |
| { |
| "first": "Oana", |
| "middle": [], |
| "last": "Cocarascu", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "Second Workshop on Fact Extraction and VERification (FEVER)", |
| "volume": "", |
| "issue": "", |
| "pages": "1--6", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "James Thorne, Andreas Vlachos, Oana Cocarascu, Christos Christodoulopoulos, and Arpit Mittal. 2019. The FEVER2.0 Shared Task. In Second Work- shop on Fact Extraction and VERification (FEVER), pages 1-6.", |
| "links": null |
| }, |
| "BIBREF38": { |
| "ref_id": "b38", |
| "title": "Fake news detection in social networks via crowd signals", |
| "authors": [ |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Tschiatschek", |
| "suffix": "" |
| }, |
| { |
| "first": "Adish", |
| "middle": [], |
| "last": "Singla", |
| "suffix": "" |
| }, |
| { |
| "first": "Manuel", |
| "middle": [ |
| "Gomez" |
| ], |
| "last": "Rodriguez", |
| "suffix": "" |
| }, |
| { |
| "first": "Arpit", |
| "middle": [], |
| "last": "Merchant", |
| "suffix": "" |
| }, |
| { |
| "first": "Andreas", |
| "middle": [], |
| "last": "Krause", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Companion Proceedings of the The Web Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "517--524", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sebastian Tschiatschek, Adish Singla, Manuel Gomez Rodriguez, Arpit Merchant, and Andreas Krause. 2018. Fake news detection in social networks via crowd signals. In Companion Pro- ceedings of the The Web Conference 2018, pages 517-524.", |
| "links": null |
| }, |
| "BIBREF39": { |
| "ref_id": "b39", |
| "title": "Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on twitter", |
| "authors": [ |
| { |
| "first": "Svitlana", |
| "middle": [], |
| "last": "Volkova", |
| "suffix": "" |
| }, |
| { |
| "first": "Kyle", |
| "middle": [], |
| "last": "Shaffer", |
| "suffix": "" |
| }, |
| { |
| "first": "Jin", |
| "middle": [ |
| "Yea" |
| ], |
| "last": "Jang", |
| "suffix": "" |
| }, |
| { |
| "first": "Nathan", |
| "middle": [], |
| "last": "Hodas", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "2", |
| "issue": "", |
| "pages": "647--653", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Svitlana Volkova, Kyle Shaffer, Jin Yea Jang, and Nathan Hodas. 2017. Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on twitter. In Proceedings of the 55th An- nual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 647- 653.", |
| "links": null |
| }, |
| "BIBREF40": { |
| "ref_id": "b40", |
| "title": "Asked and answered: Building a chatbot to address covid-19-related concerns. NEJM Catalyst Innovations in Care Delivery", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Volppkevin", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G VolppKevin et al. 2020. Asked and answered: Build- ing a chatbot to address covid-19-related concerns. NEJM Catalyst Innovations in Care Delivery.", |
| "links": null |
| }, |
| "BIBREF41": { |
| "ref_id": "b41", |
| "title": "Liar, Liar Pants on Fire\": A New Benchmark Dataset for Fake News Detection", |
| "authors": [ |
| { |
| "first": "William", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Wang", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1705.00648" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "William Yang Wang. 2017. \"Liar, Liar Pants on Fire\": A New Benchmark Dataset for Fake News Detection. arXiv preprint arXiv:1705.00648.", |
| "links": null |
| }, |
| "BIBREF42": { |
| "ref_id": "b42", |
| "title": "Learning to identify ambiguous and misleading news headlines", |
| "authors": [ |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Wei", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaojun", |
| "middle": [], |
| "last": "Wan", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1705.06031" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wei Wei and Xiaojun Wan. 2017. Learning to identify ambiguous and misleading news headlines. arXiv preprint arXiv:1705.06031.", |
| "links": null |
| }, |
| "BIBREF43": { |
| "ref_id": "b43", |
| "title": "Misinformation Related to the COVID-19 pandemic", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Wikipedia", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wikipedia. 2020. Misinformation Related to the COVID-19 pandemic. https://en.wikipedia.", |
| "links": null |
| }, |
| "BIBREF44": { |
| "ref_id": "b44", |
| "title": "org/wiki/Misinformation related to the COVID-19 pandemic", |
| "authors": [], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "org/wiki/Misinformation related to the COVID- 19 pandemic [Accessed on June 30, 2020].", |
| "links": null |
| }, |
| "BIBREF45": { |
| "ref_id": "b45", |
| "title": "A broad-coverage challenge corpus for sentence understanding through inference", |
| "authors": [ |
| { |
| "first": "Adina", |
| "middle": [], |
| "last": "Williams", |
| "suffix": "" |
| }, |
| { |
| "first": "Nikita", |
| "middle": [], |
| "last": "Nangia", |
| "suffix": "" |
| }, |
| { |
| "first": "Samuel", |
| "middle": [], |
| "last": "Bowman", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "North American Chapter of the Association for Computational Linguistics (NAACL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Adina Williams, Nikita Nangia, and Samuel Bowman. 2018. A broad-coverage challenge corpus for sen- tence understanding through inference. In North American Chapter of the Association for Computa- tional Linguistics (NAACL).", |
| "links": null |
| }, |
| "BIBREF46": { |
| "ref_id": "b46", |
| "title": "Fake news detection as natural language inference", |
| "authors": [ |
| { |
| "first": "Kai-Chou", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [], |
| "last": "Niven", |
| "suffix": "" |
| }, |
| { |
| "first": "Hung-Yu", |
| "middle": [], |
| "last": "Kao", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1907.07347" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kai-Chou Yang, Timothy Niven, and Hung-Yu Kao. 2019. Fake news detection as natural language in- ference. arXiv preprint arXiv:1907.07347.", |
| "links": null |
| }, |
| "BIBREF47": { |
| "ref_id": "b47", |
| "title": "BERTScore: Evaluating Text Generation with BERT", |
| "authors": [ |
| { |
| "first": "Tianyi", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "Varsha", |
| "middle": [], |
| "last": "Kishore", |
| "suffix": "" |
| }, |
| { |
| "first": "Felix", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Q", |
| "middle": [], |
| "last": "Kilian", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoav", |
| "middle": [], |
| "last": "Weinberger", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Artzi", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1904.09675" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q Weinberger, and Yoav Artzi. 2019. BERTScore: Evaluating Text Generation with BERT. arXiv preprint arXiv:1904.09675.", |
| "links": null |
| }, |
| "BIBREF48": { |
| "ref_id": "b48", |
| "title": "Detection and resolution of rumours in social media: A survey", |
| "authors": [ |
| { |
| "first": "Arkaitz", |
| "middle": [], |
| "last": "Zubiaga", |
| "suffix": "" |
| }, |
| { |
| "first": "Ahmet", |
| "middle": [], |
| "last": "Aker", |
| "suffix": "" |
| }, |
| { |
| "first": "Kalina", |
| "middle": [], |
| "last": "Bontcheva", |
| "suffix": "" |
| }, |
| { |
| "first": "Maria", |
| "middle": [], |
| "last": "Liakata", |
| "suffix": "" |
| }, |
| { |
| "first": "Rob", |
| "middle": [], |
| "last": "Procter", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "ACM Computing Surveys (CSUR)", |
| "volume": "51", |
| "issue": "2", |
| "pages": "1--36", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Arkaitz Zubiaga, Ahmet Aker, Kalina Bontcheva, Maria Liakata, and Rob Procter. 2018. Detection and resolution of rumours in social media: A survey. ACM Computing Surveys (CSUR), 51(2):1-36.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "text": "Confusion Matrices for stance detection task using SBERT (DA) models trained on MultiNLI. The second model uses BERTSCORE (DA) to first determine whether a misconception-tweet pair is Relevant or No Stance, and only Relevant pairs are further classified by SBERT (DA).", |
| "type_str": "figure", |
| "num": null, |
| "uris": null |
| }, |
| "TABREF0": { |
| "num": null, |
| "type_str": "table", |
| "text": "AgreeDemocrats are using the coronavirus situation to harm President Trump.65.0Coronavirus was taken from a Canadian lab or is the result of bioweapons defense research in China.60.2The media is intentionally stoking fears of COVID-19 to destabilize the Trump administration.56.3Disagree COVID-19 is only as deadly as the seasonal flu. Top Misconceptions by Class. Misconceptions with more than 80 tweets total are ranked by the percentage of tweets annotated for each class. The top three misconceptions for each class with the corresponding percentage that a paired tweet would be annotated as that respective class are shown. For example, for the misconception 'Democrats are using the coronavirus situation to harm President Trump', 65% of the tweets paired with this misinformation were annotated as Agree.", |
| "content": "<table><tr><td>Class</td><td/><td/><td>Misconception</td><td>%</td></tr><tr><td/><td/><td/><td>51.0</td></tr><tr><td/><td colspan=\"3\">Anybody in the U.S. who wants a COVID-19 test can get a test.</td><td>36.7</td></tr><tr><td/><td colspan=\"3\">The U.S. containment of the virus is 'close to airtight'.</td><td>35.4</td></tr><tr><td/><td colspan=\"3\">Acetic acid is effective against coronavirus.</td><td>100.0</td></tr><tr><td>No Stance</td><td colspan=\"3\">Cannabis protects against COVID-19.</td><td>100.0</td></tr><tr><td/><td colspan=\"3\">Clapping will kill coronavirus.</td><td>100.0</td></tr><tr><td colspan=\"2\">Class</td><td colspan=\"2\">Count Percentage</td></tr><tr><td colspan=\"2\">Agree</td><td>670</td><td>9.91 %</td></tr><tr><td colspan=\"2\">Disagree</td><td>343</td><td>5.07 %</td></tr><tr><td colspan=\"2\">No Stance</td><td>5,748</td><td>85.02 %</td></tr></table>", |
| "html": null |
| }, |
| "TABREF1": { |
| "num": null, |
| "type_str": "table", |
| "text": "Distribution of Labels in the annotations.", |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF3": { |
| "num": null, |
| "type_str": "table", |
| "text": "Misconception Retrieval Performance. We present evaluation for misinformative tweets (e.g., tweets that Agree with one or more misconceptions), as well as combined evaluation on Relevant tweets (i.e., tweets that either Agree or Disagree with one or more misconceptions.)", |
| "content": "<table><tr><td>Model</td><td colspan=\"2\">Macro Avg.</td><td/><td/><td>Agree</td><td/><td/><td>Disagree</td><td/><td colspan=\"2\">No Stance</td></tr><tr><td/><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td><td>P</td><td>R</td><td>F1</td></tr><tr><td>Trained on SNLI</td><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>Linear, Bag-of-Words</td><td colspan=\"3\">33.1 35.4 27.7</td><td colspan=\"3\">8.2 16.1 10.9</td><td colspan=\"6\">6.6 42.6 11.5 84.5 47.6 60.9</td></tr><tr><td>Linear, Avg. GloVe</td><td colspan=\"6\">32.9 30.7 28.2 13.6 26.3 17.9</td><td colspan=\"2\">2.8 16.3</td><td colspan=\"4\">4.8 82.4 49.6 61.9</td></tr><tr><td>BiLSTM</td><td colspan=\"3\">33.2 36.4 27.4</td><td colspan=\"3\">8.9 15.8 11.4</td><td colspan=\"6\">6.7 47.8 11.7 84.1 45.7 59.2</td></tr><tr><td>SBERT</td><td colspan=\"4\">32.7 30.8 26.9 11.5</td><td colspan=\"2\">9.9 10.6</td><td colspan=\"2\">4.1 31.8</td><td colspan=\"4\">7.3 82.6 50.8 62.9</td></tr><tr><td>SBERT (DA)</td><td colspan=\"6\">33.8 30.4 22.7 22.2 11.9 15.5</td><td colspan=\"2\">4.1 46.9</td><td colspan=\"4\">7.5 75.2 32.3 45.1</td></tr><tr><td>BERTSCORE (DA) + BiLSTM</td><td colspan=\"12\">44.2 45.3 43.1 28.3 15.8 20.3 14.4 32.1 19.9 90.0 88.0 89.0</td></tr><tr><td colspan=\"13\">BERTSCORE (DA) + SBERT (DA) 49.3 44.4 42.6 46.7 14.6 22.3 11.3 30.6 16.5 90.0 88.0 89.0</td></tr><tr><td>Trained on MultiNLI</td><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>Linear, Bag-of-Words</td><td colspan=\"3\">35.2 38.1 24.0</td><td colspan=\"9\">9.8 59.7 16.9 10.5 28.9 15.4 85.3 25.8 39.7</td></tr><tr><td>Linear, Avg. GloVe</td><td colspan=\"6\">35.9 40.8 26.6 15.8 68.5 25.7</td><td colspan=\"2\">4.2 21.6</td><td colspan=\"4\">7.1 87.5 32.2 47.1</td></tr><tr><td>BiLSTM</td><td colspan=\"4\">32.0 33.6 32.5 10.8</td><td>6.4</td><td>8.1</td><td>0.0</td><td>0.0</td><td colspan=\"4\">0.0 85.1 94.2 89.5</td></tr><tr><td>SBERT</td><td colspan=\"6\">36.1 40.1 32.2 17.6 31.9 22.7</td><td colspan=\"6\">6.1 37.6 10.5 84.7 50.6 63.4</td></tr><tr><td>SBERT (DA)</td><td colspan=\"6\">51.1 47.3 41.5 58.1 23.4 33.4</td><td colspan=\"6\">8.7 50.4 14.9 86.5 67.9 76.1</td></tr><tr><td>BERTSCORE (DA) + BiLSTM</td><td colspan=\"6\">39.0 44.6 41.0 27.0 45.8 34.0</td><td>0.0</td><td>0.0</td><td colspan=\"4\">0.0 90.0 88.0 89.0</td></tr><tr><td colspan=\"13\">BERTSCORE (DA) + SBERT (DA) 55.9 50.9 50.2 63.3 30.6 41.2 14.4 34.1 20.3 90.0 88.0 89.0</td></tr><tr><td>Trained on MedNLI</td><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/></tr><tr><td>Linear, Bag-of-Words</td><td colspan=\"6\">35.7 39.3 22.4 10.6 64.5 18.2</td><td colspan=\"6\">8.6 31.2 13.5 87.7 22.2 35.4</td></tr><tr><td>Linear, Avg. GloVe</td><td colspan=\"12\">39.9 50.3 28.2 13.4 74.8 22.8 12.3 48.7 19.6 94.1 27.3 42.3</td></tr><tr><td>BiLSTM</td><td colspan=\"6\">31.9 33.7 25.0 10.2 57.6 17.3</td><td>0.0</td><td>0.0</td><td colspan=\"4\">0.0 85.6 43.5 57.7</td></tr><tr><td>SBERT</td><td colspan=\"6\">35.9 37.0 16.3 10.5 87.6 18.8</td><td colspan=\"6\">8.9 12.2 10.3 88.1 11.1 19.7</td></tr><tr><td>SBERT (DA)</td><td colspan=\"6\">40.3 51.2 30.2 17.6 82.5 29.0</td><td colspan=\"6\">7.9 38.5 13.2 95.3 32.5 48.5</td></tr><tr><td>BERTSCORE (DA) + BiLSTM</td><td colspan=\"7\">43.7 44.6 41.5 27.0 44.9 33.7 14.3</td><td>0.9</td><td colspan=\"4\">1.6 90.0 88.0 89.0</td></tr><tr><td colspan=\"13\">BERTSCORE (DA) + SBERT (DA) 47.8 49.2 48.4 34.2 40.9 37.2 19.2 18.7 18.9 90.0 88.0 89.0</td></tr></table>", |
| "html": null |
| }, |
| "TABREF4": { |
| "num": null, |
| "type_str": "table", |
| "text": "Misconception Retrieval Examples.", |
| "content": "<table/>", |
| "html": null |
| }, |
| "TABREF5": { |
| "num": null, |
| "type_str": "table", |
| "text": "Stance Detection Examples. Presenting examples of cases where combining or domain adaptation lead to flipping prediction to the correct class (Agree). All models here are trained on MultiNLI.", |
| "content": "<table/>", |
| "html": null |
| } |
| } |
| } |
| } |