| { |
| "paper_id": "2021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T14:51:33.515451Z" |
| }, |
| "title": "Masking and Transformer-based Models for Hyperpartisanship Detection in News", |
| "authors": [ |
| { |
| "first": "Javier", |
| "middle": [], |
| "last": "S\u00e1nchez-Junquera", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Universitat Polit\u00e8cnica de Val\u00e8ncia", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Paolo", |
| "middle": [], |
| "last": "Rosso", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "prosso@dsic.upv.es" |
| }, |
| { |
| "first": "Manuel", |
| "middle": [], |
| "last": "Montes-Y-G\u00f3mez", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Simone", |
| "middle": [ |
| "P" |
| ], |
| "last": "Ponzetto", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Mannheim", |
| "location": {} |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Hyperpartisan news show an extreme manipulation of reality based on an underlying and extreme ideological orientation. Because of its harmful effects at reinforcing one's bias and the posterior behavior of people, hyperpartisan news detection has become an important task for computational linguists. In this paper, we evaluate two different approaches to detect hyperpartisan news. First, a text masking technique that allows us to compare style vs. topicrelated features in a different perspective from previous work. Second, the transformer-based models BERT, XLM-RoBERTa, and M-BERT, known for their ability to capture semantic and syntactic patterns in the same representation. Our results corroborate previous research on this task in that topic-related features yield better results than style-based ones, although they also highlight the relevance of using higherlength n-grams. Furthermore, they show that transformer-based models are more effective than traditional methods, but this at the cost of greater computational complexity and lack of transparency. Based on our experiments, we conclude that the beginning of the news show relevant information for the transformers at distinguishing effectively between left-wing, mainstream, and right-wing orientations.", |
| "pdf_parse": { |
| "paper_id": "2021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Hyperpartisan news show an extreme manipulation of reality based on an underlying and extreme ideological orientation. Because of its harmful effects at reinforcing one's bias and the posterior behavior of people, hyperpartisan news detection has become an important task for computational linguists. In this paper, we evaluate two different approaches to detect hyperpartisan news. First, a text masking technique that allows us to compare style vs. topicrelated features in a different perspective from previous work. Second, the transformer-based models BERT, XLM-RoBERTa, and M-BERT, known for their ability to capture semantic and syntactic patterns in the same representation. Our results corroborate previous research on this task in that topic-related features yield better results than style-based ones, although they also highlight the relevance of using higherlength n-grams. Furthermore, they show that transformer-based models are more effective than traditional methods, but this at the cost of greater computational complexity and lack of transparency. Based on our experiments, we conclude that the beginning of the news show relevant information for the transformers at distinguishing effectively between left-wing, mainstream, and right-wing orientations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Media such as radio, TV channels, and newspapers control which information spreads and how it does it. The aim is often not only to inform readers but also to influence public opinion on specific topics from a hyperpartisan perspective.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Social media, in particular, have become the default channel for many people to access information and express ideas and opinions. The most relevant and positive effect is the democratization of information and knowledge but there are also undesired effects. One of them is that social media foster information bubbles: every user may end up receiving only the information that matches his/her personal biases, beliefs, tastes and points of view. Because of this, social media are a breeding ground for the propagation of fake news: when a piece of news outrages us or matches our beliefs, we tend to share it without checking its veracity; and, on the other hand, content selection algorithms in social media give credit to this type of popularity because of the click-based economy on which their business are based. Another harmful effect is that the relative anonymity of social networks facilitates the propagation of toxic, hate and exclusion messages. Therefore, social media contribute to the misinformation and polarization of society, as we have recently witnessed in the last presidential elections in USA or the Brexit referendum. Clearly, the polarization of society and its underlying discourses are not limited to social media, but rather reflected also in political dynamics (e.g., like those found in the US Congress (Andris et al., 2015) ): even in this domain, however, social media can provide a useful signal to estimate partisanship (Hemphill et al., 2016) .", |
| "cite_spans": [ |
| { |
| "start": 1334, |
| "end": 1355, |
| "text": "(Andris et al., 2015)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 1455, |
| "end": 1478, |
| "text": "(Hemphill et al., 2016)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Closely related to the concept of controversy and the \"filter bubble effect\" is the concept of bias (Baeza-Yates, 2018) , which refers to the presentation of information according to the standpoints or interests of the journalists and the news agencies. Detecting bias is very important to help users to acquire balanced information. Moreover, how a piece of information is reported has the capacity to evoke different sentiments in the audience, which may have large social implications (especially in very controversial topics such as terror attacks and religion issues).", |
| "cite_spans": [ |
| { |
| "start": 100, |
| "end": 119, |
| "text": "(Baeza-Yates, 2018)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we approach this very broad topic by focusing on the problem of detecting hyperpartisan news, namely news written with an extreme manipulation of the reality on the basis of an underlying, typically extreme, ideology. This problem has received little attention in the context of the automatic detection of fake news, despite the potential correlation between them. Seminal work from (Potthast et al., 2018 ) presents a comparative style analysis of hyperpartisan news, evaluating features such as characters n-grams, stop words, partof-speech, readability scores, and ratios of quoted words and external links. The results indicate that a topic-based model outperforms a style-based one to separate the left, right and mainstream orientations.", |
| "cite_spans": [ |
| { |
| "start": 398, |
| "end": 420, |
| "text": "(Potthast et al., 2018", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "More recently, in (Kiesel et al., 2019) , the features that participants used in SemEval-2019 task 4 on hyperpartisan news detection have been summarized: n-grams, word embeddings, stylometry (e.g., punctuation and article structure), sentiment and emotion features, named entities, quotations, hyperlinks, and publication date. Using the same dataset from SemEval-2019, (Anthonio, 2019) evaluated features like bag-of-words, bag-of-clusters, word embeddings and contextual character-based embeddings, POS n-grams, stylistic features and the sentiment; the authors found that dense document representations work better across domains and tasks than traditional sparse representations. Finally, (Hosseinia, 2020) found effective to use personality information in hyperpartisan news detection after topic-based sub-sampling of the news training data. The datasets proposed in (Kiesel et al., 2019) were manually labeled and the largest one was labeled in a semi-automated manner via distant supervision.", |
| "cite_spans": [ |
| { |
| "start": 18, |
| "end": 39, |
| "text": "(Kiesel et al., 2019)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 371, |
| "end": 387, |
| "text": "(Anthonio, 2019)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 694, |
| "end": 711, |
| "text": "(Hosseinia, 2020)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 874, |
| "end": 895, |
| "text": "(Kiesel et al., 2019)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Instead of employing the datasets from (Kiesel et al., 2019) , we build upon previous work and use the dataset from (Potthast et al., 2018) : this way we can investigate hyperpartisan-biased news (i.e., extremely one-sided) that have been manually fact-checked by journalists from BuzzFeed, and contrast our results with what they achieved. The articles originated from 9 well-known political publishers, three each from the mainstream, the hyperpartisan left-wing, and the hyperpartisan right-wing. To detect hyperpartisanship, we aim to explore the trade-off between the performance of the models and the transparency of their results. Taking this into account, we apply two approaches diametrically opposite to each other in the text classification state of the art. On the one hand, we use three transformer-based models, which have shown outstanding performance, but high complexity and lack of transparency. On the other hand, we use a masking-based model that requires fewer computational-resources and showed a good performance in related tasks such as authorship attribution (Stamatatos, 2017a) .", |
| "cite_spans": [ |
| { |
| "start": 39, |
| "end": 60, |
| "text": "(Kiesel et al., 2019)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 116, |
| "end": 139, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 1084, |
| "end": 1103, |
| "text": "(Stamatatos, 2017a)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The masking technique transforms the original texts in a form where the textual structure is maintained, while letting the learning algorithm focus on the writing style or the topic-related information. This technique makes it possible for us to corroborate previous results that content matters more than style. Moreover, we aim to find explainable predictions of hyperpartisanship with the attention mechanism of the transformer-based models. With this purpose, we expect to derive the explanation by investigating the scores of different features used to output the final prediction. Based on this, we contrast the transparency of both approaches by comparing the relevant parts of the texts that they highlight.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of the paper is structured as follows. In Section 2 we describe our method to hyperpartisan news detection based on masking. Section 3 presents details on the dataset and the experimental setup. In Section 4 we show the obtained results and discuss about them. Finally, Section 5 concludes with some directions for future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The masking technique that we propose here for the hyperpartisan news detection task has been applied to text clustering (Granados et al., 2011) , authorship attribution (Stamatatos, 2017a) , and deception detection (S\u00e1nchez-Junquera, 2018) with encouraging results. The main idea of the proposed method is to transform the original texts to a form where the textual structure, related to a general style (or topic), is maintained while content-related (or style-related) words are masked. To this end, all the occurrences of non-desired terms are re-placed by symbols. Let W k be the set of the k most frequent words, we mask all the occurrences of a word w \u2208 W k if we want to learn a topic-related model, or we mask all w / \u2208 W k if we want to learn a style-based model. Whatever the case, the way in which we mask the terms in this work is called Distorted View with Single Asterisks and consists in replacing w with a single asterisk or a single # symbol if the term is a word or a number, respectively. For further masking methods, refer to (Stamatatos, 2017a) . Table 1 shows a fragment of an original text and the result of masking style-related information or topic-related information. With the former we obtain distorted texts that allow for learning a topicbased model; on the other hand, with the latter, it is possible to learn a style-based model. One of the options to choose the terms to be masked or maintained without masking is to take the most frequent words of the target language (Stamatatos, 2017a) . In the original text from the table, we highlight some of the most frequent words in English.", |
| "cite_spans": [ |
| { |
| "start": 121, |
| "end": 144, |
| "text": "(Granados et al., 2011)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 170, |
| "end": 189, |
| "text": "(Stamatatos, 2017a)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 1047, |
| "end": 1066, |
| "text": "(Stamatatos, 2017a)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 1503, |
| "end": 1522, |
| "text": "(Stamatatos, 2017a)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 1069, |
| "end": 1076, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Masking and Transformer-based Models 2.1 Investigating Masking for Hyperpartisanship Detection", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Transformer-based models have been trained with huge general language datasets. Such is the case of the Bidirectional Encoder Representations from Transformers (BERT). BERT is designed to pretrain deep bidirectional representations from an unlabeled text by jointly conditioning on both left and right context in all layers (Devlin et al., 2018) . This text representation allows the model to capture complex patterns going beyond merely the use of words and capturing semantic and syntactic patterns in the same representation.", |
| "cite_spans": [ |
| { |
| "start": 324, |
| "end": 345, |
| "text": "(Devlin et al., 2018)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Transformer-based Models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The framework of BERT consists of two steps: pre-training and fine-tuning. For the pre-training, the collected data included BooksCorpus (800M words) and English Wikipedia (2,500M words). The BERT BASE model has 12 layers with 12 selfattention heads, and uses 768 as hidden size, with a total of 110M parameters; and the BERT LARGE model has 24 layers with 16 self-attention heads, and uses 1024 as hidden size, with a total of 340M parameters. The vocabulary contains 30K tokens. For fine-tuning, the model is first initialized with the pre-trained parameters, and all of the parameters are fine-tuned using labeled data from the downstream task, which in our case are 1555 news annotated with the political orientation. The first token of every sequence is always a special classifica-tion token ([CLS]), which is used as the aggregate sequence representation for classification tasks. In our work, we add to the [CLS] representation two dense layers and a Softmax function to obtain the binary classification.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Transformer-based Models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "In this paper we evaluate three transformer-based models: BERT; the multilingual BERT (M-BERT) (Devlin et al., 2018) , which was pretrained on the concatenation of monolingual Wikipedia datasets from 104 languages (Wang et al., 2019; Pires et al., 2019) ; and XLM-RoBERTa, which was pretrained on 2.5TB of newly created clean CommonCrawl data in 100 languages (Ca\u00f1ete et al., 2020) .", |
| "cite_spans": [ |
| { |
| "start": 95, |
| "end": 116, |
| "text": "(Devlin et al., 2018)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 214, |
| "end": 233, |
| "text": "(Wang et al., 2019;", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 234, |
| "end": 253, |
| "text": "Pires et al., 2019)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 360, |
| "end": 381, |
| "text": "(Ca\u00f1ete et al., 2020)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Transformer-based Models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "We used the BuzzedFeed-Webis Fake News Corpus 2016 collected by (Potthast et al., 2018) whose articles were labeled with respect to three political orientations: mainstream, left-wing, and right-wing (see Table 2 ). Each article was taken from one of 9 publishers known as hyperpartisan left/right or mainstream in a period close to the US presidential elections of 2016. Therefore, the content of all the articles is related to the same topic. During initial data analysis and prototyping we identified a variety of issues with the original dataset: we cleaned the data excluding articles with empty or bogus texts, duplicates. As a result, we obtained a new dataset with 1555 articles out of 1627. 1 Following the settings of (Potthast et al., 2018), we balanced the training set using random duplicate oversampling.", |
| "cite_spans": [ |
| { |
| "start": 64, |
| "end": 87, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 205, |
| "end": 212, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experiments", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In this section, we reported the results of the masking technique from two different perspectives. In one setting, we masked topic-related information in order to maintain the predominant writing style used in each orientation. We call this approach a style-based model. With that intention we selected the k most frequent words from the target language, and then we transformed the texts by masking the occurrences of the rest of the words. In another setting, we masked style-related information to allow the system to focus only on the topic-related differences between the orientations. We call this a topic-based model. For this, we masked the k most frequent words and maintained intact the rest. After the text transformation by the masking process in both the training and test sets, we represented the documents with character n-grams and compared the results obtained with the style-based and the topic-related models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Masking Content vs. Style in Hyperpartisan News", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Text Transformation: We evaluated different values of k (k \u2208 {100, 200, ..., 5000}) for extracting the k most frequent words from English 2 . For the comparison of the results obtained by each model with the ones of the state of the art, we only showed the results fixing k = 500.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Text Representation: We used a standard bag-ofwords representation with tf weighting and extracted character 5-grams with a frequency lower than 50.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Classifiers: We compared the results obtained with Na\u00efve Bayes (NB), Support Vector Machine (SVM) and Random Forest (RF); for the three classifiers we used the versions implemented in sklearn with the parameters set by default.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Transformers: We approached the hyperparameter tuning by grid search. The best results were obtained with: learning rate = 3e \u2212 5; the batch size = 16; and the adam optimizer. Moreover, we applied a dropout value of 0.3 to the last dense layer. We have selected a value of 200 for the max length hyperparameter.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Evaluation: We performed 3-fold crossvalidation with the same configuration used in (Potthast et al., 2018) . Therefore, each fold comprised one publisher from each orientation (the classifiers did not learn a publisher's style). We used macro F 1 as the evaluation measure since the test set is unbalanced with respect to the three classes. In order to compare our results with those reported in (Potthast et al., 2018) , we also used accuracy, precision, and recall.", |
| "cite_spans": [ |
| { |
| "start": 84, |
| "end": 107, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 397, |
| "end": 420, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Baseline: Our baseline method is based on the same text representation with the character n-grams features, but without masking any word. Table 3 shows the results of the proposed method and the system from (Potthast et al., 2018) 3 in our cleaned dataset (Section 3), both considering topic and style-based methods. In order to compare our results with those reported in (Potthast et al., 2018) , we report the same measures the authors used. We also include the macro F 1 score because of the unbalance test set. For these experiments we extract the character 5-grams from the transformed texts, taking into account that as more narrow is the domain more sense has the use of longer n-grams. We follow the steps of (Stamatatos, 2017b) and set k = 500 for this comparison results. Similar to (Potthast et al., 2018) , the topic-based model achieves better results than the style-related model. However, the differences between the results of the two evaluated approaches are much higher (0.66 vs. 0.57 according to Macro F 1 ) than those obtained from the system of (Potthast et al., 2018) Table 3 : Results of the proposed masking technique (k = 500 and n = 5) applied to mask topic-related information or style-related information. NB: Naive Bayes; RF: Random Forest; SVM: Support Vector Machine. The last two rows show the results obtained by applying the system from (Potthast et al., 2018) to our cleaned dataset (Section 3).", |
| "cite_spans": [ |
| { |
| "start": 207, |
| "end": 230, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 372, |
| "end": 395, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 717, |
| "end": 736, |
| "text": "(Stamatatos, 2017b)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 793, |
| "end": 816, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 1067, |
| "end": 1090, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 1372, |
| "end": 1395, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 138, |
| "end": 145, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 1091, |
| "end": 1098, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimental Setup", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "using the SVM classifier and masking the stylerelated information (i.e., applying the topic-related model). This could be explained with the fact that all the articles are about the same political event in a very limited period of time. In line with what was already pointed out in (Potthast et al., 2018) , the left-wing orientation is harder to predict, possibly because this class is represented with fewer examples in the dataset. Another reason why our masking approach achieves better results than the system from (Potthast et al., 2018) , could be that we use a higher length of character n-grams. In fact, comparing their results against our baseline model, it is possible to note that even without masking any word, the classifier obtains better results. This suggests that the good results are due to the length of the character n-grams rather than the use of the masking technique.", |
| "cite_spans": [ |
| { |
| "start": 282, |
| "end": 305, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 520, |
| "end": 543, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results and Discussion", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The last three rows of Table 3 show the results of the transformer-based models. As we can see, these models achieved the highest results, in particular the BERT model, with a Macro F 1 = 0.86. These models are known for their ability to capture complex syntactic and semantic patterns, therefore, these results are somehow justified to be the highest compared to the masking approach. However, what is interesting at this point is the effectiveness of the models at predicting the correct orientation using just the beginning of the news (max length = 200). This is aligned to the work of (Ghanem et al., 2021) that focused on analyzing the initial part of false news articles. The authors as- sumption is that false news tend to present a unique emotional pattern for each false information type in order to trigger specific emotions to the readers; in hyperpartisan news this probably happens to gain readers' attention and sympathy. Table 4 shows the features with the highest weights from the SVM (we used scikit-learn's method to collect feature weights). It is possible to note that the mention of cnn was learned as a discriminative feature when the news from that publisher were used in the training (in the topic-based model). However, this feature is infrequent in the test set where no news from CNN publisher was included.", |
| "cite_spans": [ |
| { |
| "start": 590, |
| "end": 611, |
| "text": "(Ghanem et al., 2021)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 23, |
| "end": 30, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 937, |
| "end": 944, |
| "text": "Table 4", |
| "ref_id": "TABREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results and Discussion", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The features related to Donald Trump (donal and onal), and Hillary Clinton (llary and illar) are more frequent in one of the hyperpartisan orientation, and none of them occurs frequently in the Topic-related model left (...)which his son pretty much confirmed in a foolish statement. The content of those tax returns has been the subject of much speculation, but given Trump's long history of tax evasion and political bribery, it doesn't take much imagination to assume he's committing some kind of fraud mainstream orientation. On the other hand, the relevant features from the style-based model involve function words that are frequent in the three classes (e.g., out, you, and, of ) even if the combination between function words and other characters can lightly differ in different orientations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant Features", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Transformer-based models allow us to visualize different parts of the news according to the scores they received to obtain the final prediction. In Figure 1 we show examples of news predicted correctly by BERT (the model with the highest F 1 score). Due to space limitations, we provide fragments of six news, two per orientation. The more intense the color, the greater is the weight of attention given by the model. In the examples from 1a, the left-wing orientation remarks the names of the opposite politicians, and it is possible to see which of them is the favourite of the journalist. In particular, the leader of the right-wing (i.e., Trump) is referred in a negative way (he does not know his own words) while Hillary Clinton, the representative of the left-wing is favored by the news. Similar to this, examples 1c do the same but in the opposite direction; i.e., Hillary Clinton is put as a very negative \"character\" who loves taxes and is the most despicable liar ever. However, examples from 1c offer a comparison in which keep the reader in a neutral position. Moreover, in the second mainstream news, Trump's campaign is mentioned without describing the stance of the author whether Trump did well or not in his topic selection. This suggests that the style used to speak about the leaders can differ from the more biased (hyperpartisan) news to the less biased (mainstream).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 148, |
| "end": 156, |
| "text": "Figure 1", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Features with the Highest Attention Scores", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We can conclude that the attention mechanism of the transformers not only help in doing effective predictions, but offer some extra information that could be useful to understand some insights about hyperpartisanship. For example, the words with the highest scores can be used in other strategies to confirm the previous results that topic-based models outperform a style-based one at distinguishing left, right and mainstream orientations (Potthast et al., 2018) .", |
| "cite_spans": [ |
| { |
| "start": 440, |
| "end": 463, |
| "text": "(Potthast et al., 2018)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features with the Highest Attention Scores", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In this paper we presented initial experiments on the task of hyperpartisan news detection. In particular, we aimed to explore the trade-off between performance and transparency, and proposed a comparison of two different approaches. First, we explored the use of masking techniques to boost the performance of a lexicalized classifier. Our results corroborate previous research on the importance of content features to detect extreme content: masking, in addition, shows the benefits of reducing data sparsity for this task comparing our results with the state of the art. We evaluated different values of the parameters and see that finally our baseline model, in which we extract character 5-grams without applying any masking process, achieves the better results. This seems to indicate a strong lexical overlap between different sources with the same orientation, which, in turn, calls for more challenging datasets and task formulations to encourage the development of models covering more subtle, i.e., implicit, forms of bias. Future datasets could consider more topics and different time spans to avoid the models learn from the topic, rather than the target classes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Second, we used three transformer-based models (BERT, M-BERT, and XLM-RoBERTa) that are resource-hungrier than the masking technique, and achieved the highest results. We also presented some examples of how these models, through their attention scores, provide additional information about the relevant parts of the text for distinguishing their political orientation. Considering the high effectiveness of these models, and that they only observe the first part of the news, we will evaluate as future work how necessary is to use all the news (and not only the beginning), e.g. with the Transformer-XL model (Dai et al., 2019) . Moreover, we are motivated to take advantage of the attention scores to study in more detail the style used in hyperpartisan news in order to improve the predictions.", |
| "cite_spans": [ |
| { |
| "start": 610, |
| "end": 628, |
| "text": "(Dai et al., 2019)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The dataset will be available to the research community upon acceptance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "We use the BNC corpus (https://www.kilgarriff.co.uk/bncreadme.html) for the extraction of the most frequent words as in(Stamatatos, 2017a).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/webis-de/ACL-18", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "The work of the authors from the Universitat Polit\u00e8cnica of Val\u00e8ncia was funded by the Spanish Ministry of Science and Innovation under the research project MISMIS-FAKEnHATE on MISinformation and MIScommunication in social media: FAKE news and HATE speech (PGC2018-096212-B-C31). Experiments were carried out on the GPU cluster at PRHLT thanks to the PROM-ETEO/2019/121 (DeepPattern) research project funded by the Generalitat Valenciana.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "The rise of partisanship and Super-Cooperators in the U.S. house of representatives", |
| "authors": [ |
| { |
| "first": "Clio", |
| "middle": [], |
| "last": "Andris", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcus", |
| "middle": [ |
| "J" |
| ], |
| "last": "Hamilton", |
| "suffix": "" |
| }, |
| { |
| "first": "Mauro", |
| "middle": [], |
| "last": "Martino", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [ |
| "E" |
| ], |
| "last": "Gunning", |
| "suffix": "" |
| }, |
| { |
| "first": "John Armistead", |
| "middle": [], |
| "last": "Selden", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "PLoS ONE", |
| "volume": "10", |
| "issue": "4", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Clio Andris, David Lee, Marcus J. Hamilton, Mauro Martino, Christian E. Gunning, and John Armistead Selden. 2015. The rise of partisanship and Super- Cooperators in the U.S. house of representatives. PLoS ONE, 10(4):e0123507.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Robust document representations for hyperpartisan and fake news detection", |
| "authors": [ |
| { |
| "first": "Talita", |
| "middle": [], |
| "last": "Anthonio", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Talita Anthonio. 2019. Robust document represen- tations for hyperpartisan and fake news detection.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Master's thesis", |
| "authors": [], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Master's thesis, University of the Basque Country UPV/EHU.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Bias on the web", |
| "authors": [ |
| { |
| "first": "Ricardo", |
| "middle": [], |
| "last": "Baeza-Yates", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Communications of the ACM", |
| "volume": "61", |
| "issue": "6", |
| "pages": "54--61", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/3209581" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ricardo Baeza-Yates. 2018. Bias on the web. Commu- nications of the ACM, 61(6):54-61.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Spanish pre-trained bert model and evaluation data", |
| "authors": [ |
| { |
| "first": "Jos\u00e9", |
| "middle": [], |
| "last": "Ca\u00f1ete", |
| "suffix": "" |
| }, |
| { |
| "first": "Gabriel", |
| "middle": [], |
| "last": "Chaperon", |
| "suffix": "" |
| }, |
| { |
| "first": "Rodrigo", |
| "middle": [], |
| "last": "Fuentes", |
| "suffix": "" |
| }, |
| { |
| "first": "Jou-Hui", |
| "middle": [], |
| "last": "Ho", |
| "suffix": "" |
| }, |
| { |
| "first": "Hojin", |
| "middle": [], |
| "last": "Kang", |
| "suffix": "" |
| }, |
| { |
| "first": "Jorge", |
| "middle": [], |
| "last": "P\u00e9rez", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jos\u00e9 Ca\u00f1ete, Gabriel Chaperon, Rodrigo Fuentes, Jou- Hui Ho, Hojin Kang, and Jorge P\u00e9rez. 2020. Span- ish pre-trained bert model and evaluation data. In PML4DC at ICLR 2020.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Transformer-XL: Attentive language models beyond a fixed-length context", |
| "authors": [ |
| { |
| "first": "Zihang", |
| "middle": [], |
| "last": "Dai", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhilin", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Yiming", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "Jaime", |
| "middle": [], |
| "last": "Carbonell", |
| "suffix": "" |
| }, |
| { |
| "first": "Quoc", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruslan", |
| "middle": [], |
| "last": "Salakhutdinov", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "2978--2988", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/P19-1285" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zihang Dai, Zhilin Yang, Yiming Yang, Jaime Car- bonell, Quoc Le, and Ruslan Salakhutdinov. 2019. Transformer-XL: Attentive language models beyond a fixed-length context. In Proceedings of the 57th Annual Meeting of the Association for Computa- tional Linguistics, pages 2978-2988, Florence, Italy. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Bert: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1810.04805" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understand- ing. arXiv preprint arXiv:1810.04805.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "FakeFlow: Fake news detection by modeling the flow of affective information", |
| "authors": [ |
| { |
| "first": "Bilal", |
| "middle": [], |
| "last": "Ghanem", |
| "suffix": "" |
| }, |
| { |
| "first": "Simone", |
| "middle": [ |
| "Paolo" |
| ], |
| "last": "Ponzetto", |
| "suffix": "" |
| }, |
| { |
| "first": "Paolo", |
| "middle": [], |
| "last": "Rosso", |
| "suffix": "" |
| }, |
| { |
| "first": "Francisco", |
| "middle": [], |
| "last": "Rangel", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume", |
| "volume": "", |
| "issue": "", |
| "pages": "679--689", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bilal Ghanem, Simone Paolo Ponzetto, Paolo Rosso, and Francisco Rangel. 2021. FakeFlow: Fake news detection by modeling the flow of affective infor- mation. In Proceedings of the 16th Conference of the European Chapter of the Association for Compu- tational Linguistics: Main Volume, pages 679-689, Online. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Reducing the loss of information through annealing text distortion", |
| "authors": [ |
| { |
| "first": "Ana", |
| "middle": [], |
| "last": "Granados", |
| "suffix": "" |
| }, |
| { |
| "first": "Manuel", |
| "middle": [], |
| "last": "Cebrian", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Camacho", |
| "suffix": "" |
| }, |
| { |
| "first": "Francisco", |
| "middle": [], |
| "last": "De", |
| "suffix": "" |
| }, |
| { |
| "first": "Borja", |
| "middle": [], |
| "last": "Rodriguez", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "IEEE Transactions on Knowledge and Data Engineering", |
| "volume": "23", |
| "issue": "7", |
| "pages": "1090--1102", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ana Granados, Manuel Cebrian, David Camacho, and Francisco de Borja Rodriguez. 2011. Reducing the loss of information through annealing text distortion. IEEE Transactions on Knowledge and Data Engi- neering, 23(7):1090-1102.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "#polarscores: Measuring partisanship using social media content", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Hemphill", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Culotta", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Heston", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Journal of Information Technology & Politics", |
| "volume": "13", |
| "issue": "4", |
| "pages": "365--377", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "L. Hemphill, A. Culotta, and M. Heston. 2016. #po- larscores: Measuring partisanship using social me- dia content. Journal of Information Technology & Politics, 13(4):365-377.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Content and Stylistic Models for Authorship, Stance, and Hyperpartisan Detection", |
| "authors": [ |
| { |
| "first": "Marjan", |
| "middle": [], |
| "last": "Hosseinia", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marjan Hosseinia. 2020. Content and Stylistic Mod- els for Authorship, Stance, and Hyperpartisan De- tection. Ph.D. thesis, University of Houston.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Semeval-2019 task 4: Hyperpartisan news detection", |
| "authors": [ |
| { |
| "first": "Johannes", |
| "middle": [], |
| "last": "Kiesel", |
| "suffix": "" |
| }, |
| { |
| "first": "Maria", |
| "middle": [], |
| "last": "Mestre", |
| "suffix": "" |
| }, |
| { |
| "first": "Rishabh", |
| "middle": [], |
| "last": "Shukla", |
| "suffix": "" |
| }, |
| { |
| "first": "Emmanuel", |
| "middle": [], |
| "last": "Vincent", |
| "suffix": "" |
| }, |
| { |
| "first": "Payam", |
| "middle": [], |
| "last": "Adineh", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Corney", |
| "suffix": "" |
| }, |
| { |
| "first": "Benno", |
| "middle": [], |
| "last": "Stein", |
| "suffix": "" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Potthast", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "829--839", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/S19-2145" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Johannes Kiesel, Maria Mestre, Rishabh Shukla, Em- manuel Vincent, Payam Adineh, David Corney, Benno Stein, and Martin Potthast. 2019. Semeval- 2019 task 4: Hyperpartisan news detection. pages 829-839.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "How multilingual is multilingual bert? arXiv preprint", |
| "authors": [ |
| { |
| "first": "Telmo", |
| "middle": [], |
| "last": "Pires", |
| "suffix": "" |
| }, |
| { |
| "first": "Eva", |
| "middle": [], |
| "last": "Schlinger", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Garrette", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1906.01502" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Telmo Pires, Eva Schlinger, and Dan Garrette. 2019. How multilingual is multilingual bert? arXiv preprint arXiv:1906.01502.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "A stylometric inquiry into hyperpartisan and fake news", |
| "authors": [ |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Potthast", |
| "suffix": "" |
| }, |
| { |
| "first": "Johannes", |
| "middle": [], |
| "last": "Kiesel", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Reinartz", |
| "suffix": "" |
| }, |
| { |
| "first": "Janek", |
| "middle": [], |
| "last": "Bevendorff", |
| "suffix": "" |
| }, |
| { |
| "first": "Benno", |
| "middle": [], |
| "last": "Stein", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "231--240", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Martin Potthast, Johannes Kiesel, Kevin Reinartz, Janek Bevendorff, and Benno Stein. 2018. A stylo- metric inquiry into hyperpartisan and fake news. In Proceedings of the 56th Annual Meeting of the As- sociation for Computational Linguistics (Volume 1: Long Papers), pages 231-240.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Master's thesis, Instituto Nacional de Astrof\u00edsica, Optica y Electr\u00f3nica", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "S\u00e1nchez-Junquera", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. S\u00e1nchez-Junquera. 2018. Adaptaci\u00f3n de dominio para la detecci\u00f3n autom\u00e1tica de textos enga\u00f1osos. Master's thesis, Instituto Nacional de Astrof\u00edsica, Optica y Electr\u00f3nica. Available here (in Spanish).", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Authorship attribution using text distortion", |
| "authors": [ |
| { |
| "first": "Efstathios", |
| "middle": [], |
| "last": "Stamatatos", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1138--1149", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Efstathios Stamatatos. 2017a. Authorship attribution using text distortion. In Proceedings of the 15th Conference of the European Chapter of the Associa- tion for Computational Linguistics: Volume 1, Long Papers, volume 1, pages 1138-1149.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Authorship attribution using text distortion", |
| "authors": [ |
| { |
| "first": "Efstathios", |
| "middle": [], |
| "last": "Stamatatos", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "1138--1149", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Efstathios Stamatatos. 2017b. Authorship attribution using text distortion. In Proceedings of the 15th Conference of the European Chapter of the Associa- tion for Computational Linguistics: Volume 1, Long Papers, pages 1138-1149. Association for Computa- tional Linguistics.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Cross-lingual ability of multilingual bert: An empirical study", |
| "authors": [ |
| { |
| "first": "Zihan", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Stephen", |
| "middle": [], |
| "last": "Mayhew", |
| "suffix": "" |
| }, |
| { |
| "first": "Dan", |
| "middle": [], |
| "last": "Roth", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1912.07840" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zihan Wang, Stephen Mayhew, Dan Roth, et al. 2019. Cross-lingual ability of multilingual bert: An empir- ical study. arXiv preprint arXiv:1912.07840.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "uris": null, |
| "text": "(a) Hyperpartisan (left-wing) news. (b) Non-hyperpartisan (mainstream orientation) news.(c) Hyperpartisan (right-wing) news.", |
| "num": null, |
| "type_str": "figure" |
| }, |
| "FIGREF1": { |
| "uris": null, |
| "text": "Fragments of news (two for each political orientation) with the visualization of the attention learned by BERT. The more intense the color, the greater the weight of attention.", |
| "num": null, |
| "type_str": "figure" |
| }, |
| "TABREF0": { |
| "content": "<table><tr><td>Original text</td><td>Masking topic-related words</td><td>Masking style-related words</td></tr><tr><td/><td/><td>Officers * * Christopher * *</td></tr><tr><td/><td/><td>watching * argument * * * *</td></tr><tr><td/><td/><td>girlfriend outside * bar * * *</td></tr><tr><td/><td/><td>2015 shooting</td></tr></table>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "Officers went after Christopher Few after watching an argument between him and his girlfriend outside a bar just before the 2015 shooting * went after * Few after * an * between him and his * a * just before the # *" |
| }, |
| "TABREF1": { |
| "content": "<table><tr><td/><td colspan=\"3\">Left-wing Mainstream Right-wing</td><td/></tr><tr><td>Original data (Potthast et al., 2018)</td><td>256</td><td>826</td><td>545</td><td>1627</td></tr><tr><td>Cleaned data</td><td>252</td><td>787</td><td>516</td><td>1555</td></tr></table>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "Examples of masking style-related information or topic-related information." |
| }, |
| "TABREF2": { |
| "content": "<table/>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "Statistics of the original dataset and its subset used in this paper." |
| }, |
| "TABREF3": { |
| "content": "<table><tr><td>Masking Method</td><td colspan=\"3\">Classifier Macro F 1 Accuracy</td><td>Precision left right main left right main left right main Recall F 1</td></tr><tr><td>Baseline</td><td>NB</td><td>0.52</td><td>0.56</td><td>0.28 0.57 0.81 0.49 0.58 0.56 0.35 0.57 0.66</td></tr><tr><td>model</td><td>RF</td><td>0.56</td><td>0.62</td><td>0.28 0.61 0.80 0.36 0.72 0.63 0.32 0.66 0.70</td></tr><tr><td/><td>SVM</td><td>0.70</td><td>0.77</td><td>0.55 0.75 0.84 0.42 0.79 0.87 0.47 0.77 0.85</td></tr><tr><td colspan=\"2\">Style-based NB</td><td>0.47</td><td>0.52</td><td>0.20 0.51 0.73 0.28 0.65 0.49 0.23 0.57 0.59</td></tr><tr><td>model</td><td>RF</td><td>0.46</td><td>0.53</td><td>0.24 0.58 0.64 0.36 0.34 0.73 0.29 0.43 0.68</td></tr><tr><td/><td>SVM</td><td>0.57</td><td>0.66</td><td>0.33 0.66 0.75 0.26 0.61 0.84 0.29 0.62 0.79</td></tr><tr><td colspan=\"2\">Topic-based NB</td><td>0.54</td><td>0.60</td><td>0.26 0.63 0.74 0.36 0.62 0.65 0.29 0.62 0.69</td></tr><tr><td>model</td><td>RF</td><td>0.53</td><td>0.55</td><td>0.27 0.64 0.71 0.44 0.60 0.58 0.33 0.61 0.64</td></tr><tr><td/><td>SVM</td><td>0.66</td><td>0.74</td><td>0.48 0.73 0.81 0.38 0.78 0.82 0.42 0.75 0.82</td></tr><tr><td colspan=\"5\">System from (Potthast et al., 2018) (applied to our cleaned dataset)</td></tr><tr><td>Style</td><td>RF</td><td>0.61</td><td>0.63</td><td>0.29 0.62 0.71 0.16 0.62 0.80 0.20 0.61 0.74</td></tr><tr><td>Topic</td><td>RF</td><td>0.63</td><td>0.65</td><td>0.27 0.65 0.72 0.15 0.62 0.84 0.19 0.63 0.77</td></tr><tr><td colspan=\"2\">Transformer-based models</td><td/><td/><td/></tr><tr><td colspan=\"2\">M-BERT</td><td>0.76</td><td>0.83</td><td>0.65 0.75 0.93 0.49 0.86 0.92 0.56 0.93 0.80</td></tr><tr><td colspan=\"2\">XLM-RoBERTa</td><td>0.80</td><td>0.86</td><td>0.80 0.76 0.95 0.50 0.91 0.94 0.61 0.83 0.95</td></tr><tr><td colspan=\"2\">BERT</td><td>0.86</td><td>0.89</td><td>0.77 0.87 0.94 0.75 0.86 0.96 0.76 0.87 0.95</td></tr></table>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "(0.63 vs. 0.61). The highest scores of the masking technique were consistently achieved" |
| }, |
| "TABREF5": { |
| "content": "<table/>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "Most relevant features to each class." |
| }, |
| "TABREF6": { |
| "content": "<table><tr><td/><td>Obama proved beyond shadow * doubt * 2011 * * released * ** birth certificate (...) CNN *</td></tr><tr><td/><td>Fox News cut * *points * * presentation Networks spent * * talking * Trump * * birther (...)</td></tr><tr><td/><td>* Friday * campaign * recent deception * Wednesday * campaign advisers told reporters *</td></tr><tr><td/><td>Trump * * * releasing results * * latest medical exam</td></tr><tr><td>right</td><td>The email, which was dated March 17, 2008, and shared with POLITICO, reads: Jim, on Kenya your person in the field might look into the impact there of Obama's public comments</td></tr><tr><td/><td>about his father. I'm told by State Dept officials that Obama publicly derided his father on</td></tr><tr><td/><td>(...) Blumenthal, a longtime confidant of both Bill and Hillary Clinton, emerged as a frequent</td></tr><tr><td/><td>correspondent in the former secretary of (...)</td></tr><tr><td/><td>* email * * dated March 17 2008 * shared * POLITICO reads Jim * Kenya * * * * field * * * *</td></tr><tr><td/><td>impact * * Obama * comments * * * told * * Dept officials * Obama publicly derided * * (...)</td></tr><tr><td/><td>Blumenthal longtime confidant * * Bill * Hillary Clinton emerged * frequent correspondent * *</td></tr><tr><td/><td>former secretary * (...)</td></tr></table>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "* * son pretty * confirmed * foolish statement * content * * tax returns * * * subject * * speculation * * Trump * * * tax evasion * * bribery * doesn * * imagination * assume * committing * * * fraud mainObama proved beyond a shadow of a doubt in 2011 when he released his long-form birth certificate (...) CNN and Fox News cut away at points in the presentation. Networks spent the day talking about Trump's history as a birther (...) Before Friday, the campaign's most recent deception came Wednesday when campaign advisers told reporters that Trump would not be releasing results of his latest medical exam" |
| }, |
| "TABREF7": { |
| "content": "<table><tr><td>: Fragments of original texts and their transfor-</td></tr><tr><td>mation by masking the k most frequent terms. Some of</td></tr><tr><td>the features from Table 4 using the topic-related model</td></tr><tr><td>are highlighted.</td></tr></table>", |
| "html": null, |
| "type_str": "table", |
| "num": null, |
| "text": "" |
| } |
| } |
| } |
| } |