| { |
| "paper_id": "2021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T12:12:41.453734Z" |
| }, |
| "title": "cs english@LT-EDI-EACL2021: Hope Speech Detection Based On Fine-tuning ALBERT Model", |
| "authors": [ |
| { |
| "first": "Shi", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Yunnan University", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Bing", |
| "middle": [], |
| "last": "Kong", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Yunnan University", |
| "location": {} |
| }, |
| "email": "kongbing@ynu.edu.cn" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper mainly introduces the relevant content of the task \"Hope Speech Detection for Equality, Diversity, and Inclusion at LT-EDI 2021-EACL 2021\". A total of three language datasets were provided, and we chose the English dataset to complete this task. The specific task objective is to classify the given speech into 'Hope speech', 'Not Hope speech', and 'Not in intended language'. In terms of method, we use fine-tuned ALBERT and K fold cross-validation to accomplish this task. In the end, we achieved a good result in the rank list of the task result, and the final F1 score was 0.93, tying for first place. However, we will continue to try to improve methods to get better results in future work.", |
| "pdf_parse": { |
| "paper_id": "2021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper mainly introduces the relevant content of the task \"Hope Speech Detection for Equality, Diversity, and Inclusion at LT-EDI 2021-EACL 2021\". A total of three language datasets were provided, and we chose the English dataset to complete this task. The specific task objective is to classify the given speech into 'Hope speech', 'Not Hope speech', and 'Not in intended language'. In terms of method, we use fine-tuned ALBERT and K fold cross-validation to accomplish this task. In the end, we achieved a good result in the rank list of the task result, and the final F1 score was 0.93, tying for first place. However, we will continue to try to improve methods to get better results in future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "As we all know, we are currently facing an incompletely harmonious and secure network environment. Nowadays, the number of Internet users is very large, especially the proportion of minors is steadily increasing, which shows how important it is to create a hopeful social media environment.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Such an environment that embodies equality, tolerance, and diversity can help people who are in depression, confusion, lack of identity and other difficulties gain hope. At the same time, it also brings a better online social experience to the entire user community on social media.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Currently, the detection and classification of social media speech are mostly biased towards offensive and controversial speech detection, but this task is hope speech detection. The difference is that the former uses detection of negative speech to eliminate the impact of offensive speech on the creation of a healthy network environment. The hope speech detection task focuses on detecting hopeful words to create a good network environment, rather than detecting and deleting negative comments to encourage people and deprive individuals of their freedom of speech (Chakravarthi, 2020) . The data source of this task comes from YouTube comments. What we have to do is to divide it into three categories: 'Hope speech', 'Not hope speech' and 'Not in intended language'. And we used K-fold cross-validation and ALBERT model to complete the detection of hope speech.", |
| "cite_spans": [ |
| { |
| "start": 569, |
| "end": 589, |
| "text": "(Chakravarthi, 2020)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "At present, there are many studies on offensive and discriminatory speech detection, but there are few studies on hope speech detection. Therefore, we have looked for many related social media speech detection studies that can help us complete this task as a reference.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Marzieh Mozafari et al. used the regularization method to adjust the input text in the detection of hate and discriminatory speech and used a finetuned BERT model (Mozafari et al., 2020) . Polychronis Charitidis et al. used the most advanced deep learning architecture to detect hate speech, trained and evaluated it using annotated data sets, and proposed an overall learning architecture that combines the predictive capabilities of each classifier (Charitidis et al., 2019) .", |
| "cite_spans": [ |
| { |
| "start": 163, |
| "end": 186, |
| "text": "(Mozafari et al., 2020)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 451, |
| "end": 476, |
| "text": "(Charitidis et al., 2019)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Based on BERT's pre-trained language model, Marzieh Mozafari et al. made new fine-tuning of transfer learning and discussed the effect of using BERT to detect hate speech in a social media environment (Mozafari et al., 2019) . Shriphani Palakodety et al. performed the qualitative and quantitative analysis of this embedding space for a variety of languages, they used Active Learning to train set construction in the analysis of peaceful speech (Palakodety et al., 2019) . Zewdie Mossie et al. used Word2Vec word embedding technology for feature extraction when conducting Vulnerable community identification, and compared the effects Figure 1 : A word cloud image generated from the text marked with \"Non hope speech\" in the English training data set provided by the task organizer. Figure 2 : A word cloud image generated from the text marked with \"Hope speech\" in the English training data set provided by the task organizer.", |
| "cite_spans": [ |
| { |
| "start": 201, |
| "end": 224, |
| "text": "(Mozafari et al., 2019)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 446, |
| "end": 471, |
| "text": "(Palakodety et al., 2019)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 636, |
| "end": 644, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 785, |
| "end": 793, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "of classic classifiers with deep learning classifiers (Mossie and Wang, 2020) . Shanita Biere et al. used natural language technology to detect hate speech on social media, using a convolutional neural network (CNN) (Biere et al., 2018).", |
| "cite_spans": [ |
| { |
| "start": 54, |
| "end": 77, |
| "text": "(Mossie and Wang, 2020)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Viktor Golem et al. tried different machine learning approaches, including traditional (shallow) machine learning models, deep learning models, and a combination of both for aggressive text detection (Golem et al., 2018) . Jing MA et al. proposed a novel method that learns continuous representations of microblog events for identifying rumors. That model is based on recurrent neural networks (RNN) and that can capture the variation of contextual information of relevant posts over time (Ma et al., 2016) . When discussing the detection of hate speech, Reynaldo Gil Pons et al. proposed an Attention-based Long Short-Term Memory Network, which is a model containing a bidirectional LSTM neural network (BILSTM)(De la Pena Sarrac\u00e9n et al., 2018).", |
| "cite_spans": [ |
| { |
| "start": 200, |
| "end": 220, |
| "text": "(Golem et al., 2018)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 489, |
| "end": 506, |
| "text": "(Ma et al., 2016)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The data sets we can use are the training set and validation set in three languages(Tamil, Malayalam, and English) provided by the task organizer team(Chakravarthi, 2020; Chakravarthi and Muralidaran, 2021). We only participate in the English task, so here we only analyze the English data set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "We use the official data set for visual analysis. We can observe that some words such as \"live\", \"people\", \"black\", \"matter\" \"racist\" in the English data set appear very frequently. The appearance of these words is very suitable for the theme of the mission we participated in. However, these words mostly express some negative information. And these data mainly come from social media platforms, on the other hand, they also reflect the current status quo on social media platforms. We visualize the text data of two different labels of \"Non hope speech\" and \"Hope speech\". Figures 1 and 2 show the results of text visualization.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 575, |
| "end": 591, |
| "text": "Figures 1 and 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "There are three types of category labels that appear in the English data set. They are \"Hope speech\", \"Non hope speech\", \"Not English\". Through further analysis of the data set, most of the mark types in the English data are \"Non hope speech\" (around 90%). There is an imbalance of category labels in the data sets of the English data set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": "3" |
| }, |
| { |
| "text": "K-fold cross-validation reduces variance by averaging the results of k different group training.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "K-fold cross-validation", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The specific steps are as follows: The first step is to randomly divide the original data into k parts without repeated sampling; The second step is to extract 1 copy as the test set, and the remaining k-1 copies are applied to the training set for model training; In the third step, repeat the second step k times, so that a model is obtained after training on each training set and used to test on the corresponding test set, calculate and save the evaluation index of the mode; The fourth step is to calculate the average of k groups of test results as an estimate of model accuracy, and as the performance index of the model under current k-fold cross-validation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "K-fold cross-validation", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The reason we choose the ALBERT fine-tuning program is that ALBERT's parameters are shared. Compared with BERT, ALBERT has fewer parameters, and the corresponding model training requires less time and memory. The advantage is that when the model is large, parameter sharing will be a strong regularization method for the model, and it is not easy to overfit.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "ALBERT Fine-tuning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The related work and conclusions of Sun and others on the fine-tuning of BERT have inspired us (Sun et al., 2019) . We try to apply these conclusions to the ALBERT model. We chose to combine a portion of the output of the ALBERT encoder layer with [CLS] (the output shape is [batch size, hidden size]). We take the output of the encoder of the ALBERT model (layers 9-11) (the output shape is [batch size, sequence length, hidden size]). Then take out the data of the 0th and 2nd dimensions of the output tensor of each layer of 9-11 layers to obtain three Tensors of the same shape (shape is [batch size, hidden size]). Next, the three tensors obtained in the previous step and the [CLS] output result of the ALBERT model are spliced together (the shape is [batch size, hidden size*4]). Finally, input the results of the previous step into the classifier.", |
| "cite_spans": [ |
| { |
| "start": 95, |
| "end": 113, |
| "text": "(Sun et al., 2019)", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 682, |
| "end": 687, |
| "text": "[CLS]", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "ALBERT Fine-tuning", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "We combine the English data training set and the English data validation set provided by the task or-", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data Preprocessing", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Precision Recall F1 Score Test set 0.93 0.94 0.93 Validation set 0.94 0.95 0.94 Table 1 : The resulting score of our model method on the test set and validation set. The test set result score comes from the result score published by the task organization team on the leaderboard. The resulting score on the verification set is the result score obtained by us using the official evaluation index and the verification set.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 80, |
| "end": 87, |
| "text": "Table 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": null |
| }, |
| { |
| "text": "ganizer team in one file. The purpose is to integrate the validation set provided by the task organizer into the k-fold cross-validation data set. Then perform 5-fold cross-validation data processing on this merged file to split the data into five data sets. Each data set contains a new training set and a new validation set.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Data", |
| "sec_num": null |
| }, |
| { |
| "text": "For the English tasks, we set model training parameters as follows: epoch, batch size, maximum sequence length, and learning rate of the task are 5, 32, 60, and 3e-5, respectively. We combined the training set and validation set provided by the task manager into a new data set. Then performed k-fold cross-validation on it to get the result of the k-fold cross-validation vote. The voting method in the reasoning process is to take the average value of the probability value of k outputs as the final predicted output logical value. We choose the training language models of ALBERT-V2 version 1 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setting", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "Hope Speech Detection for Equality, Diversity, and Inclusion at LT-EDI 2021-EACL 2021 task evaluation indicators adopt F1 score weighted average. In the results announced by the task organizer team, our English scores were tied for first place on the list. This result is due to the pre-trained language model in the Transformer structure. The main reason is that the pre-trained language model has been pre-trained on a large number of English data sets. The ALBERT model is based on the Transformer model (Lan et al., 2019 ). The Transformer model has great advantages in capturing global information (Vaswani et al., 2017) . Not English 0.00 0.00 0.00 3 Table 2 : On the labeled test set data published by the task organizer team, we compare the predicted results with the real results.", |
| "cite_spans": [ |
| { |
| "start": 507, |
| "end": 524, |
| "text": "(Lan et al., 2019", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 603, |
| "end": 625, |
| "text": "(Vaswani et al., 2017)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 657, |
| "end": 664, |
| "text": "Table 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Analysis of Results", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "For the English task in this competition, we use the last three layers of the ALBERT model output to fuse with the [CLS] output. Use K-fold cross-validation on the integrated method to obtain better results. Our test results for Youtube's hopeful speech are not bad. In the final result, Precision was 0.93, Recall was 0.94, F1 Score was 0.93, and the ranking was tied for first place. But in future work, we will try to improve our method and choose to try other models, hoping to achieve better results on such tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "6" |
| }, |
| { |
| "text": "https://huggingface.co/albert-base-v2", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Hate speech detection using natural language processing techniques", |
| "authors": [], |
| "year": 2018, |
| "venue": "Master Business AnalyticsDepartment of Mathematics Faculty of Science", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shanita Biere, Sandjai Bhulai, and Master Business An- alytics. 2018. Hate speech detection using natural language processing techniques. Master Business AnalyticsDepartment of Mathematics Faculty of Sci- ence.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "HopeEDI: A multilingual hope speech detection dataset for equality, diversity, and inclusion", |
| "authors": [ |
| { |
| "first": "Chakravarthi", |
| "middle": [], |
| "last": "Bharathi Raja", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the Third Workshop on Computational Modeling of People's Opinions, Personality, and Emotion's in Social Media", |
| "volume": "", |
| "issue": "", |
| "pages": "41--53", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bharathi Raja Chakravarthi. 2020. HopeEDI: A mul- tilingual hope speech detection dataset for equality, diversity, and inclusion. In Proceedings of the Third Workshop on Computational Modeling of People's Opinions, Personality, and Emotion's in Social Me- dia, pages 41-53, Barcelona, Spain (Online). Asso- ciation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Findings of the shared task on Hope Speech Detection for Equality, Diversity, and Inclusion", |
| "authors": [ |
| { |
| "first": "Vigneshwaran", |
| "middle": [], |
| "last": "Bharathi Raja Chakravarthi", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Muralidaran", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "Proceedings of the First Workshop on Language Technology for Equality, Diversity and Inclusion", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bharathi Raja Chakravarthi and Vigneshwaran Mural- idaran. 2021. Findings of the shared task on Hope Speech Detection for Equality, Diversity, and Inclu- sion. In Proceedings of the First Workshop on Lan- guage Technology for Equality, Diversity and Inclu- sion. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Towards countering hate speech and personal attack in social media", |
| "authors": [ |
| { |
| "first": "Polychronis", |
| "middle": [], |
| "last": "Charitidis", |
| "suffix": "" |
| }, |
| { |
| "first": "Stavros", |
| "middle": [], |
| "last": "Doropoulos", |
| "suffix": "" |
| }, |
| { |
| "first": "Stavros", |
| "middle": [], |
| "last": "Vologiannidis", |
| "suffix": "" |
| }, |
| { |
| "first": "Ioannis", |
| "middle": [], |
| "last": "Papastergiou", |
| "suffix": "" |
| }, |
| { |
| "first": "Sophia", |
| "middle": [], |
| "last": "Karakeva", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1912.04106" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Polychronis Charitidis, Stavros Doropoulos, Stavros Vologiannidis, Ioannis Papastergiou, and Sophia Karakeva. 2019. Towards countering hate speech and personal attack in social media. arXiv preprint arXiv:1912.04106.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Combining shallow and deep learning for aggressive text detection", |
| "authors": [ |
| { |
| "first": "Viktor", |
| "middle": [], |
| "last": "Golem", |
| "suffix": "" |
| }, |
| { |
| "first": "Mladen", |
| "middle": [], |
| "last": "Karan", |
| "suffix": "" |
| }, |
| { |
| "first": "Jan\u0161najder", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC-2018)", |
| "volume": "", |
| "issue": "", |
| "pages": "188--198", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Viktor Golem, Mladen Karan, and Jan\u0160najder. 2018. Combining shallow and deep learning for aggressive text detection. In Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC- 2018), pages 188-198.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Albert: A lite bert for self-supervised learning of language representations", |
| "authors": [ |
| { |
| "first": "Zhenzhong", |
| "middle": [], |
| "last": "Lan", |
| "suffix": "" |
| }, |
| { |
| "first": "Mingda", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Goodman", |
| "suffix": "" |
| }, |
| { |
| "first": "Kevin", |
| "middle": [], |
| "last": "Gimpel", |
| "suffix": "" |
| }, |
| { |
| "first": "Piyush", |
| "middle": [], |
| "last": "Sharma", |
| "suffix": "" |
| }, |
| { |
| "first": "Radu", |
| "middle": [], |
| "last": "Soricut", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1909.11942" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learn- ing of language representations. arXiv preprint arXiv:1909.11942.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Detecting rumors from microblogs with recurrent neural networks", |
| "authors": [ |
| { |
| "first": "Jing", |
| "middle": [], |
| "last": "Ma", |
| "suffix": "" |
| }, |
| { |
| "first": "Wei", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| }, |
| { |
| "first": "Prasenjit", |
| "middle": [], |
| "last": "Mitra", |
| "suffix": "" |
| }, |
| { |
| "first": "Sejeong", |
| "middle": [], |
| "last": "Kwon", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Bernard", |
| "suffix": "" |
| }, |
| { |
| "first": "Kam-Fai", |
| "middle": [], |
| "last": "Jansen", |
| "suffix": "" |
| }, |
| { |
| "first": "Meeyoung", |
| "middle": [], |
| "last": "Wong", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Cha", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jing Ma, Wei Gao, Prasenjit Mitra, Sejeong Kwon, Bernard J Jansen, Kam-Fai Wong, and Meeyoung Cha. 2016. Detecting rumors from microblogs with recurrent neural networks.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Vulnerable community identification using hate speech detection on social media", |
| "authors": [ |
| { |
| "first": "Zewdie", |
| "middle": [], |
| "last": "Mossie", |
| "suffix": "" |
| }, |
| { |
| "first": "Jenq-Haur", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Information Processing & Management", |
| "volume": "57", |
| "issue": "3", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zewdie Mossie and Jenq-Haur Wang. 2020. Vulnera- ble community identification using hate speech de- tection on social media. Information Processing & Management, 57(3):102087.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "A bert-based transfer learning approach for hate speech detection in online social media", |
| "authors": [ |
| { |
| "first": "Marzieh", |
| "middle": [], |
| "last": "Mozafari", |
| "suffix": "" |
| }, |
| { |
| "first": "Reza", |
| "middle": [], |
| "last": "Farahbakhsh", |
| "suffix": "" |
| }, |
| { |
| "first": "Noel", |
| "middle": [], |
| "last": "Crespi", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "International Conference on Complex Networks and Their Applications", |
| "volume": "", |
| "issue": "", |
| "pages": "928--940", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marzieh Mozafari, Reza Farahbakhsh, and Noel Crespi. 2019. A bert-based transfer learning approach for hate speech detection in online social media. In In- ternational Conference on Complex Networks and Their Applications, pages 928-940. Springer.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Hate speech detection and racial bias mitigation in social media based on bert model", |
| "authors": [ |
| { |
| "first": "Marzieh", |
| "middle": [], |
| "last": "Mozafari", |
| "suffix": "" |
| }, |
| { |
| "first": "Reza", |
| "middle": [], |
| "last": "Farahbakhsh", |
| "suffix": "" |
| }, |
| { |
| "first": "No\u00ebl", |
| "middle": [], |
| "last": "Crespi", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "PloS one", |
| "volume": "15", |
| "issue": "8", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marzieh Mozafari, Reza Farahbakhsh, and No\u00ebl Crespi. 2020. Hate speech detection and racial bias mitiga- tion in social media based on bert model. PloS one, 15(8):e0237861.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Kashmir: A computational analysis of the voice of peace", |
| "authors": [ |
| { |
| "first": "Shriphani", |
| "middle": [], |
| "last": "Palakodety", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Ashiqur", |
| "suffix": "" |
| }, |
| { |
| "first": "Jaime", |
| "middle": [ |
| "G" |
| ], |
| "last": "Khudabukhsh", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Carbonell", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shriphani Palakodety, Ashiqur R KhudaBukhsh, and Jaime G Carbonell. 2019. Kashmir: A compu- tational analysis of the voice of peace. CoRR, abs/1909.12940.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Hate speech detection using attention-based lstm", |
| "authors": [ |
| { |
| "first": "Gretel", |
| "middle": [ |
| "Liz" |
| ], |
| "last": "De La Pena Sarrac\u00e9n", |
| "suffix": "" |
| }, |
| { |
| "first": "Reynaldo", |
| "middle": [ |
| "Gil" |
| ], |
| "last": "Pons", |
| "suffix": "" |
| }, |
| { |
| "first": "Carlos", |
| "middle": [ |
| "Enrique" |
| ], |
| "last": "Muniz Cuza", |
| "suffix": "" |
| }, |
| { |
| "first": "Paolo", |
| "middle": [], |
| "last": "Rosso", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "EVALITA Evaluation of NLP and Speech Tools for Italian", |
| "volume": "12", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gretel Liz De la Pena Sarrac\u00e9n, Reynaldo Gil Pons, Carlos Enrique Muniz Cuza, and Paolo Rosso. 2018. Hate speech detection using attention-based lstm. EVALITA Evaluation of NLP and Speech Tools for Italian, 12:235.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "How to fine-tune bert for text classification?", |
| "authors": [ |
| { |
| "first": "Chi", |
| "middle": [], |
| "last": "Sun", |
| "suffix": "" |
| }, |
| { |
| "first": "Xipeng", |
| "middle": [], |
| "last": "Qiu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yige", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Xuanjing", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "China National Conference on Chinese Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "194--206", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chi Sun, Xipeng Qiu, Yige Xu, and Xuanjing Huang. 2019. How to fine-tune bert for text classification? In China National Conference on Chinese Computa- tional Linguistics, pages 194-206. Springer.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Attention is all you need", |
| "authors": [ |
| { |
| "first": "Ashish", |
| "middle": [], |
| "last": "Vaswani", |
| "suffix": "" |
| }, |
| { |
| "first": "Noam", |
| "middle": [], |
| "last": "Shazeer", |
| "suffix": "" |
| }, |
| { |
| "first": "Niki", |
| "middle": [], |
| "last": "Parmar", |
| "suffix": "" |
| }, |
| { |
| "first": "Jakob", |
| "middle": [], |
| "last": "Uszkoreit", |
| "suffix": "" |
| }, |
| { |
| "first": "Llion", |
| "middle": [], |
| "last": "Jones", |
| "suffix": "" |
| }, |
| { |
| "first": "Aidan", |
| "middle": [ |
| "N" |
| ], |
| "last": "Gomez", |
| "suffix": "" |
| }, |
| { |
| "first": "Lukasz", |
| "middle": [], |
| "last": "Kaiser", |
| "suffix": "" |
| }, |
| { |
| "first": "Illia", |
| "middle": [], |
| "last": "Polosukhin", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1706.03762" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. arXiv preprint arXiv:1706.03762.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "text": "The fine-tuning strategies we use in English tasks. The output of layers 9-11 must be processed before being spliced with[CLS]. The processing method is to discard the first dimension data in the original data ([B, S, H] to [B, H]) to obtain a tensor output with the same shape as [CLS] (shape is[B, H]).", |
| "type_str": "figure" |
| } |
| } |
| } |
| } |