| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T14:39:23.503882Z" |
| }, |
| "title": "COVID-19 Surveillance through Twitter using Self-Supervised and Few Shot Learning", |
| "authors": [ |
| { |
| "first": "Brandon", |
| "middle": [], |
| "last": "Lwowski", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Texas at San Antonio San Antonio", |
| "location": { |
| "postCode": "78249", |
| "region": "TX", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Rad", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Texas at San Antonio San Antonio", |
| "location": { |
| "postCode": "78249", |
| "region": "TX", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Public health surveillance and tracking virus via social media can be a useful digital tool for contact tracing and preventing the spread of the virus. Nowadays, large volumes of COVID-19 tweets can quickly be processed in real-time to offer information to researchers. Nonetheless, due to the absence of labeled data for COVID-19, the preliminary supervised classifier or semi-supervised self-labeled methods will not handle non-spherical data with adequate accuracy. With the seasonal influenza and novel Coronavirus having many similar symptoms, we propose using few shot learning to fine-tune a semi-supervised model built on unlabeled COVID-19 and previously labeled influenza dataset that can provide insights into COVID-19 that have not been investigated. The experimental results show the efficacy of the proposed model with an accuracy of 86%, identification of Covid-19 related discussion using recently collected tweets.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Public health surveillance and tracking virus via social media can be a useful digital tool for contact tracing and preventing the spread of the virus. Nowadays, large volumes of COVID-19 tweets can quickly be processed in real-time to offer information to researchers. Nonetheless, due to the absence of labeled data for COVID-19, the preliminary supervised classifier or semi-supervised self-labeled methods will not handle non-spherical data with adequate accuracy. With the seasonal influenza and novel Coronavirus having many similar symptoms, we propose using few shot learning to fine-tune a semi-supervised model built on unlabeled COVID-19 and previously labeled influenza dataset that can provide insights into COVID-19 that have not been investigated. The experimental results show the efficacy of the proposed model with an accuracy of 86%, identification of Covid-19 related discussion using recently collected tweets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "The typical seasonal influenza virus and the current development of COVID-19 have multiple similarities from symptoms to how the virus is spread. Both viruses attack the respiratory system, can be spread through asymptomatic carriers, cases can range from mild to severe cases, and are transmitted by contact and/or droplets. Influenza and COVID-19 both can impact a community negatively due to the contagious nature of the virus and the high number of deaths caused by the viruses.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Public health surveillance like digital contact tracing (Ferretti et al., 2020; Ekong et al., 2020) , epidemiological studies (Salath\u00e9 et al., 2013) , event detection (Lwowski et al., 2018) , and monitoring the prevalence of vaccinations (Huang et al., 2017) can be used to help contain the virus and prevent its spread to the masses. These tools and techniques range from cellphone applications installed personal phones that track the exact spread of a virus (Ekong et al., 2020) to the development of machine learning-based techniques to study the spread of a virus using social media (Lamb et al., 2013; Corley et al., 2009 Corley et al., , 2010 Santillana et al., 2015; Broniatowski et al., 2013; Signorini et al., 2011) . Similarly, machine learning-based methods have been developed to monitor the public's view on vaccines to combat the anti-vaccine narrative (Huang et al., 2017) .", |
| "cite_spans": [ |
| { |
| "start": 56, |
| "end": 79, |
| "text": "(Ferretti et al., 2020;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 80, |
| "end": 99, |
| "text": "Ekong et al., 2020)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 126, |
| "end": 148, |
| "text": "(Salath\u00e9 et al., 2013)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 167, |
| "end": 189, |
| "text": "(Lwowski et al., 2018)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 238, |
| "end": 258, |
| "text": "(Huang et al., 2017)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 461, |
| "end": 481, |
| "text": "(Ekong et al., 2020)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 588, |
| "end": 607, |
| "text": "(Lamb et al., 2013;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 608, |
| "end": 627, |
| "text": "Corley et al., 2009", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 628, |
| "end": 649, |
| "text": "Corley et al., , 2010", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 650, |
| "end": 674, |
| "text": "Santillana et al., 2015;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 675, |
| "end": 701, |
| "text": "Broniatowski et al., 2013;", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 702, |
| "end": 725, |
| "text": "Signorini et al., 2011)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 868, |
| "end": 888, |
| "text": "(Huang et al., 2017)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Using large sources of public information from social media to mine influenza and COVID-19 data allows researchers to help gain insight about the viruses. Just using a search word like \"flu\" and \"coronavirus\" with the Twitter API will return millions of tweets with information about vaccines, rumors, symptoms, and family/friends who have contracted the virus. Classifying tweets in to smaller subsets including categories like \"Self vs Other\" and \"Awareness vs Infection\" provides a deeper understanding on how the the influenza and COVID-19 are affecting the communities. Example tweets of each category can be found in Table 1 While other researchers aim to use unsupervised learning to cluster and perform topic modeling on COVID-19 tweets (Mackey et al., 2020; Medford et al., 2020) , we decided to combine selfsupervised learning combined with few shot learning to produce more accurate predictions for specific categories.", |
| "cite_spans": [ |
| { |
| "start": 745, |
| "end": 766, |
| "text": "(Mackey et al., 2020;", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 767, |
| "end": 788, |
| "text": "Medford et al., 2020)", |
| "ref_id": "BIBREF15" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 623, |
| "end": 630, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The major road block for using deep learning models on the COVID-19 tweets is the lack of annotated data. With millions of tweets related to COVID-19 flooding social media, researchers have a difficult time performing supervised learning on the data. We propose a method to attack this problem by transferring knowledge learned in influenza data and integrating it with latent variables obtained from the unlabeled dataset of COVID-19 to preform a deeper understanding through self-", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Related -Is US using this HIV antiretroviral drug to treat corona. Seems this is very successful in treating patients in Kerala, India and they have high success rate #coronavirus #CoronavirusUSA Be able to classify the tweets into Related tweets allows researchers to filter out tweets that mention COVID-19 in order to attract consumers (Click Bait). -GREAT NEWS! Mom just woke up and asked for soup. She's eating soup and drinking water, dad is still on the vent fighting the fight #COVID19", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Category Tweet Importance", |
| "sec_num": null |
| }, |
| { |
| "text": "Awareness -U.S. Citizens that are NOT showing #COVID19 symptoms are PLEADING to be let off cruise ship that has #Coronavirus infected passengers on it. I Wonder if these same passengers spoke up when all those images of Mothers & Kids being seperated &", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Category Tweet Importance", |
| "sec_num": null |
| }, |
| { |
| "text": "Tweets that are classified as Awareness is beneficial to researchers when not wanting to look at information regarding users actually becoming infected with the virus. These topics can include perception of masks, rumors of vaccines, etc.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Category Tweet Importance", |
| "sec_num": null |
| }, |
| { |
| "text": "-#COVID19 is such a public health threat because the virus can be transmitted by individuals who are infected, but are not showing symptoms.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Category Tweet Importance", |
| "sec_num": null |
| }, |
| { |
| "text": "Infection -I'm absolutely broken! This morning I found out my bio mom (who lives in the UK) is infected; also has pneumonia. Her medical team has said to \"prepare for worst case scenario.\" Well, here we are! She's going to die alone with her entire family in another country. F YOU #COVID19", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Category Tweet Importance", |
| "sec_num": null |
| }, |
| { |
| "text": "When using Social Media for public health surveillance and contract tracing. Having a classifier that can accurately classify tweets as a positive case can be extremely beneficial in discovering hotspots for the virus as well as other people they may have came in contact with -So my 80yo old dad tested positive. He is now on a ventilator. I need all your prayers that he pulls through this. #Covid19 #CoronaVirus", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Category Tweet Importance", |
| "sec_num": null |
| }, |
| { |
| "text": "-@siggyflicker I understand that, but unless the clinical studies are throughly completed (hopefully sped up), we have to continue to be cautious -even then, unfortunately it's still not a vaccine that will prevent the contraction and spread of #COVID19", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": null |
| }, |
| { |
| "text": "Tweets aimed around the topic of Vaccines has present and future applications. In the present it allows researchers to investigate misinformation and/or the public perception of vaccines.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": null |
| }, |
| { |
| "text": "In future research we can further classify these tweets into Intent to Receive or Already has Received a vaccine. -I'm tired of hearing all the scary stuff about the virus. The news needs to include what strides we are making in treatments, a vaccine, or even a cure. #COVID-19 #coronavirus #Coron-avirusUSA (Lamsal, 2020) . Also provided in the table is a brief explanation of why the specific category is important to understanding multiple aspects of COVID-19. These categories allow researchers to study the virus at a more granular level. supervised classification. The main contribution of this paper is three-fold:", |
| "cite_spans": [ |
| { |
| "start": 308, |
| "end": 322, |
| "text": "(Lamsal, 2020)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 We propose a self-supervised learning algorithm to monitor COVID-19 Twitter using an autoencoder to learn the latent representations and then transfer the knowledge to COVID-19 Infection classifier by fine-tuning the Multi-Layer Perceptron (MLP) using few shot learning.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 We evaluate the utility of Twitter data for COVID-19 surveillance by training the 4 four binary models, in a computing environment provided by Jetstream (Stewart et al., 2015) , to classify tweets into 4 different categories, related to COVID-19, COVID-19 Infection, COVID-19 Self/Others Infection, and COVID-19 Vaccine.", |
| "cite_spans": [ |
| { |
| "start": 145, |
| "end": 177, |
| "text": "Jetstream (Stewart et al., 2015)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 Lastly but not least, we transfer a pre-trained influenza MLP classifier to fine-tune the accuracy of the self-supervised model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section, we explain the set-up of our study and motivate the core components of the proposed COVID-19 self-supervised learning with less labeled COVID-19 data. We begin with the introduction of our three datasets and data annotation strategy in Section 2.1 and follow by describing the core components of our self-supervised learning, as shown in Figure 1 . We first learn how to generate the latent representation of unlabeled COVID-19 tweets using self-supervised Convolutional Autoencoder model in Section 2.2. Subsequently, design our COVID-19 supervised downstream task with a pre-trained Influnza classification, due to symptom similarities, and fine-tuning of the model using COVID-19 Few shot learning in Section 2.3 and 2.4. We finally present our results and evaluation metrics in Section 3 followed by discussion and conclusion in Section 4. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 355, |
| "end": 363, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We make use of three datasets: Coronavirus (Lamsal, 2020) for self-supervised pre-task learning using Convolution Autoencoder, FluTrack (Lamb et al., 2013) , and FluVacc (Huang et al., 2017) for training the influenza classifiers. In order to use few-shot learning, an annotated COVID-19 dataset must be used to fine tune the overall model. The dataset collected by Lamsal et al (Lamsal, 2020) provides us with a large amount of data for analysis and predictions. a subsample of 500k tweets are used from this dataset to train the autoencoder. To obtain accurate annotations for this dataset, all tweets are shuffled and randomly sampled 25 times with a size of 100 tweets. Each sample is then distributed to 3 different annotators. Each annotator is asked to answer 4 questions about the tweet corresponding to the 4 categories we hope to classify. Only tweets where at least 2 annotators agree on the label are used for training and testing. A breakdown of each task and total number of tweets can be found in Section 3.4.", |
| "cite_spans": [ |
| { |
| "start": 136, |
| "end": 155, |
| "text": "(Lamb et al., 2013)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 170, |
| "end": 190, |
| "text": "(Huang et al., 2017)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Data Annotation", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The previous work done by Lamb et al. and Huang et al. provided annotated datasets that allow us to implement supervised deep learning models. The FluTrack dataset provides 3 different classification labels: is the tweet related to the influenza or not, is the tweet talking about awareness or infection of the flu, and finally is the tweet about the user or about someone else.The FluTrack dataset consisted of 11,990 tweets collected from years 2009 through 2012. The FluVacc dataset also has 3 classification labels: is the tweet about the vaccine or not, does the tweet contain intent to get the vaccine or not, and lastly does the tweet have information saying the user already received the vaccination or not. The FluVacc consisted of 10,000 total annotated tweets. The provided tweet labels are not mutually exclusive meaning a tweet can belong to multiple categories. In the example of \"My mom caught the flu, hopefully i dont catch it...\" can be classified as Infection and Other.", |
| "cite_spans": [ |
| { |
| "start": 26, |
| "end": 54, |
| "text": "Lamb et al. and Huang et al.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets and Data Annotation", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Most of the self-supervised techniques for latent text representation rely on Transformer architectures that predict the next token. In this study, we generate Coronavirus (Lam-sal, 2020) latent text representation by utilizing Convolutional Autoencoder. Autoencoders are a special category of deep neural networks that are deliberately programmed to make output as close as possible to input. Instead of training to predict y given input x, the network will be trained in an unsupervised approach to replicate its own input x. The autoencoder, in figure 1, is composed of 2 parts, the encoder and decoder. The job of the encoder is to compress the data into the latent space and the decoder takes the latent space as its input and attempts to reconstruct the original input. To simplify the autoencoder, we define it as a composite function, Equation 1, with the encoder E and decoder D with a loss function defined to minimize the difference between the input, X, and the output,X.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "min (X \u2212 X) = (D(E(X)) \u2212 X)", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Once an autoencoder is trained for Coronavirus dataset, the latent variable z can be used to extract important features for COVID-19 tweets. As shown in figure 1, both the encoder and decoder are convolutional neural networks. We pass the vectorized tweet through a word embedding, the word embedding layer converts every word of the tweet into an n-dimensional vector. This converts the input dimensions into 2-D that can be fed to the Convolutional Autoencoders for our task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Hadifar et al. (Hadifar et al., 2019) also uses an Autoencoder in order to help classify text. They also use the Autoencoder to pre-train the encoder, but instead of classifying text, they use KNN for clustering similar text with the learned latent space. Their claims support the use of Autoencoders to achieve a deeper understanding of short texts such as tweets.", |
| "cite_spans": [ |
| { |
| "start": 15, |
| "end": 37, |
| "text": "(Hadifar et al., 2019)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The word embedding used in our research was the GloVe 50d trained on 6 billion words from Wikipedia (Pennington et al., 2014) with a max tweet length of being 50 words. The max length of the tweet was decided by calculating 2 standard deviation greater than the mean allowing more than 95% of the tweets to not be altered in size. And tweets that are less than 50 words are post padded with 0's, representing no word being present.", |
| "cite_spans": [ |
| { |
| "start": 100, |
| "end": 125, |
| "text": "(Pennington et al., 2014)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The next step in the encoder consists of three different 1 dimensional convolutions with kernel sizes of 2,3 and 4. These three operations are grouped together in Table 2 . This allows the encoder to learn 2,3 and 4 word relationships which is important in encoding meaning and semantics into the latent space. The outputs from each of the 3 convolutional layers are passed through an activation function (ReLU) and concatenated together. One more convolutional layer is used as well as a ReLU layer before flattening the output. The output of the flattened layer is the latent space which is half the size of the original input. The encoder ends at the Flatten layer in Table 2 and thus start the process of the decoder.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 163, |
| "end": 170, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| }, |
| { |
| "start": 671, |
| "end": 678, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The output (latent space z c ) of the encoder is then used as the input to the decoder. Similar operations are performed in order to reverse the operations and reconstruct the original vector/tweet. If the decoder can accurately reconstruct the original input then the latent space has learned and encoded the right information in a compressed format. The decoder begins with the latent space being reshaped into a 2 dimensional vector. A convolution of 2,3,4 are performed on the latent space. The outputs are passed through the ReLU activation and concatenated in the same manner as the encoder. Where the decoder differs from the encoder is the final layer. The output of the final convolutional layer is passed through the softmax activation function. By minimizing the difference between the input vector and reconstructed vector we achieve our goal of generating Tweet latent representation. Once the Autoencoder is trained on Cornonavirus data,the decoder can be removed from the network, leaving the encoder and latent space to be used as its own model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Convolutional Autoencoder", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "A similar CNN architecture found in Kim et al. CNN for text classification is used to train a influenza tweet classifier (Kim, 2014) . The influenza tweets are transformed to vectors using the same word embeddings in the auto encoder. During training a latent representation is learned, Z i , from the convolutions of size 3,4 and 5 words. The latent representation is then passed to the Multi-Layer Perceptron, MLP W i , and is trained with supervision to predict the correct label Y . In Figure 1 , this can be seen on the lower section of the image. A description of each label is discussed in detail in Section 2.1. The accuracy of the influenza classifier is discussed in Section 3.4 as well as Table 3 .", |
| "cite_spans": [ |
| { |
| "start": 36, |
| "end": 46, |
| "text": "Kim et al.", |
| "ref_id": null |
| }, |
| { |
| "start": 121, |
| "end": 132, |
| "text": "(Kim, 2014)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 490, |
| "end": 498, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| }, |
| { |
| "start": 700, |
| "end": 707, |
| "text": "Table 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Influenza Classification", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "A large influence for our research comes from Bowman et al. (Bowman et al., 2015) . Bowman et al. takes the parameters of a model that was trained on one dataset and trains a new model using a portion of the new dataset. Bowman et al. suggests that by introducing a corpus that is high quality, it can be used to transfer knowledge and learn sentence meanings that can improve downstream text classification tasks.", |
| "cite_spans": [ |
| { |
| "start": 46, |
| "end": 81, |
| "text": "Bowman et al. (Bowman et al., 2015)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Few-Shot Classification", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "At this point we have 2 trained models, the autoencoder for COVID-19 tweets and the influenza tweet classifier for the 6 categories. We want to use as much knowledge learned in the COVID-19 latent representation and Influenza MLP layer to accurately predict on COVID-19 classification tasks, with limited COVID-19 labeled data. We accomplish this task with few-shot learning with a warm start. This is a similar methodology to Dirkson et al. (Dirkson and Verberne, 2019) . They use the ULMfit to transfer knowledge between health and twitter data (Howard and Ruder, 2018) . Where we differ from the ULMfit learning is we completely freeze the encoders parameters and only allow the MLP to be trained.", |
| "cite_spans": [ |
| { |
| "start": 442, |
| "end": 470, |
| "text": "(Dirkson and Verberne, 2019)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 547, |
| "end": 571, |
| "text": "(Howard and Ruder, 2018)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Few-Shot Classification", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "The same process used for the influenza classifier is used for the COVID-19 classifier. The encoder W e and latent variable z are frozen and the decoder is removed. MLP i,c is appended to z and used to classify COVID-19 tweets. If this model were trained from a cold start, the weights of MLP i,c would be randomly assigned during the first epoch of training and adjusted from there. This prevents any classification knowledge being transferred from the influenza to the COVID-19 tweets. Instead, the weights of MLP i are used to initialize the weights of MLP i,c . This warm start training allows the model to use knowledge from the the influenza training. We can then train the model with the few labeled COVID-19 tweets and fine tune the model for the downstream task of classifying all COVID-19 tweets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Few-Shot Classification", |
| "sec_num": "2.4" |
| }, |
| { |
| "text": "In table 3, the accuracy of the influenza, COVID-19 trained on a cold start, and COVID-19 trained with the weights initialized with the influenza MLP (COVID*) are given for each of the 4 tasks. The task of Received and Intent are removed from the COVID-19 classifiers since there are no vaccines currently available for the virus. Table 4 : Macro Precision, Recall, and F1 for COVID-19 Classifiers. Statistics are reported in the format of P/R/F1. COVID column is the results from training the MLP from scratch (cold start), while COVID* is the results from training the MLP by initializing the MLP with the influenza classifier weights.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 331, |
| "end": 338, |
| "text": "Table 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "3" |
| }, |
| { |
| "text": "For classifying tweets into COVID-19 Related and COVID-19 Non-Related, we took a sub sample COVID-19 Tweets and a sub sample of influenza related tweets and combined them into one dataset. We trained the classifier on 1000 influenza and 400 COVID-19 tweets but, tested on 4000 influenza and 1400 COVID-19 tweets. A training accuracy of 91% and testing accuracy of 86% were achieved as well as a precision score of 0.69, a recall score of 0.77 and a F1 score of 0.73. The same train and testing are done on the cold start model but achieved a higher train but lower test accuracy with a precision score 0.71, recall score 0.70 and a F1 score of 0.71. The COVID-19 MLP initialized with the influenza MLP outperforms the the cold start on 3 out of the 5 statistics including a 3% increase on test accuracy.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Awareness vs Infection can be seen as a binary classification problem. We classified each COVID-19 into 2 more subcategories, is the twitter user aware of the virus or are they talking about themselves or others being infected. We trained the classifier with 250 awareness and 250 infection tweets and tested on 450 tweets in each category. An accuracy of 98% on the train set and 73% on the test set were achieved as well as a precision score 0.73, a recall score of 0.72 and a F1 score 0.73. The same training and testing were carried our on the cold start model but achieved a lower training and lower testing accuracy with a precision score 0.69, recall score 0.68 and an F1 score of 0.68. The COVID-19 MLP initialized with the influenza MLP outperforms the the cold start on all accuracy statistics including a 3% increase on test accuracy.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Awareness Vs Infection", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In self vs other, we try and classify the infection tweets into further categories. We would like to clarify whether the post is about the author, or if the author is concerned about the COVID-19 infection of others. We trained the classifier on 55 tweets labeled \"Other\" and 20 tweets labeled as \"Self\". The test set consisted of 45 tweets labeled \"Other\" and 30 tweets labeled \"Self\". An accuracy of 98% on the train set and 86% on the test set were achieved as well as a precision score of 0.86, a recall score of 0.87, and a F1 score of 0.86. The same training and testing were carried our on the cold start model but achieved a lower training and lower testing accuracy with a precision score 0.81, recall score 0.76 and a F1 score of 0.77. The COVID-19 MLP initialized with the influenza MLP outperforms the the cold start on 4 out of the 5 statistics including a 5% increase on test accuracy.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Self Vs Other", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "The last category was Vaccine related tweets. We would like to clarify whether the post is about vaccine or cure, or whether it is about certain facets of the virus. The train set had 65 non vaccine related tweets and 35 vaccine related tweets. The test set had 60 non vaccine and 35 vaccine tweets. An accuracy of 97% for train and 72% on the test as well as a precision score of 0.67, a recall score of 0.65, and a F1 score of .66. The same train and testing are done on the cold start model but achieved a lower train but higher test accuracy with a precision score 0.77, recall score 0.73 and a F1 score of 0.74. The COVID-19 MLP initialized with the influenza MLP was outperformed by the cold start on 4 out of the 5 statistics including a 8% decrease on test accuracy.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Vaccine", |
| "sec_num": "3.4" |
| }, |
| { |
| "text": "Diving further into the data and investigating how the classifiers could be improved, the first step would be to improve the quality of the tweets labels. Several tweets in the related category had mislabeled gold labels. For example, \"@berniesanders thank you so very much @berniesanders for giving us hope as a nation for and end to this joke of a presidency thank you for all the fund raising and support that you have done for our country in the age of covid19 and good sir\" was given the gold label of not related to covid when the model predicted it was. The overall theme was presidential candidate Bernie Sanders, but it was still related to COVID-19. Another example, \"tell congress to put people first, demand paid sick leave for our most vulnerable workers covid-19\". This tweet should be labeled awareness but was given the gold label of infected. This tweet should be labeled as \"Awareness\" but was given the gold label of \"Infected\". With more time to weed out accurate labels and provide higher quality data to the classifier would increase overall accuracy across the board. In future research we aim to use services like Amazon Mechanical Turk to label more data for us. Our self-supervised methodology for classifying COVID-19 tweets with fewer labeled data has been developed to overcome the challenges of labeling massive COVID-19 Tweet data. At the time of this research, labeled COVID-19 datasets for supervised learning were not readily available. That being said, to achieve maximal results, providing the deep learning models with large amounts of high quality annotations for learning would be ideal. Nevertheless, with unlabeled data available and small amount of annotated tweets, our research demonstrates that we can transfer knowledge from unsupervised latent representation and high quality datasets to similar domain classifiers using selfsupervision and few shot learning.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion and Conclusion", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Lastly, our original hypothesis of influenza tweets and COVID-19 tweets being extremely close in context was not entirely true. While COVID-19 and Influenza tweets have a lot of similarities there also exists differences in the themes of the tweets. Twitter users tend to flood their timelines with posts and re-posts of the politics, rumors, misinformation and news related to COVID-19 rather than symptoms and infections.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion and Conclusion", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Looking at the results for each category, leveraging the COVID-19 self-supervised pretext training to produce COVID-19 latent representations from unlabeled data for supervise downstream COVID-19 classifier shows promising results. Although the test accuracy of classifier can still be improved, the warm start of COVID-19 outperforms the cold start model on 3 out of 4 tasks in terms of the expected accuracy , precision , recall and f1. With this experiment we show that using high quality annotated data in a similar domain, can be used with self-supervision and few shot learning to train classifiers on data where labels are limited. Rather than starting from scratch we can initialize the weights on the MLP with the knowledge learned during supervised training. We believe that the architecture and methodology provided in this research show that using self supervision and few-shot learning can overcome some of the challenges of data with limited annotations. Using the proposed model to label tweets can assist future researchers to investigate COVID-19 tweets at a more granular level.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion and Conclusion", |
| "sec_num": "4" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work is partly supported by a grant from the Intelligence Community Centers for Academic Excellence (IC CAE) Program and the Open Cloud Institute at University of Texas at San Antonio (UTSA). The authors gratefully acknowledge the use of the services of Jetstream cloud.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "A large annotated corpus for learning natural language inference", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Samuel", |
| "suffix": "" |
| }, |
| { |
| "first": "Gabor", |
| "middle": [], |
| "last": "Bowman", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [], |
| "last": "Angeli", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher", |
| "middle": [ |
| "D" |
| ], |
| "last": "Potts", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "632--642", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/D15-1075" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large anno- tated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empiri- cal Methods in Natural Language Processing, pages 632-642, Lisbon, Portugal. Association for Compu- tational Linguistics.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "National and local influenza surveillance through twitter: an analysis of the 2012-2013 influenza epidemic", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "David", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Broniatowski", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Michael", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Paul", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Dredze", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "PloS one", |
| "volume": "8", |
| "issue": "12", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David A Broniatowski, Michael J Paul, and Mark Dredze. 2013. National and local influenza surveil- lance through twitter: an analysis of the 2012-2013 influenza epidemic. PloS one, 8(12):e83672.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Text and structural data mining of influenza mentions in web and social media", |
| "authors": [ |
| { |
| "first": "Courtney", |
| "middle": [], |
| "last": "Corley", |
| "suffix": "" |
| }, |
| { |
| "first": "Diane", |
| "middle": [], |
| "last": "Cook", |
| "suffix": "" |
| }, |
| { |
| "first": "Armin", |
| "middle": [], |
| "last": "Mikler", |
| "suffix": "" |
| }, |
| { |
| "first": "Karan", |
| "middle": [], |
| "last": "Singh", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "International journal of environmental research and public health", |
| "volume": "7", |
| "issue": "2", |
| "pages": "596--615", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Courtney Corley, Diane Cook, Armin Mikler, and Karan Singh. 2010. Text and structural data min- ing of influenza mentions in web and social media. International journal of environmental research and public health, 7(2):596-615.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Monitoring influenza trends through mining social media", |
| "authors": [ |
| { |
| "first": "Courtney", |
| "middle": [], |
| "last": "Corley", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Armin", |
| "suffix": "" |
| }, |
| { |
| "first": "Karan", |
| "middle": [ |
| "P" |
| ], |
| "last": "Mikler", |
| "suffix": "" |
| }, |
| { |
| "first": "Diane", |
| "middle": [ |
| "J" |
| ], |
| "last": "Singh", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Cook", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "BIOCOMP", |
| "volume": "", |
| "issue": "", |
| "pages": "340--346", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Courtney Corley, Armin R Mikler, Karan P Singh, and Diane J Cook. 2009. Monitoring influenza trends through mining social media. In BIOCOMP, pages 340-346.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Transfer learning for health-related twitter data", |
| "authors": [ |
| { |
| "first": "Anne", |
| "middle": [], |
| "last": "Dirkson", |
| "suffix": "" |
| }, |
| { |
| "first": "Suzan", |
| "middle": [], |
| "last": "Verberne", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the Fourth Social Media Mining for Health Applications (#SMM4H) Workshop & Shared Task", |
| "volume": "", |
| "issue": "", |
| "pages": "89--92", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/W19-3212" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anne Dirkson and Suzan Verberne. 2019. Transfer learning for health-related twitter data. In Proceed- ings of the Fourth Social Media Mining for Health Applications (#SMM4H) Workshop & Shared Task, pages 89-92, Florence, Italy. Association for Com- putational Linguistics.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Covid-19 mobile positioning data contact tracing and patient privacy regulations: Exploratory search of global response strategies and the use of digital tools in nigeria", |
| "authors": [ |
| { |
| "first": "Iniobong", |
| "middle": [], |
| "last": "Ekong", |
| "suffix": "" |
| }, |
| { |
| "first": "Emeka", |
| "middle": [], |
| "last": "Chukwu", |
| "suffix": "" |
| }, |
| { |
| "first": "Martha", |
| "middle": [], |
| "last": "Chukwu", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "JMIR mHealth and uHealth", |
| "volume": "8", |
| "issue": "4", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Iniobong Ekong, Emeka Chukwu, and Martha Chukwu. 2020. Covid-19 mobile positioning data contact tracing and patient privacy regulations: Exploratory search of global response strategies and the use of digital tools in nigeria. JMIR mHealth and uHealth, 8(4):e19139.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Quantifying sars-cov-2 transmission suggests epidemic control with digital contact tracing", |
| "authors": [ |
| { |
| "first": "Luca", |
| "middle": [], |
| "last": "Ferretti", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Wymant", |
| "suffix": "" |
| }, |
| { |
| "first": "Michelle", |
| "middle": [], |
| "last": "Kendall", |
| "suffix": "" |
| }, |
| { |
| "first": "Lele", |
| "middle": [], |
| "last": "Zhao", |
| "suffix": "" |
| }, |
| { |
| "first": "Anel", |
| "middle": [], |
| "last": "Nurtay", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucie", |
| "middle": [], |
| "last": "Abeler-D\u00f6rner", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Parker", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Bonsall", |
| "suffix": "" |
| }, |
| { |
| "first": "Christophe", |
| "middle": [], |
| "last": "Fraser", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Science", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Luca Ferretti, Chris Wymant, Michelle Kendall, Lele Zhao, Anel Nurtay, Lucie Abeler-D\u00f6rner, Michael Parker, David Bonsall, and Christophe Fraser. 2020. Quantifying sars-cov-2 transmission suggests epi- demic control with digital contact tracing. Science.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "A self-training approach for short text clustering", |
| "authors": [ |
| { |
| "first": "Amir", |
| "middle": [], |
| "last": "Hadifar", |
| "suffix": "" |
| }, |
| { |
| "first": "Lucas", |
| "middle": [], |
| "last": "Sterckx", |
| "suffix": "" |
| }, |
| { |
| "first": "Thomas", |
| "middle": [], |
| "last": "Demeester", |
| "suffix": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Develder", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 4th Workshop on Representation Learning for NLP", |
| "volume": "", |
| "issue": "", |
| "pages": "194--199", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/W19-4322" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Amir Hadifar, Lucas Sterckx, Thomas Demeester, and Chris Develder. 2019. A self-training approach for short text clustering. In Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019), pages 194-199, Florence, Italy. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Universal language model fine-tuning for text classification", |
| "authors": [ |
| { |
| "first": "Jeremy", |
| "middle": [], |
| "last": "Howard", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Ruder", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "328--339", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/P18-1031" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeremy Howard and Sebastian Ruder. 2018. Universal language model fine-tuning for text classification. In Proceedings of the 56th Annual Meeting of the As- sociation for Computational Linguistics (Volume 1: Long Papers), pages 328-339, Melbourne, Australia. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Examining patterns of influenza vaccination in social media", |
| "authors": [ |
| { |
| "first": "Xiaolei", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Michael", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Smith", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Michael", |
| "suffix": "" |
| }, |
| { |
| "first": "Dmytro", |
| "middle": [], |
| "last": "Paul", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ryzhkov", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Sandra", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [ |
| "A" |
| ], |
| "last": "Quinn", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Broniatowski", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Dredze", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Workshops at the Thirty-First AAAI Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Xiaolei Huang, Michael C Smith, Michael J Paul, Dmytro Ryzhkov, Sandra C Quinn, David A Broni- atowski, and Mark Dredze. 2017. Examining pat- terns of influenza vaccination in social media. In Workshops at the Thirty-First AAAI Conference on Artificial Intelligence.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Convolutional neural networks for sentence classification", |
| "authors": [ |
| { |
| "first": "Yoon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proc. of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "1746--1751", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yoon Kim. 2014. Convolutional neural networks for sentence classification. In Proc. of EMNLP, pages 1746-1751.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Separating fact from fear: Tracking flu infections on twitter", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Lamb", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Paul", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Dredze", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "789--795", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Lamb, Michael Paul, and Mark Dredze. 2013. Separating fact from fear: Tracking flu infections on twitter. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Tech- nologies, pages 789-795.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Geospatial event detection by grouping emotion contagion in social media", |
| "authors": [ |
| { |
| "first": "Brandon", |
| "middle": [], |
| "last": "Lwowski", |
| "suffix": "" |
| }, |
| { |
| "first": "Paul", |
| "middle": [], |
| "last": "Rad", |
| "suffix": "" |
| }, |
| { |
| "first": "Kim-Kwang Raymond", |
| "middle": [], |
| "last": "Choo", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "IEEE Transactions on Big Data", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Brandon Lwowski, Paul Rad, and Kim-Kwang Ray- mond Choo. 2018. Geospatial event detection by grouping emotion contagion in social media. IEEE Transactions on Big Data.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Machine learning to detect self-reporting of symptoms, testing access, and recovery associated with covid-19 on twitter: Retrospective big data infoveillance study", |
| "authors": [ |
| { |
| "first": "Tim", |
| "middle": [], |
| "last": "Mackey", |
| "suffix": "" |
| }, |
| { |
| "first": "Vidya", |
| "middle": [], |
| "last": "Purushothaman", |
| "suffix": "" |
| }, |
| { |
| "first": "Jiawei", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "Neal", |
| "middle": [], |
| "last": "Shah", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthew", |
| "middle": [], |
| "last": "Nali", |
| "suffix": "" |
| }, |
| { |
| "first": "Cortni", |
| "middle": [], |
| "last": "Bardier", |
| "suffix": "" |
| }, |
| { |
| "first": "Bryan", |
| "middle": [], |
| "last": "Liang", |
| "suffix": "" |
| }, |
| { |
| "first": "Mingxiang", |
| "middle": [], |
| "last": "Cai", |
| "suffix": "" |
| }, |
| { |
| "first": "Raphael", |
| "middle": [], |
| "last": "Cuomo", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "JMIR Public Health and Surveillance", |
| "volume": "6", |
| "issue": "2", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tim Mackey, Vidya Purushothaman, Jiawei Li, Neal Shah, Matthew Nali, Cortni Bardier, Bryan Liang, Mingxiang Cai, and Raphael Cuomo. 2020. Ma- chine learning to detect self-reporting of symp- toms, testing access, and recovery associated with covid-19 on twitter: Retrospective big data infoveil- lance study. JMIR Public Health and Surveillance, 6(2):e19509.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "An\" infodemic\": Leveraging high-volume twitter data to understand public sentiment for the covid-19 outbreak", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Richard", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Medford", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Sameh", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Saleh", |
| "suffix": "" |
| }, |
| { |
| "first": "Trish", |
| "middle": [ |
| "M" |
| ], |
| "last": "Sumarsono", |
| "suffix": "" |
| }, |
| { |
| "first": "Christoph", |
| "middle": [ |
| "U" |
| ], |
| "last": "Perl", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Lehmann", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Richard J Medford, Sameh N Saleh, Andrew Sumar- sono, Trish M Perl, and Christoph U Lehmann. 2020. An\" infodemic\": Leveraging high-volume twitter data to understand public sentiment for the covid-19 outbreak. medRxiv.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Glove: Global vectors for word representation", |
| "authors": [ |
| { |
| "first": "Jeffrey", |
| "middle": [], |
| "last": "Pennington", |
| "suffix": "" |
| }, |
| { |
| "first": "Richard", |
| "middle": [], |
| "last": "Socher", |
| "suffix": "" |
| }, |
| { |
| "first": "Christopher D", |
| "middle": [], |
| "last": "Manning", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1532--1543", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. Glove: Global vectors for word rep- resentation. In Proceedings of the 2014 conference on empirical methods in natural language process- ing (EMNLP), pages 1532-1543.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Influenza a (h7n9) and the importance of digital epidemiology", |
| "authors": [ |
| { |
| "first": "Marcel", |
| "middle": [], |
| "last": "Salath\u00e9", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Clark", |
| "suffix": "" |
| }, |
| { |
| "first": "Sumiko", |
| "middle": [ |
| "R" |
| ], |
| "last": "Freifeld", |
| "suffix": "" |
| }, |
| { |
| "first": "Anna", |
| "middle": [ |
| "F" |
| ], |
| "last": "Mekaru", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "S" |
| ], |
| "last": "Tomasulo", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Brownstein", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "369", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marcel Salath\u00e9, Clark C Freifeld, Sumiko R Mekaru, Anna F Tomasulo, and John S Brownstein. 2013. In- fluenza a (h7n9) and the importance of digital epi- demiology. The New England journal of medicine, 369(5):401.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Combining search, social media, and traditional data sources to improve influenza surveillance", |
| "authors": [ |
| { |
| "first": "Mauricio", |
| "middle": [], |
| "last": "Santillana", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Andr\u00e9", |
| "suffix": "" |
| }, |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Nguyen", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Dredze", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Michael", |
| "suffix": "" |
| }, |
| { |
| "first": "Elaine", |
| "middle": [ |
| "O" |
| ], |
| "last": "Paul", |
| "suffix": "" |
| }, |
| { |
| "first": "John", |
| "middle": [ |
| "S" |
| ], |
| "last": "Nsoesie", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Brownstein", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "PLoS computational biology", |
| "volume": "11", |
| "issue": "10", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mauricio Santillana, Andr\u00e9 T Nguyen, Mark Dredze, Michael J Paul, Elaine O Nsoesie, and John S Brownstein. 2015. Combining search, social me- dia, and traditional data sources to improve in- fluenza surveillance. PLoS computational biology, 11(10):e1004513.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "The use of twitter to track levels of disease activity and public concern in the us during the influenza a h1n1 pandemic", |
| "authors": [ |
| { |
| "first": "Alessio", |
| "middle": [], |
| "last": "Signorini", |
| "suffix": "" |
| }, |
| { |
| "first": "Alberto", |
| "middle": [ |
| "Maria" |
| ], |
| "last": "Segre", |
| "suffix": "" |
| }, |
| { |
| "first": "Philip", |
| "middle": [ |
| "M" |
| ], |
| "last": "Polgreen", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "PloS one", |
| "volume": "6", |
| "issue": "5", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alessio Signorini, Alberto Maria Segre, and Philip M Polgreen. 2011. The use of twitter to track lev- els of disease activity and public concern in the us during the influenza a h1n1 pandemic. PloS one, 6(5):e19467.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Jetstream: a selfprovisioned, scalable science and engineering cloud environment", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Craig", |
| "suffix": "" |
| }, |
| { |
| "first": "Timothy", |
| "middle": [ |
| "M" |
| ], |
| "last": "Stewart", |
| "suffix": "" |
| }, |
| { |
| "first": "Ian", |
| "middle": [], |
| "last": "Cockerill", |
| "suffix": "" |
| }, |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Foster", |
| "suffix": "" |
| }, |
| { |
| "first": "Nirav", |
| "middle": [], |
| "last": "Hancock", |
| "suffix": "" |
| }, |
| { |
| "first": "Edwin", |
| "middle": [], |
| "last": "Merchant", |
| "suffix": "" |
| }, |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Skidmore", |
| "suffix": "" |
| }, |
| { |
| "first": "James", |
| "middle": [], |
| "last": "Stanzione", |
| "suffix": "" |
| }, |
| { |
| "first": "Steven", |
| "middle": [], |
| "last": "Taylor", |
| "suffix": "" |
| }, |
| { |
| "first": "George", |
| "middle": [], |
| "last": "Tuecke", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Turner", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "Proceedings of the 2015 XSEDE Conference: Scientific Advancements Enabled by Enhanced Cyberinfrastructure", |
| "volume": "", |
| "issue": "", |
| "pages": "1--8", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Craig A Stewart, Timothy M Cockerill, Ian Foster, David Hancock, Nirav Merchant, Edwin Skidmore, Daniel Stanzione, James Taylor, Steven Tuecke, George Turner, et al. 2015. Jetstream: a self- provisioned, scalable science and engineering cloud environment. In Proceedings of the 2015 XSEDE Conference: Scientific Advancements Enabled by Enhanced Cyberinfrastructure, pages 1-8.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "num": null, |
| "text": "The proposed pipeline of self-supervised learning for COVID-19. The COVID-19 latent features are learned through the Autoencoder Convolution network to solve a predefined pretext objective. After self-supervised pretext training finished, the learned latent variables serve as a pre-trained model and are transferred to the downstream tasks by finetuning the Multi-Layer Perceptron(MLP) using few shot learning models after transferring knowledge learned from Influenza labeled dataset. The performance on the downstream objective is used to evaluate the quality of the learned features.", |
| "uris": null |
| }, |
| "TABREF0": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "text": "Example COVID-19 Tweets from the Coronavirus Tweets Dataset", |
| "content": "<table/>" |
| }, |
| "TABREF2": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "text": "", |
| "content": "<table/>" |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "text": "Results for Influenza and COVID-19 Classifiers on the test set. Accuracies reported are in the format of (Train Acc./Test Acc. COVID column is the results from training the MLP from scratch (cold start), while COVID* is the results from training the MLP by initializing the MLP with the influenza classifier weights.", |
| "content": "<table><tr><td colspan=\"2\">Precision, Recall, F1</td><td/></tr><tr><td>Categories</td><td>COVID</td><td>COVID*</td></tr><tr><td>Related</td><td colspan=\"2\">.71/.70/.71 .69/.77/.73</td></tr><tr><td colspan=\"3\">Aware. vs Infect. .69/.68/.68 .73/.72/.73</td></tr><tr><td>Self vs Other</td><td colspan=\"2\">.81/.76/.77 .86/.87/.86</td></tr><tr><td>Vaccine</td><td colspan=\"2\">.77/.73/74 .67/.65/.66</td></tr></table>" |
| } |
| } |
| } |
| } |