| { |
| "paper_id": "2021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T10:41:31.013544Z" |
| }, |
| "title": "Forum 4.0: An Open-Source User Comment Analysis Framework", |
| "authors": [ |
| { |
| "first": "Marlo", |
| "middle": [], |
| "last": "Haering", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Jakob", |
| "middle": [ |
| "Smedegaard" |
| ], |
| "last": "Andersen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Chris", |
| "middle": [], |
| "last": "Biemann", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Wiebke", |
| "middle": [], |
| "last": "Loosen", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "w.loosen@leibniz-hbi.de" |
| }, |
| { |
| "first": "Benjamin", |
| "middle": [], |
| "last": "Milde", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Tim", |
| "middle": [], |
| "last": "Pietz", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Stoecker", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "christian.stoecker@haw-hamburg.de" |
| }, |
| { |
| "first": "Gregor", |
| "middle": [], |
| "last": "Wiedemann", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "g.wiedemann@leibniz-hbi.de" |
| }, |
| { |
| "first": "Olaf", |
| "middle": [], |
| "last": "Zukunft", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "olaf.zukunft@haw-hamburg.de" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Hamburg University of Applied Sciences", |
| "location": {} |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "With the increasing number of user comments in diverse domains, including comments on online journalism and e-commerce websites, the manual content analysis of these comments becomes time-consuming and challenging. However, research showed that user comments contain useful information for different domain experts, which is thus worth finding and utilizing. This paper introduces Forum 4.0, an opensource framework to semi-automatically analyze, aggregate, and visualize user comments based on labels defined by domain experts. We demonstrate the applicability of Forum 4.0 with comments analytics scenarios within the domains of online journalism and app stores. We outline the underlying container architecture, including the web-based user interface, the machine learning component, and the task manager for time-consuming tasks. We finally conduct machine learning experiments with simulated annotations and different sampling strategies on existing datasets from both domains to evaluate Forum 4.0's performance. Forum 4.0 achieves promising classification results (ROC-AUC \u2265 0.9 with 100 annotated samples), utilizing transformer-based embeddings with a lightweight logistic regression model. We explain how Forum 4.0's architecture is applicable for millions of user comments in real-time, yet at feasible training and classification costs.", |
| "pdf_parse": { |
| "paper_id": "2021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "With the increasing number of user comments in diverse domains, including comments on online journalism and e-commerce websites, the manual content analysis of these comments becomes time-consuming and challenging. However, research showed that user comments contain useful information for different domain experts, which is thus worth finding and utilizing. This paper introduces Forum 4.0, an opensource framework to semi-automatically analyze, aggregate, and visualize user comments based on labels defined by domain experts. We demonstrate the applicability of Forum 4.0 with comments analytics scenarios within the domains of online journalism and app stores. We outline the underlying container architecture, including the web-based user interface, the machine learning component, and the task manager for time-consuming tasks. We finally conduct machine learning experiments with simulated annotations and different sampling strategies on existing datasets from both domains to evaluate Forum 4.0's performance. Forum 4.0 achieves promising classification results (ROC-AUC \u2265 0.9 with 100 annotated samples), utilizing transformer-based embeddings with a lightweight logistic regression model. We explain how Forum 4.0's architecture is applicable for millions of user comments in real-time, yet at feasible training and classification costs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Comment sections are omnipresent in today's online environments, for example, on news websites, blogs, online shops, or app stores. In these sections, users submit their feedback and opinion, request features and information, or report issues and bugs. Also, in social media such as Twitter or Facebook, users regularly comment on specific topics, events, products, or services. In many domains, includ-ing e-commerce and journalism, users discuss with each other, read others' opinions to e.g. assess the quality of the service or the product (Springer et al., 2015; K\u00fcmpel and Springer, 2016) , and provide feedback to other users and other domain experts like the journalist , who wrote the article or the developer who created the app (Maalej et al., 2016b) .", |
| "cite_spans": [ |
| { |
| "start": 544, |
| "end": 567, |
| "text": "(Springer et al., 2015;", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 568, |
| "end": 594, |
| "text": "K\u00fcmpel and Springer, 2016)", |
| "ref_id": "BIBREF16" |
| }, |
| { |
| "start": 739, |
| "end": 761, |
| "text": "(Maalej et al., 2016b)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Even though research has criticized phenomena such as \"dark participation\" (Frischlich et al., 2019) , comments can contain constructive information for different domain experts in different fields Maalej et al., 2016a) . For example, in app development, vendors use app reviews in app stores to collect new feature ideas, bug reports, or ideas of additional user scenarios for their app . Software vendors consider the reviews to decide which bug or feature request to prioritize in the next development cycle (Martens and Maalej, 2019) . In online journalism, media outlets harness user comments to acquire a broader perspective on additional arguments, collect resonance about their articles, or identify and contact experts or persons concerned for follow-up stories . However, the quality of the comments varies significantly, and their amount is sometimes overwhelming, which makes manual monitoring and analysis a real challenge (Pagano and Maalej, 2013; Park et al., 2016a) .", |
| "cite_spans": [ |
| { |
| "start": 75, |
| "end": 100, |
| "text": "(Frischlich et al., 2019)", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 198, |
| "end": 219, |
| "text": "Maalej et al., 2016a)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 511, |
| "end": 537, |
| "text": "(Martens and Maalej, 2019)", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 936, |
| "end": 961, |
| "text": "(Pagano and Maalej, 2013;", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 962, |
| "end": 981, |
| "text": "Park et al., 2016a)", |
| "ref_id": "BIBREF24" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this work, we propose Forum 4.0, an opensource user comment analysis framework to semiautomatically analyze a large number of user comments for domain experts from various domains. Forum 4.0 leverages a combination of transfer learning (Howard and Ruder, 2018) , human-in-theloop (Bailey and Chopra, 2018) , and active learning (Settles, 2012) strategies to automatically analyze the comments' content. To enable replication and further research, we share Forum 4.0's source code, the scripts, and datasets we used for our research 1 and a video, which showcases Forum 4.0 2 .", |
| "cite_spans": [ |
| { |
| "start": 239, |
| "end": 263, |
| "text": "(Howard and Ruder, 2018)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 283, |
| "end": 308, |
| "text": "(Bailey and Chopra, 2018)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 331, |
| "end": 346, |
| "text": "(Settles, 2012)", |
| "ref_id": "BIBREF28" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We describe exemplary usage scenarios of Forum 4.0 for journalists and product managers in their respective online journalism and app development domains and introduce Forum 4.0's user interface.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Usage of Forum 4.0", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The manual effort for comment moderation in online journalism is high (Park et al., 2016b) . One the one hand, media outlets filter hate speech (Gao and Huang, 2017) , as it might negatively affect their credibility (Naab et al., 2020) . On the other hand, user comments can also be useful for different journalistic purposes (Diakopoulos, 2015) . For example, journalists can obtain new perspectives and opinions on an article, learn from users' described personal experiences, or identify potential interview partners among the commenting users . Journalists can also aggregate user comments to identify and visualize their audience's opinion on current news topics (Wang et al., 2013) . Users can also point out errors in reporting, contribute additional or missing sources and information, provide new ideas for further news, or even address the editorial team or authors directly, for example, by criticizing the article's quality .", |
| "cite_spans": [ |
| { |
| "start": 70, |
| "end": 90, |
| "text": "(Park et al., 2016b)", |
| "ref_id": "BIBREF25" |
| }, |
| { |
| "start": 144, |
| "end": 165, |
| "text": "(Gao and Huang, 2017)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 216, |
| "end": 235, |
| "text": "(Naab et al., 2020)", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 326, |
| "end": 345, |
| "text": "(Diakopoulos, 2015)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 668, |
| "end": 687, |
| "text": "(Wang et al., 2013)", |
| "ref_id": "BIBREF33" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Online Journalism", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Journalists first define a useful user comment label in Forum 4.0. Examples for such labels could be: \"criticism towards corona measures,\" or \"pros/cons regarding a legislative proposal\". Journalists or forum moderators annotate user comments regarding these labels, gradually increasing the number of training samples. Forum 4.0 trains a machine learning model using the annotated comments and classifies all other user comments. The automatic classification will improve with more annotations until it reaches sufficient precision so that journalists can conduct quantitative and qualitative analyses with the comments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Online Journalism", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "In app stores, product managers utilize user comments for multiple purposes: users report crashes and bugs in app reviews with valuable context information (e.g., device or app version), helping developers identifying and fixing them (Pagano and Maalej, 2013) . This is particularly helpful to acquire immediate feedback after a new major release or update (Guzman and Maalej, 2014) . Additionally, users suggest desired and useful app feature ideas (Maalej et al., 2016a) . Thereby, the product managers get an overview of current app issues, which they can consider for their further development. In the field of mobile learning, the product manager can utilize comments for the automatic evaluation of education apps (Haering et al., 2021) .", |
| "cite_spans": [ |
| { |
| "start": 234, |
| "end": 259, |
| "text": "(Pagano and Maalej, 2013)", |
| "ref_id": "BIBREF22" |
| }, |
| { |
| "start": 357, |
| "end": 382, |
| "text": "(Guzman and Maalej, 2014)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 450, |
| "end": 472, |
| "text": "(Maalej et al., 2016a)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 720, |
| "end": 742, |
| "text": "(Haering et al., 2021)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "App Development", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Similar to the online journalism domain, the product manager can use Forum 4.0 to first create labels for constructive app reviews. In the app development domain, useful labels include \"problems since the last app update\", \"positive/negative feedback on a certain app feature\", or \"missing or requested features\". The domain expert further annotates app reviews, compiling a training set. Forum 4.0 trains a model and classifies the other app reviews for the domain expert to analyze. Figure 1 shows Forum 4.0's user interface. The domain expert can log in to create a new label or annotate user comments. Below the title bar, the expert can select a data source containing the comments to analyze. In the figure, we selected comments from the Austrian newspaper DER STANDARD 3 , which contains the comments of the \"One Million Posts Corpus\" published by Schabus et al. (2017) . Next to the data source selector, the domain expert can create a new label or select relevant existing labels to analyze and annotate the comments.", |
| "cite_spans": [ |
| { |
| "start": 855, |
| "end": 876, |
| "text": "Schabus et al. (2017)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 485, |
| "end": 493, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "App Development", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "The pie chart shows the comment distribution among the document categories (news article or app categories). The bar chart shows the number of positive classifications for the selected labels over time with different granularity options. We train one classification model for each label and show the accuracy and the development of the F1-scores with an increasing number of training samples.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "User Interface", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "The lower part of the Forum 4.0 interface lists the actual user comments for exploration and annotation. With a full-text search, the domain expert can further filter the comment results. The list contains the comment text, the timestamp, and a column for each selected label. Each label column has two sub-columns. The first sub-column with the person symbol shows either existing human annotations when logged out or the own annotations when logged in. A logged-in user can correct the automatic classification or annotate comments as a positive or negative sample for the selected labels. The second sub-column with the robot icon shows binary labels and confidence scores. The domain expert has three sorting options for the classifications: (1) positives first, (2) negatives first, (3) uncertain first (circle with tick mark). Forum 4.0 supports finding positive samples for rare comment labels by suggesting semantically similar user comments. Thereby, Forum 4.0 employs the rapid annotation approach to quickly retrieve additional positive samples for a specific comment label.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "User Interface", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "We describe Forum 4.0's container-based architecture and its machine learning pipelines. The Comment Collector aggregates user comments from various sources, including media sites, app stores, and social media. Forum 4.0 currently contains the \"One Million Posts Corpus\" and imports comments from the Google Play store and the German news site SPIEGEL Online 4 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Architecture", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The client accessing Forum 4.0's web page requests the Reverse Proxy, which forwards the requests depending on the URL path to the responsible container. The first request loads the single page application (Flanagan and Like, 2006) from the Front-End web server, which further communicates via a restful API with the Back-End container.", |
| "cite_spans": [ |
| { |
| "start": 206, |
| "end": 231, |
| "text": "(Flanagan and Like, 2006)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Container-based Architecture", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The containers on the Docker host are only accessible from the outside through the reverse proxy for security. The Back-End provides the restful API. It invokes all machine learning, NLP, and embedding tasks via a task manager in isolated processes as they are time-consuming and would exceed the HTTP request time out. It further calculates the comment embedding index and queries the database. The Embedding Container calculates the embeddings for newly imported user comments. This container can also run on a dedicated host to calculate the embeddings with GPU support.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Container-based Architecture", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "After login, the Back-End issues a JSON web token (J\u00e1noky et al., 2018) for the Front-End. All sensitive API endpoints of the Back-End are protected and require a valid JSON web token in the request's body. Protected actions include the comment and document import, the creation of new labels, and posting annotations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Container-based Architecture", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Two essential parts of the architecture are the Model Training Pipeline (Figure 3a ) and the Comment Import Pipeline (Figure 3b) .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 72, |
| "end": 82, |
| "text": "(Figure 3a", |
| "ref_id": "FIGREF3" |
| }, |
| { |
| "start": 117, |
| "end": 128, |
| "text": "(Figure 3b)", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Machine Learning Pipelines", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The Model Training Pipeline applies supervised machine learning, and active learning strategies (Settles, 2012) to improve the comment classification continuously. To define a label and train a model for the automatic classification, the domain expert must first log in and create a new label. Domain experts can select the new label from the menu and start annotating samples. The domain expert is the human in the loop (Bailey and Chopra, 2018) , who annotates and enlarges the training set to improve the automatic classification iteratively.", |
| "cite_spans": [ |
| { |
| "start": 96, |
| "end": 111, |
| "text": "(Settles, 2012)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 421, |
| "end": 446, |
| "text": "(Bailey and Chopra, 2018)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Pipelines", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "Annotators can sort the user comments according to the uncertainty score to keep the annotation process most rewarding (Andersen et al., 2020). Forum 4.0 uses the label probability as the uncertainty value. Uncertain instances are those whose classifi- Forum 4.0 provides rapid annotation techniques to support and accelerate the collection of training samples. Forum 4.0 lists semantically similar comments to an existing comment based on the similarity of the comment embeddings. In case the annotator found a positive training example, chances are higher that semantically similar user comments are also positive user comments, which the annotator can quickly check.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Pipelines", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "We can adjust the number of required new training samples, which trigger the training of a new model. After each annotation, Forum 4.0 checks whether enough new training samples are available to invoke (re-)training of the model. The task manager executes each model training as a dedicated process, logs its training, and records the evaluation results. Forum 4.0 evaluates each model using tenfold cross-validation (Stone, 1974) to determine the classification performance. The newly trained model classifies all other user comments, which are not part of the training set, and Forum 4.0 persists its classification scores for that label.", |
| "cite_spans": [ |
| { |
| "start": 417, |
| "end": 430, |
| "text": "(Stone, 1974)", |
| "ref_id": "BIBREF32" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Pipelines", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "The Data Import Pipeline enables the import and processing of new user comments. After importing a new user comment batch, the task manager triggers the embedding process, which calculates the embeddings for the imported user comments. Forum 4.0 employs transfer learning (Howard and Ruder, 2018) by using the embeddings of wellestablished pre-trained language models, for example, BERT embeddings (Devlin et al., 2019) , as machine learning features for the classification model. Subsequently, all existing models classify the new user comment batch.", |
| "cite_spans": [ |
| { |
| "start": 272, |
| "end": 296, |
| "text": "(Howard and Ruder, 2018)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 398, |
| "end": 419, |
| "text": "(Devlin et al., 2019)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Pipelines", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To preliminary evaluate the applicability of Forum 4.0 and the performance of its machine learning models, we conducted experiments with comments from news sites and app stores. For the online journalism domain, we used the One Million Post (OMP) corpus (Schabus et al., 2017) . It consists of \u223c1M German user comments submitted to the Austrian newspaper DER STANDARD, partly annotated by forum moderators. For the app store domain, we used an existing annotated app review dataset (ARD) .", |
| "cite_spans": [ |
| { |
| "start": 254, |
| "end": 276, |
| "text": "(Schabus et al., 2017)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We used 9,336 annotated German comments (1,625 positives and 7,711 negatives) regarding OMP's \"personal story\" label. These user comments share the users' personal stories regarding the respective topic, including experiences and anecdotes. We used 6,406 annotated English app reviews (1,437 positives and 4,969 negatives) regarding the ARD's \"bug report\" label. In bug reports, users describe problems with the app that should be fixed, such as a crash, an erroneous behavior, or a performance issue.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We simulated the human annotator, who gradually annotates a batch of user comments, triggering a new training and evaluation cycle. We trained the classifier on the training set and evaluated the model on the remaining comments. We started our first training with ten samples and triggered new training for every ten new annotations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Forum 4.0 allows random sampling and uncertainty sampling for new annotations, which we compared in our experiments. With random sampling, we randomly chose and added ten new samples to our training set. With uncertainty sampling, we added the user comments for which the classifier's output is closest to 0.5. We stopped adding more user comments to the training set as soon as the balanced accuracy score converged.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We evaluated the classification model on the remaining user comments after each training, using the balanced accuracy, F1-score, and the Receiver Operating Characteristics (ROC-AUC) metrics.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Machine Learning Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "For the comment embeddings, we used two different multi-lingual pre-trained language models to embed the comments: (1) BERT (Devlin et al., 2019) is based on a transformer architecture, which learns contextual relations between sub-(word) units in a text. We used an average token embedding of the four last layers of the BERT model as the comment embeddings. (2) Sentence-BERT (S-BERT) (Reimers and Gurevych, 2019) is based on a modification of the BERT network and infers semantically meaningful sentence embeddings. We used a lightweight logistic regression model as a classifier due to performance requirements for quick updates of machine labels during human-in-the-loop coding. To assess the feasibility of our architecture, we further timed the model's training and evaluation. To mitigate the noise of our results, we performed 50 rounds for each experiment. The line plots show the average results of all rounds and the standard deviation. Figure 4 shows the balanced accuracy, ROC-AUC, and F1-scores for all our classification experiments. Overall, all classification metrics improve with increasing training data. Additionally, the uncertainty sampling strategy outperforms random sampling, and the S-BERT embeddings outperform the BERT embeddings given the same sampling strategy. All evaluation metrics significantly improve within the first 100 training samples and converge afterward.", |
| "cite_spans": [ |
| { |
| "start": 124, |
| "end": 145, |
| "text": "(Devlin et al., 2019)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 387, |
| "end": 415, |
| "text": "(Reimers and Gurevych, 2019)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 949, |
| "end": 957, |
| "text": "Figure 4", |
| "ref_id": "FIGREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Machine Learning Experiments", |
| "sec_num": "4" |
| }, |
| { |
| "text": "On the OMP dataset, we achieved a balanced accuracy of 0.86 with 100 training samples using uncertainty sampling and S-BERT embeddings. With 500 training samples, we reached 0.91. Within the first 100 training samples, S-BERT embeddings outperformed the BERT embeddings. We achieved a similar F1-score as Schabus et al. (2017) with \u223c50 training samples (0.70) and outperformed their model using 500 training samples with an F1-score of 0.82. On the app review dataset, we achieved a balanced accuracy of 0.92, a ROC-AUC of 0.96, and an F1-score of 0.85 using 500 training samples. Figure 5 shows the time measurements for training the logistic regression model. In all cases, the training size has a linear increase. Overall, the training time with the S-BERT embeddings (0.1s for 500 samples) takes a shorter time than training with the BERT embeddings (0.4s for 500 samples) on both datasets. We also measured the classification time on the remaining test set, which takes less than \u223c3ms on the OMP (\u223c8,000 test samples) and the ARD (\u223c6,000 test samples) dataset.", |
| "cite_spans": [ |
| { |
| "start": 305, |
| "end": 326, |
| "text": "Schabus et al. (2017)", |
| "ref_id": "BIBREF27" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 581, |
| "end": 589, |
| "text": "Figure 5", |
| "ref_id": "FIGREF5" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experiments Results", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Previous work in the app development domain automatically analyzed comments on apps including, app reviews (Guzman and Maalej, 2014; Dhinakaran et al., 2018; Harman et al., 2012) and tweets (Guzman et al., 2016; Williams and Mahmoud, 2017) , to understand and summarize users' needs and support development decisions . A typical analysis goal is to reduce the noisy user feedback and classify the remaining ones into bug reports, feature requests, and experience reports (Maalej et al., 2016a) . Similarly, in online journalism, previous work aimed to reduce noise and hate speech (Gao and Huang, 2017) , identify high-quality contributions (Park et al., 2016a; Diakopoulos, 2015; Wang and Diakopoulos, 2021) , summarize the audiences' sentiment (Wang et al., 2013) , or identify comments, which address journalistic aspects . Park et al. (2018) and Fast et al. (2016) developed a prototype, which supports the analysis of documents and comments regarding a custom concept based on seed terms. Forum 4.0 builds upon this previous work and features a domain-independent comment analysis framework for domain experts. Domain experts can create or reuse useful labels, annotate user comments regarding these labels, and train machinelearning models, which automatically classify the comments for further utilization.", |
| "cite_spans": [ |
| { |
| "start": 107, |
| "end": 132, |
| "text": "(Guzman and Maalej, 2014;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 133, |
| "end": 157, |
| "text": "Dhinakaran et al., 2018;", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 158, |
| "end": 178, |
| "text": "Harman et al., 2012)", |
| "ref_id": "BIBREF13" |
| }, |
| { |
| "start": 190, |
| "end": 211, |
| "text": "(Guzman et al., 2016;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 212, |
| "end": 239, |
| "text": "Williams and Mahmoud, 2017)", |
| "ref_id": "BIBREF35" |
| }, |
| { |
| "start": 471, |
| "end": 493, |
| "text": "(Maalej et al., 2016a)", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 581, |
| "end": 602, |
| "text": "(Gao and Huang, 2017)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 641, |
| "end": 661, |
| "text": "(Park et al., 2016a;", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 662, |
| "end": 680, |
| "text": "Diakopoulos, 2015;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 681, |
| "end": 708, |
| "text": "Wang and Diakopoulos, 2021)", |
| "ref_id": "BIBREF34" |
| }, |
| { |
| "start": 746, |
| "end": 765, |
| "text": "(Wang et al., 2013)", |
| "ref_id": "BIBREF33" |
| }, |
| { |
| "start": 827, |
| "end": 845, |
| "text": "Park et al. (2018)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 850, |
| "end": 868, |
| "text": "Fast et al. (2016)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We presented Forum 4.0, an open-source framework to semi-automatically analyze user comments in various domains including, online journalism and app store. Domain experts can flexibly define or reuse comment analysis dimensions as classification labels in our framework. Forum 4.0's architecture leverages state-of-the-art semantic text embeddings with a lightweight logistic regression model to address the labeling flexibility and the scalability requirements for an application to millions of user comments. Forum 4.0 starts a new model training after the domain expert annotated additional comments for the concerned label. Forum 4.0 evaluates each new model and classifies the remaining user comments for further analysis.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "We achieved promising results with our machine learning experiments in both domains with different semantic embedding and sampling strategies already after n \u2265 100 annotations with a low training time (t = 0.1s). Our evaluation suggests that Forum 4.0 can also be applied at a larger scale with millions of user comments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "https://forum40.informatik. uni-hamburg.de/git/ 2 https://forum40.informatik. uni-hamburg.de/demo.mp4", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://www.derstandard.at/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work is partly funded by the Hamburg's ahoi.digital program in the Forum 4.0 project.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgments", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Word-Level Uncertainty Estimation for Black-Box Text Classifiers using RNNs", |
| "authors": [ |
| { |
| "first": "Jakob", |
| "middle": [], |
| "last": "Smedegaard Andersen", |
| "suffix": "" |
| }, |
| { |
| "first": "Tom", |
| "middle": [], |
| "last": "Sch\u00f6ner", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 28th International Conference on Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "5541--5546", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/2020.coling-main.484" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jakob Smedegaard Andersen, Tom Sch\u00f6ner, and Walid Maalej. 2020. Word-Level Uncertainty Estimation for Black-Box Text Classifiers using RNNs. In Proceedings of the 28th International Conference on Computational Linguistics, pages 5541-5546, Barcelona, Spain (Online). International Committee on Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Fewshot text classification with pre-trained word embeddings and a human in the loop", |
| "authors": [ |
| { |
| "first": "Katherine", |
| "middle": [], |
| "last": "Bailey", |
| "suffix": "" |
| }, |
| { |
| "first": "Sunny", |
| "middle": [], |
| "last": "Chopra", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "arXiv": [ |
| "arXiv:1804.02063" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Katherine Bailey and Sunny Chopra. 2018. Few- shot text classification with pre-trained word em- beddings and a human in the loop. arXiv preprint arXiv:1804.02063.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Kenton", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Kristina", |
| "middle": [], |
| "last": "Toutanova", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "1", |
| "issue": "", |
| "pages": "4171--4186", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/N19-1423" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171-4186, Minneapolis, MN, USA. Associ- ation for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "App Review Analysis Via Active Learning: Reducing Supervision Effort without Compromising Classification Accuracy", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Venkatesh", |
| "suffix": "" |
| }, |
| { |
| "first": "Raseshwari", |
| "middle": [], |
| "last": "Dhinakaran", |
| "suffix": "" |
| }, |
| { |
| "first": "Nirav", |
| "middle": [], |
| "last": "Pulle", |
| "suffix": "" |
| }, |
| { |
| "first": "Pradeep", |
| "middle": [ |
| "K" |
| ], |
| "last": "Ajmeri", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Murukannaiah", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "2018 IEEE 26th International Requirements Engineering Conference (RE)", |
| "volume": "", |
| "issue": "", |
| "pages": "170--181", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/RE.2018.00026" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Venkatesh T. Dhinakaran, Raseshwari Pulle, Nirav Ajmeri, and Pradeep K. Murukannaiah. 2018. App Review Analysis Via Active Learning: Reducing Su- pervision Effort without Compromising Classifica- tion Accuracy. In 2018 IEEE 26th International Requirements Engineering Conference (RE), pages 170-181, Banff, AB. IEEE.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Picking the NYT picks: Editorial criteria and automation in the curation of online news comments", |
| "authors": [ |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Diakopoulos", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "ISOJ", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nicholas Diakopoulos. 2015. Picking the NYT picks: Editorial criteria and automation in the curation of online news comments. #ISOJ, page 147.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Empath: Understanding topic signals in largescale text", |
| "authors": [ |
| { |
| "first": "Ethan", |
| "middle": [], |
| "last": "Fast", |
| "suffix": "" |
| }, |
| { |
| "first": "Binbin", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "S" |
| ], |
| "last": "Bernstein", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "4647--4657", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ethan Fast, Binbin Chen, and Michael S. Bernstein. 2016. Empath: Understanding topic signals in large- scale text. In Proceedings of the 2016 CHI Con- ference on Human Factors in Computing Systems, pages 4647-4657, San Jose, CA, USA. ACM.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Javascript: The definitive guide", |
| "authors": [ |
| { |
| "first": "David", |
| "middle": [], |
| "last": "Flanagan", |
| "suffix": "" |
| }, |
| { |
| "first": "Will", |
| "middle": [], |
| "last": "Sell Like", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "David Flanagan and Will Sell Like. 2006. Javascript: The definitive guide, 5th.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Comment Sections as Targets of Dark Participation? Journalists' Evaluation and Moderation of Deviant User Comments", |
| "authors": [ |
| { |
| "first": "Lena", |
| "middle": [], |
| "last": "Frischlich", |
| "suffix": "" |
| }, |
| { |
| "first": "Svenja", |
| "middle": [], |
| "last": "Boberg", |
| "suffix": "" |
| }, |
| { |
| "first": "Thorsten", |
| "middle": [], |
| "last": "Quandt", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Journalism Studies", |
| "volume": "20", |
| "issue": "14", |
| "pages": "2014--2033", |
| "other_ids": { |
| "DOI": [ |
| "10.1080/1461670X.2018.1556320" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lena Frischlich, Svenja Boberg, and Thorsten Quandt. 2019. Comment Sections as Targets of Dark Par- ticipation? Journalists' Evaluation and Moderation of Deviant User Comments. Journalism Studies, 20(14):2014-2033.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Detecting online hate speech using context aware models", |
| "authors": [ |
| { |
| "first": "Lei", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| }, |
| { |
| "first": "Ruihong", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the International Conference Recent Advances in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "260--266", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lei Gao and Ruihong Huang. 2017. Detecting on- line hate speech using context aware models. In Proceedings of the International Conference Recent Advances in Natural Language Processing, RANLP 2017, pages 260-266, Varna, Bulgaria. INCOMA Ltd.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "A Needle in a Haystack: What Do Twitter Users Say about Software?", |
| "authors": [ |
| { |
| "first": "Emitza", |
| "middle": [], |
| "last": "Guzman", |
| "suffix": "" |
| }, |
| { |
| "first": "Rana", |
| "middle": [], |
| "last": "Alkadhi", |
| "suffix": "" |
| }, |
| { |
| "first": "Norbert", |
| "middle": [], |
| "last": "Seyff", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "2016 IEEE 24th International Requirements Engineering Conference (RE)", |
| "volume": "", |
| "issue": "", |
| "pages": "96--105", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/RE.2016.67" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emitza Guzman, Rana Alkadhi, and Norbert Seyff. 2016. A Needle in a Haystack: What Do Twitter Users Say about Software? In 2016 IEEE 24th International Requirements Engineering Conference (RE), pages 96-105, Beijing, China.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "How Do Users Like This Feature? A Fine Grained Sentiment Analysis of App Reviews", |
| "authors": [ |
| { |
| "first": "Emitza", |
| "middle": [], |
| "last": "Guzman", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2014, |
| "venue": "IEEE 22nd Int. Requirements Engineering Conf. (RE)", |
| "volume": "", |
| "issue": "", |
| "pages": "153--162", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/RE.2014.6912257" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Emitza Guzman and Walid Maalej. 2014. How Do Users Like This Feature? A Fine Grained Sentiment Analysis of App Reviews. In 2014 IEEE 22nd Int. Requirements Engineering Conf. (RE), pages 153- 162, Karlskrona, Sweden.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Automating the evaluation of education apps with app store data", |
| "authors": [ |
| { |
| "first": "Marlo", |
| "middle": [], |
| "last": "Haering", |
| "suffix": "" |
| }, |
| { |
| "first": "Muneera", |
| "middle": [], |
| "last": "Bano", |
| "suffix": "" |
| }, |
| { |
| "first": "Didar", |
| "middle": [], |
| "last": "Zowghi", |
| "suffix": "" |
| }, |
| { |
| "first": "Matthew", |
| "middle": [], |
| "last": "Kearney", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "IEEE Transactions on Learning Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "1--12", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/TLT.2021.3055121" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marlo Haering, Muneera Bano, Didar Zowghi, Matthew Kearney, and Walid Maalej. 2021. Au- tomating the evaluation of education apps with app store data. IEEE Transactions on Learning Tech- nologies (TLT), pages 1-12.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Who is Addressed in This Comment?: Automatically Classifying Meta-Comments in News Comments", |
| "authors": [ |
| { |
| "first": "Marlo", |
| "middle": [], |
| "last": "H\u00e4ring", |
| "suffix": "" |
| }, |
| { |
| "first": "Wiebke", |
| "middle": [], |
| "last": "Loosen", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proc. ACM Hum.-Comput. Interact", |
| "volume": "2", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/3274336" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Marlo H\u00e4ring, Wiebke Loosen, and Walid Maalej. 2018. Who is Addressed in This Comment?: Au- tomatically Classifying Meta-Comments in News Comments. Proc. ACM Hum.-Comput. Interact., 2(CSCW):67:1-67:20.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "App Store Mining and Analysis: MSR for App Stores", |
| "authors": [ |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Harman", |
| "suffix": "" |
| }, |
| { |
| "first": "Yue", |
| "middle": [], |
| "last": "Jia", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuanyuan", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Proceedings of the 9th IEEE Working Conference on Mining Software Repositories, MSR '12", |
| "volume": "", |
| "issue": "", |
| "pages": "108--111", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mark Harman, Yue Jia, and Yuanyuan Zhang. 2012. App Store Mining and Analysis: MSR for App Stores. In Proceedings of the 9th IEEE Working Conference on Mining Software Repositories, MSR '12, pages 108-111, Piscataway, NJ, USA. IEEE, IEEE Press.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Universal Language Model Fine-tuning for Text Classification", |
| "authors": [ |
| { |
| "first": "Jeremy", |
| "middle": [], |
| "last": "Howard", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Ruder", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
| "volume": "1", |
| "issue": "", |
| "pages": "328--339", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/P18-1031" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jeremy Howard and Sebastian Ruder. 2018. Universal Language Model Fine-tuning for Text Classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 328-339, Melbourne, Aus- tralia. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "An analysis on the revoking mechanisms for", |
| "authors": [ |
| { |
| "first": "J\u00e1nos", |
| "middle": [], |
| "last": "L\u00e1szl\u00f3 Viktor J\u00e1noky", |
| "suffix": "" |
| }, |
| { |
| "first": "P\u00e9ter", |
| "middle": [], |
| "last": "Levendovszky", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ekler", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "JSON Web Tokens. International Journal of Distributed Sensor Networks", |
| "volume": "14", |
| "issue": "9", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "10.1177/1550147718801535" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "L\u00e1szl\u00f3 Viktor J\u00e1noky, J\u00e1nos Levendovszky, and P\u00e9ter Ekler. 2018. An analysis on the revok- ing mechanisms for JSON Web Tokens. Inter- national Journal of Distributed Sensor Networks, 14(9):155014771880153.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Qualit\u00e4t kommentieren. Die Wirkung von Nutzerkommentaren auf die Wahrnehmung journalistischer Qualit\u00e4t", |
| "authors": [ |
| { |
| "first": "Anna", |
| "middle": [ |
| "Sophie" |
| ], |
| "last": "K\u00fcmpel", |
| "suffix": "" |
| }, |
| { |
| "first": "Nina", |
| "middle": [], |
| "last": "Springer", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Studies in Communication -Media", |
| "volume": "5", |
| "issue": "3", |
| "pages": "353--366", |
| "other_ids": { |
| "DOI": [ |
| "10.5771/2192-4007-2016-3-353" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Anna Sophie K\u00fcmpel and Nina Springer. 2016. Qualit\u00e4t kommentieren. Die Wirkung von Nutzerkommentaren auf die Wahrnehmung jour- nalistischer Qualit\u00e4t. Studies in Communication - Media, 5(3):353-366.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Making sense of user comments: Identifying journalists' requirements for a comment analysis framework", |
| "authors": [ |
| { |
| "first": "Wiebke", |
| "middle": [], |
| "last": "Loosen", |
| "suffix": "" |
| }, |
| { |
| "first": "Marlo", |
| "middle": [], |
| "last": "H\u00e4ring", |
| "suffix": "" |
| }, |
| { |
| "first": "Zijad", |
| "middle": [], |
| "last": "Kurtanovi\u0107", |
| "suffix": "" |
| }, |
| { |
| "first": "Lisa", |
| "middle": [], |
| "last": "Merten", |
| "suffix": "" |
| }, |
| { |
| "first": "Julius", |
| "middle": [], |
| "last": "Reimer", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "SCM Studies in Communication and Media", |
| "volume": "6", |
| "issue": "", |
| "pages": "333--364", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wiebke Loosen, Marlo H\u00e4ring, Zijad Kurtanovi\u0107, Lisa Merten, Julius Reimer, Lies van Roessel, and Walid Maalej. 2018. Making sense of user comments: Identifying journalists' requirements for a comment analysis framework. SCM Studies in Communica- tion and Media, 6(4):333-364.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "On the automatic classification of app reviews", |
| "authors": [ |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| }, |
| { |
| "first": "Zijad", |
| "middle": [], |
| "last": "Kurtanovi\u0107", |
| "suffix": "" |
| }, |
| { |
| "first": "Hadeer", |
| "middle": [], |
| "last": "Nabil", |
| "suffix": "" |
| }, |
| { |
| "first": "Christoph", |
| "middle": [], |
| "last": "Stanik", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Requirements Engineering", |
| "volume": "21", |
| "issue": "3", |
| "pages": "311--331", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Walid Maalej, Zijad Kurtanovi\u0107, Hadeer Nabil, and Christoph Stanik. 2016a. On the automatic classi- fication of app reviews. Requirements Engineering, 21(3):311-331.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Toward data-driven requirements engineering", |
| "authors": [ |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| }, |
| { |
| "first": "Maleknaz", |
| "middle": [], |
| "last": "Nayebi", |
| "suffix": "" |
| }, |
| { |
| "first": "Timo", |
| "middle": [], |
| "last": "Johann", |
| "suffix": "" |
| }, |
| { |
| "first": "Guenther", |
| "middle": [], |
| "last": "Ruhe", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "IEEE Software", |
| "volume": "33", |
| "issue": "1", |
| "pages": "48--54", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Walid Maalej, Maleknaz Nayebi, Timo Johann, and Guenther Ruhe. 2016b. Toward data-driven require- ments engineering. IEEE Software, 33(1):48-54.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Release early, release often, and watch your users' emotions: Lessons from emotional patterns", |
| "authors": [ |
| { |
| "first": "Daniel", |
| "middle": [], |
| "last": "Martens", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "IEEE Software", |
| "volume": "36", |
| "issue": "5", |
| "pages": "32--37", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Daniel Martens and Walid Maalej. 2019. Release early, release often, and watch your users' emotions: Lessons from emotional patterns. IEEE Software, 36(5):32-37.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "Comments and Credibility: How Critical User Comments Decrease Perceived News Article Credibility", |
| "authors": [ |
| { |
| "first": "Teresa", |
| "middle": [ |
| "K" |
| ], |
| "last": "Naab", |
| "suffix": "" |
| }, |
| { |
| "first": "Dominique", |
| "middle": [], |
| "last": "Heinbach", |
| "suffix": "" |
| }, |
| { |
| "first": "Marc", |
| "middle": [], |
| "last": "Ziegele", |
| "suffix": "" |
| }, |
| { |
| "first": "Marie-Theres", |
| "middle": [], |
| "last": "Grasberger", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Journalism Studies", |
| "volume": "21", |
| "issue": "6", |
| "pages": "783--801", |
| "other_ids": { |
| "DOI": [ |
| "10.1080/1461670X.2020.1724181" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Teresa K. Naab, Dominique Heinbach, Marc Ziegele, and Marie-Theres Grasberger. 2020. Comments and Credibility: How Critical User Comments De- crease Perceived News Article Credibility. Journal- ism Studies, 21(6):783-801.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "User feedback in the appstore: An empirical study", |
| "authors": [ |
| { |
| "first": "Dennis", |
| "middle": [], |
| "last": "Pagano", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "2013 21st IEEE International Requirements Engineering Conference (RE)", |
| "volume": "", |
| "issue": "", |
| "pages": "125--134", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dennis Pagano and Walid Maalej. 2013. User feedback in the appstore: An empirical study. In 2013 21st IEEE International Requirements Engineering Con- ference (RE), pages 125-134, Rio de Janeiro, Brasil.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "ConceptVector: Text visual analytics via interactive lexicon building using word embedding", |
| "authors": [ |
| { |
| "first": "Deokgun", |
| "middle": [], |
| "last": "Park", |
| "suffix": "" |
| }, |
| { |
| "first": "Seungyeon", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Jurim", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Jaegul", |
| "middle": [], |
| "last": "Choo", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Diakopoulos", |
| "suffix": "" |
| }, |
| { |
| "first": "Niklas", |
| "middle": [], |
| "last": "Elmqvist", |
| "suffix": "" |
| } |
| ], |
| "year": 2018, |
| "venue": "IEEE transactions on visualization and computer graphics", |
| "volume": "24", |
| "issue": "1", |
| "pages": "361--370", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Deokgun Park, Seungyeon Kim, Jurim Lee, Jaegul Choo, Nicholas Diakopoulos, and Niklas Elmqvist. 2018. ConceptVector: Text visual analytics via in- teractive lexicon building using word embedding. IEEE transactions on visualization and computer graphics, 24(1):361-370.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Supporting comment moderators in identifying high quality online news comments", |
| "authors": [ |
| { |
| "first": "Deokgun", |
| "middle": [], |
| "last": "Park", |
| "suffix": "" |
| }, |
| { |
| "first": "Simranjit", |
| "middle": [], |
| "last": "Sachar", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Diakopoulos", |
| "suffix": "" |
| }, |
| { |
| "first": "Niklas", |
| "middle": [], |
| "last": "Elmqvist", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, CHI '16", |
| "volume": "", |
| "issue": "", |
| "pages": "1114--1125", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/2858036.2858389" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Deokgun Park, Simranjit Sachar, Nicholas Diakopou- los, and Niklas Elmqvist. 2016a. Supporting com- ment moderators in identifying high quality online news comments. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Sys- tems, CHI '16, page 1114-1125, New York, NY, USA. Association for Computing Machinery.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "Supporting comment moderators in identifying high quality online news comments", |
| "authors": [ |
| { |
| "first": "Deokgun", |
| "middle": [], |
| "last": "Park", |
| "suffix": "" |
| }, |
| { |
| "first": "Simranjit", |
| "middle": [], |
| "last": "Sachar", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Diakopoulos", |
| "suffix": "" |
| }, |
| { |
| "first": "Niklas", |
| "middle": [], |
| "last": "Elmqvist", |
| "suffix": "" |
| } |
| ], |
| "year": 2016, |
| "venue": "Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "1114--1125", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Deokgun Park, Simranjit Sachar, Nicholas Diakopou- los, and Niklas Elmqvist. 2016b. Supporting com- ment moderators in identifying high quality online news comments. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Sys- tems, pages 1114-1125, San Jose, CA, USA. ACM.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "Sentence-BERT: Sentence embeddings using Siamese BERTnetworks", |
| "authors": [ |
| { |
| "first": "Nils", |
| "middle": [], |
| "last": "Reimers", |
| "suffix": "" |
| }, |
| { |
| "first": "Iryna", |
| "middle": [], |
| "last": "Gurevych", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "3982--3992", |
| "other_ids": { |
| "DOI": [ |
| "10.18653/v1/D19-1410" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nils Reimers and Iryna Gurevych. 2019. Sentence- BERT: Sentence embeddings using Siamese BERT- networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natu- ral Language Processing (EMNLP-IJCNLP), pages 3982-3992, Hong Kong, China. Association for Computational Linguistics.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "One million posts: A data set of German online discussions", |
| "authors": [ |
| { |
| "first": "Dietmar", |
| "middle": [], |
| "last": "Schabus", |
| "suffix": "" |
| }, |
| { |
| "first": "Marcin", |
| "middle": [], |
| "last": "Skowron", |
| "suffix": "" |
| }, |
| { |
| "first": "Martin", |
| "middle": [], |
| "last": "Trapp", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval", |
| "volume": "", |
| "issue": "", |
| "pages": "1241--1244", |
| "other_ids": { |
| "DOI": [ |
| "10.1145/3077136.3080711" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Dietmar Schabus, Marcin Skowron, and Martin Trapp. 2017. One million posts: A data set of German on- line discussions. In Proceedings of the 40th Inter- national ACM SIGIR Conference on Research and Development in Information Retrieval, pages 1241- 1244, Miyazaki, Japan. ACM Press.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Active Learning", |
| "authors": [ |
| { |
| "first": "Burr", |
| "middle": [], |
| "last": "Settles", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Burr Settles. 2012. Active Learning. Morgan & Clay- pool Publishers.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "User comments: Motives and inhibitors to write and read", |
| "authors": [ |
| { |
| "first": "Nina", |
| "middle": [], |
| "last": "Springer", |
| "suffix": "" |
| }, |
| { |
| "first": "Ines", |
| "middle": [], |
| "last": "Engelmann", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Pfaffinger", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "18", |
| "issue": "", |
| "pages": "798--815", |
| "other_ids": { |
| "DOI": [ |
| "10.1080/1369118X.2014.997268" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nina Springer, Ines Engelmann, and Christian Pfaffin- ger. 2015. User comments: Motives and inhibitors to write and read. Information, Communication & Society, 18(7):798-815.", |
| "links": null |
| }, |
| "BIBREF30": { |
| "ref_id": "b30", |
| "title": "Classifying multilingual user feedback using traditional machine learning and deep learning", |
| "authors": [ |
| { |
| "first": "Christoph", |
| "middle": [], |
| "last": "Stanik", |
| "suffix": "" |
| }, |
| { |
| "first": "Marlo", |
| "middle": [], |
| "last": "Haering", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "IEEE 27th International Requirements Engineering Conference Workshops", |
| "volume": "", |
| "issue": "", |
| "pages": "220--226", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christoph Stanik, Marlo Haering, and Walid Maalej. 2019. Classifying multilingual user feedback using traditional machine learning and deep learning. In IEEE 27th International Requirements Engineering Conference Workshops, pages 220-226, Jeju Island, South Korea.", |
| "links": null |
| }, |
| "BIBREF31": { |
| "ref_id": "b31", |
| "title": "Requirements intelligence with OpenReq analytics", |
| "authors": [ |
| { |
| "first": "Christoph", |
| "middle": [], |
| "last": "Stanik", |
| "suffix": "" |
| }, |
| { |
| "first": "Walid", |
| "middle": [], |
| "last": "Maalej", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "2019 IEEE 27th International Requirements Engineering Conference (RE)", |
| "volume": "", |
| "issue": "", |
| "pages": "482--483", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christoph Stanik and Walid Maalej. 2019. Require- ments intelligence with OpenReq analytics. In 2019 IEEE 27th International Requirements Engineering Conference (RE), pages 482-483, Jeju Island, South Korea.", |
| "links": null |
| }, |
| "BIBREF32": { |
| "ref_id": "b32", |
| "title": "Cross-Validatory Choice and Assessment of Statistical Predictions", |
| "authors": [ |
| { |
| "first": "Mervyn", |
| "middle": [], |
| "last": "Stone", |
| "suffix": "" |
| } |
| ], |
| "year": 1974, |
| "venue": "Journal of the Royal Statistical Society: Series B (Methodological)", |
| "volume": "36", |
| "issue": "2", |
| "pages": "111--133", |
| "other_ids": { |
| "DOI": [ |
| "10.1111/j.2517-6161.1974.tb00994.x" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mervyn Stone. 1974. Cross-Validatory Choice and As- sessment of Statistical Predictions. Journal of the Royal Statistical Society: Series B (Methodological), 36(2):111-133.", |
| "links": null |
| }, |
| "BIBREF33": { |
| "ref_id": "b33", |
| "title": "SentiView: Sentiment Analysis and Visualization for Internet Popular Topics", |
| "authors": [ |
| { |
| "first": "Changbo", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Zhao", |
| "middle": [], |
| "last": "Xiao", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuhua", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| }, |
| { |
| "first": "Yanru", |
| "middle": [], |
| "last": "Xu", |
| "suffix": "" |
| }, |
| { |
| "first": "Aoying", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Kang", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "IEEE Transactions on Human-Machine Systems", |
| "volume": "43", |
| "issue": "6", |
| "pages": "620--630", |
| "other_ids": { |
| "DOI": [ |
| "10.1109/THMS.2013.2285047" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Changbo Wang, Zhao Xiao, Yuhua Liu, Yanru Xu, Aoying Zhou, and Kang Zhang. 2013. SentiView: Sentiment Analysis and Visualization for Internet Popular Topics. IEEE Transactions on Human- Machine Systems, 43(6):620-630.", |
| "links": null |
| }, |
| "BIBREF34": { |
| "ref_id": "b34", |
| "title": "The Role of New York Times Picks in Comment Quality and Engagement", |
| "authors": [ |
| { |
| "first": "Yixue", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "Nicholas", |
| "middle": [], |
| "last": "Diakopoulos", |
| "suffix": "" |
| } |
| ], |
| "year": 2021, |
| "venue": "Hawaii International Conference on System Sciences", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": { |
| "DOI": [ |
| "10.24251/HICSS.2021.356" |
| ] |
| }, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yixue Wang and Nicholas Diakopoulos. 2021. The Role of New York Times Picks in Comment Quality and Engagement. In Hawaii International Confer- ence on System Sciences, page to appear, Hawaii.", |
| "links": null |
| }, |
| "BIBREF35": { |
| "ref_id": "b35", |
| "title": "Mining Twitter feeds for software user requirements", |
| "authors": [ |
| { |
| "first": "Grant", |
| "middle": [], |
| "last": "Williams", |
| "suffix": "" |
| }, |
| { |
| "first": "Anas", |
| "middle": [], |
| "last": "Mahmoud", |
| "suffix": "" |
| } |
| ], |
| "year": 2017, |
| "venue": "2017 IEEE 25th International Requirements Engineering Conference (RE)", |
| "volume": "", |
| "issue": "", |
| "pages": "Por-- tugal", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Grant Williams and Anas Mahmoud. 2017. Mining Twitter feeds for software user requirements. In 2017 IEEE 25th International Requirements Engi- neering Conference (RE), pages 1-10, Lisbon, Por- tugal.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Main user interface of Forum 4.0." |
| }, |
| "FIGREF1": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Forum 4.0 is composed of containers, interacting with each other via a restful API.Figure 2outlines a UML deployment diagram." |
| }, |
| "FIGREF2": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Forum 4.0's container architecture." |
| }, |
| "FIGREF3": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Machine Learning Pipelines cation is the least confident, i.e. P (c|d) \u2248 0.5 for comment d belonging to class c." |
| }, |
| "FIGREF4": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Balanced accuracy (top), ROC-AUC (center), and F1-scores (bottom) for all classification experiments on the OMP (left column) and the ARD (right column)." |
| }, |
| "FIGREF5": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "Training time of the logistic regression model." |
| } |
| } |
| } |
| } |