ACL-OCL / Base_JSON /prefixP /json /P15 /P15-1008.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "P15-1008",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T09:08:58.791844Z"
},
"title": "Weakly Supervised Models of Aspect-Sentiment for Online Course Discussion Forums",
"authors": [
{
"first": "Arti",
"middle": [],
"last": "Ramesh",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Maryland",
"location": {
"settlement": "College Park"
}
},
"email": "artir@cs.umd.edu"
},
{
"first": "Shachi",
"middle": [
"H"
],
"last": "Kumar",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of California",
"location": {
"settlement": "Santa Cruz"
}
},
"email": ""
},
{
"first": "James",
"middle": [],
"last": "Foulds",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of California",
"location": {
"settlement": "Santa Cruz"
}
},
"email": "jfoulds@ucsc.edu"
},
{
"first": "Lise",
"middle": [],
"last": "Getoor",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of California",
"location": {
"settlement": "Santa Cruz"
}
},
"email": "getoor@ucsc.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Massive open online courses (MOOCs) are redefining the education system and transcending boundaries posed by traditional courses. With the increase in popularity of online courses, there is a corresponding increase in the need to understand and interpret the communications of the course participants. Identifying topics or aspects of conversation and inferring sentiment in online course forum posts can enable instructor interventions to meet the needs of the students, rapidly address course-related issues, and increase student retention. Labeled aspect-sentiment data for MOOCs are expensive to obtain and may not be transferable between courses, suggesting the need for approaches that do not require labeled data. We develop a weakly supervised joint model for aspectsentiment in online courses, modeling the dependencies between various aspects and sentiment using a recently developed scalable class of statistical relational models called hinge-loss Markov random fields. We validate our models on posts sampled from twelve online courses, each containing an average of 10,000 posts, and demonstrate that jointly modeling aspect with sentiment improves the prediction accuracy for both aspect and sentiment.",
"pdf_parse": {
"paper_id": "P15-1008",
"_pdf_hash": "",
"abstract": [
{
"text": "Massive open online courses (MOOCs) are redefining the education system and transcending boundaries posed by traditional courses. With the increase in popularity of online courses, there is a corresponding increase in the need to understand and interpret the communications of the course participants. Identifying topics or aspects of conversation and inferring sentiment in online course forum posts can enable instructor interventions to meet the needs of the students, rapidly address course-related issues, and increase student retention. Labeled aspect-sentiment data for MOOCs are expensive to obtain and may not be transferable between courses, suggesting the need for approaches that do not require labeled data. We develop a weakly supervised joint model for aspectsentiment in online courses, modeling the dependencies between various aspects and sentiment using a recently developed scalable class of statistical relational models called hinge-loss Markov random fields. We validate our models on posts sampled from twelve online courses, each containing an average of 10,000 posts, and demonstrate that jointly modeling aspect with sentiment improves the prediction accuracy for both aspect and sentiment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Massive Open Online Courses (MOOCs) have emerged as a powerful medium for imparting education to a wide geographical population. Discussion forums are the primary means of communication between MOOC participants (students, TAs, and instructors). Due to the open nature of these courses, they attract people from all over the world leading to large numbers of participants and hence, large numbers of posts in the discussion forums. In the courses we worked with, we found that over the course of the class there were typically over 10,000 posts.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Within this slew of posts, there are valuable problem-reporting posts that identify issues such as broken links, audio-visual glitches, and inaccuracies in the course materials. Automatically identifying these reported problems is important for several reasons: i) it is time-consuming for instructors to manually screen through all of the posts due to the highly skewed instructor-tostudent ratio in MOOCs, ii) promptly addressing issues could help improve student retention, and iii) future iterations of the course could benefit from identifying technical and logistical issues currently faced by students. In this paper, we investigate the problem of determining the fine-grained topics of posts (which we refer to as \"MOOC aspects\") and the sentiment toward them, which can potentially be used to improve the course.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "While aspect-sentiment has been widely studied, the MOOC discussion forum scenario presents a unique set of challenges. Labeled data are expensive to obtain, and posts containing finegrained aspects occur infrequently in courses and differ across courses, thereby making it expensive to get sufficient coverage of all labels. Few distinct aspects occur per course, and only 5-10% of posts in a course are relevant. Hence, getting labels for fine-grained labels involves mining and annotating posts from a large number of courses. Further, creating and sharing labeled data is difficult as data from online courses is governed by IRB regula-tions. Privacy restrictions are another reason why unsupervised/weakly-supervised methods can be helpful. Lastly, to design a system capable of identifying all possible MOOC aspects across courses, we need to develop a system that is not fine-tuned to any particular course, but can adapt seamlessly across courses.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "To this end, we develop a weakly supervised system for detecting aspect and sentiment in MOOC forum posts and validate its effectiveness on posts sampled from twelve MOOC courses. Our system can be applied to any MOOC discussion forum with no or minimal modifications.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Our contributions in this paper are as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "\u2022 We show how to encode weak supervision in the form of seed words to extract extract course-specific features in MOOCs using SeededLDA, a seeded variation of topic modeling (Jagarlamudi et al., 2012) .",
"cite_spans": [
{
"start": 174,
"end": 200,
"text": "(Jagarlamudi et al., 2012)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "\u2022 Building upon our SeededLDA approach, we develop a joint model for aspects and sentiment using the hinge-loss Markov random field (HL-MRF) probabilistic modeling framework. This framework is especially well-suited for this problem because of its ability to combine information from multiple features and jointly reason about aspect and sentiment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "\u2022 To validate the effectiveness of our system, we construct a labeled evaluation dataset by sampling posts from twelve MOOC courses, and annotating these posts with fine-grained MOOC aspects and sentiment via crowdsourcing. The annotation captures finegrained aspects of the course such as content, grading, deadlines, audio and video of lectures and sentiment (i.e., positive, negative, and neutral) toward the aspect in the post.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "\u2022 We demonstrate that the proposed HL-MRF model can predict fine-grained aspects and sentiment and outperforms the model based only on SeededLDA.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "To the best of our knowledge, the problem of predicting aspect and sentiment in MOOC forums has not yet been addressed in the literature. We review prior work in related areas here.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Aspect-Sentiment in Online Reviews It is valuable to identify the sentiment of online reviews towards aspects such as hotel cleanliness and cellphone screen brightness, and sentiment analysis at the aspect-level has been studied extensively in this context (Liu and Zhang, 2012) . Several of these methods use latent Dirichlet allocation topic models (Blei et al., 2003) and variants of it for detecting aspect and sentiment (Lu et al., 2011; Lin and He, 2009) . Liu and Zhang (2012) provide a comprehensive survey of techniques for aspect and sentiment analysis. Here, we discuss works that are closely related to ours. Titov and McDonald (2008) emphasize the importance of an unsupervised approach for aspect detection. However, the authors also indicate that standard LDA (Blei et al., 2003) methods capture global topics and not necessarily pertinent aspects -a challenge that we address in this work. Brody and Elhadad (2010) , Titov and McDonald (2008) , and Jo and Oh (2011) apply variations of LDA at the sentence level for online reviews. We find that around 90% of MOOC posts have only one aspect, which makes sentence-level aspect modeling inappropriate for our domain.",
"cite_spans": [
{
"start": 257,
"end": 278,
"text": "(Liu and Zhang, 2012)",
"ref_id": "BIBREF16"
},
{
"start": 351,
"end": 370,
"text": "(Blei et al., 2003)",
"ref_id": "BIBREF2"
},
{
"start": 425,
"end": 442,
"text": "(Lu et al., 2011;",
"ref_id": "BIBREF18"
},
{
"start": 443,
"end": 460,
"text": "Lin and He, 2009)",
"ref_id": "BIBREF14"
},
{
"start": 463,
"end": 483,
"text": "Liu and Zhang (2012)",
"ref_id": "BIBREF16"
},
{
"start": 621,
"end": 646,
"text": "Titov and McDonald (2008)",
"ref_id": "BIBREF24"
},
{
"start": 775,
"end": 794,
"text": "(Blei et al., 2003)",
"ref_id": "BIBREF2"
},
{
"start": 906,
"end": 930,
"text": "Brody and Elhadad (2010)",
"ref_id": "BIBREF3"
},
{
"start": 933,
"end": 958,
"text": "Titov and McDonald (2008)",
"ref_id": "BIBREF24"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Most previous approaches for sentiment rely on manually constructed lexicons of strongly positive and negative words (Fahrni and Klenner, 2008; Brody and Elhadad, 2010) . These methods are effective in an online review context, however sentiment in MOOC forum posts is often implicit, and not necessarily indicated by standard lexicons. For example, the post \"Where is my certificate? Waiting over a month for it.\" expresses negative sentiment toward the certificate aspect, but does not include any typical negative sentiment words. In our work, we use a data-driven model-based approach to discover domain-specific lexicon information guided by small sets of seed words.",
"cite_spans": [
{
"start": 117,
"end": 143,
"text": "(Fahrni and Klenner, 2008;",
"ref_id": "BIBREF6"
},
{
"start": 144,
"end": 168,
"text": "Brody and Elhadad, 2010)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "There has also been substantial work on joint models for aspect and sentiment (Kim et al., 2013; Diao et al., 2014; Zhao et al., 2010; Lin et al., 2012) , and we adopt such an approach in this paper. Kim et al. (2013) use a hierarchical aspectsentiment model and evaluate it for online reviews. Mukherjee and Liu (2012) use seed words for discovering aspect-based sentiment topics. Drawing on the ideas of Mukherjee and Liu (2012) and Kim et al. (2013) , we propose a statistical relational learning approach that combines the advantages of seed words, aspect hierarchy, and flat aspect-sentiment relationships. It is important to note that a broad majority of the previous work on aspect sentiment focuses on the specific challenges of online review data. As discussed in detail above, MOOC forum data have substantially different properties, and our approach is the first to be designed particularly for this domain.",
"cite_spans": [
{
"start": 78,
"end": 96,
"text": "(Kim et al., 2013;",
"ref_id": "BIBREF12"
},
{
"start": 97,
"end": 115,
"text": "Diao et al., 2014;",
"ref_id": "BIBREF5"
},
{
"start": 116,
"end": 134,
"text": "Zhao et al., 2010;",
"ref_id": "BIBREF25"
},
{
"start": 135,
"end": 152,
"text": "Lin et al., 2012)",
"ref_id": "BIBREF15"
},
{
"start": 200,
"end": 217,
"text": "Kim et al. (2013)",
"ref_id": "BIBREF12"
},
{
"start": 295,
"end": 319,
"text": "Mukherjee and Liu (2012)",
"ref_id": "BIBREF19"
},
{
"start": 406,
"end": 430,
"text": "Mukherjee and Liu (2012)",
"ref_id": "BIBREF19"
},
{
"start": 435,
"end": 452,
"text": "Kim et al. (2013)",
"ref_id": "BIBREF12"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "Learning Analytics In another line of research, there is a growing body of work on the analysis of online courses. Regarding MOOC forum data, Stump et al. (2013) propose a framework for taxonomically categorizing forum posts, leveraging manual annotations. We differ from their approach in that we develop an automatic system to predict MOOC forum categories without using labeled training data. Ramesh et al. (2014b) categorize forum posts into three broad categories in order to predict student engagement. Unlike this method, our system is capable of fine-grained categorization and of identifying aspects in MOOCS. Chaturvedi et al. (2014) focus on predicting instructor intervention using lexicon features and thread features. In contrast, our system is capable of predicting fine MOOC aspects and sentiment of discussion forum posts and thus provides a more informed analysis of MOOC posts.",
"cite_spans": [
{
"start": 142,
"end": 161,
"text": "Stump et al. (2013)",
"ref_id": "BIBREF23"
},
{
"start": 396,
"end": 417,
"text": "Ramesh et al. (2014b)",
"ref_id": "BIBREF22"
},
{
"start": 619,
"end": 643,
"text": "Chaturvedi et al. (2014)",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "2"
},
{
"text": "MOOC participants primarily communicate through discussion forums, consisting of posts, which are short pieces of text. Table 1 provides examples of posts in MOOC forums. Posts 1 and 2 report issues and feedback for the course, while post 3 is a social interaction message. Our goal is to distinguish problem-reporting posts such as 1 and 2 from social posts such as 3, and to identify the issues that are being discussed.",
"cite_spans": [],
"ref_spans": [
{
"start": 120,
"end": 127,
"text": "Table 1",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Problem Setting and Data",
"sec_num": "3"
},
{
"text": "We formalize this task as an aspect-sentiment prediction problem (Liu and Zhang, 2012) . The issues reported in MOOC forums can be related to the different elements of the course such as lectures and quizzes, which are referred to as aspects. The aspects are selected based on MOOC domain expertise and inspiration from Stump et al. (2013) , aiming to cover common concerns that could benefit from intervention. The task is to predict these COARSE-ASPECT FINE-ASPECT Table 2 : Descriptions of coarse and fine aspects.",
"cite_spans": [
{
"start": 65,
"end": 86,
"text": "(Liu and Zhang, 2012)",
"ref_id": "BIBREF16"
},
{
"start": 320,
"end": 339,
"text": "Stump et al. (2013)",
"ref_id": "BIBREF23"
}
],
"ref_spans": [
{
"start": 467,
"end": 474,
"text": "Table 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Problem Setting and Data",
"sec_num": "3"
},
{
"text": "aspects for each post, along with the sentiment polarity toward the aspect, which we code as positive, negative, or neutral. The negative-sentiment posts, along with their aspects, allow us to identify potentially correctable issues in the course.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Problem Setting and Data",
"sec_num": "3"
},
{
"text": "As labels are expensive in this scenario, we formulate the task as a weakly supervised prediction problem. In our work, we assume that a post has at most one fine-grained aspect, as we found that this was true for 90% of the posts in our data. This property is due in part to the brevity of forum posts, which are much shorter documents than those considered in other aspect-sentiment scenarios such as product reviews.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Problem Setting and Data",
"sec_num": "3"
},
{
"text": "While we do not require labeled data, our approaches allow the analyst to instead relatively easily encode a small amount of domain knowledge by seeding the models with a few words relating to each aspect of interest. Hence, we refer to our approach as weakly supervised. Our models can further make use of hierarchical structure between the aspects. The proposed approach is flexible, allowing the aspect seeds and hierarchy to be selected for a given MOOC domain. For the purposes of this study, we represent the MOOC aspects with a two-level hierarchy. We identify a list of nine fine-grained aspects, which are grouped into four coarse topics. The coarse aspects consist of LECTURE, QUIZ, CERTIFICATE, and SOCIAL topics. Table 2 provides a description of each of the aspects and also gives the number of posts in each aspect category after annotation.",
"cite_spans": [],
"ref_spans": [
{
"start": 725,
"end": 732,
"text": "Table 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Aspect Hierarchy",
"sec_num": "3.1"
},
{
"text": "As both LECTURE and QUIZ are key coarselevel aspects in online courses, and more nuanced aspect information for these is important to facilitate instructor interventions, we identify fine-grained aspects for these coarse aspects.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Aspect Hierarchy",
"sec_num": "3.1"
},
{
"text": "For LECTURE we identify LECTURE-CONTENT, LECTURE-VIDEO, LECTURE-AUDIO, LECTURE-SUBTITLES, and LECTURE-LECTURER as fine aspects. For QUIZ, we identify the fine aspects QUIZ-CONTENT, QUIZ-GRADING, QUIZ-DEADLINES, and QUIZ-SUBMISSION. We use the label SOCIAL to refer to social interaction posts that do not mention a problem-related aspect.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Aspect Hierarchy",
"sec_num": "3.1"
},
{
"text": "We construct a dataset by sampling posts from MOOC courses to capture the variety of aspects discussed in online courses. We include courses from different disciplines (business, technology, history, and the sciences) to ensure broad coverage of aspects. Although we adopt an approach that does not require labeled data for training, which is important for most practical MOOC scenarios, in order to validate our methods we obtain labels for the sampled posts using Crowdflower, 1 an online crowd-sourcing annotation platform. Each post was annotated by at least 3 annotators. Crowdflower calculates confidence in labels by computing trust scores for annotators using test questions. Kolhatkar et al. (2013) provide a detailed analysis of Crowdflower trust calculations and the relationship to inter-annotator agreement. We follow their recommendations and retain only labels with confidence > 0.5.",
"cite_spans": [
{
"start": 684,
"end": 707,
"text": "Kolhatkar et al. (2013)",
"ref_id": "BIBREF13"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Dataset",
"sec_num": "3.2"
},
{
"text": "In this section, we develop models and featureextraction techniques to address the challenges of aspect-sentiment prediction for MOOC forums. We present two weakly-supervised methodsfirst, using a seeded topic modeling approach (Jagarlamudi et al., 2012) to identify aspects and sentiment. Second, building upon this method, we then introduce a more powerful statistical relational model which reasons over the seeded LDA predictions as well as sentiment side-information to encode hierarchy information and correlations between sentiment and aspect.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Aspect-Sentiment Prediction Models",
"sec_num": "4"
},
{
"text": "Topic models (Blei et al., 2003) , which identify latent semantic themes from text corpora, have previously been successfully used to discover aspects for sentiment analysis (Diao et al., 2014) . By equating the topics, i.e. discrete distributions over words, with aspects and/or sentiment polarities, topic models can recover aspect-sentiment predictions. In the MOOC context we are specifically interested in problems with the courses, rather than general topics which may be identified by a topic model, such as the topics of the course material.",
"cite_spans": [
{
"start": 13,
"end": 32,
"text": "(Blei et al., 2003)",
"ref_id": "BIBREF2"
},
{
"start": 174,
"end": 193,
"text": "(Diao et al., 2014)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Seeded LDA Model",
"sec_num": "4.1"
},
{
"text": "To guide the topic model to identify aspects of interest, we use SeededLDA (Jagarlamudi et al., 2012) , a variant of LDA which allows an analyst to \"seed\" topics by providing key words that should belong to the topics.",
"cite_spans": [
{
"start": 75,
"end": 101,
"text": "(Jagarlamudi et al., 2012)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Seeded LDA Model",
"sec_num": "4.1"
},
{
"text": "We construct SeededLDA models by providing a set of seed words for each of the coarse and fine aspects in the aspect hierarchy of Table 2 . We also seed topics for positive, negative and neutral sentiment polarities. The seed words for coarse topics are provided in Table 3 , and fine aspects in Table 4. For the sentiment topics (Table 5) , the seed words for the topic positive are positive words often found in online courses such as thank, congratulations, learn, and interest. Similarly, the seed words for the negative topic are negative in the context of online courses, such as difficult, error, issue, problem, and misunderstand.",
"cite_spans": [],
"ref_spans": [
{
"start": 130,
"end": 137,
"text": "Table 2",
"ref_id": null
},
{
"start": 266,
"end": 273,
"text": "Table 3",
"ref_id": null
},
{
"start": 330,
"end": 339,
"text": "(Table 5)",
"ref_id": null
}
],
"eq_spans": [],
"section": "Seeded LDA Model",
"sec_num": "4.1"
},
{
"text": "Additionally, we also use SeededLDA for isolating some common problems in online courses that are associated with sentiment, such as difficulty, availability, correctness, and coursespecific seed words from the syllabus as described in Table 6 . Finally, having inferred the Seed-edLDA model from the data set, for each post p we predict the most likely aspect and the most likely sentiment polarity according to the post's inferred distribution over topics \u03b8 (p) .",
"cite_spans": [
{
"start": 460,
"end": 463,
"text": "(p)",
"ref_id": null
}
],
"ref_spans": [
{
"start": 236,
"end": 243,
"text": "Table 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "Seeded LDA Model",
"sec_num": "4.1"
},
{
"text": "In our experiments, we tokenize and stem the posts using NLTK toolkit (Loper and Bird, 2002) , and use a stop word list tuned to online course discussion forums. The topic model Dirichlet hyperparameters are set to \u03b1 = 0.01, \u03b2 = 0.01 in our experiments. For SeededLDA models corresponding to the seed sets in Tables 3, 4, and 5, the number of topics is equal to the number of seeded topics. For SeededLDA models corresponding to the seed words in Tables 6 and 3, we use 10 topics, allowing for some unseeded topics that are not captured by the seed words.",
"cite_spans": [
{
"start": 70,
"end": 92,
"text": "(Loper and Bird, 2002)",
"ref_id": "BIBREF17"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Seeded LDA Model",
"sec_num": "4.1"
},
{
"text": "The approach described in the previous section automatically identifies user-seeded aspects and sentiment, but it does not make further use of struc-LECTURE: lectur, video, download, volum, low, headphon, sound, audio, transcript, subtitl, slide, note QUIZ: quiz, assignment, question, midterm,exam, submiss, answer, grade, score, grad, midterm, due, deadlin CERTIFICATE: certif, score, signatur, statement, final, course, pass, receiv, coursera, accomplish, fail SOCIAL: name, course, introduction, stud, group, everyon, student Table 3 : Seed words for coarse aspects LECTURE-VIDEO: video, problem, download, play, player, watch, speed, length, long, fast, slow, render, qualiti LECTURE-AUDIO: volum, low, headphon, sound, audio, hear, maximum, troubl, qualiti, high, loud, heard LECTURE-LECTURER: professor, fast, speak, pace, follow, speed, slow, accent, absorb, quick, slowli LECTURE-SUBTITLES: transcript, subtitl, slide, note, lectur, difficult, pdf LECTURE-CONTENT: typo, error, mistak, wrong, right, incorrect, mistaken QUIZ-CONTENT: question, challeng, difficult, understand, typo, error, mistak, quiz, assignment QUIZ-SUBMISSION: submiss, submit, quiz, error, unabl, resubmit QUIZ-GRADING: answer, question, answer, grade, assignment, quiz, respons ,mark, wrong, score QUIZ-DEADLINE: due, deadlin, miss, extend, late Table 4 : Seed words for fine aspects POSITIVE: interest, excit, thank, great, happi, glad, enjoy, forward, insight, opportun, clear, fantast, fascin, learn, hope, congratul NEGATIVE: problem, difficult, error, issu, unabl, misunderstand, terribl, bother, hate, bad, wrong, mistak, fear, troubl NEUTRAL: coursera, class, hello, everyon, greet, nam, meet, group, studi, request, join, introduct, question, thank Table 5 : Seed words for sentiment DIFFICULTY: difficult, understand, ambigu, disappoint, hard, follow, mislead, difficulti, challeng, clear CONTENT: typo, error, mistak, wrong, right, incorrect, mistaken, score AVAILABILITY: avail, nowher, find, access, miss, view, download, broken, link, bad, access, deni, miss, permiss COURSE-1: develop, eclips, sdk, softwar, hardware, accuser, html, platform, environ, lab, ide, java, COURSE-2: protein, food, gene, vitamin, evolut, sequenc, chromosom, genet, speci, peopl, popul, evolv, mutat, ancestri COURSE-3: compani, product, industri, strategi, decision, disrupt, technolog, market Table 6 : Seed words for sentiment specific to online courses ture or dependencies between these values, or any additional side-information. To address this, we propose a more powerful approach using hingeloss Markov random fields (HL-MRFs), a scalable class of continuous, conditional graphical models (Bach et al., 2013) . HL-MRFs have achieved state-of-the-art performance in many domains including knowledge graph identification (Pujara et al., 2013) , understanding engagements in MOOCs (Ramesh et al., 2014a) , biomedicine and multirelational link prediction (Fakhraei et al., 2014) , and modelling social trust . These models can be specified using Probabilistic Soft Logic (PSL) (Bach et al., 2015) , a weighted first order logical templating language. An example of a PSL rule is",
"cite_spans": [
{
"start": 2671,
"end": 2690,
"text": "(Bach et al., 2013)",
"ref_id": "BIBREF0"
},
{
"start": 2801,
"end": 2822,
"text": "(Pujara et al., 2013)",
"ref_id": "BIBREF20"
},
{
"start": 2860,
"end": 2882,
"text": "(Ramesh et al., 2014a)",
"ref_id": "BIBREF21"
},
{
"start": 2933,
"end": 2956,
"text": "(Fakhraei et al., 2014)",
"ref_id": "BIBREF7"
},
{
"start": 3055,
"end": 3074,
"text": "(Bach et al., 2015)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [
{
"start": 530,
"end": 537,
"text": "Table 3",
"ref_id": null
},
{
"start": 1328,
"end": 1335,
"text": "Table 4",
"ref_id": null
},
{
"start": 1739,
"end": 1746,
"text": "Table 5",
"ref_id": null
},
{
"start": 2368,
"end": 2375,
"text": "Table 6",
"ref_id": null
}
],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "\u03bb : P (a) \u2227 Q(a, b) \u2192 R(b),",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "where P, Q, and R are predicates, a and b are variables, and \u03bb is the weight associated with the rule. The weight of the rule indicates its importance in the HL-MRF probabilistic model, which defines a probability density function of the form",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "P (Y|X) \u221d exp \u2212 M r=1 \u03bb r \u03c6 r (Y, X) \u03c6 r (Y, X) = (max{l r (Y, X), 0}) \u03c1r ,",
"eq_num": "(1)"
}
],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "where \u03c6 r (Y, X) is a hinge-loss potential corresponding to an instantiation of a rule, and is specified by a linear function l r and optional exponent \u03c1 r \u2208 {1, 2}. For example, in our MOOC aspectsentiment model, if P and F denote post P and fine aspect F, then we have predicates SEEDLDA-FINE(P, F) to denote the value corresponding to topic F in SeededLDA, and FINE-ASPECT(P, F) is the target variable denoting the fine aspect of the post P. A PSL rule to encode that the SeededLDA topic F suggests that aspect F is present is \u03bb : SEEDLDA-FINE(P, F ) \u2192 FINE-ASPECT(P, F ).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "We can generate more complex rules connecting the different features and target variables, e.g.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "\u03bb : SEEDLDA-FINE(P, F ) \u2227 SENTIMENT(P, S)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "\u2192 FINE-ASPECT(P, F ).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "This rule encodes a dependency between SENTI-MENT and FINE-ASPECT, namely that the Seed-edLDA topic and a strong sentiment score increase the probability of the fine aspect. The HL-MRF model uses these rules to encode domain knowledge about dependencies among the predicates. The continuous value representation further helps in understanding the confidence of predictions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Hinge-loss Markov Random Fields",
"sec_num": "4.2"
},
{
"text": "Probabilistic Soft Logic (PSL-Joint)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Joint Aspect-Sentiment Prediction using",
"sec_num": "4.3"
},
{
"text": "In this section, we describe our joint approach to predicting aspect and sentiment in online discussion forums, leveraging the strong dependence between aspect and sentiment. We present a system designed using HL-MRFs which combines different features, accounting for their respective uncertainty, and encodes the dependencies between aspect and sentiment in the MOOC context. Table 7 provides some representative rules from our model. 2 The rules can be classified into two broad categories-1) rules that combine multiple features, and 2) rules that encode the dependencies between aspect and sentiment.",
"cite_spans": [],
"ref_spans": [
{
"start": 377,
"end": 384,
"text": "Table 7",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Joint Aspect-Sentiment Prediction using",
"sec_num": "4.3"
},
{
"text": "The first set of rules in Table 7 combine different features extracted from the post. SEEDLDA-FINE, SEEDLDA-COARSE and SEEDLDA-SENTIMENT-COURSE predicates in rules refer to SeededLDA posterior distributions using coarse, fine, and course-specific sentiment seed words respectively. The strength of our model comes from its ability to encode different combinations of features and weight them according to their importance. The first rule in Table 7 combines the SeededLDA features from both SEEDLDA-FINE and SEEDLDA-COARSE to predict the fine aspect. Interpreting the rule, the fine aspect of the post is more likely to be LECTURE-LECTURER if the coarse Seed-edLDA score for the post is LECTURE, and the fine SeededLDA score for the post is LECTURE-LECTURER. Similarly, the second rule provides combinations of some of the other features used by the model-two different SeededLDA scores for sentiment, as indicated by seed words in Tables 5 and 6. The third rule states that certain fine aspects occur together with certain values of sentiment more than others. In online courses, posts that discuss grading usually talk about grievances and issues. The rule captures that QUIZ-GRADING occurs with negative sentiment in most cases.",
"cite_spans": [],
"ref_spans": [
{
"start": 26,
"end": 33,
"text": "Table 7",
"ref_id": "TABREF3"
},
{
"start": 441,
"end": 448,
"text": "Table 7",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Combining Features",
"sec_num": "4.3.1"
},
{
"text": "Full model available at https://github.com/artir/ramesh-acl15",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "2",
"sec_num": null
},
{
"text": "In addition to combining features, we also encode rules to capture the taxonomic dependence between coarse and fine aspects, and the dependence between aspect and sentiment ( Table 7 , bottom). Rules 4 and 5 encode pair-wise dependency between FINE-ASPECT and SENTIMENT, and COARSE-ASPECT and FINE-ASPECT respectively. Rule 4 uses the SeededLDA value for QUIZ-DEADLINES to predict both SENTIMENT, and FINE-ASPECT jointly. This together with other rules for predicting SENTIMENT and FINE-ASPECT individually creates a constrained satisfaction problem, forcing aspect and sentiment to agree with each other. Rule 5 is similar to rule 4, capturing the taxonomic relationship between target variables COARSE-ASPECT and FINE-ASPECT.",
"cite_spans": [],
"ref_spans": [
{
"start": 175,
"end": 182,
"text": "Table 7",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "Encoding Dependencies Between Aspect and Sentiment",
"sec_num": "4.3.2"
},
{
"text": "Thus, by using conjunctions to combine features and appropriately weighting these rules, we account for the uncertainties in the underlying features and make them more robust. The combination of these two different types of weighted rules, referred to below as PSL-Joint, is able to reason collectively about aspect and sentiment.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Encoding Dependencies Between Aspect and Sentiment",
"sec_num": "4.3.2"
},
{
"text": "In this section, we present the quantitative and qualitative results of our models on the annotated MOOC dataset. Our models do not require labeled data for training; we use the label annotations only for evaluation. Tables 8 -11 show the results for the SeededLDA and PSL-Joint models. Statistically significant differences, evaluated using a paired t-test with a rejection threshold of 0.01, are typed in bold.",
"cite_spans": [],
"ref_spans": [
{
"start": 217,
"end": 229,
"text": "Tables 8 -11",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Empirical Evaluation",
"sec_num": "5"
},
{
"text": "For SeededLDA, we use the seed words for coarse, fine, and sentiment given in Tables 3 -5 . After training the model, we use the SeededLDA multinomial posterior distribution to predict the target variables. We use the maximum value in the posterior for the distribution over topics for each post to obtain predictions for coarse aspect, fine aspect, and sentiment. We then calculate precision, recall and F1 values comparing with our ground truth labels.",
"cite_spans": [],
"ref_spans": [
{
"start": 78,
"end": 89,
"text": "Tables 3 -5",
"ref_id": null
}
],
"eq_spans": [],
"section": "SeededLDA for Aspect-Sentiment",
"sec_num": "5.1"
},
{
"text": "SEEDLDA-FINE(POST, LECTURE-LECTURER) \u2227 SEEDLDA-COARSE(POST, LECTURE) \u2192 FINE-ASPECT(POST, LECTURE-LECTURER) SEEDLDA-SENTIMENT-COURSE(POST, NEGATIVE) \u2227 SEEDLDA-SENTIMENT(POST, NEGATIVE) \u2192 SENTIMENT(POST, NEGATIVE) SEEDLDA-SENTIMENT-COURSE(POST, NEGATIVE) \u2227 SEEDLDA-FINE(POST, QUIZ-GRADING) \u2192 FINE-ASPECT(POST, QUIZ-GRADING) Encoding dependencies between aspect and sentiment SEEDLDA-FINE(POST, QUIZ-DEADLINES) \u2227 SENTIMENT(POST, NEGATIVE) \u2192 FINE-ASPECT(POST, QUIZ-DEADLINES) SEEDLDA-FINE(POST, QUIZ-SUBMISSION) \u2227 FINE-ASPECT(POST, QUIZ-SUBMISSION) \u2192 COARSE-ASPECT(POST, QUIZ) Tables 8 and 9 give the results for the fine aspects under LECTURE and QUIZ. PSL-JOINT performs better than SEEDEDLDA in most cases, without suffering any statistically significant losses. Notable cases include the increase in scores for LECTURE-LECTURER, LECTURE-SUBTITLES, LECTURE-CONTENT, QUIZ-CONTENT, QUIZ-GRADING, and QUIZ-DEADLINES, for which the scores increase by a large margin over Seed-edLDA. We observe that for LECTURE-CONTENT and QUIZ-CONTENT, the increase in scores is more significant than others with SeededLDA performing very poorly. Since both lecture and quiz content have the same kind of words related to the course material, SeededLDA is not able to distinguish between these two aspects. We found that in 63% of these missed predictions, Seed-edLDA predicts LECTURE-CONTENT, instead of QUIZ-CONTENT, and vice versa. In contrast, PSL-Joint uses both coarse and fine SeededLDA scores and captures the dependency between a coarse aspect and its corresponding fine aspect. Therefore, PSL-Joint is able to distinguish between LECTURE-CONTENT and QUIZ-CONTENT. In the next section, we present some examples of posts that SEEDEDLDA misclassified but were predicted correctly by PSL-Joint. Table 10 presents results for the coarse-aspects. We observe that PSL-Joint performs better than SeededLDA for all classes. In particular for CER-TIFICATE and QUIZ, PSL-Joint exhibits a marked increase in scores when compared to SeededLDA. This is also true for sentiment, for which the scores for NEUTRAL and NEGATIVE sentiment show significant improvement (Table 11) . ",
"cite_spans": [
{
"start": 308,
"end": 321,
"text": "QUIZ-GRADING)",
"ref_id": null
}
],
"ref_spans": [
{
"start": 573,
"end": 587,
"text": "Tables 8 and 9",
"ref_id": "TABREF4"
},
{
"start": 1780,
"end": 1788,
"text": "Table 10",
"ref_id": "TABREF1"
},
{
"start": 2138,
"end": 2148,
"text": "(Table 11)",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "Rules combining features",
"sec_num": null
},
{
"text": "As the second lecture video told me I started windows telnet and connected to the virtual device. Then I typed the same command for sending an sms that the lecture video told me to. The phone received a message all right and I was able to open it but the message itself seems to be written with some strange characters. Table 12 presents some examples of posts that PSL-Joint predicted correctly, and which Seed-edLDA misclassified. The first two examples illustrate that PSL can predict the subtle difference between LECTURE-CONTENT and QUIZ-CONTENT. Particularly notable is the third example, which contains mention of both subtitles and audio, but the negative sentiment is associated with audio rather than subtitles. PSL-Joint predicts the fine aspect as LECTURE-AUDIO, even though the underlying SeededLDA feature has a high score for LECTURE-SUBTITLES. This example illustrates the strength of the joint reasoning approach in PSL-Joint. Finally, in the last example, the post mentions starting a group to discuss videos. This is an ambiguous post containing the keyword video, while it is in reality a social post about starting a group. PSL-Joint is able to predict this because it uses both the sentiment scores associated with the post and the SeededLDA scores for fine aspect, and infers that social posts are generally positive. So, combining the feature values for social aspect and positive sentiment, it is able to predict the fine aspect as SOCIAL correctly. The continuous valued output predictions produced by PSL-Joint allow us to rank the predicted variables by output prediction value. Analyzing the predictions for posts that PSL-Joint misclassified, we observe that for four out of nine fine aspects, more than 70% of the time the correct label is in the top three predictions. And, for all fine aspects, the correct label is found in the top 3 predictions around 40% of the time. Thus, using the top three predictions made by PSL-Joint, we can understand the fine aspect of the post to a great extent. Table 13 gives some examples of posts for which the second best prediction by PSL-Joint is the correct label. For these examples, we found that PSL-Joint misses the correct prediction by a small margin(< 0.2). Since our evaluation scheme only considers the maximum value to determine the scores, these examples were treated as misclassified.",
"cite_spans": [],
"ref_spans": [
{
"start": 320,
"end": 328,
"text": "Table 12",
"ref_id": "TABREF1"
},
{
"start": 2026,
"end": 2034,
"text": "Table 13",
"ref_id": "TABREF1"
}
],
"eq_spans": [],
"section": "LECTURE-CONTENT QUIZ-CONTENT LECTURE-CONTENT",
"sec_num": null
},
{
"text": "In our 3275 annotated posts, the instructor replied to 787 posts. Of these, 699 posts contain a mention of some MOOC aspect. PSL-Joint predicts 97.8% from those as having an aspect and 46.9% as the correct aspect. This indicates that PSL-Joint is capable of identifying the most important posts, i.e. those that the instructor replied to, with high accuracy. PSL-Joint's MOOC aspect predictions can potentially be used by the instructor to select a subset of posts to address in order to cover the main reported issues. We found in our data that some fine aspects, such as CERTIFICATE, have a higher percentage of instructor replies than others, such as QUIZ-GRADING. Using our system, instructors can sample from multiple aspect cate-gories, thereby making sure that all categories of problems receive attention.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Understanding Instructor Intervention using PSL-Joint Predictions",
"sec_num": "5.4"
},
{
"text": "In this paper, we developed a weakly supervised joint probabilistic model (PSL-Joint) for predicting aspect-sentiment in online courses. Our model provides the ability to conveniently encode domain information in the form of seed words, and weighted logical rules capturing the dependencies between aspects and sentiment. We validated our approach on an annotated dataset of MOOC posts sampled from twelve courses. We compared our PSL-Joint probabilistic model to a simpler SeededLDA approach, and demonstrated that PSL-Joint produced statistically significantly better results, exhibiting a 3-5 times improvement in F1 score in most cases over a system using only SeededLDA. As further shown by our qualitative results and instructor reply information, our system can potentially be used for understanding student requirements and issues, identifying posts for instructor intervention, increasing student retention, and improving future iterations of the course.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "http://www.crowdflower.com/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "Acknowledgements This work was supported by NSF grant IIS1218488, and IARPA via DoI/NBC contract number D12PC00337. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright annotation thereon. Disclaimer: The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of IARPA, DoI/NBC, or the U.S. Government.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "acknowledgement",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Hinge-loss Markov random fields: Convex inference for structured prediction",
"authors": [
{
"first": "H",
"middle": [],
"last": "Stephen",
"suffix": ""
},
{
"first": "Bert",
"middle": [],
"last": "Bach",
"suffix": ""
},
{
"first": "Ben",
"middle": [],
"last": "Huang",
"suffix": ""
},
{
"first": "Lise",
"middle": [],
"last": "London",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Getoor",
"suffix": ""
}
],
"year": 2013,
"venue": "Proceedings of the Conference on Uncertainty in Artificial Intelligence (UAI)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Stephen H. Bach, Bert Huang, Ben London, and Lise Getoor. 2013. Hinge-loss Markov random fields: Convex inference for structured prediction. In Pro- ceedings of the Conference on Uncertainty in Artifi- cial Intelligence (UAI).",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Hinge-loss Markov random fields and probabilistic soft logic",
"authors": [
{
"first": "S",
"middle": [
"H"
],
"last": "Bach",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Broecheler",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Huang",
"suffix": ""
},
{
"first": "L",
"middle": [],
"last": "Getoor",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {
"arXiv": [
"arXiv:1505.04406"
]
},
"num": null,
"urls": [],
"raw_text": "S. H. Bach, M. Broecheler, B. Huang, and L. Getoor. 2015. Hinge-loss Markov random fields and proba- bilistic soft logic. arXiv:1505.04406 [cs.LG].",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Latent Dirichlet allocation",
"authors": [
{
"first": "David",
"middle": [
"M"
],
"last": "Blei",
"suffix": ""
},
{
"first": "Andrew",
"middle": [
"Y"
],
"last": "Ng",
"suffix": ""
},
{
"first": "Michael",
"middle": [
"I"
],
"last": "Jordan",
"suffix": ""
}
],
"year": 2003,
"venue": "Journal of Machine Learning Research",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent Dirichlet allocation. Journal of Ma- chine Learning Research (JMLR).",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "An unsupervised aspect-sentiment model for online reviews",
"authors": [
{
"first": "Samuel",
"middle": [],
"last": "Brody",
"suffix": ""
},
{
"first": "Noemie",
"middle": [],
"last": "Elhadad",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of Human Language Technologies: Conference of the North American Chapter of the Association for Computational Linguistics (HLT)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Samuel Brody and Noemie Elhadad. 2010. An unsu- pervised aspect-sentiment model for online reviews. In Proceedings of Human Language Technologies: Conference of the North American Chapter of the Association for Computational Linguistics (HLT).",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Predicting instructor's intervention in mooc forums",
"authors": [
{
"first": "Snigdha",
"middle": [],
"last": "Chaturvedi",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Goldwasser",
"suffix": ""
},
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Snigdha Chaturvedi, Dan Goldwasser, and Hal Daum\u00e9 III. 2014. Predicting instructor's interven- tion in mooc forums. In Proceedings of the Annual Meeting of the Association for Computational Lin- guistics (ACL).",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Jointly modeling aspects, ratings and sentiments for movie recommendation (JMARS)",
"authors": [
{
"first": "Qiming",
"middle": [],
"last": "Diao",
"suffix": ""
},
{
"first": "Minghui",
"middle": [],
"last": "Qiu",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Chao-Yuan",
"suffix": ""
},
{
"first": "Alexander",
"middle": [
"J"
],
"last": "Wu",
"suffix": ""
},
{
"first": "Jing",
"middle": [],
"last": "Smola",
"suffix": ""
},
{
"first": "Chong",
"middle": [],
"last": "Jiang",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Wang",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Qiming Diao, Minghui Qiu, Chao-Yuan Wu, Alexan- der J. Smola, Jing Jiang, and Chong Wang. 2014. Jointly modeling aspects, ratings and sentiments for movie recommendation (JMARS). In Proceedings of the SIGKDD International Conference on Knowl- edge Discovery and Data Mining (KDD).",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Old wine or warm beer: Target-specific sentiment analysis of adjectives",
"authors": [
{
"first": "Angela",
"middle": [],
"last": "Fahrni",
"suffix": ""
},
{
"first": "Manfred",
"middle": [],
"last": "Klenner",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of the Symposium on Affective Language in Human and Machine (AISB)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Angela Fahrni and Manfred Klenner. 2008. Old wine or warm beer: Target-specific sentiment analysis of adjectives. In Proceedings of the Symposium on Af- fective Language in Human and Machine (AISB).",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Network-based drug-target interaction prediction with probabilistic soft logic",
"authors": [
{
"first": "Shobeir",
"middle": [],
"last": "Fakhraei",
"suffix": ""
},
{
"first": "Bert",
"middle": [],
"last": "Huang",
"suffix": ""
},
{
"first": "Louiqa",
"middle": [],
"last": "Raschid",
"suffix": ""
},
{
"first": "Lise",
"middle": [],
"last": "Getoor",
"suffix": ""
}
],
"year": 2014,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Shobeir Fakhraei, Bert Huang, Louiqa Raschid, and Lise Getoor. 2014. Network-based drug-target interaction prediction with probabilistic soft logic.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "A flexible framework for probabilistic models of social trust",
"authors": [
{
"first": "Bert",
"middle": [],
"last": "Huang",
"suffix": ""
},
{
"first": "Angelika",
"middle": [],
"last": "Kimmig",
"suffix": ""
},
{
"first": "Lise",
"middle": [],
"last": "Getoor",
"suffix": ""
},
{
"first": "Jennifer",
"middle": [],
"last": "Golbeck",
"suffix": ""
}
],
"year": 2013,
"venue": "Proceedings of the International Conference on Social Computing, Behavioral-Cultural Modeling, & Prediction (SBP)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bert Huang, Angelika Kimmig, Lise Getoor, and Jen- nifer Golbeck. 2013. A flexible framework for probabilistic models of social trust. In Proceed- ings of the International Conference on Social Com- puting, Behavioral-Cultural Modeling, & Prediction (SBP).",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Incorporating lexical priors into topic models",
"authors": [
{
"first": "Jagadeesh",
"middle": [],
"last": "Jagarlamudi",
"suffix": ""
},
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
},
{
"first": "Raghavendra",
"middle": [],
"last": "Udupa",
"suffix": ""
}
],
"year": 2012,
"venue": "Proceedings of the European Chapter of the Association for Computational Linguistics (EACL)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jagadeesh Jagarlamudi, Hal Daum\u00e9, III, and Raghavendra Udupa. 2012. Incorporating lex- ical priors into topic models. In Proceedings of the European Chapter of the Association for Computational Linguistics (EACL).",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Aspect and sentiment unification model for online review analysis",
"authors": [
{
"first": "Y",
"middle": [],
"last": "Jo",
"suffix": ""
},
{
"first": "A",
"middle": [
"H"
],
"last": "Oh",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the International Conference on Web Search and Data Mining (WSDM)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Y. Jo and A.H. Oh. 2011. Aspect and sentiment unifi- cation model for online review analysis. In Proceed- ings of the International Conference on Web Search and Data Mining (WSDM).",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "A hierarchical aspect-sentiment model for online reviews",
"authors": [
{
"first": "Suin",
"middle": [],
"last": "Kim",
"suffix": ""
},
{
"first": "Jianwen",
"middle": [],
"last": "Zhang",
"suffix": ""
},
{
"first": "Zheng",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Alice",
"middle": [],
"last": "Oh",
"suffix": ""
},
{
"first": "Shixia",
"middle": [],
"last": "Liu",
"suffix": ""
}
],
"year": 2013,
"venue": "Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Suin Kim, Jianwen Zhang, Zheng Chen, Alice Oh, and Shixia Liu. 2013. A hierarchical aspect-sentiment model for online reviews. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI).",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "Annotating anaphoric shell nouns with their antecedents",
"authors": [
{
"first": "Varada",
"middle": [],
"last": "Kolhatkar",
"suffix": ""
},
{
"first": "Heike",
"middle": [],
"last": "Zinsmeister",
"suffix": ""
},
{
"first": "Graeme",
"middle": [],
"last": "Hirst",
"suffix": ""
}
],
"year": 2013,
"venue": "Linguistic Annotation Workshop and Interoperability with Discourse",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Varada Kolhatkar, Heike Zinsmeister, and Graeme Hirst. 2013. Annotating anaphoric shell nouns with their antecedents. In Linguistic Annotation Work- shop and Interoperability with Discourse.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Joint sentiment/topic model for sentiment analysis",
"authors": [
{
"first": "Chenghua",
"middle": [],
"last": "Lin",
"suffix": ""
},
{
"first": "Yulan",
"middle": [],
"last": "He",
"suffix": ""
}
],
"year": 2009,
"venue": "Proceedings of the Conference on Information and Knowledge Management (CIKM)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chenghua Lin and Yulan He. 2009. Joint senti- ment/topic model for sentiment analysis. In Pro- ceedings of the Conference on Information and Knowledge Management (CIKM).",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "Weakly supervised joint sentiment-topic detection from text",
"authors": [
{
"first": "Chenghua",
"middle": [],
"last": "Lin",
"suffix": ""
},
{
"first": "Yulan",
"middle": [],
"last": "He",
"suffix": ""
},
{
"first": "R",
"middle": [],
"last": "Everson",
"suffix": ""
},
{
"first": "S",
"middle": [],
"last": "Ruger",
"suffix": ""
}
],
"year": 2012,
"venue": "IEEE Transactions on Knowledge and Data Engineering",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Chenghua Lin, Yulan He, R. Everson, and S. Ruger. 2012. Weakly supervised joint sentiment-topic de- tection from text. IEEE Transactions on Knowledge and Data Engineering.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "A survey of opinion mining and sentiment analysis",
"authors": [
{
"first": "Bing",
"middle": [],
"last": "Liu",
"suffix": ""
},
{
"first": "Lei",
"middle": [],
"last": "Zhang",
"suffix": ""
}
],
"year": 2012,
"venue": "Mining Text Data",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bing Liu and Lei Zhang. 2012. A survey of opinion mining and sentiment analysis. In Mining Text Data.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "NLTK: The natural language toolkit",
"authors": [
{
"first": "Edward",
"middle": [],
"last": "Loper",
"suffix": ""
},
{
"first": "Steven",
"middle": [],
"last": "Bird",
"suffix": ""
}
],
"year": 2002,
"venue": "Proceedings of the ACL Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and Computational Linguistics (ETMTNLP)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Edward Loper and Steven Bird. 2002. NLTK: The natural language toolkit. In Proceedings of the ACL Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and Compu- tational Linguistics (ETMTNLP).",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "Multi-aspect sentiment analysis with topic models",
"authors": [
{
"first": "Bin",
"middle": [],
"last": "Lu",
"suffix": ""
},
{
"first": "Myle",
"middle": [],
"last": "Ott",
"suffix": ""
},
{
"first": "Claire",
"middle": [],
"last": "Cardie",
"suffix": ""
},
{
"first": "Benjamin",
"middle": [
"K"
],
"last": "Tsou",
"suffix": ""
}
],
"year": 2011,
"venue": "Proceedings of the International Conference on Data Mining Workshops (ICDMW)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bin Lu, Myle Ott, Claire Cardie, and Benjamin K. Tsou. 2011. Multi-aspect sentiment analysis with topic models. In Proceedings of the International Conference on Data Mining Workshops (ICDMW).",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Aspect extraction through semi-supervised modeling",
"authors": [
{
"first": "Arjun",
"middle": [],
"last": "Mukherjee",
"suffix": ""
},
{
"first": "Bing",
"middle": [],
"last": "Liu",
"suffix": ""
}
],
"year": 2012,
"venue": "Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Arjun Mukherjee and Bing Liu. 2012. Aspect extrac- tion through semi-supervised modeling. In Proceed- ings of the Annual Meeting of the Association for Computational Linguistics (ACL).",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "Knowledge graph identification",
"authors": [
{
"first": "Jay",
"middle": [],
"last": "Pujara",
"suffix": ""
},
{
"first": "Hui",
"middle": [],
"last": "Miao",
"suffix": ""
},
{
"first": "Lise",
"middle": [],
"last": "Getoor",
"suffix": ""
},
{
"first": "William",
"middle": [],
"last": "Cohen",
"suffix": ""
}
],
"year": 2013,
"venue": "International Semantic Web Conference (ISWC)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jay Pujara, Hui Miao, Lise Getoor, and William Cohen. 2013. Knowledge graph identification. In Interna- tional Semantic Web Conference (ISWC).",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Learning latent engagement patterns of students in online courses",
"authors": [
{
"first": "Arti",
"middle": [],
"last": "Ramesh",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Goldwasser",
"suffix": ""
},
{
"first": "Bert",
"middle": [],
"last": "Huang",
"suffix": ""
},
{
"first": "Hal",
"middle": [],
"last": "Daume",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
},
{
"first": "Lise",
"middle": [],
"last": "Getoor",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Arti Ramesh, Dan Goldwasser, Bert Huang, Hal Daume III, and Lise Getoor. 2014a. Learning latent engagement patterns of students in online courses. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI).",
"links": null
},
"BIBREF22": {
"ref_id": "b22",
"title": "Understanding MOOC discussion forums using seeded lda",
"authors": [
{
"first": "Arti",
"middle": [],
"last": "Ramesh",
"suffix": ""
},
{
"first": "Dan",
"middle": [],
"last": "Goldwasser",
"suffix": ""
},
{
"first": "Bert",
"middle": [],
"last": "Huang",
"suffix": ""
},
{
"first": "Hal",
"middle": [],
"last": "Daum\u00e9",
"suffix": ""
},
{
"first": "Iii",
"middle": [],
"last": "",
"suffix": ""
},
{
"first": "Lise",
"middle": [],
"last": "Getoor",
"suffix": ""
}
],
"year": 2014,
"venue": "ACL Workshop on Innovative Use of NLP for Building Educational Applications (BEA)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Arti Ramesh, Dan Goldwasser, Bert Huang, Hal Daum\u00e9 III, and Lise Getoor. 2014b. Understand- ing MOOC discussion forums using seeded lda. In ACL Workshop on Innovative Use of NLP for Build- ing Educational Applications (BEA).",
"links": null
},
"BIBREF23": {
"ref_id": "b23",
"title": "Development of a framework to classify MOOC discussion forum posts: Methodology and challenges",
"authors": [
{
"first": "Glenda",
"middle": [
"S"
],
"last": "Stump",
"suffix": ""
},
{
"first": "Jennifer",
"middle": [],
"last": "Deboer",
"suffix": ""
},
{
"first": "Jonathan",
"middle": [],
"last": "Whittinghill",
"suffix": ""
},
{
"first": "Lori",
"middle": [],
"last": "Breslow",
"suffix": ""
}
],
"year": 2013,
"venue": "NIPS Workshop on Data Driven Education",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Glenda S. Stump, Jennifer DeBoer, Jonathan Whit- tinghill, and Lori Breslow. 2013. Development of a framework to classify MOOC discussion forum posts: Methodology and challenges. In NIPS Work- shop on Data Driven Education.",
"links": null
},
"BIBREF24": {
"ref_id": "b24",
"title": "Modeling online reviews with multi-grain topic models",
"authors": [
{
"first": "Ivan",
"middle": [],
"last": "Titov",
"suffix": ""
},
{
"first": "Ryan",
"middle": [],
"last": "Mcdonald",
"suffix": ""
}
],
"year": 2008,
"venue": "Proceedings of the International Conference on World Wide Web",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ivan Titov and Ryan McDonald. 2008. Modeling on- line reviews with multi-grain topic models. In Pro- ceedings of the International Conference on World Wide Web (WWW).",
"links": null
},
"BIBREF25": {
"ref_id": "b25",
"title": "Jointly modeling aspects and opinions with a maxEnt-LDA hybrid",
"authors": [
{
"first": "Jing",
"middle": [],
"last": "Wayne Xin Zhao",
"suffix": ""
},
{
"first": "Hongfei",
"middle": [],
"last": "Jiang",
"suffix": ""
},
{
"first": "Xiaoming",
"middle": [],
"last": "Yan",
"suffix": ""
},
{
"first": "",
"middle": [],
"last": "Li",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Wayne Xin Zhao, Jing Jiang, Hongfei Yan, and Xiaom- ing Li. 2010. Jointly modeling aspects and opinions with a maxEnt-LDA hybrid. In Proceedings of the Conference on Empirical Methods in Natural Lan- guage Processing (EMNLP).",
"links": null
}
},
"ref_entries": {
"TABREF0": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Post 1: I have not received the midterm.</td></tr><tr><td>Post 2:</td></tr></table>",
"text": "No lecture subtitles week, will they be uploaded? Post 3: I am ... and I am looking forward to learn more ...",
"num": null
},
"TABREF1": {
"type_str": "table",
"html": null,
"content": "<table/>",
"text": "Example posts from MOOC forums. Aspect words are highlighted in bold.",
"num": null
},
"TABREF3": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Model</td><td colspan=\"3\">LECTURE-CONTENT</td><td colspan=\"3\">LECTURE-VIDEO</td><td colspan=\"3\">LECTURE-AUDIO</td><td colspan=\"3\">LECTURE-LECTURER</td><td colspan=\"3\">LECTURE-SUBTITLES</td></tr><tr><td/><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td></tr><tr><td>SEEDEDLDA</td><td>0.137</td><td>0.057</td><td>0.08</td><td>0.156</td><td>0.256</td><td>0.240</td><td>0.684</td><td>0.684</td><td>0.684</td><td>0.037</td><td>0.159</td><td>0.06</td><td>0.289</td><td>0.631</td><td>0.397</td></tr><tr><td>PSL-JOINT</td><td>0.407</td><td>0.413</td><td>0.410</td><td>0.411</td><td>0.591</td><td>0.485</td><td>0.635</td><td>0.537</td><td>0.582</td><td>0.218</td><td>0.623</td><td>0.323</td><td>0.407</td><td>0.53</td><td>0.461</td></tr></table>",
"text": "Representative rules from PSL-Joint model",
"num": null
},
"TABREF4": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Model</td><td colspan=\"3\">QUIZ-CONTENT</td><td colspan=\"3\">QUIZ-SUBMISSION</td><td colspan=\"3\">QUIZ-DEADLINES</td><td colspan=\"3\">QUIZ-GRADING</td></tr><tr><td/><td>Prec</td><td>Rec.</td><td>F1</td><td>Prec</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td></tr><tr><td>SEEDEDLDA</td><td>0.042</td><td>0.006</td><td>0.011</td><td>0.485</td><td>0.398</td><td>0.437</td><td>0.444</td><td>0.141</td><td>0.214</td><td>0.524</td><td>0.508</td><td>0.514</td></tr><tr><td>PSL-JOINT</td><td>0.324</td><td>0.405</td><td>0.36</td><td>0.521</td><td>0.347</td><td>0.416</td><td>0.667</td><td>0.563</td><td>0.611</td><td>0.572</td><td>0.531</td><td>0.550</td></tr></table>",
"text": "Precision, recall and F1 scores for LECTURE fine aspects",
"num": null
},
"TABREF5": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Model</td><td/><td>LECTURE</td><td/><td/><td>QUIZ</td><td/><td/><td>CERTIFICATE</td><td/><td/><td>SOCIAL</td><td/></tr><tr><td/><td>Prec</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td></tr><tr><td>SEEDEDLDA</td><td>0.597</td><td>0.673</td><td>0.632</td><td>0.752</td><td>0.583</td><td>0.657</td><td>0.315</td><td>0.845</td><td>0.459</td><td>0.902</td><td>0.513</td><td>0.654</td></tr><tr><td>PSL-JOINT</td><td>0.563</td><td>0.715</td><td>0.630</td><td>0.724</td><td>0.688</td><td>0.706</td><td>0.552</td><td>0.711</td><td>0.621</td><td>0.871</td><td>0.530</td><td>0.659</td></tr></table>",
"text": "Precision, recall and F1 scores for QUIZ fine aspects",
"num": null
},
"TABREF6": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Model</td><td/><td>POSITIVE</td><td/><td/><td>NEGATIVE</td><td/><td/><td>NEUTRAL</td><td/></tr><tr><td/><td>Prec</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td><td>Prec.</td><td>Rec.</td><td>F1</td></tr><tr><td>SEEDEDLDA</td><td>0.104</td><td>0.721</td><td>0.182</td><td>0.650</td><td>0.429</td><td>0.517</td><td>0.483</td><td>0.282</td><td>0.356</td></tr><tr><td>PSL-JOINT</td><td>0.114</td><td>0.544</td><td>0.189</td><td>0.571</td><td>0.666</td><td>0.615</td><td>0.664</td><td>0.322</td><td>0.434</td></tr></table>",
"text": "Precision, recall and F1 scores for coarse aspects",
"num": null
},
"TABREF7": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>: Precision, recall and F1 scores for sentiment</td></tr><tr><td>5.2 PSL for Joint Aspect-Sentiment</td></tr><tr><td>(PSL-Joint)</td></tr></table>",
"text": "",
"num": null
},
"TABREF8": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Correct Label</td><td>PSL</td><td>SeededLDA</td><td>Post</td></tr><tr><td>QUIZ-CONTENT</td><td>QUIZ-CONTENT</td><td>LECTURE-CONTENT</td><td>There is a typo or other mistake in the assignment instructions (e.g. es-</td></tr><tr><td/><td/><td/><td>sential information omitted) Type ID: programming-content Problem ID:</td></tr><tr><td/><td/><td/><td>programming-mistake Browser: Chrome 32 OS: Windows 7</td></tr><tr><td>QUIZ-CONTENT</td><td>QUIZ-CONTENT</td><td>LECTURE-CONTENT</td><td/></tr><tr><td>LECTURE-AUDIO</td><td>LECTURE-AUDIO</td><td>LECTURE-SUBTITLES</td><td/></tr><tr><td>SOCIAL</td><td>SOCIAL</td><td>LECTURE-VIDEO</td><td>Let's start a group for discussing the lecture videos.</td></tr></table>",
"text": "There is a typo or other mistake on the page (e.g. factual error information omitted) Week 4 Quiz Question 6: Question 6 When a user clicks on a View that has registered to show a Context Menu which one of the following methods will be called?Thanks for the suggestion about downloading the video and referring to the subtitles. I will give that a try but I would also like to point out that what the others are saying is true for me too: The audio is just barely audible even when the volume on my computer is set to 100%.",
"num": null
},
"TABREF9": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>Correct Label</td><td>Predicted Label</td><td>Second</td><td>Post</td></tr><tr><td/><td/><td>Prediction</td><td/></tr><tr><td>LECTURE-CONTENT</td><td>QUIZ-CONTENT</td><td>LECTURE-CONTENT</td><td>I have a difference of opinion to the answer for Question 6 too. It differs from</td></tr><tr><td/><td/><td/><td>what is presented in lecture 1.</td></tr><tr><td>SOCIAL</td><td>LECTURE-SUBTITLES</td><td>SOCIAL</td><td>Hello guys!!! I am ... The course materials are extraordinary. The subtitles are</td></tr><tr><td/><td/><td/><td>really helpful! Thanks to instructors for giving us all a wonderful opportunity.</td></tr></table>",
"text": "Example posts that PSL-Joint predicted correctly, but were misclassified by SeededLDA",
"num": null
},
"TABREF10": {
"type_str": "table",
"html": null,
"content": "<table><tr><td>5.3 Interpreting PSL-Joint Predictions</td></tr></table>",
"text": "Example posts whose second-best prediction is correct",
"num": null
}
}
}
}