| { |
| "paper_id": "P09-1028", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T08:53:15.960387Z" |
| }, |
| "title": "A Non-negative Matrix Tri-factorization Approach to Sentiment Classification with Lexical Prior Knowledge", |
| "authors": [ |
| { |
| "first": "Tao", |
| "middle": [], |
| "last": "Li", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Florida International University", |
| "location": {} |
| }, |
| "email": "taoli@cs.fiu.edu" |
| }, |
| { |
| "first": "Yi", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Florida International University", |
| "location": {} |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Vikas", |
| "middle": [], |
| "last": "Sindhwani", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Florida International University", |
| "location": {} |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Sentiment classification refers to the task of automatically identifying whether a given piece of text expresses positive or negative opinion towards a subject at hand. The proliferation of user-generated web content such as blogs, discussion forums and online review sites has made it possible to perform large-scale mining of public opinion. Sentiment modeling is thus becoming a critical component of market intelligence and social media technologies that aim to tap into the collective wisdom of crowds. In this paper, we consider the problem of learning high-quality sentiment models with minimal manual supervision. We propose a novel approach to learn from lexical prior knowledge in the form of domain-independent sentimentladen terms, in conjunction with domaindependent unlabeled data and a few labeled documents. Our model is based on a constrained non-negative tri-factorization of the term-document matrix which can be implemented using simple update rules. Extensive experimental studies demonstrate the effectiveness of our approach on a variety of real-world sentiment prediction tasks.", |
| "pdf_parse": { |
| "paper_id": "P09-1028", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Sentiment classification refers to the task of automatically identifying whether a given piece of text expresses positive or negative opinion towards a subject at hand. The proliferation of user-generated web content such as blogs, discussion forums and online review sites has made it possible to perform large-scale mining of public opinion. Sentiment modeling is thus becoming a critical component of market intelligence and social media technologies that aim to tap into the collective wisdom of crowds. In this paper, we consider the problem of learning high-quality sentiment models with minimal manual supervision. We propose a novel approach to learn from lexical prior knowledge in the form of domain-independent sentimentladen terms, in conjunction with domaindependent unlabeled data and a few labeled documents. Our model is based on a constrained non-negative tri-factorization of the term-document matrix which can be implemented using simple update rules. Extensive experimental studies demonstrate the effectiveness of our approach on a variety of real-world sentiment prediction tasks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Web 2.0 platforms such as blogs, discussion forums and other such social media have now given a public voice to every consumer. Recent surveys have estimated that a massive number of internet users turn to such forums to collect recommendations for products and services, guiding their own choices and decisions by the opinions that other consumers have publically expressed. Gleaning insights by monitoring and analyzing large amounts of such user-generated data is thus becoming a key competitive differentiator for many companies. While tracking brand perceptions in traditional media is hardly a new challenge, handling the unprecedented scale of unstructured user-generated web content requires new methodologies. These methodologies are likely to be rooted in natural language processing and machine learning techniques.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Automatically classifying the sentiment expressed in a blog around selected topics of interest is a canonical machine learning task in this discussion. A standard approach would be to manually label documents with their sentiment orientation and then apply off-the-shelf text classification techniques. However, sentiment is often conveyed with subtle linguistic mechanisms such as the use of sarcasm and highly domain-specific contextual cues. This makes manual annotation of sentiment time consuming and error-prone, presenting a bottleneck in learning high quality models. Moreover, products and services of current focus, and associated community of bloggers with their idiosyncratic expressions, may rapidly evolve over time causing models to potentially lose performance and become stale. This motivates the problem of learning robust sentiment models from minimal supervision.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In their seminal work, (Pang et al., 2002) demonstrated that supervised learning significantly outperformed a competing body of work where hand-crafted dictionaries are used to assign sentiment labels based on relative frequencies of positive and negative terms. As observed by (Ng et al., 2006) , most semi-automated dictionary-based approaches yield unsatisfactory lexicons, with either high coverage and low precision or vice versa. However, the treatment of such dictionaries as forms of prior knowledge that can be incorporated in machine learning models is a relatively less explored topic; even lesser so in conjunction with semi-supervised models that attempt to utilize un-labeled data. This is the focus of the current paper.", |
| "cite_spans": [ |
| { |
| "start": 23, |
| "end": 42, |
| "text": "(Pang et al., 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 278, |
| "end": 295, |
| "text": "(Ng et al., 2006)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Our models are based on a constrained nonnegative tri-factorization of the term-document matrix, which can be implemented using simple update rules. Treated as a set of labeled features, the sentiment lexicon is incorporated as one set of constraints that enforce domain-independent prior knowledge. A second set of constraints introduce domain-specific supervision via a few document labels. Together these constraints enable learning from partial supervision along both dimensions of the term-document matrix, in what may be viewed more broadly as a framework for incorporating dual-supervision in matrix factorization models. We provide empirical comparisons with several competing methodologies on four, very different domains -blogs discussing enterprise software products, political blogs discussing US presidential candidates, amazon.com product reviews and IMDB movie reviews. Results demonstrate the effectiveness and generality of our approach.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of the paper is organized as follows. We begin by discussing related work in Section 2. Section 3 gives a quick background on Nonnegative Matrix Tri-factorization models. In Section 4, we present a constrained model and computational algorithm for incorporating lexical knowledge in sentiment analysis. In Section 5, we enhance this model by introducing document labels as additional constraints. Section 6 presents an empirical study on four datasets. Finally, Section 7 concludes this paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "We point the reader to a recent book (Pang and Lee, 2008) for an in-depth survey of literature on sentiment analysis. In this section, we briskly cover related work to position our contributions appropriately in the sentiment analysis and machine learning literature.", |
| "cite_spans": [ |
| { |
| "start": 37, |
| "end": 57, |
| "text": "(Pang and Lee, 2008)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Methods focussing on the use and generation of dictionaries capturing the sentiment of words have ranged from manual approaches of developing domain-dependent lexicons (Das and Chen, 2001) to semi-automated approaches (Hu and Liu, 2004; Zhuang et al., 2006; Kim and Hovy, 2004) , and even an almost fully automated approach (Turney, 2002) . Most semi-automated approaches have met with limited success (Ng et al., 2006) and supervised learning models have tended to outperform dictionary-based classification schemes (Pang et al., 2002) . A two-tier scheme (Pang and Lee, 2004) where sentences are first classified as subjective versus objective, and then applying the sentiment classifier on only the subjective sentences further improves performance. Results in these papers also suggest that using more sophisticated linguistic models, incorporating parts-of-speech and n-gram language models, do not improve over the simple unigram bag-of-words representation. In keeping with these findings, we also adopt a unigram text model. A subjectivity classification phase before our models are applied may further improve the results reported in this paper, but our focus is on driving the polarity prediction stage with minimal manual effort.", |
| "cite_spans": [ |
| { |
| "start": 168, |
| "end": 188, |
| "text": "(Das and Chen, 2001)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 218, |
| "end": 236, |
| "text": "(Hu and Liu, 2004;", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 237, |
| "end": 257, |
| "text": "Zhuang et al., 2006;", |
| "ref_id": null |
| }, |
| { |
| "start": 258, |
| "end": 277, |
| "text": "Kim and Hovy, 2004)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 324, |
| "end": 338, |
| "text": "(Turney, 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 402, |
| "end": 419, |
| "text": "(Ng et al., 2006)", |
| "ref_id": null |
| }, |
| { |
| "start": 517, |
| "end": 536, |
| "text": "(Pang et al., 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 567, |
| "end": 577, |
| "text": "Lee, 2004)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In this regard, our model brings two interrelated but distinct themes from machine learning to bear on this problem: semi-supervised learning and learning from labeled features. The goal of the former theme is to learn from few labeled examples by making use of unlabeled data, while the goal of the latter theme is to utilize weak prior knowledge about term-class affinities (e.g., the term \"awful\" indicates negative sentiment and therefore may be considered as a negatively labeled feature). Empirical results in this paper demonstrate that simultaneously attempting both these goals in a single model leads to improvements over models that focus on a single goal. (Goldberg and Zhu, 2006) adapt semi-supervised graph-based methods for sentiment analysis but do not incorporate lexical prior knowledge in the form of labeled features. Most work in machine learning literature on utilizing labeled features has focused on using them to generate weakly labeled examples that are then used for standard supervised learning: (Schapire et al., 2002) propose one such framework for boosting logistic regression; (Wu and Srihari, 2004) build a modified SVM and (Liu et al., 2004) use a combination of clustering and EM based methods to instantiate similar frameworks. By contrast, we incorporate lexical knowledge directly as constraints on our matrix factorization model. In recent work, Druck et al. (Druck et al., 2008) constrain the predictions of a multinomial logistic regression model on unlabeled instances in a Generalized Expectation formulation for learning from labeled features. Unlike their approach which uses only unlabeled instances, our method uses both labeled and unlabeled documents in conjunction with labeled and unlabeled words.", |
| "cite_spans": [ |
| { |
| "start": 668, |
| "end": 692, |
| "text": "(Goldberg and Zhu, 2006)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 1024, |
| "end": 1047, |
| "text": "(Schapire et al., 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 1398, |
| "end": 1418, |
| "text": "(Druck et al., 2008)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The matrix tri-factorization models explored in this paper are closely related to the models proposed recently in (Li et al., 2008; Sindhwani et al., 2008) . Though, their techniques for proving algorithm convergence and correctness can be readily adapted for our models, (Li et al., 2008) do not incorporate dual supervision as we do. On the other hand, while (Sindhwani et al., 2008) do incorporate dual supervision in a non-linear kernelbased setting, they do not enforce non-negativity or orthogonality -aspects of matrix factorization models that have shown benefits in prior empirical studies, see e.g., (Ding et al., 2006) .", |
| "cite_spans": [ |
| { |
| "start": 114, |
| "end": 131, |
| "text": "(Li et al., 2008;", |
| "ref_id": null |
| }, |
| { |
| "start": 132, |
| "end": 155, |
| "text": "Sindhwani et al., 2008)", |
| "ref_id": null |
| }, |
| { |
| "start": 272, |
| "end": 289, |
| "text": "(Li et al., 2008)", |
| "ref_id": null |
| }, |
| { |
| "start": 361, |
| "end": 385, |
| "text": "(Sindhwani et al., 2008)", |
| "ref_id": null |
| }, |
| { |
| "start": 610, |
| "end": 629, |
| "text": "(Ding et al., 2006)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We also note the very recent work of (Sindhwani and Melville, 2008) which proposes a dualsupervision model for semi-supervised sentiment analysis. In this model, bipartite graph regularization is used to diffuse label information along both sides of the term-document matrix. Conceptually, their model implements a co-clustering assumption closely related to Singular Value Decomposition (see also (Dhillon, 2001; Zha et al., 2001 ) for more on this perspective) while our model is based on Non-negative Matrix Factorization. In another recent paper (Sandler et al., 2008) , standard regularization models are constrained using graphs of word co-occurences. These are very recently proposed competing methodologies, and we have not been able to address empirical comparisons with them in this paper.", |
| "cite_spans": [ |
| { |
| "start": 398, |
| "end": 413, |
| "text": "(Dhillon, 2001;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 414, |
| "end": 430, |
| "text": "Zha et al., 2001", |
| "ref_id": null |
| }, |
| { |
| "start": 550, |
| "end": 572, |
| "text": "(Sandler et al., 2008)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Finally, recent efforts have also looked at transfer learning mechanisms for sentiment analysis, e.g., see (Blitzer et al., 2007) . While our focus is on single-domain learning in this paper, we note that cross-domain variants of our model can also be orthogonally developed.", |
| "cite_spans": [ |
| { |
| "start": 107, |
| "end": 129, |
| "text": "(Blitzer et al., 2007)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Our proposed models are based on non-negative matrix Tri-factorization (Ding et al., 2006) . In these models, an m \u00d7 n term-document matrix X is approximated by three factors that specify soft membership of terms and documents in one of kclasses:", |
| "cite_spans": [ |
| { |
| "start": 71, |
| "end": 90, |
| "text": "(Ding et al., 2006)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "X \u2248 FSG T .", |
| "eq_num": "(1)" |
| } |
| ], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "where F is an m \u00d7 k non-negative matrix representing knowledge in the word space, i.e., i-th row of F represents the posterior probability of word i belonging to the k classes, G is an n \u00d7 k nonnegative matrix representing knowledge in document space, i.e., the i-th row of G represents the posterior probability of document i belonging to the k classes, and S is an k \u00d7 k nonnegative matrix providing a condensed view of X. The matrix factorization model is similar to the probabilistic latent semantic indexing (PLSI) model (Hofmann, 1999) . In PLSI, X is treated as the joint distribution between words and documents by the scaling", |
| "cite_spans": [ |
| { |
| "start": 526, |
| "end": 541, |
| "text": "(Hofmann, 1999)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "X \u2192X = X/ \u2211 i j X i j thus \u2211 i jXi j = 1).X is factorized as X \u2248 W SD T , \u2211 k W ik = 1, \u2211 k D jk = 1, \u2211 k S kk = 1.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "(2) where X is the m \u00d7 n word-document semantic matrix, X = W SD, W is the word classconditional probability, and D is the document class-conditional probability and S is the class probability distribution.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "PLSI provides a simultaneous solution for the word and document class conditional distribution. Our model provides simultaneous solution for clustering the rows and the columns of X. To avoid ambiguity, the orthogonality conditions", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "F T F = I, G T G = I.", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "can be imposed to enforce each row of F and G to possess only one nonzero entry. Approximating the term-document matrix with a tri-factorization while imposing non-negativity and orthogonality constraints gives a principled framework for simultaneously clustering the rows (words) and columns (documents) of X. In the context of coclustering, these models return excellent empirical performance, see e.g., (Ding et al., 2006) . Our goal now is to bias these models with constraints incorporating (a) labels of features (coming from a domain-independent sentiment lexicon), and (b) labels of documents for the purposes of domainspecific adaptation. These enhancements are addressed in Sections 4 and 5 respectively.", |
| "cite_spans": [ |
| { |
| "start": 406, |
| "end": 425, |
| "text": "(Ding et al., 2006)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Basic Matrix Factorization Model", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We used a sentiment lexicon generated by the IBM India Research Labs that was developed for other text mining applications (Ramakrishnan et al., 2003) . It contains 2,968 words that have been human-labeled as expressing positive or negative sentiment. In total, there are 1,267 positive (e.g. \"great\") and 1,701 negative (e.g., \"bad\") unique terms after stemming. We eliminated terms that were ambiguous and dependent on context, such as \"dear\" and \"fine\". It should be noted, that this list was constructed without a specific domain in mind; which is further motivation for using training examples and unlabeled data to learn domain specific connotations.", |
| "cite_spans": [ |
| { |
| "start": 123, |
| "end": 150, |
| "text": "(Ramakrishnan et al., 2003)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Lexical Knowledge", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Lexical knowledge in the form of the polarity of terms in this lexicon can be introduced in the matrix factorization model. By partially specifying term polarities via F, the lexicon influences the sentiment predictions G over documents.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Incorporating Lexical Knowledge", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Let F 0 represent prior knowledge about sentimentladen words in the lexicon, i.e., if word i is a positive word (F 0 ) i1 = 1 while if it is negative (F 0 ) i2 = 1. Note that one may also use soft sentiment polarities though our experiments are conducted with hard assignments. This information is incorporated in the tri-factorization model via a squared loss term,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representing Knowledge in Word Space", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "min F,G,S X \u2212 FSG T 2 + \u03b1Tr (F \u2212 F 0 ) T C 1 (F \u2212 F 0 )", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Representing Knowledge in Word Space", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "where the notation Tr(A) means trace of the matrix A. Here, \u03b1 > 0 is a parameter which determines the extent to which we enforce F \u2248 F 0 , C 1 is a m \u00d7 m diagonal matrix whose entry (C 1 ) ii = 1 if the category of the i-th word is known (i.e., specified by the i-th row of F 0 ) and (C 1 ) ii = 0 otherwise. The squared loss terms ensure that the solution for F in the otherwise unsupervised learning problem be close to the prior knowledge F 0 . Note that if C 1 = I, then we know the class orientation of all the words and thus have a full specification of F 0 , Eq.(4) is then reduced to", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representing Knowledge in Word Space", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "min F,G,S X \u2212 FSG T 2 + \u03b1 F \u2212 F 0 2 (5)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representing Knowledge in Word Space", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The above model is generic and it allows certain flexibility. For example, in some cases, our prior knowledge on F 0 is not very accurate and we use smaller \u03b1 so that the final results are not dependent on F 0 very much, i.e., the results are mostly unsupervised learning results. In addition, the introduction of C 1 allows us to incorporate partial knowledge on word polarity information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representing Knowledge in Word Space", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The optimization problem in Eq.( 4) can be solved using the following update rules", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "G jk \u2190 G jk (X T FS) jk (GG T X T FS) jk ,", |
| "eq_num": "(6)" |
| } |
| ], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "S ik \u2190 S ik (F T XG) ik (F T FSG T G) ik . (7) F ik \u2190 F ik (XGS T + \u03b1C 1 F 0 ) ik (FF T XGS T + \u03b1C 1 F) ik .", |
| "eq_num": "(8)" |
| } |
| ], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The algorithm consists of an iterative procedure using the above three rules until convergence. We call this approach Matrix Factorization with Lexical Knowledge (MFLK) and outline the precise steps in the table below.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Algorithm 1 Matrix Factorization with Lexical Knowledge (MFLK) begin 1. Initialization:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Initialize F = F 0 G to K-means clustering results, S = (F T F) \u22121 F T XG(G T G) \u22121 . 2. Iteration:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Update G: fixing F, S, updating G Update F: fixing S, G, updating F Update S: fixing F, G, updating S end", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Updating F, G, S using the rules above leads to an asymptotic convergence to a local minima. This can be proved using arguments similar to (Ding et al., 2006) . We outline the proof of correctness for updating F since the squared loss term that involves F is a new component in our models.", |
| "cite_spans": [ |
| { |
| "start": 139, |
| "end": 158, |
| "text": "(Ding et al., 2006)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Algorithm Correctness and Convergence", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "The above iterative algorithm converges.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "Theorem 2 At convergence, the solution satisfies the Karuch, Kuhn, Tucker optimality condition, i.e., the algorithm converges correctly to a local optima. Theorem 1 can be proved using the standard auxiliary function approach used in (Lee and Seung, 2001 ).", |
| "cite_spans": [ |
| { |
| "start": 243, |
| "end": 254, |
| "text": "Seung, 2001", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "Proof of Theorem 2. Following the theory of constrained optimization (Nocedal and Wright, 1999), we minimize the following function", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "L(F) = ||X \u2212FSG T || 2 +\u03b1Tr (F \u2212 F 0 ) T C 1 (F \u2212 F0)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "Note that the gradient of L is,", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "\u2202L \u2202F = \u22122XGS T + 2FSG T GS T + 2\u03b1C 1 (F \u2212 F 0 ).", |
| "eq_num": "(9)" |
| } |
| ], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "The KKT complementarity condition for the nonnegativity of F ik gives", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "[\u22122XGS T + FSG T GS T + 2\u03b1C 1 (F \u2212 F 0 )] ik F ik = 0.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "(10) This is the fixed point relation that local minima for F must satisfy. Given an initial guess of F, the successive update of F using Eq.(8) will converge to a local minima. At convergence, we have", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "F ik = F ik (XGS T + \u03b1C 1 F 0 ) ik (FF T XGS T + \u03b1C 1 F) ik .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "which is equivalent to the KKT condition of Eq.(10). The correctness of updating rules for G in Eq.(6) and S in Eq. 7have been proved in (Ding et al., 2006) . -Note that we do not enforce exact orthogonality in our updating rules since this often implies softer class assignments.", |
| "cite_spans": [ |
| { |
| "start": 137, |
| "end": 156, |
| "text": "(Ding et al., 2006)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theorem 1", |
| "sec_num": null |
| }, |
| { |
| "text": "So far our models have made no demands on human effort, other than unsupervised collection of the term-document matrix and a one-time effort in compiling a domain-independent sentiment lexicon. We now assume that a few documents are manually labeled for the purposes of capturing some domain-specific connotations leading to a more domain-adapted model. The partial labels on documents can be described using G 0 where (G 0 ) i1 = 1 if the document expresses positive sentiment, and (G 0 ) i2 = 1 for negative sentiment. As with F 0 , one can also use soft sentiment labeling for documents, though our experiments are conducted with hard assignments. Therefore, the semi-supervised learning with lexical knowledge can be described as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semi-Supervised Learning With Lexical Knowledge", |
| "sec_num": "5" |
| }, |
| { |
| "text": "min F,G,S X \u2212 FSG T 2 + \u03b1Tr (F \u2212 F 0 ) T C 1 (F \u2212 F 0 ) + \u03b2Tr (G \u2212 G 0 ) T C 2 (G \u2212 G 0 )", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semi-Supervised Learning With Lexical Knowledge", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Where \u03b1 > 0, \u03b2 > 0 are parameters which determine the extent to which we enforce F \u2248 F 0 and G \u2248 G 0 respectively, C 1 and C 2 are diagonal matrices indicating the entries of F 0 and G 0 that correspond to labeled entities. The squared loss terms ensure that the solution for F, G, in the otherwise unsupervised learning problem, be close to the prior knowledge F 0 and G 0 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semi-Supervised Learning With Lexical Knowledge", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The optimization problem in Eq.( 4) can be solved using the following update rules", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "G jk \u2190 G jk (X T FS + \u03b2C 2 G 0 ) jk (GG T X T FS + \u03b2GG T C 2 G 0 ) jk (11) S ik \u2190 S ik (F T XG) ik (F T FSG T G) ik .", |
| "eq_num": "(12)" |
| } |
| ], |
| "section": "Computational Algorithm", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "F ik \u2190 F ik (XGS T + \u03b1C 1 F 0 ) ik (FF T XGS T + \u03b1C 1 F) ik .", |
| "eq_num": "(13)" |
| } |
| ], |
| "section": "Computational Algorithm", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Thus the algorithm for semi-supervised learning with lexical knowledge based on our matrix factorization framework, referred as SSMFLK, consists of an iterative procedure using the above three rules until convergence. The correctness and convergence of the algorithm can also be proved using similar arguments as what we outlined earlier for MFLK in Section 4.3.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "A quick word about computational complexity. The term-document matrix is typically very sparse with z nm non-zero entries while k is typically also much smaller than n, m. By using sparse matrix multiplications and avoiding dense intermediate matrices, the updates can be very efficiently and easily implemented. In particular, updating F, S, G each takes O(k 2 (m + n) + kz) time per iteration which scales linearly with the dimensions and density of the data matrix. Empirically, the number of iterations before practical convergence is usually very small (less than 100). Thus, computationally our approach scales to large datasets even though our experiments are run on relatively small-sized datasets.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Computational Algorithm", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "Four different datasets are used in our experiments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets Description", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "Movies Reviews: This is a popular dataset in sentiment analysis literature (Pang et al., 2002) . It consists of 1000 positive and 1000 negative movie reviews drawn from the IMDB archive of the rec.arts.movies.reviews newsgroups.", |
| "cite_spans": [ |
| { |
| "start": 75, |
| "end": 94, |
| "text": "(Pang et al., 2002)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Datasets Description", |
| "sec_num": "6.1" |
| }, |
| { |
| "text": "The data set is targeted at detecting sentiment around enterprise software, specifically pertaining to the IBM Lotus brand (Sindhwani and Melville, 2008) . An unlabeled set of blog posts was created by randomly sampling 2000 posts from a universe of 14,258 blogs that discuss issues relevant to Lotus software. In addition to this unlabeled set, 145 posts were chosen for manual labeling. These posts came from 14 individual blogs, 4 of which are actively posting negative content on the brand, with the rest tending to write more positive or neutral posts. The data was collected by downloading the latest posts from each blogger's RSS feeds, or accessing the blog's archives. Manual labeling resulted in 34 positive and 111 negative examples.", |
| "cite_spans": [ |
| { |
| "start": 123, |
| "end": 153, |
| "text": "(Sindhwani and Melville, 2008)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Lotus blogs:", |
| "sec_num": null |
| }, |
| { |
| "text": "For our second blog domain, we used data gathered from 16,742 political blogs, which contain over 500,000 posts. As with the Lotus dataset, an unlabeled set was created by randomly sampling 2000 posts. 107 posts were chosen for labeling. A post was labeled as having positive or negative sentiment about a specific candidate (Barack Obama or Hillary Clinton) if it explicitly mentioned the candidate in positive or negative terms. This resulted in 49 positively and 58 negatively labeled posts. Amazon Reviews: The dataset contains product reviews taken from Amazon.com from 4 product types: Kitchen, Books, DVDs, and Electronics (Blitzer et al., 2007) . The dataset contains about 4000 positive reviews and 4000 negative reviews and can be obtained from http://www.cis.upenn. edu/\u02dcmdredze/datasets/sentiment/.", |
| "cite_spans": [ |
| { |
| "start": 630, |
| "end": 652, |
| "text": "(Blitzer et al., 2007)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Political candidate blogs:", |
| "sec_num": null |
| }, |
| { |
| "text": "For all datasets, we picked 5000 words with highest document-frequency to generate the vocabulary. Stopwords were removed and a normalized term-frequency representation was used. Genuinely unlabeled posts for Political and Lotus were used for semi-supervised learning experiments in section 6.3; they were not used in section 6.2 on the effect of lexical prior knowledge. In the experiments, we set \u03b1, the parameter determining the extent to which to enforce the feature labels, to be 1/2, and \u03b2, the corresponding parameter for enforcing document labels, to be 1.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Political candidate blogs:", |
| "sec_num": null |
| }, |
| { |
| "text": "Of course, one can remove all burden on human effort by simply using unsupervised tech-niques. Our interest in the first set of experiments is to explore the benefits of incorporating a sentiment lexicon over unsupervised approaches. Does a one-time effort in compiling a domainindependent dictionary and using it for different sentiment tasks pay off in comparison to simply using unsupervised methods? In our case, matrix tri-factorization and other co-clustering methods form the obvious unsupervised baseline for comparison and so we start by comparing our method (MFLK) with the following methods:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Lexical Knowledge", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "\u2022 Four document clustering methods: Kmeans, Tri-Factor Nonnegative Matrix Factorization (TNMF) (Ding et al., 2006) , Information-Theoretic Co-clustering (ITCC) (Dhillon et al., 2003) , and Euclidean Co-clustering algorithm (ECC) (Cho et al., 2004) . These methods do not make use of the sentiment lexicon.", |
| "cite_spans": [ |
| { |
| "start": 95, |
| "end": 114, |
| "text": "(Ding et al., 2006)", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 160, |
| "end": 182, |
| "text": "(Dhillon et al., 2003)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 229, |
| "end": 247, |
| "text": "(Cho et al., 2004)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Lexical Knowledge", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "\u2022 Feature Centroid (FC): This is a simple dictionary-based baseline method. Recall that each word can be expressed as a \"bagof-documents\" vector. In this approach, we compute the centroids of these vectors, one corresponding to positive words and another corresponding to negative words. This yields a two-dimensional representation for documents, on which we then perform K-means clustering. Figure 1 shows the experimental results on four datasets using accuracy as the performance measure. The results are obtained by averaging 20 runs. It can be observed that our MFLK method can effectively utilize the lexical knowledge to improve the quality of sentiment prediction. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 393, |
| "end": 401, |
| "text": "Figure 1", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Lexical Knowledge", |
| "sec_num": "6.2" |
| }, |
| { |
| "text": "We also investigate the effects of the size of the sentiment lexicon on the performance of our model. Figure 2 shows results with random subsets of the lexicon of increasing size. We observe that generally the performance increases as more and more lexical supervision is provided. Robustness to Vocabulary Size High dimensionality and noise can have profound impact on the comparative performance of clustering and semi-supervised learning algorithms. We simulate scenarios with different vocabulary sizes by selecting words based on information gain. It should, however, be kept in mind that in a truely unsupervised setting document labels are unavailable and therefore information gain cannot be practically computed. Figure 3 and Figure 4 show results for Lotus and Amazon datasets respectively and are representative of performance on other datasets. MLFK tends to retain its position as the best performing method even at different vocabulary sizes. ITCC performance is also noteworthy given that it is a completely unsupervised method. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 102, |
| "end": 110, |
| "text": "Figure 2", |
| "ref_id": "FIGREF2" |
| }, |
| { |
| "start": 722, |
| "end": 730, |
| "text": "Figure 3", |
| "ref_id": null |
| }, |
| { |
| "start": 735, |
| "end": 743, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Size of Sentiment Lexicon", |
| "sec_num": null |
| }, |
| { |
| "text": "Supervision We now assume that together with labeled features from the sentiment lexicon, we also have access to a few labeled documents. The natural question is whether the presence of lexical constraints leads to better semi-supervised models. In this section, we compare our method (SSMFLK) with the following three semi-supervised approaches: (1) The algorithm proposed in (Zhou et al., 2003) which conducts semi-supervised learning with local and global consistency (Consistency Method); (2) Zhu et al.'s harmonic Gaussian field method coupled with the Class Mass Normalization (Harmonic-CMN) (Zhu et al., 2003) ; and (3) Green's function learning algorithm (Green's Function) proposed in (Ding et al., 2007) .", |
| "cite_spans": [ |
| { |
| "start": 377, |
| "end": 396, |
| "text": "(Zhou et al., 2003)", |
| "ref_id": null |
| }, |
| { |
| "start": 598, |
| "end": 616, |
| "text": "(Zhu et al., 2003)", |
| "ref_id": null |
| }, |
| { |
| "start": 694, |
| "end": 713, |
| "text": "(Ding et al., 2007)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Dual", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "We also compare the results of SSMFLK with those of two supervised classification methods: Support Vector Machine (SVM) and Naive Bayes. Both of these methods have been widely used in sentiment analysis. In particular, the use of SVMs in (Pang et al., 2002) initially sparked interest in using machine learning methods for sentiment classification. Note that none of these competing methods utilizes lexical knowledge.", |
| "cite_spans": [ |
| { |
| "start": 238, |
| "end": 257, |
| "text": "(Pang et al., 2002)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Dual", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "The results are presented in Figure 5 , Figure 6 , Figure 7 , and Figure 8 . We note that our SSMFLK method either outperforms all other methods over the entire range of number of labeled documents (Movies, Political), or ultimately outpaces other methods (Lotus, Amazon) as a few document labels come in. Figure 6 : Accuracy results with increasing number of labeled documents on Lotus dataset experiments, we additionally introduced labeled documents for domain-specific adjustments. Between these experiments, we can now look for words that switch sentiment polarity. These words are interesting because their domain-specific connotation differs from their lexical orientation. For amazon reviews, the following words switched polarity from positive to negative: fan, important, learning, cons, fast, feature, happy, memory, portable, simple, small, work while the following words switched polarity from negative to positive: address, finish, lack, mean, budget, rent, throw. Note that words like fan, memory probably refer to product or product components (i.e., computer fan and memory) in the amazon review context but have a very different connotation say in the context of movie reviews where they probably refer to movie fanfare and memorable performances. We were surprised to see happy switch polarity! Two examples of its negative-sentiment usage are:", |
| "cite_spans": [ |
| { |
| "start": 771, |
| "end": 857, |
| "text": "important, learning, cons, fast, feature, happy, memory, portable, simple, small, work", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 29, |
| "end": 37, |
| "text": "Figure 5", |
| "ref_id": null |
| }, |
| { |
| "start": 40, |
| "end": 48, |
| "text": "Figure 6", |
| "ref_id": null |
| }, |
| { |
| "start": 51, |
| "end": 59, |
| "text": "Figure 7", |
| "ref_id": null |
| }, |
| { |
| "start": 66, |
| "end": 74, |
| "text": "Figure 8", |
| "ref_id": "FIGREF5" |
| }, |
| { |
| "start": 306, |
| "end": 314, |
| "text": "Figure 6", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Dual", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "I ended up buying a Samsung and I couldn't be more happy and BORING, not one single exciting thing about this book. I was happy when my lunch break ended so I could go back to work and stop reading. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Sentiment Analysis with Dual", |
| "sec_num": "6.3" |
| }, |
| { |
| "text": "The primary contribution of this paper is to propose and benchmark new methodologies for sentiment analysis. Non-negative Matrix Factorizations constitute a rich body of algorithms that have found applicability in a variety of machine learning applications: from recommender systems to document clustering. We have shown how to build effective sentiment models by appropriately constraining the factors using lexical prior knowledge and document annotations. To more effectively utilize unlabeled data and induce domain-specific adaptation of our models, several extensions are possible: facilitating learning from related domains, incorporating hyperlinks between documents, incorporating synonyms or co-occurences between words etc. As a topic of vigorous current activity, there are several very recently proposed competing methodologies for sentiment analysis that we would like to benchmark against. These are topics for future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "Acknowledgement: The work of T. Li is partially supported by NSF grants DMS-0844513 and CCF-0830659. We would also like to thank Prem Melville and Richard Lawrence for their support.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Blitzer", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Dredze", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "440--447", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Blitzer, M. Dredze, and F. Pereira. 2007. Biogra- phies, bollywood, boom-boxes and blenders: Do- main adaptation for sentiment classification. In Pro- ceedings of ACL, pages 440-447.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Minimum sum squared residue co-clustering of gene expression data", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Cho", |
| "suffix": "" |
| }, |
| { |
| "first": "I", |
| "middle": [], |
| "last": "Dhillon", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Guan", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Sra", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of The 4th SIAM Data Mining Conference", |
| "volume": "", |
| "issue": "", |
| "pages": "22--24", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H. Cho, I. Dhillon, Y. Guan, and S. Sra. 2004. Mini- mum sum squared residue co-clustering of gene ex- pression data. In Proceedings of The 4th SIAM Data Mining Conference, pages 22-24, April.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Yahoo! for amazon: Extracting market sentiment from stock message boards", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Das", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 8th Asia Pacific Finance Association (APFA)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Das and M. Chen. 2001. Yahoo! for amazon: Extracting market sentiment from stock message boards. In Proceedings of the 8th Asia Pacific Fi- nance Association (APFA).", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Information-theoretical co-clustering", |
| "authors": [ |
| { |
| "first": "I", |
| "middle": [ |
| "S" |
| ], |
| "last": "Dhillon", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Mallela", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [ |
| "S" |
| ], |
| "last": "Modha", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of ACM SIGKDD", |
| "volume": "", |
| "issue": "", |
| "pages": "89--98", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "I. S. Dhillon, S. Mallela, and D. S. Modha. 2003. Information-theoretical co-clustering. In Proceed- ings of ACM SIGKDD, pages 89-98.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Co-clustering documents and words using bipartite spectral graph partitioning", |
| "authors": [ |
| { |
| "first": "I", |
| "middle": [ |
| "S" |
| ], |
| "last": "Dhillon", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of ACM SIGKDD", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "I. S. Dhillon. 2001. Co-clustering documents and words using bipartite spectral graph partitioning. In Proceedings of ACM SIGKDD.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Orthogonal nonnegative matrix tri-factorizations for clustering", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Ding", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Peng", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Park", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of ACM SIGKDD", |
| "volume": "", |
| "issue": "", |
| "pages": "126--135", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. Ding, T. Li, W. Peng, and H. Park. 2006. Orthogo- nal nonnegative matrix tri-factorizations for cluster- ing. In Proceedings of ACM SIGKDD, pages 126- 135.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "A learning framework using green's function and kernel regularization with application to recommender system", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Ding", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Jin", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [ |
| "D" |
| ], |
| "last": "Simon", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of ACM SIGKDD", |
| "volume": "", |
| "issue": "", |
| "pages": "260--269", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. Ding, R. Jin, T. Li, and H.D. Simon. 2007. A learning framework using green's function and ker- nel regularization with application to recommender system. In Proceedings of ACM SIGKDD, pages 260-269.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Learning from labeled features using generalized expectation criteria", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Druck", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Mann", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "SIGIR", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "G. Druck, G. Mann, and A. McCallum. 2008. Learn- ing from labeled features using generalized expecta- tion criteria. In SIGIR.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Seeing stars when there aren't many stars: Graph-based semisupervised learning for sentiment categorization", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Goldberg", |
| "suffix": "" |
| }, |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "HLT-NAACL 2006: Workshop on Textgraphs", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Goldberg and X. Zhu. 2006. Seeing stars when there aren't many stars: Graph-based semi- supervised learning for sentiment categorization. In HLT-NAACL 2006: Workshop on Textgraphs.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Probabilistic latent semantic indexing", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Hofmann", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceeding of SIGIR", |
| "volume": "", |
| "issue": "", |
| "pages": "50--57", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "T. Hofmann. 1999. Probabilistic latent semantic in- dexing. Proceeding of SIGIR, pages 50-57.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Mining and summarizing customer reviews", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Hu", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Liu", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "KDD", |
| "volume": "", |
| "issue": "", |
| "pages": "168--177", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Hu and B. Liu. 2004. Mining and summarizing customer reviews. In KDD, pages 168-177.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Determining the sentiment of opinions", |
| "authors": [ |
| { |
| "first": "S.-M", |
| "middle": [], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of International", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S.-M. Kim and E. Hovy. 2004. Determining the sen- timent of opinions. In Proceedings of International", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "uris": null, |
| "num": null, |
| "text": "Accuracy results on four datasets", |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "uris": null, |
| "num": null, |
| "text": "MFLK accuracy as size of sentiment lexicon (i.e., number of words in the lexicon) increases on the four datasets", |
| "type_str": "figure" |
| }, |
| "FIGREF3": { |
| "uris": null, |
| "num": null, |
| "text": "Accuracy results on Lotus dataset with increasing vocabulary size Accuracy results on Amazon dataset with increasing vocabulary size", |
| "type_str": "figure" |
| }, |
| "FIGREF4": { |
| "uris": null, |
| "num": null, |
| "text": "Domain-Specific ConnotationsIn our first set of experiments, we incorporated the sentiment lexicon in our models and learnt the sentiment orientation of words and documents via F, G factors respectively. In the second set of", |
| "type_str": "figure" |
| }, |
| "FIGREF5": { |
| "uris": null, |
| "num": null, |
| "text": "of documents labeled as a fraction of the original set of labeled Accuracy results with increasing number of labeled documents on Amazon dataset", |
| "type_str": "figure" |
| } |
| } |
| } |
| } |