{ "paper_id": "P09-1029", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T08:54:16.381944Z" }, "title": "Discovering the Discriminative Views: Measuring Term Weights for Sentiment Analysis", "authors": [ { "first": "Jungi", "middle": [], "last": "Kim", "suffix": "", "affiliation": { "laboratory": "", "institution": "Pohang University of Science and Technology", "location": { "settlement": "Pohang", "country": "Republic of Korea" } }, "email": "" }, { "first": "Jin-Ji", "middle": [], "last": "Li", "suffix": "", "affiliation": { "laboratory": "", "institution": "Pohang University of Science and Technology", "location": { "settlement": "Pohang", "country": "Republic of Korea" } }, "email": "" }, { "first": "Jong-Hyeok", "middle": [], "last": "Lee", "suffix": "", "affiliation": { "laboratory": "", "institution": "Pohang University of Science and Technology", "location": { "settlement": "Pohang", "country": "Republic of Korea" } }, "email": "jhlee@postech.ac.kr" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "This paper describes an approach to utilizing term weights for sentiment analysis tasks and shows how various term weighting schemes improve the performance of sentiment analysis systems. Previously, sentiment analysis was mostly studied under data-driven and lexicon-based frameworks. Such work generally exploits textual features for fact-based analysis tasks or lexical indicators from a sentiment lexicon. We propose to model term weighting into a sentiment analysis system utilizing collection statistics, contextual and topicrelated characteristics as well as opinionrelated properties. Experiments carried out on various datasets show that our approach effectively improves previous methods.", "pdf_parse": { "paper_id": "P09-1029", "_pdf_hash": "", "abstract": [ { "text": "This paper describes an approach to utilizing term weights for sentiment analysis tasks and shows how various term weighting schemes improve the performance of sentiment analysis systems. Previously, sentiment analysis was mostly studied under data-driven and lexicon-based frameworks. Such work generally exploits textual features for fact-based analysis tasks or lexical indicators from a sentiment lexicon. We propose to model term weighting into a sentiment analysis system utilizing collection statistics, contextual and topicrelated characteristics as well as opinionrelated properties. Experiments carried out on various datasets show that our approach effectively improves previous methods.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "With the explosion in the amount of commentaries on current issues and personal views expressed in weblogs on the Internet, the field of studying how to analyze such remarks and sentiments has been increasing as well. The field of opinion mining and sentiment analysis involves extracting opinionated pieces of text, determining the polarities and strengths, and extracting holders and targets of the opinions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Much research has focused on creating testbeds for sentiment analysis tasks.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Most notable and widely used are Multi-Perspective Question Answering (MPQA) and Movie-review datasets. MPQA is a collection of newspaper articles annotated with opinions and private states at the subsentence level (Wiebe et al., 2003) . Movie-review dataset consists of positive and negative reviews from the Internet Movie Database (IMDb) archive (Pang et al., 2002) .", "cite_spans": [ { "start": 215, "end": 235, "text": "(Wiebe et al., 2003)", "ref_id": "BIBREF27" }, { "start": 349, "end": 368, "text": "(Pang et al., 2002)", "ref_id": "BIBREF20" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Evaluation workshops such as TREC and NT-CIR have recently joined in this new trend of research and organized a number of successful meetings. At the TREC Blog Track meetings, researchers have dealt with the problem of retrieving topically-relevant blog posts and identifying documents with opinionated contents (Ounis et al., 2008) . NTCIR Multilingual Opinion Analysis Task (MOAT) shared a similar mission, where participants are provided with a number of topics and a set of relevant newspaper articles for each topic, and asked to extract opinion-related properties from enclosed sentences (Seki et al., 2008) .", "cite_spans": [ { "start": 312, "end": 332, "text": "(Ounis et al., 2008)", "ref_id": "BIBREF18" }, { "start": 594, "end": 613, "text": "(Seki et al., 2008)", "ref_id": "BIBREF22" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Previous studies for sentiment analysis belong to either the data-driven approach where an annotated corpus is used to train a machine learning (ML) classifier, or to the lexicon-based approach where a pre-compiled list of sentiment terms is utilized to build a sentiment score function.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "This paper introduces an approach to the sentiment analysis tasks with an emphasis on how to represent and evaluate the weights of sentiment terms. We propose a number of characteristics of good sentiment terms from the perspectives of informativeness, prominence, topic-relevance, and semantic aspects using collection statistics, contextual information, semantic associations as well as opinion-related properties of terms. These term weighting features constitute the sentiment analysis model in our opinion retrieval system. We test our opinion retrieval system with TREC and NT-CIR datasets to validate the effectiveness of our term weighting features. We also verify the effectiveness of the statistical features used in datadriven approaches by evaluating an ML classifier with labeled corpora.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Representing text with salient features is an important part of a text processing task, and there exists many works that explore various features for text analysis systems (Sebastiani, 2002; Forman, 2003) . Sentiment analysis task have also been using various lexical, syntactic, and statistical features (Pang and Lee, 2008) . Pang et al. (2002) employed n-gram and POS features for ML methods to classify movie-review data. Also, syntactic features such as the dependency relationship of words and subtrees have been shown to effectively improve the performances of sentiment analysis (Kudo and Matsumoto, 2004; Gamon, 2004; Matsumoto et al., 2005; Ng et al., 2006) .", "cite_spans": [ { "start": 172, "end": 190, "text": "(Sebastiani, 2002;", "ref_id": "BIBREF21" }, { "start": 191, "end": 204, "text": "Forman, 2003)", "ref_id": "BIBREF3" }, { "start": 305, "end": 325, "text": "(Pang and Lee, 2008)", "ref_id": "BIBREF19" }, { "start": 328, "end": 346, "text": "Pang et al. (2002)", "ref_id": "BIBREF20" }, { "start": 587, "end": 613, "text": "(Kudo and Matsumoto, 2004;", "ref_id": "BIBREF7" }, { "start": 614, "end": 626, "text": "Gamon, 2004;", "ref_id": "BIBREF4" }, { "start": 627, "end": 650, "text": "Matsumoto et al., 2005;", "ref_id": "BIBREF11" }, { "start": 651, "end": 667, "text": "Ng et al., 2006)", "ref_id": "BIBREF16" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "While these features are usually employed by data-driven approaches, there are unsupervised approaches for sentiment analysis that make use of a set of terms that are semantically oriented toward expressing subjective statements (Yu and Hatzivassiloglou, 2003) . Accordingly, much research has focused on recognizing terms' semantic orientations and strength, and compiling sentiment lexicons (Hatzivassiloglou and Mckeown, 1997; Turney and Littman, 2003; Kamps et al., 2004; Whitelaw et al., 2005; Esuli and Sebastiani, 2006) .", "cite_spans": [ { "start": 229, "end": 260, "text": "(Yu and Hatzivassiloglou, 2003)", "ref_id": "BIBREF32" }, { "start": 393, "end": 429, "text": "(Hatzivassiloglou and Mckeown, 1997;", "ref_id": "BIBREF5" }, { "start": 430, "end": 455, "text": "Turney and Littman, 2003;", "ref_id": "BIBREF24" }, { "start": 456, "end": 475, "text": "Kamps et al., 2004;", "ref_id": "BIBREF6" }, { "start": 476, "end": 498, "text": "Whitelaw et al., 2005;", "ref_id": "BIBREF26" }, { "start": 499, "end": 526, "text": "Esuli and Sebastiani, 2006)", "ref_id": "BIBREF1" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "Interestingly, there are conflicting conclusions about the usefulness of the statistical features in sentiment analysis tasks (Pang and Lee, 2008) . Pang et al. (2002) presents empirical results indicating that using term presence over term frequency is more effective in a data-driven sentiment classification task. Such a finding suggests that sentiment analysis may exploit different types of characteristics from the topical tasks, that, unlike fact-based text analysis tasks, repetition of terms does not imply a significance on the overall sentiment. On the other hand, Wiebe et al. (2004) have noted that hapax legomena (terms that only appear once in a collection of texts) are good signs for detecting subjectivity. Other works have also exploited rarely occurring terms for sentiment analysis tasks (Dave et al., 2003; Yang et al., 2006 ).", "cite_spans": [ { "start": 126, "end": 146, "text": "(Pang and Lee, 2008)", "ref_id": "BIBREF19" }, { "start": 149, "end": 167, "text": "Pang et al. (2002)", "ref_id": "BIBREF20" }, { "start": 576, "end": 595, "text": "Wiebe et al. (2004)", "ref_id": "BIBREF28" }, { "start": 809, "end": 828, "text": "(Dave et al., 2003;", "ref_id": "BIBREF0" }, { "start": 829, "end": 846, "text": "Yang et al., 2006", "ref_id": "BIBREF31" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "The opinion retrieval task is a relatively recent issue that draws both the attention of IR and NLP communities. Its task is to find relevant documents that also contain sentiments about a given topic. Generally, the opinion retrieval task has been approached as a two-stage task: first, retrieving topically relevant documents, then reranking the documents by the opinion scores . This approach is also appropriate for evaluation systems such as NTCIR MOAT that assumes that the set of topically relevant documents are already known in advance. On the other hand, there are also some interesting works on modeling the topic and sentiment of documents in a unified way (Mei et al., 2007; Zhang and Ye, 2008) .", "cite_spans": [ { "start": 669, "end": 687, "text": "(Mei et al., 2007;", "ref_id": "BIBREF12" }, { "start": 688, "end": 707, "text": "Zhang and Ye, 2008)", "ref_id": "BIBREF34" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "In this section, we describe the characteristics of terms that are useful in sentiment analysis, and present our sentiment analysis model as part of an opinion retrieval system and an ML sentiment classifier.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Term Weighting and Sentiment Analysis", "sec_num": "3" }, { "text": "This section examines the qualities of useful terms for sentiment analysis tasks and corresponding features. For the sake of organization, we categorize the sources of features into either global or local knowledge, and either topic-independent or topic-dependent knowledge. Topic-independently speaking, a good sentiment term is discriminative and prominent, such that the appearance of the term imposes greater influence on the judgment of the analysis system. The rare occurrence of terms in document collections has been regarded as a very important feature in IR methods, and effective IR models of today, either explicitly or implicitly, accommodate this feature as an Inverse Document Frequency (IDF) heuristic (Fang et al., 2004) . Similarly, prominence of a term is recognized by the frequency of the term in its local context, formulated as Term Frequency (TF) in IR.", "cite_spans": [ { "start": 718, "end": 737, "text": "(Fang et al., 2004)", "ref_id": "BIBREF2" } ], "ref_spans": [], "eq_spans": [], "section": "Characteristics of Good Sentiment Terms", "sec_num": "3.1" }, { "text": "If a topic of the text is known, terms that are relevant and descriptive of the subject should be regarded to be more useful than topically-irrelevant and extraneous terms. One way of measuring this is using associations between the query and terms. Statistical measures of associations between terms include estimations by the co-occurrence in the whole collection, such as Point-wise Mutual Information (PMI) and Latent Semantic Analysis (LSA). Another method is to use proximal information of the query and the word, using syntactic structure such as dependency relations of words that provide the graphical representation of the text (Mullen and Collier, 2004) . The minimum spans of words in such graph may represent their associations in the text. Also, the distance between words in the local context or in the thesauruslike dictionaries such as WordNet may be approximated as such measure.", "cite_spans": [ { "start": 638, "end": 664, "text": "(Mullen and Collier, 2004)", "ref_id": "BIBREF14" } ], "ref_spans": [], "eq_spans": [], "section": "Characteristics of Good Sentiment Terms", "sec_num": "3.1" }, { "text": "The goal of an opinion retrieval system is to find a set of opinionated documents that are relevant to a given topic. We decompose the opinion retrieval system into two tasks: the topical retrieval task and the sentiment analysis task. This two-stage approach for opinion retrieval has been taken by many systems and has been shown to perform well . The topic and the sentiment aspects of the opinion retrieval task are modeled separately, and linearly combined together to produce a list of topically-relevant and opinionated documents as below.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "Score OpRet (D, Q) = \u03bb\u2022Score rel (D, Q)+(1\u2212\u03bb)\u2022Scoreop(D, Q)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "The topic-relevance model Score rel may be substituted by any IR system that retrieves relevant documents for the query Q. For tasks such as NTCIR MOAT, relevant documents are already known in advance and it becomes unnecessary to estimate the relevance degree of the documents. We focus on modeling the sentiment aspect of the opinion retrieval task, assuming that the topicrelevance of documents is provided in some way.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "To assign documents with sentiment degrees, we estimate the probability of a document D to generate a query Q and to possess opinions as indicated by a random variable Op. 1 Assuming uniform prior probabilities of documents D, query Q, and Op, and conditional independence between Q and Op, the opinion score function reduces to estimating the generative probability of Q and Op given D.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "Scoreop(D, Q) \u2261 p(D | Op, Q) \u221d p(Op, Q | D)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "If we regard that the document D is represented as a bag of words and that the words are uniformly distributed, then", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "p(Op, Q | D) = X w\u2208D p(Op, Q | w) \u2022 p(w | D) = X w\u2208D p(Op | w) \u2022 p(Q | w) \u2022 p(w | D)", "eq_num": "(1)" } ], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "Equation 1 consists of three factors: the probability of a word to be opinionated (P (Op|w)), the likelihood of a query given a word (P (Q|w)), and the probability of a document generating a word (P (w|D)). Intuitively speaking, the probability of a document embodying topically related opinion is estimated by accumulating the probabilities of all words from the document to have sentiment meanings and associations with the given query.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "In the following sections, we assess the three factors of the sentiment models from the perspectives of term weighting.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retrieval Model", "sec_num": "3.2" }, { "text": "Modeling the sentiment of a word has been a popular approach in sentiment analysis. There are many publicly available lexicon resources. The size, format, specificity, and reliability differ in all these lexicons. For example, lexicon sizes range from a few hundred to several hundred thousand. Some lexicons assign real number scores to indicate sentiment orientations and strengths (i.e. probabilities of having positive and negative sentiments) (Esuli and Sebastiani, 2006) while other lexicons assign discrete classes (weak/strong, positive/negative) . There are manually compiled lexicons (Stone et al., 1966) while some are created semi-automatically by expanding a set of seed terms (Esuli and Sebastiani, 2006) .", "cite_spans": [ { "start": 448, "end": 476, "text": "(Esuli and Sebastiani, 2006)", "ref_id": "BIBREF1" }, { "start": 594, "end": 614, "text": "(Stone et al., 1966)", "ref_id": "BIBREF23" }, { "start": 690, "end": 718, "text": "(Esuli and Sebastiani, 2006)", "ref_id": "BIBREF1" } ], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "The goal of this paper is not to create or choose an appropriate sentiment lexicon, but rather it is to discover useful term features other than the sentiment properties. For this reason, one sentiment lexicon, namely SentiWordNet, is utilized throughout the whole experiment.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "SentiWordNet is an automatically generated sentiment lexicon using a semi-supervised method (Esuli and Sebastiani, 2006) . It consists of Word-Net synsets, where each synset is assigned three probability scores that add up to 1: positive, negative, and objective.", "cite_spans": [ { "start": 92, "end": 120, "text": "(Esuli and Sebastiani, 2006)", "ref_id": "BIBREF1" } ], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "These scores are assigned at sense level (synsets in WordNet), and we use the following equations to assess the sentiment scores at the word level.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "p(P os | w) = max s\u2208synset(w) SW N P os (s) p(N eg | w) = max s\u2208synset(w) SW N N eg (s) p(Op | w) = max (p(P os | w), p(N eg | w))", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "where synset(w) is the set of synsets of w and SW N P os (s), SW N N eg (s) are positive and negative scores of a synset in SentiWordNet. We assess the subjective score of a word as the maximum value of the positive and the negative scores, because a word has either a positive or a negative sentiment in a given context. The word sentiment model can also make use of other types of sentiment lexicons. The sub-jectivity lexicon used in OpinionFinder 2 is compiled from several manually and automatically built resources. Each word in the lexicon is tagged with the strength (strong/weak) and polarity (Positive/Negative/Neutral). The word sentiment can be modeled as below.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "P (P os|w) = 8 > < > :", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "1.0 if w is Positive and Strong 0.5 if w is Positive and Weak 0.0 otherwise", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "P (Op | w) = max (p(P os | w), p(N eg | w))", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Sentiment Model", "sec_num": "3.2.1" }, { "text": "If a topic is given in the sentiment analysis, terms that are closely associated with the topic should be assigned heavy weighting. For example, sentiment words such as scary and funny are more likely to be associated with topic words such as book and movie than grocery or refrigerator.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "In the topic association model, p(Q | w) is estimated from the associations between the word w and a set of query terms Q.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "p(Q | w) = P q\u2208Q Asc-Score(q, w) | Q | \u221d X q\u2208Q Asc-Score(q, w)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Asc-Score(q, w) is the association score between q and w, and | Q | is the number of query words.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "To measure associations between words, we employ statistical approaches using document collections such as LSA and PMI, and local proximity features using the distance in dependency trees or texts.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Latent Semantic Analysis (LSA) (Landauer and Dumais, 1997) creates a semantic space from a collection of documents to measure the semantic relatedness of words. Point-wise Mutual Information (PMI) is a measure of associations used in information theory, where the association between two words is evaluated with the joint and individual distributions of the two words. PMI-IR (Turney, 2001) uses an IR system and its search operators to estimate the probabilities of two terms and their conditional probabilities. Equations for association scores using LSA and PMI are given below.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Asc-ScoreLSA(w1, w2) = 1 + LSA(w1, w2) 2", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Asc-ScoreP M I (w1, w2) = 1 + P M I-IR(w1, w2) 2", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "For the experimental purpose, we used publicly available online demonstrations for LSA and PMI. For LSA, we used the online demonstration mode from the Latent Semantic Analysis page from the University of Colorado at Boulder. 3 For PMI, we used the online API provided by the CogWorks Lab at the Rensselaer Polytechnic Institute. 4 Word associations between two terms may also be evaluated in the local context where the terms appear together. One way of measuring the proximity of terms is using the syntactic structures. Given the dependency tree of the text, we model the association between two terms as below.", "cite_spans": [ { "start": 330, "end": 331, "text": "4", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Asc-Score DT P (w1, w2) = (", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "1.0 min. span in dep. tree \u2264 Dsyn 0.5 otherwise where, D syn is arbitrarily set to 3.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Another way is to use co-occurrence statistics as below.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Asc-Score W P (w1, w2) = ( 1.0 if distance betweenw1andw2 \u2264 K 0.5 otherwise", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "where K is the maximum window size for the co-occurrence and is arbitrarily set to 3 in our experiments.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "The statistical approaches may suffer from data sparseness problems especially for named entity terms used in the query, and the proximal clues cannot sufficiently cover all term-query associations. To avoid assigning zero probabilities, our topic association models assign 0.5 to word pairs with no association and 1.0 to words with perfect association.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Note that proximal features using co-occurrence and dependency relationships were used in previous work. For opinion retrieval tasks, Yang et al. (2006) and Zhang and Ye (2008) used the cooccurrence of a query word and a sentiment word within a certain window size. Mullen and Collier (2004) manually annotated named entities in their dataset (i.e. title of the record and name of the artist for music record reviews), and utilized presence and position features in their ML approach.", "cite_spans": [ { "start": 134, "end": 152, "text": "Yang et al. (2006)", "ref_id": "BIBREF31" }, { "start": 157, "end": 176, "text": "Zhang and Ye (2008)", "ref_id": "BIBREF34" } ], "ref_spans": [], "eq_spans": [], "section": "Topic Association Model", "sec_num": "3.2.2" }, { "text": "Our word generation model p(w | d) evaluates the prominence and the discriminativeness of a word w in a document d. These issues correspond to the core issues of traditional IR tasks. IR models, such as Vector Space (VS), probabilistic models such as BM25, and Language Modeling (LM), albeit in different forms of approach and measure, employ heuristics and formal modeling approaches to effectively evaluate the relevance of a term to a document (Fang et al., 2004) . Therefore, we estimate the word generation model with popular IR models' the relevance scores of a document d given w as a query. 5", "cite_spans": [ { "start": 447, "end": 466, "text": "(Fang et al., 2004)", "ref_id": "BIBREF2" } ], "ref_spans": [], "eq_spans": [], "section": "Word Generation Model", "sec_num": "3.2.3" }, { "text": "p(w | d) \u2261 IR-SCORE(w, d)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Generation Model", "sec_num": "3.2.3" }, { "text": "In our experiments, we use the Vector Space model with Pivoted Normalization (VS), Probabilistic model (BM25), and Language modeling with Dirichlet Smoothing (LM). ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Generation Model", "sec_num": "3.2.3" }, { "text": "V SP N (w, d) = 1 + ln(1 + ln(c(w, d))) (1 \u2212 s) + s \u2022 | d | avgdl \u2022 ln N + 1 df (w) BM 25(w, d) = ln N \u2212 df (w) + 0.5 df (w) + 0.5 \u2022 (k 1 + 1) \u2022 c(w, d) k 1 \" (1 \u2212 b) + b |d| avgdl \" + c(w, d) LM DI(w, d) = ln 1 + c(w, d) \u00b5 \u2022 c(w, C ) ! + ln \u00b5 | d | +\u00b5 c(w, d) is the frequency of w in d, | d | is", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Generation Model", "sec_num": "3.2.3" }, { "text": "To verify the effectiveness of our term weighting schemes in experimental settings of the datadriven approach, we carry out a set of simple experiments with ML classifiers. Specifically, we explore the statistical term weighting features of the word generation model with Support Vector machine (SVM), faithfully reproducing previous work as closely as possible (Pang et al., 2002) . Each instance of train and test data is represented as a vector of features. We test various combinations of the term weighting schemes listed below.", "cite_spans": [ { "start": 362, "end": 381, "text": "(Pang et al., 2002)", "ref_id": "BIBREF20" } ], "ref_spans": [], "eq_spans": [], "section": "Data-driven Approach", "sec_num": "3.3" }, { "text": "\u2022 PRESENCE: binary indicator for the presence of a term ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Data-driven Approach", "sec_num": "3.3" }, { "text": "Our experiments consist of an opinion retrieval task and a sentiment classification task. We use MPQA and movie-review corpora in our experiments with an ML classifier. For the opinion retrieval task, we use the two datasets used by TREC blog track and NTCIR MOAT evaluation workshops.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experiment", "sec_num": "4" }, { "text": "The opinion retrieval task at TREC Blog Track consists of three subtasks: topic retrieval, opinion retrieval, and polarity retrieval. Opinion and polarity retrieval subtasks use the relevant documents retrieved at the topic retrieval stage. On the other hand, the NTCIR MOAT task aims to find opinionated sentences given a set of documents that are already hand-assessed to be relevant to the topic. Task (Nam et al., 2009) . While our sentiment analysis model uses the entire relevant portion of the blog posts, further stopword removal and stemming is done for the blog retrieval system.", "cite_spans": [ { "start": 405, "end": 423, "text": "(Nam et al., 2009)", "ref_id": "BIBREF15" } ], "ref_spans": [ { "start": 400, "end": 404, "text": "Task", "ref_id": null } ], "eq_spans": [], "section": "Experiment", "sec_num": "4" }, { "text": "For the relevance retrieval model, we faithfully reproduce the passage-based language model with pseudo-relevance feedback .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retieval", "sec_num": "4.1" }, { "text": "We use in total 100 topics from TREC 2007 and 2008 blog opinion retrieval tasks (07:901-950 and 08:1001-1050). We use the topics from Blog 07 to optimize the parameter for linearly combining the retrieval and opinion models, and use Blog 08 topics as our test data. Topics are extracted only from the Title field, using the Porter stemmer and a stopword list. ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Opinion Retieval", "sec_num": "4.1" }, { "text": "Retrieval performances using different combinations of term weighting features are presented in Table 1 . Using only the word sentiment model is set as our baseline.", "cite_spans": [], "ref_spans": [ { "start": 96, "end": 103, "text": "Table 1", "ref_id": "TABREF2" } ], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.1.2" }, { "text": "First, each feature of the word generation and topic association models are tested; all features of the models improve over the baseline. We observe that the features of our word generation model is more effective than those of the topic association model. Among the features of the word generation model, the most improvement was achieved with BM 25, improving the MAP by 2.27%.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.1.2" }, { "text": "Features of the topic association model show only moderate improvements over the baseline. We observe that these features generally improve P@10 performance, indicating that they increase the accuracy of the sentiment analysis system. PMI out-performed LSA for all evaluation measures. Among the topic association models, PMI performs the best in MAP and R-prec, while WP achieved the biggest improvement in P@10.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.1.2" }, { "text": "Since BM25 performs the best among the word generation models, its combination with other features was investigated. Combinations of BM25 with the topic association models all improve the performance of the baseline and BM25. This demonstrates that the word generation model and the topic association model are complementary to each other.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.1.2" }, { "text": "The best MAP was achieved with BM25, PMI, and WP (+2.44% over the baseline). We observe that PMI and WP also complement each other.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.1.2" }, { "text": "Another set of experiments for our opinion analysis model was carried out on the NTCIR-7 MOAT English corpus. The English opinion corpus for NTCIR MOAT consists of newspaper articles from the Mainichi Daily News, Korea Times, Xinhua News, Hong Kong Standard, and the Straits Times. It is a collection of documents manually assessed for relevance to a set of queries from NTCIR-7 Advanced Cross-lingual Information Access (ACLIA) task. The corpus consists of 167 documents, or 4,711 sentences for 14 test topics. Each sentence is manually tagged with opinionatedness, polarity, and relevance to the topic by three annotators from a pool of six annotators. For preprocessing, no removal or stemming is performed on the data. Each sentence was processed with the Stanford English parser 6 to produce a dependency parse tree. Only the Title fields of the topics were used.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setting", "sec_num": "4.2.1" }, { "text": "For performance evaluations of opinion and polarity detection, we use precision, recall, and Fmeasure, the same measure used to report the official results at the NTCIR MOAT workshop. There are lenient and strict evaluations depending on the agreement of the annotators; if two out of three annotators agreed upon an opinion or polarity annotation then it is used during the lenient evaluation, similarly three out of three agreements are used during the strict evaluation. We present the performances using the lenient evaluation only, for the two evaluations generally do not show much difference in relative performance changes.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setting", "sec_num": "4.2.1" }, { "text": "Since MOAT is a classification task, we use a threshold parameter to draw a boundary between opinionated and non-opinionated sentences. We report the performance of our system using the NTCIR-7 dataset, where the threshold parameter is optimized using the NTCIR-6 dataset.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setting", "sec_num": "4.2.1" }, { "text": "We present the performance of our sentiment analysis system in Table 2 . As in the experiments with Similarly to the TREC experiments, the features of the word generation model perform exceptionally better than that of the topic association model. The best performing feature of the word generation model is VS, achieving a 4.21% improvement over the baseline's f-measure. Interestingly, this is the tied top performing f-measure over all combinations of our features.", "cite_spans": [], "ref_spans": [ { "start": 63, "end": 70, "text": "Table 2", "ref_id": "TABREF3" } ], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.2.2" }, { "text": "While LSA and DTP show mild improvements, PMI performed worse than baseline, with higher precision but a drop in recall. DTP was the best performing topic association model. When combining the best performing feature of the word generation model (VS) with the features of the topic association model, LSA, PMI and DTP all performed worse than or as well as the VS in f-measure evaluation. LSA and DTP improves precision slightly, but with a drop in recall. PMI shows the opposite tendency. The best performing system was achieved using VS, LSA and DTP at both precision and f-measure evaluations.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.2.2" }, { "text": "To test our SVM classifier, we perform the classification task. Movie Review polarity dataset 7 was .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setting", "sec_num": "4.3.1" }, { "text": "To closely reproduce the experiment with the best performance carried out in (Pang et al., 2002) using SVM, we use unigram with the presence feature. We test various combinations of our features applicable to the task. For evaluation, we use ten-fold cross-validation accuracy.", "cite_spans": [ { "start": 77, "end": 96, "text": "(Pang et al., 2002)", "ref_id": "BIBREF20" } ], "ref_spans": [], "eq_spans": [], "section": "Experimental Setting", "sec_num": "4.3.1" }, { "text": "We present the sentiment classification performances in Table 3 .", "cite_spans": [], "ref_spans": [ { "start": 56, "end": 63, "text": "Table 3", "ref_id": "TABREF4" } ], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.3.2" }, { "text": "As observed by Pang et al. (2002) , using the raw tf drops the accuracy of the sentiment classification (-13.92%) of movie-review data. Using the raw idf feature worsens the accuracy even more (-25.42%). Normalized tf-variants show improvements over tf but are worse than presence. Normalized idf features produce slightly better accuracy results than the baseline. Finally, combining any normalized tf and idf features improved the baseline (high 83% \u223c low 85%). The best combination was BM25.TF\u2022VS.IDF.", "cite_spans": [ { "start": 15, "end": 33, "text": "Pang et al. (2002)", "ref_id": "BIBREF20" } ], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.3.2" }, { "text": "MPQA corpus reveals similar but somewhat uncertain tendency.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Result", "sec_num": "4.3.2" }, { "text": "Overall, the opinion retrieval and the sentiment analysis models achieve improvements using our proposed features. Especially, the features of the word generation model improve the overall performances drastically. Its effectiveness is also verified with a data-driven approach; the accuracy of a sentiment classifier trained on a polarity dataset was improved by various combinations of normalized tf and idf statistics.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "4.4" }, { "text": "Differences in effectiveness of VS, BM25, and LM come from parameter tuning and corpus differences. For the TREC dataset, BM25 performed better than the other models, and for the NTCIR dataset, VS performed better.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "4.4" }, { "text": "Our features of the topic association model show mild improvement over the baseline performance in general. PMI and LSA, both modeling the semantic associations between words, show different behaviors on the datasets. For the NT-CIR dataset, LSA performed better, while PMI is more effective for the TREC dataset. We believe that the explanation lies in the differences between the topics for each dataset. In general, the NTCIR topics are general descriptive words such as \"regenerative medicine\", \"American economy after the 911 terrorist attacks\", and \"lawsuit brought against Microsoft for monopolistic practices.\" The TREC topics are more namedentity-like terms such as \"Carmax\", \"Wikipedia primary source\", \"Jiffy Lube\", \"Starbucks\", and \"Windows Vista.\" We have experimentally shown that LSA is more suited to finding associations between general terms because its training documents are from a general domain. 9 Our PMI measure utilizes a web search engine, which covers a variety of named entity terms.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "4.4" }, { "text": "Though the features of our topic association model, WP and DTP, were evaluated on different datasets, we try our best to conjecture the differences. WP on TREC dataset shows a small improvement of MAP compared to other topic association features, while the precision is improved the most when this feature is used alone. The DTP feature displays similar behavior with precision. It also achieves the best f-measure over other topic association features. DTP achieves higher relative improvement (3.99% F-measure verse 2.32% MAP), and is more effective for improving the performance in combination with LSA and PMI. 9 TASA Corpus, http://lsa.colorado.edu/spaces.html", "cite_spans": [ { "start": 615, "end": 616, "text": "9", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "Discussion", "sec_num": "4.4" }, { "text": "In this paper, we proposed various term weighting schemes and how such features are modeled in the sentiment analysis task. Our proposed features include corpus statistics, association measures using semantic and local-context proximities. We have empirically shown the effectiveness of the features with our proposed opinion retrieval and sentiment analysis models.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion", "sec_num": "5" }, { "text": "There exists much room for improvement with further experiments with various term weighting methods and datasets. Such methods include, but by no means limited to, semantic similarities between word pairs using lexical resources such as WordNet (Miller, 1995) and data-driven methods with various topic-dependent term weighting schemes on labeled corpus with topics such as MPQA.", "cite_spans": [ { "start": 245, "end": 259, "text": "(Miller, 1995)", "ref_id": "BIBREF13" } ], "ref_spans": [], "eq_spans": [], "section": "Conclusion", "sec_num": "5" }, { "text": "Throughout this paper, Op indicates Op = 1.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "http://www.cs.pitt.edu/mpqa/", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "http://lsa.colorado.edu/, default parameter settings for the semantic space (TASA, 1st year college level) and number of factors (300).4 http://cwl-projects.cogsci.rpi.edu/msr/, PMI-IR with the Google Search Engine.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "With proper assumptions and derivations, p(w | d) can be derived to language modeling approaches. Refer to(Zhai and Lafferty, 2004).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "http://nlp.stanford.edu/software/lex-parser.shtml", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "http://www.cs.cornell.edu/people/pabo/movie-reviewdata/", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "http://www.cs.pitt.edu/mpqa/databaserelease/", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "This work was supported in part by MKE & IITA through IT Leading R&D Support Project and in part by the BK 21 Project in 2009.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgments", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Mining the peanut gallery: Opinion extraction and semantic classification of product reviews", "authors": [ { "first": "Kushal", "middle": [], "last": "Dave", "suffix": "" }, { "first": "Steve", "middle": [], "last": "Lawrence", "suffix": "" }, { "first": "David", "middle": [ "M" ], "last": "Pennock", "suffix": "" } ], "year": 2003, "venue": "Proceedings of WWW", "volume": "", "issue": "", "pages": "519--528", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kushal Dave, Steve Lawrence, and David M. Pennock. 2003. Mining the peanut gallery: Opinion extraction and seman- tic classification of product reviews. In Proceedings of WWW, pages 519-528.", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "Sentiwordnet: A publicly available lexical resource for opinion mining", "authors": [ { "first": "Andrea", "middle": [], "last": "Esuli", "suffix": "" }, { "first": "Fabrizio", "middle": [], "last": "Sebastiani", "suffix": "" } ], "year": 2006, "venue": "Proceedings of the 5th Conference on Language Resources and Evaluation (LREC'06)", "volume": "", "issue": "", "pages": "417--422", "other_ids": {}, "num": null, "urls": [], "raw_text": "Andrea Esuli and Fabrizio Sebastiani. 2006. Sentiword- net: A publicly available lexical resource for opinion min- ing. In Proceedings of the 5th Conference on Language Resources and Evaluation (LREC'06), pages 417-422, Geneva, IT.", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "A formal study of information retrieval heuristics", "authors": [ { "first": "Hui", "middle": [], "last": "Fang", "suffix": "" }, { "first": "Tao", "middle": [], "last": "Tao", "suffix": "" }, { "first": "Chengxiang", "middle": [], "last": "Zhai", "suffix": "" } ], "year": 2004, "venue": "SIGIR '04: Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval", "volume": "", "issue": "", "pages": "49--56", "other_ids": {}, "num": null, "urls": [], "raw_text": "Hui Fang, Tao Tao, and ChengXiang Zhai. 2004. A formal study of information retrieval heuristics. In SIGIR '04: Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval, pages 49-56, New York, NY, USA. ACM.", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "An extensive empirical study of feature selection metrics for text classification", "authors": [ { "first": "George", "middle": [], "last": "Forman", "suffix": "" } ], "year": 2003, "venue": "Journal of Machine Learning Research", "volume": "3", "issue": "", "pages": "1289--1305", "other_ids": {}, "num": null, "urls": [], "raw_text": "George Forman. 2003. An extensive empirical study of fea- ture selection metrics for text classification. Journal of Machine Learning Research, 3:1289-1305.", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "Sentiment classification on customer feedback data: noisy data, large feature vectors, and the role of linguistic analysis", "authors": [ { "first": "Michael", "middle": [], "last": "Gamon", "suffix": "" } ], "year": 2004, "venue": "Proceedings of the International Conference on Computational Linguistics (COL-ING)", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Michael Gamon. 2004. Sentiment classification on customer feedback data: noisy data, large feature vectors, and the role of linguistic analysis. In Proceedings of the Inter- national Conference on Computational Linguistics (COL- ING).", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Predicting the semantic orientation of adjectives", "authors": [ { "first": "Vasileios", "middle": [], "last": "Hatzivassiloglou", "suffix": "" }, { "first": "Kathleen", "middle": [ "R" ], "last": "Mckeown", "suffix": "" } ], "year": 1997, "venue": "Proceedings of the 35th Annual Meeting of the Association for Computational Linguistics (ACL'97)", "volume": "", "issue": "", "pages": "174--181", "other_ids": {}, "num": null, "urls": [], "raw_text": "Vasileios Hatzivassiloglou and Kathleen R. Mckeown. 1997. Predicting the semantic orientation of adjectives. In Pro- ceedings of the 35th Annual Meeting of the Association for Computational Linguistics (ACL'97), pages 174-181, madrid, ES.", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "Using wordnet to measure semantic orientation of adjectives", "authors": [ { "first": "Jaap", "middle": [], "last": "Kamps", "suffix": "" }, { "first": "Maarten", "middle": [], "last": "Marx", "suffix": "" }, { "first": "Robert", "middle": [ "J" ], "last": "Mokken", "suffix": "" }, { "first": "Maarten", "middle": [], "last": "De Rijke", "suffix": "" } ], "year": 2004, "venue": "Proceedings of the 4th International Conference on Language Resources and Evaluation (LREC'04)", "volume": "", "issue": "", "pages": "1115--1118", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jaap Kamps, Maarten Marx, Robert J. Mokken, and Maarten De Rijke. 2004. Using wordnet to measure se- mantic orientation of adjectives. In Proceedings of the 4th International Conference on Language Resources and Evaluation (LREC'04), pages 1115-1118, Lisbon, PT.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "A boosting algorithm for classification of semi-structured text", "authors": [ { "first": "Taku", "middle": [], "last": "Kudo", "suffix": "" }, { "first": "Yuji", "middle": [], "last": "Matsumoto", "suffix": "" } ], "year": 2004, "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Taku Kudo and Yuji Matsumoto. 2004. A boosting algorithm for classification of semi-structured text. In Proceedings of the Conference on Empirical Methods in Natural Lan- guage Processing (EMNLP).", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "A solution to plato's problem: The latent semantic analysis theory of acquisition, induction, and representation of knowledge", "authors": [ { "first": "K", "middle": [], "last": "Thomas", "suffix": "" }, { "first": "Susan", "middle": [ "T" ], "last": "Landauer", "suffix": "" }, { "first": "", "middle": [], "last": "Dumais", "suffix": "" } ], "year": 1997, "venue": "Psychological Review", "volume": "104", "issue": "2", "pages": "211--240", "other_ids": {}, "num": null, "urls": [], "raw_text": "Thomas K. Landauer and Susan T. Dumais. 1997. A solution to plato's problem: The latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychological Review, 104(2):211-240, April.", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "Kle at trec 2008 blog track: Blog post and feed retrieval", "authors": [ { "first": "Yeha", "middle": [], "last": "Lee", "suffix": "" }, { "first": "", "middle": [], "last": "Seung-Hoon", "suffix": "" }, { "first": "Jungi", "middle": [], "last": "Na", "suffix": "" }, { "first": "Sang-Hyob", "middle": [], "last": "Kim", "suffix": "" }, { "first": "Hun", "middle": [], "last": "Nam", "suffix": "" }, { "first": "Jung", "middle": [], "last": "Young", "suffix": "" }, { "first": "Jong-Hyeok", "middle": [], "last": "Lee", "suffix": "" } ], "year": 2008, "venue": "Proceedings of TREC-08", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yeha Lee, Seung-Hoon Na, Jungi Kim, Sang-Hyob Nam, Hun young Jung, and Jong-Hyeok Lee. 2008. Kle at trec 2008 blog track: Blog post and feed retrieval. In Proceed- ings of TREC-08.", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "The TREC Blogs06 collection: creating and analysing a blog test collection", "authors": [ { "first": "Craig", "middle": [], "last": "Macdonald", "suffix": "" }, { "first": "Iadh", "middle": [], "last": "Ounis", "suffix": "" } ], "year": 2006, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Craig Macdonald and Iadh Ounis. 2006. The TREC Blogs06 collection: creating and analysing a blog test collection. Technical Report TR-2006-224, Department of Computer Science, University of Glasgow.", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "Sentiment classification using word subsequences and dependency sub-trees", "authors": [ { "first": "Shotaro", "middle": [], "last": "Matsumoto", "suffix": "" }, { "first": "Hiroya", "middle": [], "last": "Takamura", "suffix": "" }, { "first": "Manabu", "middle": [], "last": "Okumura", "suffix": "" } ], "year": 2005, "venue": "Proceedings of PAKDD'05, the 9th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Shotaro Matsumoto, Hiroya Takamura, and Manabu Oku- mura. 2005. Sentiment classification using word sub- sequences and dependency sub-trees. In Proceedings of PAKDD'05, the 9th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining.", "links": null }, "BIBREF12": { "ref_id": "b12", "title": "Topic sentiment mixture: Modeling facets and opinions in weblogs", "authors": [ { "first": "Qiaozhu", "middle": [], "last": "Mei", "suffix": "" }, { "first": "Xu", "middle": [], "last": "Ling", "suffix": "" }, { "first": "Matthew", "middle": [], "last": "Wondra", "suffix": "" }, { "first": "Hang", "middle": [], "last": "Su", "suffix": "" }, { "first": "Chengxiang", "middle": [], "last": "Zhai", "suffix": "" } ], "year": 2007, "venue": "Proceedings of WWW", "volume": "", "issue": "", "pages": "171--180", "other_ids": {}, "num": null, "urls": [], "raw_text": "Qiaozhu Mei, Xu Ling, Matthew Wondra, Hang Su, and ChengXiang Zhai. 2007. Topic sentiment mixture: Mod- eling facets and opinions in weblogs. In Proceedings of WWW, pages 171-180, New York, NY, USA. ACM Press.", "links": null }, "BIBREF13": { "ref_id": "b13", "title": "Wordnet: a lexical database for english", "authors": [ { "first": "A", "middle": [], "last": "George", "suffix": "" }, { "first": "", "middle": [], "last": "Miller", "suffix": "" } ], "year": 1995, "venue": "Commun. ACM", "volume": "38", "issue": "11", "pages": "39--41", "other_ids": {}, "num": null, "urls": [], "raw_text": "George A. Miller. 1995. Wordnet: a lexical database for english. Commun. ACM, 38(11):39-41.", "links": null }, "BIBREF14": { "ref_id": "b14", "title": "Sentiment analysis using support vector machines with diverse information sources", "authors": [ { "first": "Tony", "middle": [], "last": "Mullen", "suffix": "" }, { "first": "Nigel", "middle": [], "last": "Collier", "suffix": "" } ], "year": 2004, "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", "volume": "", "issue": "", "pages": "412--418", "other_ids": {}, "num": null, "urls": [], "raw_text": "Tony Mullen and Nigel Collier. 2004. Sentiment analysis using support vector machines with diverse information sources. In Proceedings of the Conference on Empiri- cal Methods in Natural Language Processing (EMNLP), pages 412-418, July. Poster paper.", "links": null }, "BIBREF15": { "ref_id": "b15", "title": "Diffpost: Filtering non-relevant content based on content difference between two consecutive blog posts", "authors": [ { "first": "Sang-Hyob", "middle": [], "last": "Nam", "suffix": "" }, { "first": "Seung-Hoon", "middle": [], "last": "Na", "suffix": "" }, { "first": "Yeha", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Jong-Hyeok", "middle": [], "last": "Lee", "suffix": "" } ], "year": 2009, "venue": "ECIR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Sang-Hyob Nam, Seung-Hoon Na, Yeha Lee, and Jong- Hyeok Lee. 2009. Diffpost: Filtering non-relevant con- tent based on content difference between two consecutive blog posts. In ECIR.", "links": null }, "BIBREF16": { "ref_id": "b16", "title": "Examining the role of linguistic knowledge sources in the automatic identification and classification of reviews", "authors": [ { "first": "Vincent", "middle": [], "last": "Ng", "suffix": "" }, { "first": "Sajib", "middle": [], "last": "Dasgupta", "suffix": "" }, { "first": "S", "middle": [ "M" ], "last": "", "suffix": "" } ], "year": 2006, "venue": "Proceedings of the COLING/ACL Main Conference Poster Sessions", "volume": "", "issue": "", "pages": "611--618", "other_ids": {}, "num": null, "urls": [], "raw_text": "Vincent Ng, Sajib Dasgupta, and S. M. Niaz Arifin. 2006. Examining the role of linguistic knowledge sources in the automatic identification and classification of reviews. In Proceedings of the COLING/ACL Main Conference Poster Sessions, pages 611-618, Sydney, Australia, July. Associ- ation for Computational Linguistics.", "links": null }, "BIBREF17": { "ref_id": "b17", "title": "Overview of the trec-2006 blog track", "authors": [ { "first": "I", "middle": [], "last": "Ounis", "suffix": "" }, { "first": "M", "middle": [], "last": "De Rijke", "suffix": "" }, { "first": "C", "middle": [], "last": "Macdonald", "suffix": "" }, { "first": "G", "middle": [ "A" ], "last": "Mishne", "suffix": "" }, { "first": "I", "middle": [], "last": "Soboroff", "suffix": "" } ], "year": 2006, "venue": "Proceedings of TREC-06", "volume": "", "issue": "", "pages": "15--27", "other_ids": {}, "num": null, "urls": [], "raw_text": "I. Ounis, M. de Rijke, C. Macdonald, G. A. Mishne, and I. Soboroff. 2006. Overview of the trec-2006 blog track. In Proceedings of TREC-06, pages 15-27, November.", "links": null }, "BIBREF18": { "ref_id": "b18", "title": "Overview of the trec-2008 blog track", "authors": [ { "first": "I", "middle": [], "last": "Ounis", "suffix": "" }, { "first": "C", "middle": [], "last": "Macdonald", "suffix": "" }, { "first": "I", "middle": [], "last": "Soboroff", "suffix": "" } ], "year": 2008, "venue": "Proceedings of TREC-08", "volume": "", "issue": "", "pages": "15--27", "other_ids": {}, "num": null, "urls": [], "raw_text": "I. Ounis, C. Macdonald, and I. Soboroff. 2008. Overview of the trec-2008 blog track. In Proceedings of TREC-08, pages 15-27, November.", "links": null }, "BIBREF19": { "ref_id": "b19", "title": "Opinion mining and sentiment analysis. Foundations and Trends in Information Retrieval", "authors": [ { "first": "Bo", "middle": [], "last": "Pang", "suffix": "" }, { "first": "Lillian", "middle": [], "last": "Lee", "suffix": "" } ], "year": 2008, "venue": "", "volume": "2", "issue": "", "pages": "1--135", "other_ids": {}, "num": null, "urls": [], "raw_text": "Bo Pang and Lillian Lee. 2008. Opinion mining and sen- timent analysis. Foundations and Trends in Information Retrieval, 2(1-2):1-135.", "links": null }, "BIBREF20": { "ref_id": "b20", "title": "Thumbs up? Sentiment classification using machine learning techniques", "authors": [ { "first": "Bo", "middle": [], "last": "Pang", "suffix": "" }, { "first": "Lillian", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Shivakumar", "middle": [], "last": "Vaithyanathan", "suffix": "" } ], "year": 2002, "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", "volume": "", "issue": "", "pages": "79--86", "other_ids": {}, "num": null, "urls": [], "raw_text": "Bo Pang, Lillian Lee, and Shivakumar Vaithyanathan. 2002. Thumbs up? Sentiment classification using machine learning techniques. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 79-86.", "links": null }, "BIBREF21": { "ref_id": "b21", "title": "Machine learning in automated text categorization", "authors": [ { "first": "Fabrizio", "middle": [], "last": "Sebastiani", "suffix": "" } ], "year": 2002, "venue": "ACM Computing Surveys", "volume": "34", "issue": "1", "pages": "1--47", "other_ids": {}, "num": null, "urls": [], "raw_text": "Fabrizio Sebastiani. 2002. Machine learning in automated text categorization. ACM Computing Surveys, 34(1):1-47.", "links": null }, "BIBREF22": { "ref_id": "b22", "title": "Overview of multilingual opinion analysis task at ntcir-7", "authors": [ { "first": "Yohei", "middle": [], "last": "Seki", "suffix": "" }, { "first": "David", "middle": [ "Kirk" ], "last": "Evans", "suffix": "" }, { "first": "Lun-Wei", "middle": [], "last": "Ku", "suffix": "" }, { "first": "Le", "middle": [], "last": "Sun", "suffix": "" }, { "first": "Hsin-Hsi", "middle": [], "last": "Chen", "suffix": "" }, { "first": "Noriko", "middle": [], "last": "Kando", "suffix": "" } ], "year": 2007, "venue": "Evaluation of Information Access Technologies: Information Retrieval, Question Answering and Cross-Lingual Information Access", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yohei Seki, David Kirk Evans, Lun-Wei Ku, Le Sun, Hsin- Hsi Chen, and Noriko Kando. 2008. Overview of mul- tilingual opinion analysis task at ntcir-7. In Proceedings of The 7th NTCIR Workshop (2007/2008) -Evaluation of Information Access Technologies: Information Retrieval, Question Answering and Cross-Lingual Information Ac- cess.", "links": null }, "BIBREF23": { "ref_id": "b23", "title": "The General Inquirer: A Computer Approach to Content Analysis", "authors": [ { "first": "Philip", "middle": [ "J" ], "last": "Stone", "suffix": "" }, { "first": "Dexter", "middle": [ "C" ], "last": "Dunphy", "suffix": "" }, { "first": "Marshall", "middle": [ "S" ], "last": "Smith", "suffix": "" }, { "first": "Daniel", "middle": [ "M" ], "last": "Ogilvie", "suffix": "" } ], "year": 1966, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Philip J. Stone, Dexter C. Dunphy, Marshall S. Smith, and Daniel M. Ogilvie. 1966. The General Inquirer: A Com- puter Approach to Content Analysis. MIT Press, Cam- bridge, USA.", "links": null }, "BIBREF24": { "ref_id": "b24", "title": "Measuring praise and criticism: Inference of semantic orientation from association", "authors": [ { "first": "D", "middle": [], "last": "Peter", "suffix": "" }, { "first": "Michael", "middle": [ "L" ], "last": "Turney", "suffix": "" }, { "first": "", "middle": [], "last": "Littman", "suffix": "" } ], "year": 2003, "venue": "ACM Transactions on Information Systems", "volume": "21", "issue": "4", "pages": "315--346", "other_ids": {}, "num": null, "urls": [], "raw_text": "Peter D. Turney and Michael L. Littman. 2003. Measur- ing praise and criticism: Inference of semantic orientation from association. ACM Transactions on Information Sys- tems, 21(4):315-346.", "links": null }, "BIBREF25": { "ref_id": "b25", "title": "Mining the web for synonyms: Pmiir versus lsa on toefl", "authors": [ { "first": "D", "middle": [], "last": "Peter", "suffix": "" }, { "first": "", "middle": [], "last": "Turney", "suffix": "" } ], "year": 2001, "venue": "EMCL '01: Proceedings of the 12th European Conference on Machine Learning", "volume": "", "issue": "", "pages": "491--502", "other_ids": {}, "num": null, "urls": [], "raw_text": "Peter D. Turney. 2001. Mining the web for synonyms: Pmi- ir versus lsa on toefl. In EMCL '01: Proceedings of the 12th European Conference on Machine Learning, pages 491-502, London, UK. Springer-Verlag.", "links": null }, "BIBREF26": { "ref_id": "b26", "title": "Using appraisal groups for sentiment analysis", "authors": [ { "first": "Casey", "middle": [], "last": "Whitelaw", "suffix": "" }, { "first": "Navendu", "middle": [], "last": "Garg", "suffix": "" }, { "first": "Shlomo", "middle": [], "last": "Argamon", "suffix": "" } ], "year": 2005, "venue": "Proceedings of the 14th ACM international conference on Information and knowledge management (CIKM'05)", "volume": "", "issue": "", "pages": "625--631", "other_ids": {}, "num": null, "urls": [], "raw_text": "Casey Whitelaw, Navendu Garg, and Shlomo Argamon. 2005. Using appraisal groups for sentiment analysis. In Proceedings of the 14th ACM international conference on Information and knowledge management (CIKM'05), pages 625-631, Bremen, DE.", "links": null }, "BIBREF27": { "ref_id": "b27", "title": "Recognizing and organizing opinions expressed in the world press", "authors": [ { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" }, { "first": "E", "middle": [], "last": "Breck", "suffix": "" }, { "first": "Christopher", "middle": [], "last": "Buckley", "suffix": "" }, { "first": "Claire", "middle": [], "last": "Cardie", "suffix": "" }, { "first": "P", "middle": [], "last": "Davis", "suffix": "" }, { "first": "B", "middle": [], "last": "Fraser", "suffix": "" }, { "first": "Diane", "middle": [], "last": "Litman", "suffix": "" }, { "first": "D", "middle": [], "last": "Pierce", "suffix": "" }, { "first": "Ellen", "middle": [], "last": "Riloff", "suffix": "" }, { "first": "Theresa", "middle": [], "last": "Wilson", "suffix": "" }, { "first": "D", "middle": [], "last": "Day", "suffix": "" }, { "first": "Mark", "middle": [], "last": "Maybury", "suffix": "" } ], "year": 2003, "venue": "Proceedings of the 2003 AAAI Spring Symposium on New Directions in Question Answering", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Janyce Wiebe, E. Breck, Christopher Buckley, Claire Cardie, P. Davis, B. Fraser, Diane Litman, D. Pierce, Ellen Riloff, Theresa Wilson, D. Day, and Mark Maybury. 2003. Rec- ognizing and organizing opinions expressed in the world press. In Proceedings of the 2003 AAAI Spring Sympo- sium on New Directions in Question Answering.", "links": null }, "BIBREF28": { "ref_id": "b28", "title": "Learning subjective language", "authors": [ { "first": "Janyce", "middle": [ "M" ], "last": "Wiebe", "suffix": "" }, { "first": "Theresa", "middle": [], "last": "Wilson", "suffix": "" }, { "first": "Rebecca", "middle": [], "last": "Bruce", "suffix": "" }, { "first": "Matthew", "middle": [], "last": "Bell", "suffix": "" }, { "first": "Melanie", "middle": [], "last": "Martin", "suffix": "" } ], "year": 2004, "venue": "Computational Linguistics", "volume": "30", "issue": "3", "pages": "277--308", "other_ids": {}, "num": null, "urls": [], "raw_text": "Janyce M. Wiebe, Theresa Wilson, Rebecca Bruce, Matthew Bell, and Melanie Martin. 2004. Learning subjec- tive language. Computational Linguistics, 30(3):277-308, September.", "links": null }, "BIBREF29": { "ref_id": "b29", "title": "Annotating expressions of opinions and emotions in language", "authors": [ { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" }, { "first": "Theresa", "middle": [], "last": "Wilson", "suffix": "" }, { "first": "Claire", "middle": [], "last": "Cardie", "suffix": "" } ], "year": 2005, "venue": "Language Resources and Evaluation", "volume": "39", "issue": "2/3", "pages": "164--210", "other_ids": {}, "num": null, "urls": [], "raw_text": "Janyce Wiebe, Theresa Wilson, and Claire Cardie. 2005. Annotating expressions of opinions and emotions in language. Language Resources and Evaluation, 39(2/3):164-210.", "links": null }, "BIBREF30": { "ref_id": "b30", "title": "Recognizing contextual polarity in phrase-level sentiment analysis", "authors": [ { "first": "Theresa", "middle": [], "last": "Wilson", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" }, { "first": "Paul", "middle": [], "last": "Hoffmann", "suffix": "" } ], "year": 2005, "venue": "Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing (HLT-EMNLP'05)", "volume": "", "issue": "", "pages": "347--354", "other_ids": {}, "num": null, "urls": [], "raw_text": "Theresa Wilson, Janyce Wiebe, and Paul Hoffmann. 2005. Recognizing contextual polarity in phrase-level sentiment analysis. In Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing (HLT-EMNLP'05), pages 347-354, Vancouver, CA.", "links": null }, "BIBREF31": { "ref_id": "b31", "title": "WIDIT in TREC-2006 Blog track", "authors": [ { "first": "Kiduk", "middle": [], "last": "Yang", "suffix": "" }, { "first": "Ning", "middle": [], "last": "Yu", "suffix": "" }, { "first": "Alejandro", "middle": [], "last": "Valerio", "suffix": "" }, { "first": "Hui", "middle": [], "last": "Zhang", "suffix": "" } ], "year": 2006, "venue": "Proceedings of TREC", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kiduk Yang, Ning Yu, Alejandro Valerio, and Hui Zhang. 2006. WIDIT in TREC-2006 Blog track. In Proceedings of TREC.", "links": null }, "BIBREF32": { "ref_id": "b32", "title": "Towards answering opinion questions: Separating facts from opinions and identifying the polarity of opinion sentences", "authors": [ { "first": "Hong", "middle": [], "last": "Yu", "suffix": "" }, { "first": "Vasileios", "middle": [], "last": "Hatzivassiloglou", "suffix": "" } ], "year": 2003, "venue": "Proceedings of 2003 Conference on the Empirical Methods in Natural Language Processing (EMNLP'03)", "volume": "", "issue": "", "pages": "129--136", "other_ids": {}, "num": null, "urls": [], "raw_text": "Hong Yu and Vasileios Hatzivassiloglou. 2003. Towards an- swering opinion questions: Separating facts from opinions and identifying the polarity of opinion sentences. In Pro- ceedings of 2003 Conference on the Empirical Methods in Natural Language Processing (EMNLP'03), pages 129- 136, Sapporo, JP.", "links": null }, "BIBREF33": { "ref_id": "b33", "title": "A study of smoothing methods for language models applied to information retrieval", "authors": [ { "first": "Chengxiang", "middle": [], "last": "Zhai", "suffix": "" }, { "first": "John", "middle": [], "last": "Lafferty", "suffix": "" } ], "year": 2004, "venue": "ACM Trans. Inf. Syst", "volume": "22", "issue": "2", "pages": "179--214", "other_ids": {}, "num": null, "urls": [], "raw_text": "Chengxiang Zhai and John Lafferty. 2004. A study of smoothing methods for language models applied to infor- mation retrieval. ACM Trans. Inf. Syst., 22(2):179-214.", "links": null }, "BIBREF34": { "ref_id": "b34", "title": "A generation model to unify topic relevance and lexicon-based sentiment for opinion retrieval", "authors": [ { "first": "Min", "middle": [], "last": "Zhang", "suffix": "" }, { "first": "Xingyao", "middle": [], "last": "Ye", "suffix": "" } ], "year": 2008, "venue": "SIGIR '08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval", "volume": "", "issue": "", "pages": "411--418", "other_ids": {}, "num": null, "urls": [], "raw_text": "Min Zhang and Xingyao Ye. 2008. A generation model to unify topic relevance and lexicon-based sentiment for opinion retrieval. In SIGIR '08: Proceedings of the 31st annual international ACM SIGIR conference on Research and development in information retrieval, pages 411-418, New York, NY, USA. ACM.", "links": null } }, "ref_entries": { "FIGREF0": { "num": null, "uris": null, "type_str": "figure", "text": "the number of unique terms in d, avgdl is the average | d | of all documents, N is the number of documents in the collection, df (w) is the number of documents with w, C is the entire collection, and k 1 and b are constants 2.0 and 0.75." }, "TABREF0": { "html": null, "type_str": "table", "num": null, "content": "", "text": "TF: term frequency \u2022 VS.TF: normalized tf as in VS \u2022 BM25.TF: normalized tf as in BM25 \u2022 IDF: inverse document frequency \u2022 VS.IDF: normalized idf as in VS \u2022 BM25.IDF: normalized idf as in BM25" }, "TABREF2": { "html": null, "type_str": "table", "num": null, "content": "
ModelMAPR-prec P@10
TOPIC REL.0.4052 0.4366 0.6440
BASELINE0.4141 0.4534 0.6440
VS0.4196 0.4542 0.6600
BM250.4235 \u2020 0.4579 0.6600
LM0.4158 0.4520 0.6560
PMI0.4177 0.4538 0.6620
LSA0.4155 0.4526 0.6480
WP0.4165 0.4533 0.6640
BM25\u2022PMI0.4238 \u2020 0.4575 0.6600
BM25\u2022LSA0.4237 \u2020 0.4578 0.6600
BM25\u2022WP0.4237 \u2020 0.4579 0.6600
BM25\u2022PMI\u2022WP 0.4242 \u2020 0.4574 0.6620
BM25\u2022LSA\u2022WP 0.4238 \u2020 0.4576 0.6580
", "text": "Performance of opinion retrieval models using Blog 08 topics. The linear combination parameter \u03bb is optimized on Blog 07 topics. \u2020 indicates statistical significance at the 1% level over the baseline." }, "TABREF3": { "html": null, "type_str": "table", "num": null, "content": "
Opinionated
ModelPrecision Recall F-Measure
BASELINE0.3050.8660.451
VS0.3310.8070.470
BM250.3270.7950.464
LM0.3250.7940.461
LSA0.3150.8060.453
PMI0.3420.6030.436
DTP0.3220.7780.455
VS\u2022LSA0.3350.7690.466
VS\u2022PMI0.3110.8330.453
VS\u2022DTP0.3420.7450.469
VS\u2022LSA\u2022DTP0.3490.7190.470
VS\u2022PMI\u2022DTP0.3280.7730.461
the TREC dataset, using only the word sentiment
model is used as our baseline.
", "text": "Performance of the Sentiment Analysis System on NTCIR7 dataset. System parameters are optimized for F-measure using NTCIR6 dataset with lenient evaluations." }, "TABREF4": { "html": null, "type_str": "table", "num": null, "content": "
Accuracy
FeaturesMovie-review MPQA
PRESENCE82.676.8
TF71.176.5
VS.TF81.376.7
BM25.TF81.477.9
IDF61.661.8
VS.IDF83.677.9
BM25.IDF83.677.8
VS.TF\u2022VS.IDF83.877.9
BM25.TF\u2022BM25.IDF84.177.7
BM25.TF\u2022VS.IDF85.177.7
first introduced by Pang et al. (2002) to test various
ML-based methods for sentiment classification. It
is a balanced dataset of 700 positive and 700 neg-
ative reviews, collected from the Internet Movie
Database (IMDb) archive. MPQA Corpus 8 con-
tains 535 newspaper articles manually annotated
at sentence and subsentence level for opinions and
other private states
", "text": "Average ten-fold cross-validation accuracies of polarity classification task with SVM." } } } }