{ "paper_id": "D12-1006", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T16:24:19.613893Z" }, "title": "Detecting Subgroups in Online Discussions by Modeling Positive and Negative Relations among Participants", "authors": [ { "first": "Ahmed", "middle": [], "last": "Hassan", "suffix": "", "affiliation": { "laboratory": "", "institution": "Microsoft Research Redmond", "location": { "region": "WA" } }, "email": "hassanam@microsoft.com" }, { "first": "Amjad", "middle": [], "last": "Abu-Jbara", "suffix": "", "affiliation": { "laboratory": "", "institution": "University of Michigan", "location": { "settlement": "Ann Arbor", "region": "MI" } }, "email": "" }, { "first": "Dragomir", "middle": [], "last": "Radev", "suffix": "", "affiliation": { "laboratory": "", "institution": "University of Michigan", "location": { "settlement": "Ann Arbor", "region": "MI" } }, "email": "radev@umich.edu" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "A mixture of positive (friendly) and negative (antagonistic) relations exist among users in most social media applications. However, many such applications do not allow users to explicitly express the polarity of their interactions. As a result most research has either ignored negative links or was limited to the few domains where such relations are explicitly expressed (e.g. Epinions trust/distrust). We study text exchanged between users in online communities. We find that the polarity of the links between users can be predicted with high accuracy given the text they exchange. This allows us to build a signed network representation of discussions; where every edge has a sign: positive to denote a friendly relation, or negative to denote an antagonistic relation. We also connect our analysis to social psychology theories of balance. We show that the automatically predicted networks are consistent with those theories. Inspired by that, we present a technique for identifying subgroups in discussions by partitioning singed networks representing them.", "pdf_parse": { "paper_id": "D12-1006", "_pdf_hash": "", "abstract": [ { "text": "A mixture of positive (friendly) and negative (antagonistic) relations exist among users in most social media applications. However, many such applications do not allow users to explicitly express the polarity of their interactions. As a result most research has either ignored negative links or was limited to the few domains where such relations are explicitly expressed (e.g. Epinions trust/distrust). We study text exchanged between users in online communities. We find that the polarity of the links between users can be predicted with high accuracy given the text they exchange. This allows us to build a signed network representation of discussions; where every edge has a sign: positive to denote a friendly relation, or negative to denote an antagonistic relation. We also connect our analysis to social psychology theories of balance. We show that the automatically predicted networks are consistent with those theories. Inspired by that, we present a technique for identifying subgroups in discussions by partitioning singed networks representing them.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "Most online communities involve a mixture of positive and negative relations between users. Positive relations may indicate friendship, agreement, or approval. Negative relations usually indicate antagonism, opposition, or disagreement.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Most of the research on relations in social media applications has almost exclusively focused on positive links between individuals (e.g. friends, fans, followers, etc.). We think that one of the main reasons, of why the interplay of positive and negative links did not receive enough attention, is the lack of a notion for explicitly expressing negative interactions. Recently, this problem has received increasing attention. However, all studies have been limited to a handful of datasets from applications that allow users to explicitly label relations as either positive or negative (e.g. trust/distrust on Epinion (Leskovec et al., 2010b) and friends/foes on Slashdot (Kunegis et al., 2009) ).", "cite_spans": [ { "start": 619, "end": 643, "text": "(Leskovec et al., 2010b)", "ref_id": "BIBREF24" }, { "start": 673, "end": 695, "text": "(Kunegis et al., 2009)", "ref_id": "BIBREF21" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Predicting positive/negative relations between discussants is related to another well studied problem, namely debate stance recognition. The objective of this problem is to identify which participants are supporting and which are opposing the topic being discussed. This line of work does not pay enough attention to the relations between participants, rather it focuses on participant's stance toward the topic. It also assumes that every participant either supports or opposes the topic being discussed. This is a simplistic view that ignore the nature of complex topics that has many aspects involved which may result in more than two subgroups with different opinions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "In this work, we apply Natural Language Processing techniques to text correspondences exchanged between individuals to identify the underlying signed social structure in online communities. We present a method for identifying user attitude and for automatically constructing a signed social network representation of discussions. We apply the proposed methods to a large set of discussion posts. We evaluate the performance using a manually labeled dataset. We also conduct a large scale evaluation by showing that predicted links are consistent with the principals of social psychology theories, namely the Structural Balance Theory (Heider, 1946 ). The balance theory has been shown to hold both theoretically (Heider, 1946) and empirically (Leskovec et al., 2010c) for a variety of social community settings. Finally, we present a method for identifying subgroups in online discussions by identifying groups with high density of intra-group positive relations and high density of inter-group negative relations. This method is capable of identifying subgroups even if the community splits into more than two subgroups which is more general than stance recognition which assumes that only two groups exist. The input to our algorithm is a set of text correspondences exchanged between users (e.g. posts or comments). The output is a signed network where edges signify the existence of an interaction between two users. The resulting network has polarity associated with every edge. Edge polarity is a means for indicating positive or negative affinity between two individuals. Figure 1 shows a signed network representation for a subset of posts from a long discussion thread. The thread discussed the November 2010 Wikileaks cable release. We notice that participants split into two groups, one supporting and one opposing the leak. We also notice that most negative edges are between groups, and most positive edges are within groups. It is worth mentioning that networks generated from larger datasets (i.e. with thousands of posts) have much more noise compared to this example.", "cite_spans": [ { "start": 634, "end": 647, "text": "(Heider, 1946", "ref_id": "BIBREF18" }, { "start": 712, "end": 726, "text": "(Heider, 1946)", "ref_id": "BIBREF18" }, { "start": 743, "end": 767, "text": "(Leskovec et al., 2010c)", "ref_id": "BIBREF25" } ], "ref_spans": [ { "start": 1579, "end": 1587, "text": "Figure 1", "ref_id": "FIGREF0" } ], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "The rest of the paper is structured as follows. In section 2, we review some of the related prior work on mining sentiment from text, mining online discussions, extracting social networks from text, and analyzing signed social networks. We define our problem and explain our approach in Section 3. Section 4 describes our dataset. Results and discussion are presented in Section 5. We present a method for identifying subgroups in online discussions in Section 3.3. We conclude in Section 6.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "In this section, we survey several lines of research that are related to our work.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "Our general goal of mining attitude from one individual toward another makes our work related to a huge body of work on sentiment analysis. One such line of research is the well-studied problem of identifying the polarity of individual words (Hatzivassiloglou and McKeown, 1997; Turney and Littman, 2003; Kim and Hovy, 2004; Takamura et al., 2005) . Subjectivity analysis is yet another research line that is closely related to our general goal of mining attitude. The objective of subjectivity analysis is to identify text that presents opinion as opposed to objective text that presents factual information (Wiebe, 2000; Hatzivassiloglou and Wiebe, 2000; Banea et al., 2008; Riloff and Wiebe, 2003) . Our work is different from subjectivity analysis because we are not only interested in discriminating between opinions and facts. Rather, we are interested in identifying the polarity of interactions between individuals. Our method is not restricted to phrases or words, rather it generalizes this to identifying the polarity of an interaction between two individuals based on several posts they exchange.", "cite_spans": [ { "start": 242, "end": 278, "text": "(Hatzivassiloglou and McKeown, 1997;", "ref_id": "BIBREF16" }, { "start": 279, "end": 304, "text": "Turney and Littman, 2003;", "ref_id": "BIBREF35" }, { "start": 305, "end": 324, "text": "Kim and Hovy, 2004;", "ref_id": "BIBREF19" }, { "start": 325, "end": 347, "text": "Takamura et al., 2005)", "ref_id": "BIBREF32" }, { "start": 609, "end": 622, "text": "(Wiebe, 2000;", "ref_id": "BIBREF37" }, { "start": 623, "end": 656, "text": "Hatzivassiloglou and Wiebe, 2000;", "ref_id": "BIBREF17" }, { "start": 657, "end": 676, "text": "Banea et al., 2008;", "ref_id": "BIBREF1" }, { "start": 677, "end": 700, "text": "Riloff and Wiebe, 2003)", "ref_id": "BIBREF31" } ], "ref_spans": [], "eq_spans": [], "section": "Mining Sentiment from Text", "sec_num": "2.1" }, { "text": "Perhaps the closest work to this paper is the work on stance classification. We notice that most of these methods focus on the polarity of the written text assuming that anyone using positive text belongs to one group and anyone using negative text belongs to another. This works well for single-aspect topics or entities like the ones used in (Tan et al., 2011) (e.g. Obama, Sara Palin, Lakers, etc.) . In this simple notion of topics, it is safe to assume that text polarity is a good enough discriminator. This unfortunately is not the case in online discussions about complex topics having many aspects (e.g. abortion, health care, etc.). In such complex topics, people use positive and negative text targeting different aspects of the topic, for example in the health care bill topic, discussants expressed their opinion regarding many aspects including: the enlarged coverage, the insurance premiums, Obama, socialism, etc. This shows that simply looking at text polarity is not enough to identify groups. Tan et al. (2011) studied how twitter following relations can be used to improve stance classification. Their main hypothesis is that connected users are more likely to hold similar opinions. This may be correct for the twitter following relations, but it is not necessarily correct for open discussions where no such relations exist. The only criterion that can be used to connect discussants is how often they reply to each other's posts. We will show later that while many people reply to people with similar opinions, many others reply to people with different opinions as well. Thomas et al. (2006) address the same problem of determining support and opposition as applied to congressional floor-debates. They assess the agreement/disagreement between different speakers by training a text classifier and applying it to a window surrounding the names of other speakers. They construct their training data by assuming that if two speaker have the same vote, then every reference connecting them is an agreement and vice versa. We believe this will result in a very noisy training/testing set and hence we decided to recruit human annotators to create a training set. We found out that many instances with references to other discussants were labeled as neither agreement nor disagreement regardless of whether the discussants have similar or opposing positions. We will use this system as a baseline and will show that the existence of positive/negative words close to a person name does not necessarily show agreement or disagreement with that person.", "cite_spans": [ { "start": 344, "end": 401, "text": "(Tan et al., 2011) (e.g. Obama, Sara Palin, Lakers, etc.)", "ref_id": null }, { "start": 1012, "end": 1029, "text": "Tan et al. (2011)", "ref_id": "BIBREF33" }, { "start": 1595, "end": 1615, "text": "Thomas et al. (2006)", "ref_id": "BIBREF34" } ], "ref_spans": [], "eq_spans": [], "section": "Stance Classification", "sec_num": "2.2" }, { "text": "Hassan et al. (2010) use a language model based approach for identifying agreement and disagreement sentences in discussions. This work is limited to sentences. It does not consider the overall relation between participants. It also does not consider subgroup detection. We will use this method as a baseline for one of our components and will show that the proposed method outperforms it.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Stance Classification", "sec_num": "2.2" }, { "text": "Murakami and Raymond (2010) present another method for stance recognition. They use a small number of hand crafted rules to identify agreement and disagreement interactions. Hand crafted rules usually result in systems with very low recall causing them to miss many agreement/disagreement instances (they report 0.26 recall at the 0.56 precision level). We present a machine learning system to solve this problem and achieve much better performance. Park et al. (2011) propose a method for finding news articles with different views on contentious issues. Mohit et al. (2008) present a set of heuristics for including disagreement information in a minimum cut stance classification framework. Galley et al. (2004) show the value of using durational and structural features for identifying agreement and disagreement in spoken conver-sational speech. They use features like duration of spurts, speech rate, speaker overlap, etc. which are not applicable to written language.", "cite_spans": [ { "start": 450, "end": 468, "text": "Park et al. (2011)", "ref_id": "BIBREF30" }, { "start": 556, "end": 575, "text": "Mohit et al. (2008)", "ref_id": "BIBREF3" }, { "start": 693, "end": 713, "text": "Galley et al. (2004)", "ref_id": "BIBREF13" } ], "ref_spans": [], "eq_spans": [], "section": "Stance Classification", "sec_num": "2.2" }, { "text": "Our approach is different from agreement/disagreement identification because we not only study sentiment at the local sentiment level but also at the global level that takes into consideration many posts exchanged between participants to build a signed network representation of the discussion. Research on debate stance recognition attempts to perform classification under the \"supporting vs. opposing\" paradigm. However such simple view might not always be accurate for discussions on more complex topics with many aspects. After building the signed network representation of discussions, we present a method that can detect how the large group could split into many subgroups (not necessarily two) with coherent opinions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Stance Classification", "sec_num": "2.2" }, { "text": "Little work has been done on the front of extracting social relations between individuals from text. Elson et al. (2010) present a method for extracting social networks from nineteenth-century British novels and serials. They link two characters based on whether they are in conversation or not. McCallum et al. (2007) explored the use of structured data such as email headers for social network construction. Gruzd and Hyrthonthwaite (2008) explored the use of post text in discussions to study interaction patterns in e-learning communities. Extracting social power relations from natural language (i.e. who influences whom) has been studied in (Bramsen et al., 2011; Danescu-Niculescu-Mizil et al., 2011) .", "cite_spans": [ { "start": 296, "end": 318, "text": "McCallum et al. (2007)", "ref_id": "BIBREF28" }, { "start": 410, "end": 441, "text": "Gruzd and Hyrthonthwaite (2008)", "ref_id": null }, { "start": 647, "end": 669, "text": "(Bramsen et al., 2011;", "ref_id": "BIBREF4" }, { "start": 670, "end": 707, "text": "Danescu-Niculescu-Mizil et al., 2011)", "ref_id": "BIBREF10" } ], "ref_spans": [], "eq_spans": [], "section": "Extracting Social Networks from Text", "sec_num": "2.3" }, { "text": "Our work is related to this line of research because we employ natural language processing techniques to reveal embedded social structures. Despite similarities, our work is uniquely characterized by the fact that we extract signed social networks with both positive and negative links from text.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Extracting Social Networks from Text", "sec_num": "2.3" }, { "text": "Most of the work on social networks analysis has only focused on positive interactions. A few recent papers have taken the signs of edges into account. Brzozowski et al. (2008) study the positive and negative relationships between users of Essembly. Essembly is an ideological social network that distinguishes between ideological allies and nemeses. Kunegis et al. (2009) analyze user relationships in the Slashdot technology news site. Slashdot allows users of the website to tag other users as friends or foes, providing positive and negative endorsements. Leskovec et al. (2010b) study signed social networks generated from Slashdot, Epinions, and Wikipedia. They also connect their analysis to theories of signed networks from social psychology. A similar study used the same datasets for predicting positive and negative links given their context (Leskovec et al., 2010a) .", "cite_spans": [ { "start": 152, "end": 176, "text": "Brzozowski et al. (2008)", "ref_id": "BIBREF5" }, { "start": 351, "end": 372, "text": "Kunegis et al. (2009)", "ref_id": "BIBREF21" }, { "start": 560, "end": 583, "text": "Leskovec et al. (2010b)", "ref_id": "BIBREF24" }, { "start": 853, "end": 877, "text": "(Leskovec et al., 2010a)", "ref_id": "BIBREF23" } ], "ref_spans": [], "eq_spans": [], "section": "Signed Social Networks", "sec_num": "2.4" }, { "text": "All this work has been limited to analyzing a handful of datasets for which an explicit notion of both positive and negative relations exists. Our work goes beyond this limitation by leveraging the power of natural language processing to automate the discovery of signed social networks using the text embedded in the network.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Signed Social Networks", "sec_num": "2.4" }, { "text": "The research presented in this paper extends this previous work in a number of ways: (i) we present a method based on linguistic analysis that finds instances of showing positive or negative attitude between participants (ii) we propose a technique for representing discussions as signed networks where a sign is associated with every edge to denote whether the relation is friendly or antagonistic (iii) we evaluate the proposed methods using human annotated data and also conduct a large scale evaluation based on social psychology theories; (iv) finally we present a method for identifying subgroups that globally splits the community involved in the discussion by utilizing the dynamics of the local interactions between participants.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Signed Social Networks", "sec_num": "2.4" }, { "text": "To build a signed network representation of discussants, we start by trying to identify sentences that show positive or negative attitude from the writer to the addressee. The first step toward identifying attitude is to identify words with positive/negative semantic orientation. The semantic orientation or polarity of a word indicates the direction the word deviates from the norm (Lehrer, 1974) . We use Opinion-Finder (Wilson et al., 2005a) to identify words with positive or negative semantic orientation. The polarity of a word is also affected by the context where the word appears. For example, a positive word that appears in a negated context should have a negative polarity. Other polarized words sometimes appear as neutral words in some contexts. To identify contex-tual polarity of words, a large set of features is used including words, sentences, structure, and other features similar to the method described in (Wilson et al., 2005b) .", "cite_spans": [ { "start": 384, "end": 398, "text": "(Lehrer, 1974)", "ref_id": "BIBREF22" }, { "start": 423, "end": 445, "text": "(Wilson et al., 2005a)", "ref_id": "BIBREF38" }, { "start": 929, "end": 951, "text": "(Wilson et al., 2005b)", "ref_id": "BIBREF39" } ], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "Our overall objective is to find the direct attitude between participants. Hence after identifying the semantic orientation of individual words, we move on to predicting which polarized expressions target the addressee and which do not.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "Text polarity alone cannot be used to identify attitude between participants. Sentences that show an attitude are different from subjective sentences. Subjective sentences are sentences used to express opinions, evaluations, and speculations (Riloff and Wiebe, 2003) . While every sentence that shows an attitude is a subjective sentence, not every subjective sentence shows an attitude toward the recipient.", "cite_spans": [ { "start": 242, "end": 266, "text": "(Riloff and Wiebe, 2003)", "ref_id": "BIBREF31" } ], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "In this method, we address the problem of identifying sentences with attitude as a relation detection problem in a supervised learning setting. We study sentences that has mentions to the addressee and polarized expressions (negative/positive words or phrases). Mentions could either be names of other participants or second person pronouns (you, your, yours) used in text posted as a reply to another participant. Reply structure (i.e. who replies to whom) is readily available in many discussion forums. In cases where reply structure is not available, we can use a method like the one in (Lin et al., 2009) to recover it.", "cite_spans": [ { "start": 591, "end": 609, "text": "(Lin et al., 2009)", "ref_id": "BIBREF26" } ], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "We predict whether the mention is related to the polarized expression or not. We regard the mention and the polarized expression as two entities and try to learn a classifier that predicts whether the two entities are related or not.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "The text connecting the two entities offers a very condensed representation of the information needed to assess whether they are related or not. For example the two sentences \"you are completely unqualified\" and \"you know what, he is unqualified ...\" show two different ways the words \"you\", and \"unqualified\" could appear in a sentence. In the first case the polarized word \"unqualified\" refers to the word \"you\". In the second case, the two words are not related. The information in the shortest path between two entities in a dependency tree can be used to assert whether a relationship exists between them (Bunescu and Mooney, 2005).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "The sequence of words connecting the two entities is a very good predictor of whether they are related or not. However, these paths are completely lexicalized and consequently their performance will be limited by data sparseness. To alleviate this problem, we use higher levels of generalization to represent the path connecting the two tokens. These representations are the part-of-speech tags, and the shortest path in a dependency graph connecting the two tokens. We represent every sentence with several representations at different levels of generalization. For example, the sentence \"your ideas are very inspiring\" will be represented using lexical, polarity, part-of-speech, and dependency information as follows:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "LEX: \"YOUR ideas are very POS\" POS: \"YOUR NNS VBP RB JJ POS\" DEP: \"YOUR poss nsubj POS\"", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "The set of features we use are the set of unigrams, and bigrams representing the words, part-of-speech tags, and dependency relations connecting the two entities. For example the following features will be set for the previous example:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "YOUR ideas, YOUR NNS, YOUR poss, poss nsubj, ...., etc.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "We use Support Vector Machines (SVM) as a learning system because it is good with handling high dimensional feature spaces.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Attitude from Text", "sec_num": "3.1" }, { "text": "In this subsection, we describe the procedure we used to build the signed network given the component we described in the previous subsection. This procedure consists of two main steps. The first is building the network without signs, and the second is assigning signs to different edges.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Extracting the Signed Network", "sec_num": "3.2" }, { "text": "To build the network, we parse our data to identify different threads, posts and senders. Every sender is represented with a node in the network. An edge connects two nodes if there exists an interaction between the corresponding participants. We add a directed edge A \u2192 B, if A replies to B's posts at least n times in m different threads. We set m, and n to 2 in all of our experiments. The interaction information (i.e. who replies to whom) can be extracted directly from the thread structure. Alternatively, as mentioned earlier, we can use a method similar to the one presented in (Lin et al., 2009) to recover the reply structure if it is not readily available.", "cite_spans": [ { "start": 586, "end": 604, "text": "(Lin et al., 2009)", "ref_id": "BIBREF26" } ], "ref_spans": [], "eq_spans": [], "section": "Extracting the Signed Network", "sec_num": "3.2" }, { "text": "Once we build the network, we move to the more challenging task in which we associate a sign with every edge. We have shown in the previous section how sentences with positive and negative attitude can be extracted from text. Unfortunately the sign of an interaction cannot be trivially inferred from the polarity of sentences. For example, a single negative sentence written by A and directed to B does not mean that the interaction between A and B is negative. One way to solve this problem would be to compare the number of negative sentences to positive sentences in all posts between A and B and classify the interaction according to the plurality value. We will show later, in our experiments section, that such a simplistic method does not perform well in predicting the sign of an interaction.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Extracting the Signed Network", "sec_num": "3.2" }, { "text": "As a result, we decided to pose the problem as a classical supervised learning problem. We came up with a set of features that we think are good predictors of the interaction sign, and we trained a classifier using those features on a labeled dataset. Our features include numbers and percentages of positive/negative sentences per post, posts per thread, and so on. A sentence is labeled as positive/negative if a relation has been detected in this sentence between a mention referring to the addressee and a positive/negative expression. A post is considered positive/negative based on the majority of relations detected in it. We use two sets of features. The first set is related to A only or B only. The second set is related to the interactions between A and B. The features are summarized in Table 1 .", "cite_spans": [], "ref_spans": [ { "start": 799, "end": 806, "text": "Table 1", "ref_id": "TABREF1" } ], "eq_spans": [], "section": "Extracting the Signed Network", "sec_num": "3.2" }, { "text": "In any discussion, different subgroups may emerge. Members of every subgroup usually have a common focus (positive or negative) toward the topic being discussed. Each member of a group is more likely to show positive attitude to members of the same group, and negative attitude to members of opposing groups. The signed network representation could prove to be very useful for identifying those subgroups. To detect subgroups in a discussion thread, we would like to partition the corresponding signed network such that positive intra-group links and negative inter-group links are dense.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "3.3" }, { "text": "This problem is related to the constrained clustering (Wagstaff et al., 2001 ) and the correlation clustering problem (Bansal et al., 2004) . In constrained clustering, a pairwise similarity metric (which is not available in our domain), and a set of mustlink/cannot-link constraints are used with a standard data clustering algorithm. Correlation clustering operates in a scenario where given a signed graph G = (V, E) where the edge label indicates whether two nodes are similar (+) or different (-), the task is to cluster the vertices so that similar objects are grouped together. Bansal et. al (2004) proved NPhardness and gave constant-factor approximation algorithms for the special case in which the graph is complete (full information) and every edge has weight +1 or -1 which is not the case in our network. Alternatively, we can use a greedy optimization algorithm to find partitions. A criterion function for a local optimization partitioning procedure is constructed such that positive links are dense within groups and negative links are dense between groups.", "cite_spans": [ { "start": 54, "end": 76, "text": "(Wagstaff et al., 2001", "ref_id": "BIBREF36" }, { "start": 118, "end": 139, "text": "(Bansal et al., 2004)", "ref_id": "BIBREF2" }, { "start": 585, "end": 605, "text": "Bansal et. al (2004)", "ref_id": "BIBREF2" } ], "ref_spans": [], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "3.3" }, { "text": "For any potential partition C, we seek to optimize the following function:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "3.3" }, { "text": "P (C) = \u03b1 n +(1\u2212\u03b1) p", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "3.3" }, { "text": "where n is the number of negative links between nodes in the same subgroup, p is the number of positive links between nodes in different subgroups, and \u03b1 is a trade factor that represents the importance of the two terms. We set \u03b1 to 0.5 in all our experiments.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "3.3" }, { "text": "Clusters are selected such that: C * = arg min P (C). A greedy optimization framework is used to minimize P (C). Initially, nodes are randomly partitioned into t different clusters and the criterion function P is evaluated for that cluster. Every cluster has a set of neighbors in the cluster space. A neighbor cluster is obtained by moving one node from one cluster to another, or by exchanging two nodes in two different clusters. Neighbor partitions are evaluated, and if one with a lower value for the criterion function is found, it is set as the current partition. This greedy procedure is repeated with random restarts until a minimal solution is found. To determine the number of subgroups t, we select t that minimizes the optimization function P (C). In all experiments we used an upper limit of t = 5. This technique was able to identify the correct number of subgroups in 77% of the times. In the rest of the cases, the number was different from the correct number by at most 1 except for a single case where it was 2.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "3.3" }, { "text": "Our data consists of a large amount of discussion threads collected from online discussion forums. We collected around 41, 000 topics (threads) and 1.2M posts from the period between the end of 2008 and the end of 2010. All threads were in English and had 5 posts or more. They covered 11 different domains including: politics, religion, science, etc. The average number of participants per domain is 1320 and per topic is 52. The data was tokenized, sentencesplit, and part-of-speech tagged with the OpenNLP toolkit. It was parsed with the Stanford parser (Klein and Manning, 2003) .", "cite_spans": [ { "start": 557, "end": 582, "text": "(Klein and Manning, 2003)", "ref_id": "BIBREF20" } ], "ref_spans": [], "eq_spans": [], "section": "Signed Network Extraction", "sec_num": "4.1" }, { "text": "We randomly selected around 5300 posts (1000 interactions), and asked human annotators to label them. Our annotators were instructed to read all the posts exchanged between two participants and decide whether the interaction between them is positive or negative. We used Amazon Mechanical Turk for annotations. Following previous work (Callison-Burch, 2009; Akkaya et al., 2010) , we took several precautions to maintain data integrity. We restricted annotators to those based in the US to maintain an acceptable level of English fluency. We also restricted annotators to those who have more than 95% approval rate for all previous work. Moreover, we asked three different annotators to label every interaction. The label was computed by taking the majority vote among the three annotators. We refer to this data as the Interactions Dataset.", "cite_spans": [ { "start": 335, "end": 357, "text": "(Callison-Burch, 2009;", "ref_id": "BIBREF7" }, { "start": 358, "end": 378, "text": "Akkaya et al., 2010)", "ref_id": "BIBREF0" } ], "ref_spans": [], "eq_spans": [], "section": "Signed Network Extraction", "sec_num": "4.1" }, { "text": "We ran a different annotation task where we selected sentences including mentions referring to discussants (names or pronouns) and polarized expressions. Annotators were asked to select sentences where the polarized attribute is referring to the mention and hence show a positive or negative attitude toward other discussion participants. This resulted in a set of 5000 manually annotated sentences. We refer to this data as the Sentences Dataset.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Signed Network Extraction", "sec_num": "4.1" }, { "text": "We asked three different annotators to label every instance. The kappa measure between the three groups of annotations was 0.62 for the Interactions Dataset and 0.64 for the Sentences Dataset. To better assess the quality of the annotations, we asked a trained annotator to label 10% of the data. We measured the agreement between the expert annotator ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Signed Network Extraction", "sec_num": "4.1" }, { "text": "We used a dataset of more than 42 topics and approximately 9000 posts collected from two political forums (Createdebate 1 and Politicalforum 2 ). The forum administrators ran a poll asking participants to select their stance from a set of possible answers and hence the dataset was self-labeled with respect to groups. We also used a set of discussions from the Wikipedia discussion section. When a topic on Wikipedia is disputed, the editors of that topic start a discussion about it. We collected 117 Wikipedia discussion threads. The threads contain a total of 1,867 posts. The discussions were annotated by an expert annotator (a professor in sociolinguistics, not an author of the paper) who was instructed to read each of the Wikipedia discussion threads in its entirety and determine whether the discussants split into subgroups, in which case he was asked to identify the subgroup membership for each discussant. In total, we had 159 topics with an average of approximately 500 posts, 60 participants and 2.7 subgroups per topic. Examples of the topics include: Arizona immigration law, airport security, oil spill, evolution, Ireland partitions, abortion and many others.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Sub-group Detection", "sec_num": "4.2" }, { "text": "We performed experiments on the data described in the previous section. We trained and tested the sentence with the attitude detection classifiers described in Section 3.1 using the Sentences Dataset.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Results and Discussion", "sec_num": "5" }, { "text": "We also trained and tested the interaction sign classifier described in Section 3.2 using the Interactions Dataset. We built one signed social network for every domain (e.g. politics, economics, etc.). We decided to build a network for every domain as opposed to one single network because the relation between any two individuals may vary across domains (e.g. politics vs. science). In the rest of this section, we will describe the experiments we did to assess the performance of the sentences with attitude detection and interaction sign prediction steps. In addition to classical evaluation, we evaluate our results using the structural balance theory which has been shown to hold both theoretically (Heider, 1946) and empirically (Leskovec et al., 2010c) . We validate our results by showing that the automatically extracted networks mostly agree with the theory. We evaluated the approach using the structural balance theory because it presents a global (pertaining to relations between multiple edges) and largescale (used millions of posts and thousands of users) evaluation of the results as opposed to traditional evaluation which is local in nature (only considers one edge at a time) and smaller in scale (used thousands of posts).", "cite_spans": [ { "start": 704, "end": 718, "text": "(Heider, 1946)", "ref_id": "BIBREF18" }, { "start": 735, "end": 759, "text": "(Leskovec et al., 2010c)", "ref_id": "BIBREF25" } ], "ref_spans": [], "eq_spans": [], "section": "Results and Discussion", "sec_num": "5" }, { "text": "We compare the proposed methods to two baselines. The first baseline is based on the work of (Thomas et al., 2006) . We used the speaker agreement component presented in (Thomas et al., 2006) as a baseline. The speaker agreement component is one step in their approach. In this component, they used an SVM classifier trained using a window of text surrounding references to other speakers to predict agreement/disagreement between speakers.", "cite_spans": [ { "start": 93, "end": 114, "text": "(Thomas et al., 2006)", "ref_id": "BIBREF34" }, { "start": 170, "end": 191, "text": "(Thomas et al., 2006)", "ref_id": "BIBREF34" } ], "ref_spans": [], "eq_spans": [], "section": "Identifying Sentences with Attitude", "sec_num": "5.1" }, { "text": "We build an SVM text classifier trained on the sentence at which the mention referring to the other participant occurred. We refer to this baseline as the Text Classification approach. The second baselines adopts the language model approach presented in (Hassan et al., 2010) . Two language models are trained using a stream of words, part-of-speech tags, and dependency relations, one for sentences that show an attitude and one for sentences that do not. New sentences are classified based on generation likelihoods. We refer to this baseline as the Language Models approach.", "cite_spans": [ { "start": 254, "end": 275, "text": "(Hassan et al., 2010)", "ref_id": "BIBREF15" } ], "ref_spans": [], "eq_spans": [], "section": "Identifying Sentences with Attitude", "sec_num": "5.1" }, { "text": "We tested this component using the Sentences Dataset described in Section 4. We compared the performance of the proposed method and the two baselines. Table 5 compares the precision, recall, F1, and accuracy for the three methods. The text classification based approach does much worse than others. The reasons is that it ignores the structure and uses much less information (part-of-speech tags and dependency trees are not used) compared to the other methods. Additionally, the short length of the sentences compared to what is typical in text classification may have had a bad effect on the performance. Both other models try to learn the characteristics of the path connecting the mention and the polarized expression. We notice that optimizing the weights for unigram and bigrams features using SVM results in a better performance compared to language models because it does not have the constraints imposed by the former model on the learned weights.", "cite_spans": [], "ref_spans": [ { "start": 151, "end": 158, "text": "Table 5", "ref_id": "TABREF7" } ], "eq_spans": [], "section": "Identifying Sentences with Attitude", "sec_num": "5.1" }, { "text": "We evaluated the importance of the feature types (i.e. dependency vs. pos tags vs words) by measuring the chi-squared statistic for every feature with respect to the class. Dependency features were most helpful, but other types of features helped improve the performance as well.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Identifying Sentences with Attitude", "sec_num": "5.1" }, { "text": "We used the relation detection classifier described in Section 3.1 to find sentences with positive and negative attitude. The output of this classifier was used to compute the features described in Section 3.2, which were used to train a classifier that predicts the sign of an interaction between any two individuals.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Interaction Sign Classifier", "sec_num": "5.2" }, { "text": "We used both Support Vector Machines (SVM) and logistic regression to train the sign interaction classifier. We report several performance metrics for them in Table 2 . We notice that the SVM classifier performs better with an accuracy of 83.5% and an F-measure of 81%. All results were computed using 10 fold cross validation on the labeled data. To better assess the performance of the proposed classifier, we compare it to a baseline that labels the relation as negative if the percentage of negative sentences exceeds a particular threshold, otherwise it is labeled as positive. The thresholds were empirically estimated using a separate development set. The accuracy of this baseline is only 71%.", "cite_spans": [], "ref_spans": [ { "start": 159, "end": 166, "text": "Table 2", "ref_id": "TABREF3" } ], "eq_spans": [], "section": "Interaction Sign Classifier", "sec_num": "5.2" }, { "text": "To better assess the performance of the proposed classifier, we compare it to three baselines. The first is a random baseline that predicts an interaction as positive with probability p that equals the proportion of positive instances to all instances in the training set. The second classifier (Thresh-Num) labels the edge as negative if the number of negative instances exceeds a threshold T n . The third classifier (Thresh-Perc) labels the edge as negative if the percentage of negative instances to all instances exceeds a threshold T p . The cutoff thresholds were estimated using a separate development set.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Interaction Sign Classifier", "sec_num": "5.2" }, { "text": "The 3 baselines were tested using the entire labeled dataset. The SVM classifier was tested using 10 fold cross validation. The accuracy of the random classifier, the two based on a cut off number and percentage , and the SVM classifier are shown in Table 3 . We notice that the random classifier performs worst, and the classifier based on percentage cutoff outperforms the one based on number cutoff. The SVM classifier significantly outperforms all other classifiers. We tried to train a classifier using both the number and percentage of negative and positive posts. The improvement over using the baseline using the percentage of negative posts was not statistically significant.", "cite_spans": [], "ref_spans": [ { "start": 250, "end": 257, "text": "Table 3", "ref_id": "TABREF4" } ], "eq_spans": [], "section": "Interaction Sign Classifier", "sec_num": "5.2" }, { "text": "We evaluated the importance of the features listed in Table 1 by measuring the chi-squared statistic for every feature with respect to the class. We found out that the features describing the interaction between the two participants are more informative than the ones describing individuals characteristics. The later features are still helpful though and they improve the performance by a statistically significant amount. We also noticed that all features based on percentages are more informative than those based on counts. The most informative features are: percentage of negative posts per tread, percentage of negative sentences per post, percentage of positive posts per thread, number of negative posts, and discussion domain.", "cite_spans": [], "ref_spans": [ { "start": 54, "end": 61, "text": "Table 1", "ref_id": "TABREF1" } ], "eq_spans": [], "section": "Interaction Sign Classifier", "sec_num": "5.2" }, { "text": "The structural balance theory is a psychological theory that tries to explain the dynamics of signed social interactions. It has been shown to hold both theoretically (Heider, 1946) and empirically (Leskovec et al., 2010c) . In this section, we study the agreement between the theory and our automatically extracted networks. The theory has its origins in the work of Heider (1946) . It was then formalized in a graph theoretic form by (Cartwright and Harary, 1956) . The theory is based on the principles that \"the friend of my friend is my friend\", \"the enemy of my friend is my enemy\", \"the friend of my enemy is my enemy\", and variations on these. The structural balance theory states that triangles that have an odd number of positive signs (+ + + and + --) are balanced, while triangles that have an even number of positive signs (---and + + -) are not. In this section, we compare the predictions of edge signs made by our system to the structural balance theory by counting the frequencies of different types of triangles in the predicted network. Table 4 shows the frequency of every type of triangle for 10 different domains. To better understand these numbers, we compare them to the frequencies of triangles in a set of random networks. We shuffle the signs for all edges on every network keeping the fractions of positive and negative edges constant. We repeat shuffling for 1000 times and report the average.", "cite_spans": [ { "start": 167, "end": 181, "text": "(Heider, 1946)", "ref_id": "BIBREF18" }, { "start": 198, "end": 222, "text": "(Leskovec et al., 2010c)", "ref_id": "BIBREF25" }, { "start": 368, "end": 381, "text": "Heider (1946)", "ref_id": "BIBREF18" }, { "start": 436, "end": 465, "text": "(Cartwright and Harary, 1956)", "ref_id": "BIBREF8" } ], "ref_spans": [], "eq_spans": [], "section": "Structural Balance Theory", "sec_num": "5.3" }, { "text": "We find that the all-positive triangle (+ + +) is overrepresented in the generated network compared to chance across all domains. We also see that the triangle with two positive edges (+ + \u2212), and the all-negative triangle (\u2212 \u2212 \u2212) are underrepresented compared to chance across all domains. The tri-angle with a single positive edge is slightly overrepresented in most but not all of the topics compared to chance. This shows that the predicted networks mostly agree with the structural balance theory. The slightly non standard behavior of the triangle with one positive edge could be explained in light of the weak balance theory. In this theory, Davis (1967) states that this triangle, which corresponds to the \"enemy of enemy is my friend\" proposition, holds only if the network can be partitioned into exactly two subsets, but not when there are more than two. In general, the percentage of balanced triangles in the predicted networks is higher than in the shuffled networks, and hence the balanced triangles are significantly overrepresented compared to chance showing that our automatically constructed network is similar to explicit signed networks in that they both mostly agree with the balance theory.", "cite_spans": [ { "start": 649, "end": 661, "text": "Davis (1967)", "ref_id": "BIBREF11" } ], "ref_spans": [], "eq_spans": [], "section": "Structural Balance Theory", "sec_num": "5.3" }, { "text": "We compare the performance of the sub-group detection method to three baselines. The first baseline uses graph clustering (GC) to partition a network based on the frequency of interaction between participants. We build a graph where each node represents a participant. Edges link participants if they exchange posts, and edge weights are based on the number of posts exchanged. The second baseline (TC) is based on the premise that participants with similar text are more likely to belong to the same subgroup. We measure text similarity by computing the cosine similarity between the tf-idf representations of the text in a high dimensional vector space. We tried two methods for partitioning those graphs: spectral partitioning (Luxburg, 2007) and a hierarchical agglomeration algorithm which works by greedily optimizing the modularity for graphs (Clauset et al., 2004) . The third baseline is based on stance classification approaches (e.g. (Tan et al., 2011) ). In this baseline we put all the participants who use more positive text in one subgroup and the participants who use more negative text in another subgroup. Text polarity is identified using the method described in Section 3.1. Table 6 shows the average purity (Purity), entropy (Entropy), Normalizes Mutual Information (NMI), and Rand Index (RandIndex) values of the method based on signed networks and the baselines using different partitioning algorithms. The differences in the results shown in the table are statistically significant at the 0.05 level (as indicated by a 2-tailed Figure 2 : A signed network representing participants in a discussion about the \"Health Care Reform Bill\". Blue (dark) nodes represent participants with the bill, Yellow (light) nodes represent participants against the bill, red (solid) edges represent negative attitude, while green (dashed) edges represent positive attitude. Table 6 : Comparison of the sub-group detection method to baseline systems paired t-test).", "cite_spans": [ { "start": 730, "end": 745, "text": "(Luxburg, 2007)", "ref_id": "BIBREF27" }, { "start": 850, "end": 872, "text": "(Clauset et al., 2004)", "ref_id": "BIBREF9" }, { "start": 945, "end": 963, "text": "(Tan et al., 2011)", "ref_id": "BIBREF33" } ], "ref_spans": [ { "start": 1195, "end": 1202, "text": "Table 6", "ref_id": null }, { "start": 1552, "end": 1560, "text": "Figure 2", "ref_id": null }, { "start": 1880, "end": 1887, "text": "Table 6", "ref_id": null } ], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "5.4" }, { "text": "We notice that partitioning the signed network that was automatically extracted from text results in significantly better partitions on the three datasets as indicated by the higher Purity, NMI, and RandIndex and the lower Entropy values it achieves. We believe that the first two baselines performed poorly because the interaction frequency and the text similarity are not key factors in identifying subgroup structures. Many people would respond to people they disagree with more, while others would mainly respond to people they agree with most of the time. Also, people in opposing subgroups tend to use very similar text when discussing the same topic and hence text clustering does not work as well. The baseline that classifies the stance of discussants based on the polarity of their text performed bad too because it overlooks the fact that most of the discussed topics in our datasets have multiple aspects and a discussant may use both positive and negative text targeting different aspects of the topic. An example of a signed network and the corresponding subgtoups as extracted from real data is showm in Figure 2 .", "cite_spans": [], "ref_spans": [ { "start": 1119, "end": 1127, "text": "Figure 2", "ref_id": null } ], "eq_spans": [], "section": "Sub-Group Detection", "sec_num": "5.4" }, { "text": "In this paper, we have shown that natural language processing techniques can be reliably used to extract signed social networks from text correspondences. We believe that this work brings us closer to understanding the relation between language use and social interactions and opens the door to further research efforts that go beyond standard social network analysis by studying the interplay of positive and negative connections. We rigorously evaluated the proposed methods on labeled data and connected our analysis to social psychology theories to show that our predictions mostly agree with them. Finally, we presented potential applications that benefit from the automatically extracted signed network.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusions", "sec_num": "6" }, { "text": "www.createdebate.com 2 www.politicalforum.com", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "This research was funded in part by the Office of the Director of National Intelligence, Intelligence Advanced Research Projects Activity. All statements of fact, opinion or conclusions contained herein are those of the authors and should not be construed as representing the official views or policies of IARPA, the ODNI or the U.S. Government", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgments", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Amazon mechanical turk for subjectivity word sense disambiguation", "authors": [ { "first": "Cem", "middle": [], "last": "Akkaya", "suffix": "" }, { "first": "Alexander", "middle": [], "last": "Conrad", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" }, { "first": "Rada", "middle": [], "last": "Mihalcea", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the NAACL HLT 2010 Workshop on Creating Speech and Language Data with Amazon's Mechanical Turk, CSLDAMT '10", "volume": "", "issue": "", "pages": "195--203", "other_ids": {}, "num": null, "urls": [], "raw_text": "Cem Akkaya, Alexander Conrad, Janyce Wiebe, and Rada Mihalcea. 2010. Amazon mechanical turk for subjectivity word sense disambiguation. In Proceed- ings of the NAACL HLT 2010 Workshop on Creating Speech and Language Data with Amazon's Mechani- cal Turk, CSLDAMT '10, pages 195-203.", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "A bootstrapping method for building subjectivity lexicons for languages with scarce resources", "authors": [ { "first": "Carmen", "middle": [], "last": "Banea", "suffix": "" }, { "first": "Rada", "middle": [], "last": "Mihalcea", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" } ], "year": 2008, "venue": "LREC'08", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Carmen Banea, Rada Mihalcea, and Janyce Wiebe. 2008. A bootstrapping method for building subjec- tivity lexicons for languages with scarce resources. In LREC'08.", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "Correlation Clustering. Machine Learning", "authors": [ { "first": "Nikhil", "middle": [], "last": "Bansal", "suffix": "" }, { "first": "Avrim", "middle": [], "last": "Blum", "suffix": "" }, { "first": "Shuchi", "middle": [], "last": "Chawla", "suffix": "" } ], "year": 2004, "venue": "", "volume": "56", "issue": "", "pages": "89--113", "other_ids": {}, "num": null, "urls": [], "raw_text": "Nikhil Bansal, Avrim Blum, and Shuchi Chawla. 2004. Correlation Clustering. Machine Learning, 56(1):89- 113.", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "The power of negative thinking: Exploiting label disagreement in the min-cut classification framework", "authors": [ { "first": "Mohit", "middle": [], "last": "Bansal", "suffix": "" }, { "first": "Claire", "middle": [], "last": "Cardie", "suffix": "" }, { "first": "Lillian", "middle": [], "last": "Lee", "suffix": "" } ], "year": 2008, "venue": "Proceedings of the 23rd International Conference on Computational Linguistics: Posters", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Mohit Bansal, Claire Cardie, and Lillian Lee. 2008. The power of negative thinking: Exploiting label dis- agreement in the min-cut classification framework. In Proceedings of the 23rd International Conference on Computational Linguistics: Posters.", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "Extracting social power relationships from natural language", "authors": [ { "first": "Philip", "middle": [], "last": "Bramsen", "suffix": "" }, { "first": "Martha", "middle": [], "last": "Escobar-Molano", "suffix": "" }, { "first": "Ami", "middle": [], "last": "Patel", "suffix": "" }, { "first": "Rafael", "middle": [], "last": "Alonso", "suffix": "" } ], "year": 2011, "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", "volume": "1", "issue": "", "pages": "773--782", "other_ids": {}, "num": null, "urls": [], "raw_text": "Philip Bramsen, Martha Escobar-Molano, Ami Patel, and Rafael Alonso. 2011. Extracting social power rela- tionships from natural language. In Proceedings of the 49th Annual Meeting of the Association for Compu- tational Linguistics: Human Language Technologies - Volume 1, pages 773-782.", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Friends and foes: ideological social networking", "authors": [ { "first": "Michael", "middle": [ "J" ], "last": "Brzozowski", "suffix": "" }, { "first": "Tad", "middle": [], "last": "Hogg", "suffix": "" }, { "first": "Gabor", "middle": [], "last": "Szabo", "suffix": "" } ], "year": 2008, "venue": "Proceeding of the twenty-sixth annual SIGCHI conference on Human factors in computing systems", "volume": "", "issue": "", "pages": "817--820", "other_ids": {}, "num": null, "urls": [], "raw_text": "Michael J. Brzozowski, Tad Hogg, and Gabor Szabo. 2008. Friends and foes: ideological social network- ing. In Proceeding of the twenty-sixth annual SIGCHI conference on Human factors in computing systems, pages 817-820, New York, NY, USA.", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "A shortest path dependency kernel for relation extraction", "authors": [ { "first": "C", "middle": [], "last": "Razvan", "suffix": "" }, { "first": "Raymond", "middle": [ "J" ], "last": "Bunescu", "suffix": "" }, { "first": "", "middle": [], "last": "Mooney", "suffix": "" } ], "year": 2005, "venue": "Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing, HLT '05", "volume": "", "issue": "", "pages": "724--731", "other_ids": {}, "num": null, "urls": [], "raw_text": "Razvan C. Bunescu and Raymond J. Mooney. 2005. A shortest path dependency kernel for relation extrac- tion. In Proceedings of the conference on Human Language Technology and Empirical Methods in Nat- ural Language Processing, HLT '05, pages 724-731, Stroudsburg, PA, USA. Association for Computational Linguistics.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Fast, cheap, and creative: evaluating translation quality using amazon's mechanical turk", "authors": [ { "first": "Chris", "middle": [], "last": "Callison", "suffix": "" }, { "first": "-", "middle": [], "last": "Burch", "suffix": "" } ], "year": 2009, "venue": "Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing", "volume": "1", "issue": "", "pages": "286--295", "other_ids": {}, "num": null, "urls": [], "raw_text": "Chris Callison-Burch. 2009. Fast, cheap, and creative: evaluating translation quality using amazon's mechan- ical turk. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 1 -Volume 1, EMNLP '09, pages 286-295.", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "Structure balance: A generalization of heiders theory", "authors": [ { "first": "Dorwin", "middle": [], "last": "Cartwright", "suffix": "" }, { "first": "Frank", "middle": [], "last": "Harary", "suffix": "" } ], "year": 1956, "venue": "Psych. Rev", "volume": "63", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Dorwin Cartwright and Frank Harary. 1956. Structure balance: A generalization of heiders theory. Psych. Rev., 63.", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "Finding community structure in very large networks", "authors": [ { "first": "Aaron", "middle": [], "last": "Clauset", "suffix": "" }, { "first": "E", "middle": [ "J" ], "last": "Mark", "suffix": "" }, { "first": "Cristopher", "middle": [], "last": "Newman", "suffix": "" }, { "first": "", "middle": [], "last": "Moore", "suffix": "" } ], "year": 2004, "venue": "Phys. Rev. E", "volume": "70", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Aaron Clauset, Mark E. J. Newman, and Cristopher Moore. 2004. Finding community structure in very large networks. Phys. Rev. E, 70:066111.", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "Echoes of power: Language effects and power differences in social interaction", "authors": [ { "first": "Cristian", "middle": [], "last": "Danescu-Niculescu-Mizil", "suffix": "" }, { "first": "Lillian", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Bo", "middle": [], "last": "Pang", "suffix": "" }, { "first": "Jon", "middle": [ "M" ], "last": "Kleinberg", "suffix": "" } ], "year": 2011, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Cristian Danescu-Niculescu-Mizil, Lillian Lee, Bo Pang, and Jon M. Kleinberg. 2011. Echoes of power: Lan- guage effects and power differences in social interac- tion. CoRR.", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "Clustering and structural balance in graphs", "authors": [ { "first": "J", "middle": [ "A" ], "last": "Davis", "suffix": "" } ], "year": 1967, "venue": "Human Relations", "volume": "20", "issue": "", "pages": "181--187", "other_ids": {}, "num": null, "urls": [], "raw_text": "J. A. Davis. 1967. Clustering and structural balance in graphs. Human Relations, 20:181-187.", "links": null }, "BIBREF12": { "ref_id": "b12", "title": "Extracting social networks from literary fiction", "authors": [ { "first": "David", "middle": [], "last": "Elson", "suffix": "" }, { "first": "Nicholas", "middle": [], "last": "Dames", "suffix": "" }, { "first": "Kathleen", "middle": [], "last": "Mckeown", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "138--147", "other_ids": {}, "num": null, "urls": [], "raw_text": "David Elson, Nicholas Dames, and Kathleen McKeown. 2010. Extracting social networks from literary fiction. In Proceedings of the 48th Annual Meeting of the Asso- ciation for Computational Linguistics, pages 138-147, Uppsala, Sweden, July.", "links": null }, "BIBREF13": { "ref_id": "b13", "title": "Identifying agreement and disagreement in conversational speech: use of bayesian networks to model pragmatic dependencies", "authors": [ { "first": "Michel", "middle": [], "last": "Galley", "suffix": "" }, { "first": "Kathleen", "middle": [], "last": "Mckeown", "suffix": "" }, { "first": "Julia", "middle": [], "last": "Hirschberg", "suffix": "" }, { "first": "Elizabeth", "middle": [], "last": "Shriberg", "suffix": "" } ], "year": 2004, "venue": "Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, ACL '04", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Michel Galley, Kathleen McKeown, Julia Hirschberg, and Elizabeth Shriberg. 2004. Identifying agree- ment and disagreement in conversational speech: use of bayesian networks to model pragmatic dependen- cies. In Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics, ACL '04, Stroudsburg, PA, USA. Association for Computational Linguistics.", "links": null }, "BIBREF14": { "ref_id": "b14", "title": "Automated discovery and analysis of social networks from threaded discussions", "authors": [ { "first": "Anatoliy", "middle": [], "last": "Gruzd", "suffix": "" }, { "first": "Caroline", "middle": [], "last": "Haythornthwaite", "suffix": "" } ], "year": 2008, "venue": "Proceedings of the International Network of Social Network Analysis (IN-SNA)", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Anatoliy Gruzd and Caroline Haythornthwaite. 2008. Automated discovery and analysis of social networks from threaded discussions. In Proceedings of the In- ternational Network of Social Network Analysis (IN- SNA), St. Pete Beach, Florida.", "links": null }, "BIBREF15": { "ref_id": "b15", "title": "What's with the attitude?: identifying sentences with attitude in online discussions", "authors": [ { "first": "Ahmed", "middle": [], "last": "Hassan", "suffix": "" }, { "first": "Vahed", "middle": [], "last": "Qazvinian", "suffix": "" }, { "first": "Dragomir", "middle": [], "last": "Radev", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "1245--1255", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ahmed Hassan, Vahed Qazvinian, and Dragomir Radev. 2010. What's with the attitude?: identifying sentences with attitude in online discussions. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing, pages 1245-1255.", "links": null }, "BIBREF16": { "ref_id": "b16", "title": "Predicting the semantic orientation of adjectives", "authors": [ { "first": "Vasileios", "middle": [], "last": "Hatzivassiloglou", "suffix": "" }, { "first": "Kathleen", "middle": [ "R" ], "last": "Mckeown", "suffix": "" } ], "year": 1997, "venue": "EACL'97", "volume": "", "issue": "", "pages": "174--181", "other_ids": {}, "num": null, "urls": [], "raw_text": "Vasileios Hatzivassiloglou and Kathleen R. McKeown. 1997. Predicting the semantic orientation of adjec- tives. In EACL'97, pages 174-181.", "links": null }, "BIBREF17": { "ref_id": "b17", "title": "Effects of adjective orientation and gradability on sentence subjectivity", "authors": [ { "first": "Vasileios", "middle": [], "last": "Hatzivassiloglou", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" } ], "year": 2000, "venue": "COLING", "volume": "", "issue": "", "pages": "299--305", "other_ids": {}, "num": null, "urls": [], "raw_text": "Vasileios Hatzivassiloglou and Janyce Wiebe. 2000. Ef- fects of adjective orientation and gradability on sen- tence subjectivity. In COLING, pages 299-305.", "links": null }, "BIBREF18": { "ref_id": "b18", "title": "Attitudes and cognitive organization", "authors": [ { "first": "Fritz", "middle": [], "last": "Heider", "suffix": "" } ], "year": 1946, "venue": "Journal of Psychology", "volume": "21", "issue": "", "pages": "107--112", "other_ids": {}, "num": null, "urls": [], "raw_text": "Fritz Heider. 1946. Attitudes and cognitive organization. Journal of Psychology, 21:107-112.", "links": null }, "BIBREF19": { "ref_id": "b19", "title": "Determining the sentiment of opinions", "authors": [ { "first": "Min", "middle": [], "last": "Soo", "suffix": "" }, { "first": "Eduard", "middle": [], "last": "Kim", "suffix": "" }, { "first": "", "middle": [], "last": "Hovy", "suffix": "" } ], "year": 2004, "venue": "COLING", "volume": "", "issue": "", "pages": "1367--1373", "other_ids": {}, "num": null, "urls": [], "raw_text": "Soo-Min Kim and Eduard Hovy. 2004. Determining the sentiment of opinions. In COLING, pages 1367-1373.", "links": null }, "BIBREF20": { "ref_id": "b20", "title": "Accurate unlexicalized parsing", "authors": [ { "first": "Dan", "middle": [], "last": "Klein", "suffix": "" }, { "first": "Christopher", "middle": [ "D" ], "last": "Manning", "suffix": "" } ], "year": 2003, "venue": "ACL'03", "volume": "", "issue": "", "pages": "423--430", "other_ids": {}, "num": null, "urls": [], "raw_text": "Dan Klein and Christopher D. Manning. 2003. Accurate unlexicalized parsing. In ACL'03, pages 423-430.", "links": null }, "BIBREF21": { "ref_id": "b21", "title": "The slashdot zoo: mining a social network with negative edges", "authors": [ { "first": "J\u00e9r\u00f4me", "middle": [], "last": "Kunegis", "suffix": "" }, { "first": "Andreas", "middle": [], "last": "Lommatzsch", "suffix": "" }, { "first": "Christian", "middle": [], "last": "Bauckhage", "suffix": "" } ], "year": 2009, "venue": "Proceedings of the 18th international conference on World wide web", "volume": "", "issue": "", "pages": "741--750", "other_ids": {}, "num": null, "urls": [], "raw_text": "J\u00e9r\u00f4me Kunegis, Andreas Lommatzsch, and Christian Bauckhage. 2009. The slashdot zoo: mining a so- cial network with negative edges. In Proceedings of the 18th international conference on World wide web, pages 741-750, New York, NY, USA.", "links": null }, "BIBREF22": { "ref_id": "b22", "title": "Semantic fields and lezical structure", "authors": [ { "first": "Adrienne", "middle": [], "last": "Lehrer", "suffix": "" } ], "year": 1974, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Adrienne Lehrer. 1974. Semantic fields and lezical struc- ture. North Holland, Amsterdam and New York.", "links": null }, "BIBREF23": { "ref_id": "b23", "title": "Predicting positive and negative links in online social networks", "authors": [ { "first": "Jure", "middle": [], "last": "Leskovec", "suffix": "" }, { "first": "Daniel", "middle": [], "last": "Huttenlocher", "suffix": "" }, { "first": "Jon", "middle": [], "last": "Kleinberg", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the 19th international conference on World wide web", "volume": "", "issue": "", "pages": "641--650", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jure Leskovec, Daniel Huttenlocher, and Jon Kleinberg. 2010a. Predicting positive and negative links in online social networks. In Proceedings of the 19th interna- tional conference on World wide web, pages 641-650, New York, NY, USA.", "links": null }, "BIBREF24": { "ref_id": "b24", "title": "Signed networks in social media", "authors": [ { "first": "Jure", "middle": [], "last": "Leskovec", "suffix": "" }, { "first": "Daniel", "middle": [], "last": "Huttenlocher", "suffix": "" }, { "first": "Jon", "middle": [], "last": "Kleinberg", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the 28th international conference on Human factors in computing systems", "volume": "", "issue": "", "pages": "1361--1370", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jure Leskovec, Daniel Huttenlocher, and Jon Kleinberg. 2010b. Signed networks in social media. In Proceed- ings of the 28th international conference on Human factors in computing systems, pages 1361-1370, New York, NY, USA.", "links": null }, "BIBREF25": { "ref_id": "b25", "title": "Signed networks in social media", "authors": [ { "first": "Jure", "middle": [], "last": "Leskovec", "suffix": "" }, { "first": "Daniel", "middle": [], "last": "Huttenlocher", "suffix": "" }, { "first": "Jon", "middle": [], "last": "Kleinberg", "suffix": "" } ], "year": 2010, "venue": "CHI 2010", "volume": "", "issue": "", "pages": "1361--1370", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jure Leskovec, Daniel Huttenlocher, and Jon Kleinberg. 2010c. Signed networks in social media. In CHI 2010, pages 1361-1370, New York, NY, USA. ACM.", "links": null }, "BIBREF26": { "ref_id": "b26", "title": "Simultaneously modeling semantics and structure of threaded discussions: a sparse coding approach and its applications", "authors": [ { "first": "Chen", "middle": [], "last": "Lin", "suffix": "" }, { "first": "Jiang-Ming", "middle": [], "last": "Yang", "suffix": "" }, { "first": "Rui", "middle": [], "last": "Cai", "suffix": "" }, { "first": "Xin-Jing", "middle": [], "last": "Wang", "suffix": "" }, { "first": "Wei", "middle": [], "last": "Wang", "suffix": "" } ], "year": 2009, "venue": "SIGIR '09", "volume": "", "issue": "", "pages": "131--138", "other_ids": {}, "num": null, "urls": [], "raw_text": "Chen Lin, Jiang-Ming Yang, Rui Cai, Xin-Jing Wang, and Wei Wang. 2009. Simultaneously modeling se- mantics and structure of threaded discussions: a sparse coding approach and its applications. In SIGIR '09, pages 131-138.", "links": null }, "BIBREF27": { "ref_id": "b27", "title": "A tutorial on spectral clustering", "authors": [ { "first": "Ulrike", "middle": [], "last": "Luxburg", "suffix": "" } ], "year": 2007, "venue": "Statistics and Computing", "volume": "17", "issue": "", "pages": "395--416", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ulrike Luxburg. 2007. A tutorial on spectral clustering. Statistics and Computing, 17:395-416, December.", "links": null }, "BIBREF28": { "ref_id": "b28", "title": "Topic and role discovery in social networks with experiments on enron and academic email", "authors": [ { "first": "Andrew", "middle": [], "last": "Mccallum", "suffix": "" }, { "first": "Xuerui", "middle": [], "last": "Wang", "suffix": "" }, { "first": "Andr\u00e9s", "middle": [], "last": "Corrada-Emmanuel", "suffix": "" } ], "year": 2007, "venue": "J. Artif. Int. Res", "volume": "30", "issue": "", "pages": "249--272", "other_ids": {}, "num": null, "urls": [], "raw_text": "Andrew McCallum, Xuerui Wang, and Andr\u00e9s Corrada- Emmanuel. 2007. Topic and role discovery in so- cial networks with experiments on enron and academic email. J. Artif. Int. Res., 30:249-272, October.", "links": null }, "BIBREF29": { "ref_id": "b29", "title": "Support or oppose?: classifying positions in online debates from reply activities and opinion expressions", "authors": [ { "first": "Akiko", "middle": [], "last": "Murakami", "suffix": "" }, { "first": "Rudy", "middle": [], "last": "Raymond", "suffix": "" } ], "year": 2010, "venue": "Proceedings of the 23rd International Conference on Computational Linguistics: Posters", "volume": "", "issue": "", "pages": "869--875", "other_ids": {}, "num": null, "urls": [], "raw_text": "Akiko Murakami and Rudy Raymond. 2010. Support or oppose?: classifying positions in online debates from reply activities and opinion expressions. In Proceed- ings of the 23rd International Conference on Compu- tational Linguistics: Posters, pages 869-875.", "links": null }, "BIBREF30": { "ref_id": "b30", "title": "Contrasting opposing views of news articles on contentious issues", "authors": [ { "first": "Souneil", "middle": [], "last": "Park", "suffix": "" }, { "first": "Kyungsoon", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Junehwa", "middle": [], "last": "Song", "suffix": "" } ], "year": 2011, "venue": "Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies", "volume": "1", "issue": "", "pages": "340--349", "other_ids": {}, "num": null, "urls": [], "raw_text": "Souneil Park, KyungSoon Lee, and Junehwa Song. 2011. Contrasting opposing views of news articles on con- tentious issues. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguis- tics: Human Language Technologies -Volume 1, pages 340-349.", "links": null }, "BIBREF31": { "ref_id": "b31", "title": "Learning extraction patterns for subjective expressions", "authors": [ { "first": "Ellen", "middle": [], "last": "Riloff", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" } ], "year": 2003, "venue": "EMNLP'03", "volume": "", "issue": "", "pages": "105--112", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ellen Riloff and Janyce Wiebe. 2003. Learning extraction patterns for subjective expressions. In EMNLP'03, pages 105-112.", "links": null }, "BIBREF32": { "ref_id": "b32", "title": "Extracting semantic orientations of words using spin model", "authors": [ { "first": "Hiroya", "middle": [], "last": "Takamura", "suffix": "" }, { "first": "Takashi", "middle": [], "last": "Inui", "suffix": "" }, { "first": "Manabu", "middle": [], "last": "Okumura", "suffix": "" } ], "year": 2005, "venue": "ACL'05", "volume": "", "issue": "", "pages": "133--140", "other_ids": {}, "num": null, "urls": [], "raw_text": "Hiroya Takamura, Takashi Inui, and Manabu Okumura. 2005. Extracting semantic orientations of words using spin model. In ACL'05, pages 133-140.", "links": null }, "BIBREF33": { "ref_id": "b33", "title": "User-level sentiment analysis incorporating social networks", "authors": [ { "first": "Chenhao", "middle": [], "last": "Tan", "suffix": "" }, { "first": "Lillian", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Jie", "middle": [], "last": "Tang", "suffix": "" }, { "first": "Long", "middle": [], "last": "Jiang", "suffix": "" }, { "first": "Ming", "middle": [], "last": "Zhou", "suffix": "" }, { "first": "Ping", "middle": [], "last": "Li", "suffix": "" } ], "year": 2011, "venue": "Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining", "volume": "", "issue": "", "pages": "1397--1405", "other_ids": {}, "num": null, "urls": [], "raw_text": "Chenhao Tan, Lillian Lee, Jie Tang, Long Jiang, Ming Zhou, and Ping Li. 2011. User-level sentiment anal- ysis incorporating social networks. In Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining, KDD '11, pages 1397-1405.", "links": null }, "BIBREF34": { "ref_id": "b34", "title": "Get out the vote: Determining support or opposition from congressional floor-debate transcripts", "authors": [ { "first": "Matt", "middle": [], "last": "Thomas", "suffix": "" }, { "first": "Bo", "middle": [], "last": "Pang", "suffix": "" }, { "first": "Lillian", "middle": [], "last": "Lee", "suffix": "" } ], "year": 2006, "venue": "Proceedings of EMNLP", "volume": "", "issue": "", "pages": "327--335", "other_ids": {}, "num": null, "urls": [], "raw_text": "Matt Thomas, Bo Pang, and Lillian Lee. 2006. Get out the vote: Determining support or opposition from con- gressional floor-debate transcripts. In In Proceedings of EMNLP, pages 327-335.", "links": null }, "BIBREF35": { "ref_id": "b35", "title": "Measuring praise and criticism: Inference of semantic orientation from association", "authors": [ { "first": "Peter", "middle": [], "last": "Turney", "suffix": "" }, { "first": "Michael", "middle": [], "last": "Littman", "suffix": "" } ], "year": 2003, "venue": "ACM Transactions on Information Systems", "volume": "21", "issue": "", "pages": "315--346", "other_ids": {}, "num": null, "urls": [], "raw_text": "Peter Turney and Michael Littman. 2003. Measuring praise and criticism: Inference of semantic orientation from association. ACM Transactions on Information Systems, 21:315-346.", "links": null }, "BIBREF36": { "ref_id": "b36", "title": "Constrained k-means clustering with background knowledge", "authors": [ { "first": "Kiri", "middle": [], "last": "Wagstaff", "suffix": "" }, { "first": "Claire", "middle": [], "last": "Cardie", "suffix": "" }, { "first": "Seth", "middle": [], "last": "Rogers", "suffix": "" }, { "first": "Stefan", "middle": [], "last": "Schr\u00f6dl", "suffix": "" } ], "year": 2001, "venue": "Proceedings of the Eighteenth International Conference on Machine Learning", "volume": "", "issue": "", "pages": "577--584", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kiri Wagstaff, Claire Cardie, Seth Rogers, and Stefan Schr\u00f6dl. 2001. Constrained k-means clustering with background knowledge. In Proceedings of the Eigh- teenth International Conference on Machine Learning, pages 577-584.", "links": null }, "BIBREF37": { "ref_id": "b37", "title": "Learning subjective adjectives from corpora", "authors": [ { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" } ], "year": 2000, "venue": "Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Artificial Intelligence", "volume": "", "issue": "", "pages": "735--740", "other_ids": {}, "num": null, "urls": [], "raw_text": "Janyce Wiebe. 2000. Learning subjective adjectives from corpora. In Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Ar- tificial Intelligence, pages 735-740.", "links": null }, "BIBREF38": { "ref_id": "b38", "title": "Opinionfinder: a system for subjectivity analysis", "authors": [ { "first": "Theresa", "middle": [], "last": "Wilson", "suffix": "" }, { "first": "Paul", "middle": [], "last": "Hoffmann", "suffix": "" }, { "first": "Swapna", "middle": [], "last": "Somasundaran", "suffix": "" }, { "first": "Jason", "middle": [], "last": "Kessler", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" }, { "first": "Yejin", "middle": [], "last": "Choi", "suffix": "" }, { "first": "Claire", "middle": [], "last": "Cardie", "suffix": "" }, { "first": "Ellen", "middle": [], "last": "Riloff", "suffix": "" }, { "first": "Siddharth", "middle": [], "last": "Patwardhan", "suffix": "" } ], "year": 2005, "venue": "Proceedings of HLT/EMNLP on Interactive Demonstrations, HLT-Demo '05", "volume": "", "issue": "", "pages": "34--35", "other_ids": {}, "num": null, "urls": [], "raw_text": "Theresa Wilson, Paul Hoffmann, Swapna Somasun- daran, Jason Kessler, Janyce Wiebe, Yejin Choi, Claire Cardie, Ellen Riloff, and Siddharth Patward- han. 2005a. Opinionfinder: a system for subjectiv- ity analysis. In Proceedings of HLT/EMNLP on Inter- active Demonstrations, HLT-Demo '05, pages 34-35, Stroudsburg, PA, USA. Association for Computational Linguistics.", "links": null }, "BIBREF39": { "ref_id": "b39", "title": "Recognizing contextual polarity in phraselevel sentiment analysis", "authors": [ { "first": "Theresa", "middle": [], "last": "Wilson", "suffix": "" }, { "first": "Janyce", "middle": [], "last": "Wiebe", "suffix": "" }, { "first": "Paul", "middle": [], "last": "Hoffmann", "suffix": "" } ], "year": 2005, "venue": "HLT/EMNLP'05, Vancouver", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Theresa Wilson, Janyce Wiebe, and Paul Hoffmann. 2005b. Recognizing contextual polarity in phrase- level sentiment analysis. In HLT/EMNLP'05, Vancou- ver, Canada.", "links": null }, "BIBREF40": { "ref_id": "b40", "title": "Community mining from signed social networks", "authors": [ { "first": "Bo", "middle": [], "last": "Yang", "suffix": "" }, { "first": "William", "middle": [], "last": "Cheung", "suffix": "" }, { "first": "Jiming", "middle": [], "last": "Liu", "suffix": "" } ], "year": 2007, "venue": "IEEE Trans. on Knowl. and Data Eng", "volume": "19", "issue": "10", "pages": "1333--1348", "other_ids": {}, "num": null, "urls": [], "raw_text": "Bo Yang, William Cheung, and Jiming Liu. 2007. Com- munity mining from signed social networks. IEEE Trans. on Knowl. and Data Eng., 19(10):1333-1348.", "links": null } }, "ref_entries": { "FIGREF0": { "num": null, "uris": null, "text": "An example showing a signed social network along with evidence from text that justifies edge signs.", "type_str": "figure" }, "FIGREF1": { "num": null, "uris": null, "text": "Participant Features Number of posts per month for A (B) Percentage of positive posts per month for A (B) Percentage of negative posts per month for A (B) gender Interaction Features Percentage/number of positive (negative) sentences per post Percentage/number of positive (negative) posts per thread Discussion Domain (e.g. politics, science, etc.)", "type_str": "figure" }, "FIGREF2": { "num": null, "uris": null, "text": "+) (+ + \u2212) (+ \u2212 \u2212) (\u2212 \u2212 \u2212) (+ + +) (+ + \u2212) (+ \u2212 \u2212) (\u2212 \u2212 \u2212) abortion", "type_str": "figure" }, "TABREF1": { "num": null, "text": "Features used by the Interaction Sign Classifier.", "html": null, "type_str": "table", "content": "" }, "TABREF3": { "num": null, "text": "Interaction sign classifier performance.", "html": null, "type_str": "table", "content": "
ClassifierRandom Thresh-Num Thresh-Perc.SVM
Accuracy65%69%71%83.5%
" }, "TABREF4": { "num": null, "text": "A comparison of different sign interaction classifiers. and the majority label from Mechanical Turk. The kappa measure was 0.69 for the Interactions Dataset and 0.67 for the Sentences Dataset.", "html": null, "type_str": "table", "content": "" }, "TABREF6": { "num": null, "text": "Percentage of different types of triangles in the extracted networks vs. the random networks.", "html": null, "type_str": "table", "content": "
MethodAccuracy Precision RecallF1
Text Classification60.461.160.260.6
Language Models80.381.079.480.2
Relation Extraction82.382.382.382.3
" }, "TABREF7": { "num": null, "text": "Comparison of attitude identification methods.", "html": null, "type_str": "table", "content": "" } } } }