<<

Sentiment Analysis of Figurative Language using a Sense Disambiguation Approach

Vassiliki Rentoumi∗† [email protected]

George Giannakopoulos∗† [email protected]

Vangelis Karkaletsis∗ [email protected] George A. Vouros† [email protected]

Abstract polarity to the enclosing sentences, exploiting other In this paper we propose a methodology for sen- contextual features as well. timent analysis of figurative language which ap- In Section 2 we briefly present related work while in plies Word Sense Disambiguation and, through Section 3 we present the theoretical background of this an n-gram graph based method, assigns polar- work, together with evidence for the conjectures made. ity to word senses. Polarity assigned to senses, Section 4 presents a detailed description of the overall combined with contextual valence shifters, is methodology and of the specific methods used. Section exploited for further assigning polarity to sen- tences, using Hidden Markov Models. Evalua- 5 details the evaluation of the specific techniques used tion results using the corpus of the Affective Text as well as the overall evaluation of the system. Section task of SemEval’07, are presented together with 6 concludes this article with a brief presentation of a comparison with other state-of-the-art meth- future research. ods, showing that the proposed method provides promising results, and positive evidence support- ing our conjecture: figurative language conveys 2 Related Work sentiment. Sentiment analysis aims to determine the exact polar- ity of a subjective expression. Specifically in [3] there 1 Introduction is an effort using semi-supervised methods to determine orientation of subjective terms Metaphors and expansions are very common phenom- by exploiting information given in glosses provided by ena in everyday language. Exploring such cases, we WordNet. In particular this approach is based on the aim at revealing new semantic aspects of figurative lan- assumption that terms with similar orientation tend guage. Detection of sentiments and implicit polarity, to have similar glosses. Therefore, by means of glosses is within our focus. We thus propose a methodology classification authors aim to classify the terms de- for sentiment analysis that can be valuable in detect- scribed by these glosses. Moreover in [16] the authors ing new semantic aspects of language. Recent stud- try to achieve a classification of phrases and sentences ies have shown that subjectivity is a language prop- into positive/negative, by exploiting their context. In erty which is directly related to word senses [14]. We our approach we exploit the context where figurative believe that subjectivity and thus, senses related to expressions appear and perform disambiguation to re- meanings that convey subjectivity, are valuable indica- veal their senses which are considered as indicators of tors of sentiment. In order to prove this, we are led to sentiment. the exploration of non-literal senses. Work presented There are contextual elements in discourse that in [11] provides evidence that non-literal senses, such could modify the valence of bearing opinion, as metaphors and expanded senses, tend to indicate thus affecting the overall polarity of a sentence. These subjectivity, triggering polarity. In order to capture contextual valence shifters are studied in [9]. the polarity that figurative language conveys [12], it is Words, as shown in [14] can be assigned a subjective necessary to resolve ambiguities and detect the cases (with polarity) sense or an objective (neutral) sense. where words have non-literal meaning. Detecting this In this paper we support that we need to relate word property for words, we can further assign polarity to senses with polarity, rather than the words themselves. the enclosing context, as it is shown in [12]. Towards It has also been shown through an empirical evalua- this goal, other elements in discourse such as valence tion in [11], that especially metaphorical and expanded shifters [9] affect evaluative terms such as figurative ex- senses are strongly polar. Recent approaches follow pressions and modify the polarity of the whole context. this trend by developing sense tagged lists [4]. In this paper we introduce a methodology for polar- The methodology we propose aims to perform sen- ity detection that applies word sense disambiguation timent analysis on figurative language, detecting the (WSD), exploits the assessed word senses and assigns writer’s attitude. The suggested approach shows: (a) the necessity of WSD for assigning polarity to fig- ∗National Centre for Scientific Research ”Demokritos”, Inst. of Informatics and Telecommunications, Greece urative expressions, (b) that figurative expressions †University of the Aegean, Dpt. of Information and Com- combined with valence shifters drive the polarity of munication System Enginneering, Greece the sentence in which they appear, (c) that it seems 370

International Conference RANLP 2009 - Borovets, Bulgaria, pages 370–375 promising even for the cases where the language is not evaluative term such as “to soar”, under the scope of figurative. “fail” will be neutralized, “fail” preserves its negative orientation and propagates it to the whole sentence. Moreover, in example (b) additional valence shifters 3 Theoretical Background and are presented which are used as expanded senses, and they act as polarity reversers. These are the verbs Corpus Study “halt”, “end” and “stop”. In the following examples we meet two more cate- We claim that metaphors and expansions drive the gories of valence modifiers: the diminishers and the polarity of the whole sentence, as they constitute in- intensifiers: (c)“Tsunami fears ease after quake”, (d) tense subjective expressions. The author, as shown in “New Orleans violence sparks tourism fears”. In ex- the following examples, transmits his opinion: (a)“Ice ample (c) “ease” functions as a valence diminisher for storm cripples central U.S”, (b)“Woman fights to “fears”, as it means “ to calm down”. The polarity of keep drunken driver in jail”. In (a) the author uses the whole sentence becomes less negative. In example “cripple” to express implicitly his frustration about (d), “spark” is used with its expanded sense denoting the physical disaster. The same is happening in (b) “to trigger a reaction”, thus strengthening the evalu- with the expanded sense of “fight”, where the writer ative term it modifies. expresses indirectly a positive attitude towards this There are words that always act as valence shifters, woman. while certain polysemous words, act as valence shifters We consider figurative language as the language when they are used under a specific sense (i.e. as ex- which digresses from literal meanings. We con- panded senses or metaphors). We manually compiled jecture that expanded senses and metaphors, being a list of 40 valence shifters derived from our corpus. part of figurative language, can be used as expres- It contains common valence shifters (e.g. negations) sive subjective elements since they display sentiment and words used as such on a per context basis. In the implicitly [12], [15]. To provide evidence for this, near future we intend to exploit a WSD approach in we used the corpus of the Affective Text task of 1 order to detect the specific word senses of non literal Sem Eval ´07 comprising 1000 newspaper headlines expressions that act as valence shifters. as it contains strongly personal and figurative lan- guage. We extracted headlines containing metaphor- ical and expanded expressions, using criteria inspired by Lakoff [6]. Lakoff’s theory follows the principle 4 The Proposed Method For of “from more concrete to more abstract meaning”. Sentiment Analysis For this reason, we mostly considered as metaphorical those headlines whose word senses do not coincide with Our methodology involves three steps:(a) disambigua- the default reading. The “default reading” of a word tion of word senses (WSD). (b) assignment of polarity is the first sense that comes to mind in the absence to word senses, based on the results derived from the of contextual clues, and it usually coincides with the WSD step. (c) polarity detection on a sentence level, literal one: the more concrete connotation of a word. by exploiting polarities of word senses and contextual In contrast, headlines containing figurative language cues such as valence shifters. The specific methods invoke a deduction to a more abstract reading. that implement these steps are presented in more de- We manually extracted from the corpus 277 head- tail in the subsequent sections. lines in total2; 190 containing expanded senses (95 pos- itive and 95 negative) and 87 containing metaphors (39 positive and 48 negative). These are annotated, 4.1 First Step: Word Sense Disam- as described in [13], according to a valence scale in biguation (WSD) which 0 represents a neutral opinion, -100 a negative 3 opinion, and 100 a positive opinion. The average po- For WSD we chose an algorithm , [8] that assigns to every word in a sentence the sense that is most closely larity assigned to headlines containing expansions and 4 metaphors is above 40 which provides evidence that related to the WordNet senses of its neighbouring figurative language conveys significant polarity. words, revealing the meaning of that word. We used We consider that in the headlines, there can be con- a context window of 8 words, as the mean length of textual indicators, referred to as “valence shifters”, each headline consists of 8-10 words. This WSD algo- that can strengthen, weaken or even reverse the po- rithm performs disambiguation for every word of each larity evoked by metaphors and expansions. We headline of our corpus, taking as input a headline and first examine valence shifters that reverse the polar- a relatedness measure [8]. Given such a measure, it ity of a sentence. Let us consider the following ex- computes similarity score for word sense pairs, created amples: (a)“Blazing Angels” for PS3 fail to soar, using every sense of a target word and every sense of (b)“Stop/halt/end, violent nepal strikes”. In exam- its neighbouring words. The score of a sense of a target ple (a) we observe, as is also claimed in [9], that “fail” word is the sum of the maximum individual scores of has a negative connotation. On the other hand “to that sense with the senses of the neighbouring words. soar” in this context, has a positive connotation. The The algorithm then assigns the sense with the highest score to the target word. The algorithm supports sev- 1 http://www.cse.unt.edu/~rada/affectivetext/ eral WordNet based similarity measures, and among 2 The SemEval 07 corpus subset, annotated with metaphors and expansions can be downloaded from: 3 http://www.d.umn.edu/~tpederse/senserelate.html http://www.iit.demokritos.gr/~vrentoumi/corpus.zip 4 http://wordnet.princeton.edu 371 these, Gloss Vector (GV) performs best for non lit- n-grams label the vertices vG V G of the graph. The eral verbs and nouns [11]. GV creates a co-occurence (directed) edges are labeled by∈ the concatenation of matrix of words. Each cell in this matrix indicates the the labels of the vertices they connect in the direction number of times the words represented by the row and of the connection. The edges eG EG connecting the column occur together in a WordNet gloss. Every the n-grams indicate proximity of these∈ n-grams in the word existing within a WordNet gloss is represented in text within a given window Dwin of the original text a multi-dimensional space by treating its correspond- [5]. The edges are weighted by measuring the number ing row as a vector. A context vector is created for of co-occurrences of the vertices’ n-grams within the each word in the gloss, using its corresponding row in window Dwin. Subsequent paragraphs explain how the the co-occurence matrix. Then the gloss of each word models of polarity are generated from the General In- sense is represented by a GV that is the average of all quirer and how mappings of WordNet senses to these these context vectors. In order to measure similarity models are calculated by exploiting n-gram graph rep- between two word senses, the cosine similarity of their resentations. corresponding gloss vectors is calculated. The input to the algorithm is the corpus enriched with Part-of- 4.2.1 Constructing models using n-gram Speech (POS) tags performed by the Stanford POS Graphs tagger 5. To compute models of polarity using n-gram graphs, we have used two sets of positive and negative exam- 4.2 Second Step: Sense Level Polarity ples of words and definitions provided by the General Assignment Inquirer (GI). To represent a text set using n-gram This step detects polarity of the senses computed dur- graphs, we have implemented an update/merge opera- ing the first step. To do this, WordNet senses associ- tor between n-gram graphs of the same rank. Specifi- ated with words in the corpus are mapped to models of cally, given two graphs, G1 and G2, each representing positive or negative polarity. These models are learned a subset of the set of texts, we create a single graph that represents the merging of the two text subsets: by exploiting corresponding examples from the Gen- u u u u eral Inquirer (GI). GI 6 is a containing update(G1,G2) G = (E ,V , L, W ), such that Eu = EG EG,≡ where EG,EG are the edge sets of 1915 words labeled “positive” and 2291 labeled “neg- 1 ∪ 2 1 2 ative”. Results from preliminary experiments (Sec- G1,G2 correspondingly. The weights of the resulting graph’s edges are calcu- tion 5) show that GI provides correct information con- i 1 2 1 cerning polarity of non literal senses. On the other lated as follows: W (e) = W (e)+(W (e) W (e)) l. The factor l [0, 1] is called the learning− factor:× the hand, SentiWordNet [4] is a resource for opinion min- ∈ ing assigning every synset in WordNet three scores, higher the value of learning factor, the higher the im- which represent the probability for a sense of a word pact of the second graph to the first graph. The model to be used in a positive, negative or objective manner. construction process for each class (e.g. of the posi- Our method assumed binary classification of polarity tive/negative polarity class) comprises the initializa- while SentiWordNet performs ternary polarity classi- tion of a graph with the first document of a class, and fication. In the training procedure for SentiWordNet’s the subsequent update of this initial graph with the construction a seed list consisting of positive and neg- graphs of the other documents in the class using the ative synsets was compiled and was expanded itera- union operator. As we need the model of a class to tively through WordNet lexical relations. Our method hold the average weights of all the individual graphs uses GI’s positive and negative terms together with contributing to this model, functioning as a represen- their definitions, instead. Moreover, after experimen- tative graph for the class documents, the i-th graph tal evaluation with various WordNet features in order that updates the class graph (model) uses a learning i 1 to detect sense level polarity for non literal senses, we factor of l = −i , i > 1. decided that the best representative combination in When the model for each class is created, we can judging the polarity for non literal senses is the com- determine the class of a test document by computing bination which consists of synsets and GlossExamples the similarity of the test document n-gram graph to (GES) of non literal senses. On the other hand Sen- the models of the classes: the class whose model is the tiWordNet uses synsets and the whole gloss of every most similar to the test document graph, is the class WordNet sense, in order to judge sense level polarity. of the document. More specifically, for every sense To compute the models of positive and negative po- x of the test set, the set of its synonyms (synsets) larity and produce mappings of senses to these models, and Gloss Example Sentences (GES) extracted from we adopt a graph based method based on character n- WordNet, are being used for the construction of the grams [5], which takes into account contextual (neigh- corresponding n-gram graph X for this sense. bourhood) and sub-word information. A (character) n-gram Sn contained in a text T can 4.2.2 Graph Similarity be any substring of length n of the original text. The n-gram graph is a graph G = V G,EG, L, W , where To represent a character sequence or text we use a V G is the set of vertices, EG is{ the set of edges,} L is set of n-gram graphs, for various n-gram ranks (i.e. a function assigning a label to each vertex and edge, lengths), instead of a single n-gram graph. and W is a function assigning a weight to every edge. To compare two graph sets G1, G2 (one representing a sense and the other the model of a polarity class) we 5 http://nlp.stanford.edu/software/tagger.shtml first use the Value Similarity (VS) for every n-gram 6 http://www.wjh.harvard.edu/~inquirer/ rank [5], indicating how many of the edges contained in 372 graph Gi of rank n are also contained in graph Gj also 5 Experimental Results of rank n, considering also the weights of the matching edges. In this measure each matching edge e having We evaluated the performance of the whole system i i VR(e) (Table 4), but also the distinct steps comprising our weight we in graph G contributes max( Gi , Gj ) to the | | | | method in order to verify our initial hypotheses, (a) sum, while not matching edges do not contribute i.e. i i WSD helps in polarity detection of non literal sen- if an edge e / G then we = 0. The ValueRatio (VR) ∈ i j tences and (b) the polarity of figurative language ex- min(we,we) scaling factor is defined as VR(e) = i j . Thus, pressions drives the overall polarity of the sentences max(we,we) the full equation for VS is: where these are present. To evaluate the WSD method selected, we manu- i j min(we,we) ally annotated the metaphorical and expanded cases e Gi max(wi ,wj ) VS(Gi,Gj) = ∈ e e (1) with their corresponding senses in WordNet, indicated Pmax( Gi , Gj ) by their synsets and glosses. Two annotators were | | | | instructed to assign the most appropriate senses de- VS is a measure converging to 1 for graphs that share rived from WordNet according to the semantics of each their edges and have identical edge weights. headline’s context and a third one refereed any dis- O The overall similarity VS of the sets G1, G2 is com- agreement In order to evaluate the polarity assignment puted as the weighted sum of the VS over all ranks: to senses, we manually aligned the senses of metaphors and expansions indicated by the WSD step, with the r VSr O r [Lmin,LMAX] corresponding senses existing in GI, in order to assign VS (G1, G2) = ∈ × (2) r to them the polarity provided by the latter. P r [Lmin,LMAX] ∈ In assigning polarities to senses, three annotators where VSr is the VS measureP for extracted graphs of participated. Two of them mapped senses to GI, and rank r in G, and Lmin, LMAX are arbitrary chosen the third refereed any disagreement. The annota- minimum and maximum n-gram ranks. For our task tors aligned metaphorical and expanded senses from we used Lmin= 3 and LMAX= 5. WordNet, considering synsets and GES, with the cor- responding senses from GI and took into account the 4.3 Third Step: Sentence Level Polar- polarity orientation (pos/neg) assigned to these senses. As synsets and GES denote the contextual use of the ity Detection given sense, they can also reveal its polarity orien- For the sentence level polarity detection we train two tation in a given context. For each corpus subset HMMs [10] - one for the positive, and one for the nega- (metaphors and expansions) there were two sets of tive cases. The reason behind the choice of HMMs was senses, one extracted manually and one using auto- that they take under consideration transitions among matic WSD. The annotators performed the alignment observations which constitute sequences. In our case of all four of these sense sets with GI, which was ex- the POS of a word combined with the word’s polar- ploited in the experimental evaluation. The results ity constitutes an observation. This information is for the four polarity alignment tasks concern disagree- provided by the POS tagging, and the graph based ment upon polarity alignment between annotators. In polarity assignment method upon metaphorical and particular for metaphors, annotators disagreed in 10 expanded senses of the input sentences. The transi- senses for manual and 13 senses for automatic dis- tions among these observations yield the polarity of ambiguation, out of a total of 128 senses. Moreover the sentential sequences. Structured models have been for expansions annotators disagreed in 20 senses for exploited for polarity detection showing promising re- manual and 24 senses for automatic disambiguation, sults [2]. To exploit valence shifters, these are man- out of a total of 243 senses. In preliminary research ually annotated in the corpus: they are assigned a we performed an extra alignment with GI in order to predefined value depending on whether they revert, detect if the figurative senses investigated are polar. strengthen or weaken the polarity, in order to be inte- Results show us that according to GI, the majority of grated in the HMM. metaphors (positive: 38.28%, negative: 35.15%) and This choice of features is based on the assumption expansions (positive: 31.27% negative: 37.8%) are po- that polarity of sentences is implied by patterns of lar. This verifies our initial hypothesis concerning the parts of speech appearance, by the polarity assigned to polarity of metaphors and expansions. specific senses in the specific context, and by the pres- ence of valence shifters types in a sentence. We need 5.1 Evaluation of WSD in Polarity De- to emphasize that the sequences are constructed using tection only non literal senses and valence shifters (if present), as we believe that these sequences enclose and deter- We first defined a baseline WSD method. In this mine the polarity of the sentences in which they par- method all senses were assigned the first sense given by ticipate. Having trained two HMM’s, one for positive WordNet. Since WordNet ranks the senses depending and one for negative cases, we determine the polarity on their frequency, the first sense indicates the most of each headline sentence by means of the maximum common sense. We then compared the performance of likelihood of the judged observations given by each the baseline method against a method without WSD HMM. In order to evaluate this method of classifying for the polarity detection process and we observed headlines containing metaphors and expanded senses that the former performs better than the latter. In into positive and negative, we have used a 10-fold cross Table 1 results are presented, in terms of recall and validation method for each of the two subsets. precision (prec), concerning polarity classification of 373 headlines containing metaphors (Met) and expansions GVbasedWSD nonWSD baselineWSD (Exp), with WSD (GVbasedWSD/baselineWSD) and recall prec recall prec recall prec without the use of WSD (nonWSD). The results pre- Met 72.6 76.5 46.75 48.5 54.20 57.00 sented in Table 1 are based on automatic WSD and Exp 67.9 68.0 48.1 48.0 63.12 63.00 Sense level polarity assignment. It is deduced that even crude WSD, like the baseline method, could help Table 1: Polarity classification results of headlines con- the polarity classification of non literal sentences. Ta- taining metaphors (Met) and expansions (Exp) with or ble 1 shows that polarity detection using the GV based without the use of WSD WSD method performs much better than the one with- out WSD (nonWSD), for both categories of headlines. WSD manual auto manual We further performed t-tests [7], to verify if the per- Sense Pol manual(a) manual(b) n-gram graphs recall prec recall prec recall prec formance boost when GV based WSD is exploited is Met 78.5 81.5 71.00 74.5 62.57 66.00 statistically significant over that when WSD is absent. Exp 79.1 79.5 75.36 75.5 62.53 62.95 Indeed, the method exploiting GV based WSD is bet- ter - within the 1% statistical error threshold (p-value Table 2: Evaluation of GV based WSD (auto) and 0.01). n-gram graphs steps, Sense Pol: sense level polarity, Table 1 shows that polarity classification using GV manual(a): manual polarity assignment on the manu- based WSD outperforms the one using baseline WSD ally disambiguated senses, manual(b): manual polarity for both subsets of headlines. For metaphors, the GV assignment on the automatically disambiguated senses based WSD method is better than the one using base- line WSD, within the 1% statistical error threshold (p- value 0.01). On the other hand, for expanded senses above lead to the deduction that WSD helps indeed to we cannot support within 5% statistical error that GV improve sentiment analysis, even though the “exact” based WSD performs better than baseline (p-value sense is not always correct. 0.20). The above results verify our initial hypothe- sis that WSD is valuable in the polarity classification of headlines containing non literal senses. 5.2 Evaluation of n-gram graphs for In order to evaluate the GV based WSD method, sense level polarity assignment we first compared the output of the GV based WSD step with that of the manually performed WSD. This In order to evaluate the n-gram graph method used for comparison is performed for both metaphorical and ex- sense level polarity assignment, we first compare the panded cases. We detected that in the WSD process output of n-gram graphs, with that of the manual pro- for metaphorical senses, GV based WSD had a 49.3% cedure. The n-gram graphs scored 60.15% success for success and for expanded senses 45%. The mediocre metaphorical senses, and 67.07% for expanded senses. performance of GV based WSD is attributed to the We present in Table 2 the performance of the sys- fact that disambiguation of metaphors and expansions tem, with n-gram graphs (manual/n-gram graphs) and is itself a difficult task. Additionally, the restricted with manual sense level polarity assignment (man- context of a headline makes the process even more dif- ual/manual(a)). The significant drop of the perfor- ficult. In order to find out to which extent the errors mance when n-gram graphs are used, led to the as- introduced in the disambiguation step can affect the sumption that sense level polarity assignment errors performance of the whole system, we compare two ver- affect our system’s performance because they change sions of our system. These versions are differentiated the input of the decisive HMM component. by the automation of different components of the sys- tem. 5.3 Metaphors and Expansions: In Table 2, we see the performance of both versions Enough for sentence level polarity in terms of recall and precision for polarity classifi- cation of headlines containing metaphors and expan- detection sions. The first version is based on manual WSD We also performed experiments to verify that figu- and manual sense level polarity assignment (man- rative language expressions represent the polarity of ual/manual(a)), and the second is based on auto- the sentences in which they appear. For that we per- matic WSD and manual polarity assignment upon formed two more experiments - all steps of which are the senses that the automatic WSD method indicated performed automatically - one for metaphors and one (auto/manual(b)). Both versions use HMM models for expansions, where we trained HMMs with input for headlines classification. We can also deduce from these results, that errors introduced during automatic WSD do not affect the system significantly. This is at- headlines with met headlines with exp tributed to the fact that the prototypical core sense of met all words exp all words a word remains, even if the word can be semantically rec prec rec prec rec prec rec prec expanded, acquiring elements from relative semantic 72.6 76.5 55.9 57.45 67.89 68.0 59.4 59.8 fields. This core sense can bear a very subtle polar- ity orientation, which becomes stronger and eventually Table 3: Evaluation of the system for polarity classi- gets activated as the word sense digresses (as in the fication of headlines containing metaphors and expan- cases of expansions and metaphors) from the core one. sions (headlines with met/headlines with exp) using There also exists the rare case when the polarity of only non literal expressions (metaphors(met) and ex- a word is reversed because of a semantic change. The pansions(exp)) vs using all words 374 Our System CLaC CLaC - NB through experimental evaluation that WSD is valuable head. met. head. exp. 1000 headlines 1000 headlines in polarity classification of sentences containing figu- rec prec rec prec rec prec rec prec rec prec rative expressions. Moreover, we showed that polarity 72.6 76.5 67.9 68.0 55.60 55.65 9.20 61.42 66.38 31.18 orientation hidden in figurative expressions prevails in sentences where such expressions are present and com- Table 4: Evaluation of the performance of the system bined with contextual valence shifters, can lead us to assessing the overall polarity for the sentence. So far evaluation results of our methodology, seem compa- sequences containing all the words of each headline rable with the state-of-the art methodologies tested instead of only the non literal expressions. In Table upon the same data set. Testing our methodology in 3 the system’s performance for polarity classification, a more extended corpus is our next step. in terms of recall and precision, is presented, for each Acknowledgments. subset. Results are shown for the cases when all words We would like to thank Giota Antonakaki, Ilias Zavit- of the sentence are used and for the cases where only sanos, Anastasios Skarlatidis, Kostas Stamatakis and Jim the non literal expressions are used. The experiments Tsarouhas for their invaluable help. with only the non literal expressions have much better results. This verifies our initial hypothesis that the po- larity of figurative language expressions can represent References the polarity of the sentence in which they appear. [1] A. Andreevskaia and S. Bergler. Clac and clac-nb: Knowledge- based and corpus-based approaches to sentiment tagging. Pro- ceedings of the Fourth International Workshop on Semantic 5.4 System Evaluation, Comparison Evaluations (SemEval-2007), pages 119–120, 2007. with state-of-the-art systems [2] Y. Choi, E. Breck, and C. Cardie. Joint extraction of entities and relations for opinion recognition. In Proc. EMNLP, 2006.

Table 4 presents results for our system as well as two [3] A. Esuli and F. Sebastiani. Determining the semantic orienta- state-of-the-art systems CLaC and CLaC-NB [1], in tion of terms through gloss analysis. Proc. CIKM-2005, 2005. terms of recall and precision, compared to the Gold [4] A. Esuli and F. Sebastiani. Sentiwordnet: A publicly avail- Standard polarity annotation. For our system we able lexical resource for opinion mining. Proceedings of LREC, present in Table 4 three sets of results, for headlines pages 417–422, 2006. containing metaphors (head. met), expanded senses [5] G. Giannakopoulos, V. Karkaletsis, G. Vouros, and P. Stam- atopoulos. Summarization system evaluation revisited: N-gram (head. exp) and for the whole corpus (1000 head- graphs. ACM Transactions on Speech and Language Process- lines). The last set of results is presented in order to ing (TSLP), 5(3), 2008. have a comparison with the two other systems under [6] G. Lakoff and M. Johnson. Metaphors we live by. Cultural a common data set. As mentioned in the beginning of Metaphors: Readings, Research Translations, and Commen- this paper, our system aims to fill the gap for polarity tary, 2001. detection in sentences containing figurative language. [7] M. O’Mahony. Sensory evaluation of food: Statistical methods The results for the individual cases (exp. and met.) and procedures. CRC Press, 1986. show that the system performs well under these cir- [8] T. Pedersen, S. Banerjee, and S. Patwardhan. Maximizing se- cumstances. This leads to the result that the specific mantic relatedness to perform word sense disambiguation. Su- combination of GV based WSD, n-gram graphs and percomputing institute research report umsi, 25, 2005. HMMs works well for these two subsets. As results [9] L. Polanyi and A. Zaenen. Contextual valence shifters. Com- puting Attitude and Affect in Text: Theory and Applications, in Tables 4 and 2 show, the performance of the over- pages 1–9, 2004. all method, compared to configurations where some of [10] L. Rabiner. A tutorial on hidden Markov models and selected the steps are performed manually, is very promising. applications inspeech recognition. Proceedings of the IEEE, When our system is applied to the overall corpus 77(2):257–286, 1989. (Table 4), although the peak values are lower than [11] V. Rentoumi, V. Karkaletsis, G. Vouros, and A. Mozer. Senti- the two other systems, they are high enough to spark ment analysis exploring metaphorical and idiomatic senses: A further research. We can see that although precision word sense disambiguation approach. Proccedings of Interna- tional Workshop on Computational Aspects of Affectual and in the CLaC system is quite high, it suffers a low re- Emotional Interaction (CAFFEi 2008), 2008. call value. The authors attribute this to the fact that [12] E. Riloff and J. Wiebe. Learning extraction patterns for sub- the system was based on an unsupervised knowledge- jective expressions. Proceedings of the Conference on Empiri- based approach in which they aimed to achieve results cal Methods in Natural Language Processing (EMNLP-2003), of higher quality, thus missing a lot of sentiment bear- pages 105–112, 2003. ing headlines [1]. On the contrary CLaC-NB has a [13] C. Strapparava and R. Mihalcea. SemEval-2007 Task 14: Af- fective Text. In Proceedings of the thInternational Workshop high recall and relatively low precision. This system is on Semantic Evaluations (SemEval 2007), Prague, Czeck Re- based on supervised machine learning and the authors public, pages 70–74, 2007. attributed this behaviour to the lack of significant cor- [14] J. Wiebe and R. Mihalcea. Word sense and subjectivity. Pro- pus to train the statistical classifier. The strength of ceedings of the 21st International Conference on Computa- tional Linguistics and the 44th annual meeting of the ACL, our system is that we achieved relatively high values pages 1065–1072, 2006. in both recall and precision. [15] J. Wiebe, T. Wilson, and C. Cardie. Annotating expressions of opinions and in language. Language Resources and Evaluation, 39(2):165–210, 2005.

6 Conclusions and Future Work [16] T. Wilson, J. Wiebe, and P. Hoffmann. Recognizing contex- tual polarity in phrase-level sentiment analysis. Proceedings of This paper presents a new methodology for polar- HLT/EMNLP 2005, 2005. ity classification of non literal sentences. We showed 375