<<

Dependency-Based Word Embeddings

Omer Levy∗ and Yoav Goldberg Department Bar-Ilan University Ramat-Gan, Israel omerlevy,yoav.goldberg @gmail.com { }

Abstract are represented as a very high dimensional but sparse vectors in which each entry is a measure While continuous word embeddings are of the association between the word and a particu- gaining popularity, current models are lar context (see (Turney and Pantel, 2010; Baroni based solely on linear contexts. In this and Lenci, 2010) for a comprehensive survey). work, we generalize the skip-gram model In some works, the dimensionality of the sparse with negative sampling introduced by word-context vectors is reduced, using techniques Mikolov et al. to include arbitrary con- such as SVD (Bullinaria and Levy, 2007) or LDA texts. In particular, we perform exper- (Ritter et al., 2010; Seaghdha,´ 2010; Cohen et iments with dependency-based contexts, al., 2012). Most recently, it has been proposed and show that they produce markedly to represent words as dense vectors that are de- different embeddings. The dependency- rived by various training methods inspired from based embeddings are less topical and ex- neural-network language modeling (Bengio et al., hibit more functional similarity than the 2003; Collobert and Weston, 2008; Mnih and original skip-gram embeddings. Hinton, 2008; Mikolov et al., 2011; Mikolov et al., 2013b). These representations, referred to as 1 Introduction “neural embeddings” or “word embeddings”, have been shown to perform well across a variety of Word representation is central to natural language tasks (Turian et al., 2010; Collobert et al., 2011; processing. The default approach of represent- Socher et al., 2011; Al-Rfou et al., 2013). ing words as discrete and distinct symbols is in- Word embeddings are easy to work with be- sufficient for many tasks, and suffers from poor cause they enable efficient computation of word generalization. For example, the symbolic repre- similarities through low-dimensional matrix op- sentation of the words “pizza” and “hamburger” erations. Among the state-of-the-art word- are completely unrelated: even if we know that embedding methods is the skip-gram with nega- the word “pizza” is a good argument for the verb tive sampling model (SKIPGRAM), introduced by “eat”, we cannot infer that “hamburger” is also Mikolov et al. (2013b) and implemented in the a good argument. We thus seek a representation software.1 Not only does it produce that captures semantic and syntactic similarities useful word representations, but it is also very ef- between words. A very common paradigm for ac- ficient to train, works in an online fashion, and quiring such representations is based on the distri- scales well to huge copora (billions of words) as butional hypothesis of Harris (1954), stating that well as very large word and context vocabularies. words in similar contexts have similar meanings. Previous work on neural word embeddings take Based on the distributional hypothesis, many the contexts of a word to be its linear context – methods of deriving word representations were ex- words that precede and follow the target word, typ- plored in the NLP community. On one end of the ically in a window of k tokens to each side. How- spectrum, words are grouped into clusters based ever, other types of contexts can be explored too. on their contexts (Brown et al., 1992; Uszkor- In this work, we generalize the SKIP- eit and Brants, 2008). On the other end, words GRAM model, and move from linear bag-of-words ∗ Supported by the European Community’s Seventh contexts to arbitrary word contexts. Specifically, Framework Programme (FP7/2007-2013) under grant agree- ment no. 287923 (EXCITEMENT). 1code..com/p/word2vec/

302 Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Short Papers), pages 302–308, Baltimore, Maryland, USA, June 23-25 2014. c 2014 Association for Computational Linguistics following work in sparse vector-space models a large body of text. Consider a word-context pair (Lin, 1998; Pado´ and Lapata, 2007; Baroni and (w, c). Did this pair come from the data? We de- Lenci, 2010), we experiment with syntactic con- note by p(D = 1 w, c) the probability that (w, c) | texts that are derived from automatically produced came from the data, and by p(D = 0 w, c) = | dependency parse-trees. 1 p(D = 1 w, c) the probability that (w, c) did − | The different kinds of contexts produce no- not. The distribution is modeled as: ticeably different embeddings, and induce differ- 1 p(D = 1 w, c) = vw vc ent word similarities. In particular, the bag-of- | 1+e− · words nature of the contexts in the “original” where v and v (each a d-dimensional vector) are SKIPGRAM model yield broad topical similari- w c the model parameters to be learned. We seek to ties, while the dependency-based contexts yield maximize the log-probability of the observed pairs more functional similarities of a cohyponym na- belonging to the data, leading to the objective: ture. This effect is demonstrated using both quali- tative and quantitative analysis (Section 4). 1 arg maxvw,vc (w,c) D log 1+e vc vw The neural word-embeddings are considered ∈ − · P opaque, in the sense that it is hard to assign mean- This objective admits a trivial solution in which ings to the dimensions of the induced represen- p(D = 1 w, c) = 1 for every pair (w, c). This can | tation. In Section 5 we show that the SKIP- be easily achieved by setting v = v and v v = c w c· w GRAM model does allow for some introspection K for all c, w, where K is large enough number. by querying it for contexts that are “activated by” a In order to prevent the trivial solution, the ob- target word. This allows us to peek into the learned jective is extended with (w, c) pairs for which representation and explore the contexts that are p(D = 1 w, c) must be low, i.e. pairs which are | found by the learning process to be most discrim- not in the data, by generating the set D0 of ran- inative of particular words (or groups of words). dom (w, c) pairs (assuming they are all incorrect), To the best of our knowledge, this is the first work yielding the negative-sampling training objective: to suggest such an analysis of discriminatively- trained word-embedding models. arg maxvw,vc (w,c) D p(D = 1 c, w) (w,c) D p(D = 0 c, w) ∈ | ∈ 0 | Q Q  2 The Skip-Gram Model which can be rewritten as:

Our departure point is the skip-gram neural em- arg maxvw,vc (w,c) D log σ(vc vw) + (w,c) D log σ( vc vw) ∈ · ∈ 0 − · bedding model introduced in (Mikolov et al., P P  x 2013a) trained using the negative-sampling pro- where σ(x) = 1/(1+e ). The objective is trained cedure presented in (Mikolov et al., 2013b). In in an online fashion using stochastic-gradient up- dates over the corpus D D . this section we summarize the model and train- ∪ 0 ing objective following the derivation presented by The negative samples D0 can be constructed in Goldberg and Levy (2014), and highlight the ease various ways. We follow the method proposed by Mikolov et al.: for each (w, c) D we construct of incorporating arbitrary contexts in the model. ∈ In the skip-gram model, each word w W is n samples (w, c1),..., (w, cn), where n is a hy- ∈ associated with a vector v Rd and similarly perparameter and each cj is drawn according to its w ∈ each context c C is represented as a vector unigram distribution raised to the 3/4 power. ∈ v Rd, where W is the words vocabulary, C Optimizing this objective makes observed c ∈ is the contexts vocabulary, and d is the embed- word-context pairs have similar embeddings, ding dimensionality. The entries in the vectors while scattering unobserved pairs. Intuitively, are latent, and treated as parameters to be learned. words that appear in similar contexts should have Loosely speaking, we seek parameter values (that similar embeddings, though we have not yet found is, vector representations for both words and con- a formal proof that SKIPGRAM does indeed max- texts) such that the dot product v v associated imize the dot product of similar words. w · c with “good” word-context pairs is maximized. 3 Embedding with Arbitrary Contexts More specifically, the negative-sampling objec- tive assumes a dataset D of observed (w, c) pairs In the SKIPGRAM embedding algorithm, the con- of words w and the contexts c, which appeared in texts of a word w are the words surrounding it

303 prep amod nsubj in the text. The context vocabulary C is thus dobj pobj identical to the word vocabulary W . However, Australian scientist discovers star with telescope this restriction is not required by the model; con- prep with amod nsubj texts need not correspond to words, and the num- dobj ber of context-types can be substantially larger Australian scientist discovers star telescope than the number of word-types. We generalize WORDCONTEXTS SKIPGRAM by replacing the bag-of-words con- 1 australian scientist/amod− texts with arbitrary contexts. 1 scientist australian/amod, discovers/nsubj− In this paper we experiment with dependency- discovers scientist/nsubj, star/dobj, telescope/prep with 1 based syntactic contexts. Syntactic contexts cap- star discovers/dobj− 1 ture different information than bag-of-word con- telescope discovers/prep with− texts, as we demonstrate using the sentence “Aus- Figure 1: Dependency-based context extraction example. tralian scientist discovers star with telescope”. Top: preposition relations are collapsed into single arcs, making telescope a direct modifier of discovers. Bottom: the Linear Bag-of-Words Contexts This is the contexts extracted for each word in the sentence. context used by word2vec and many other neu- ral embeddings. Using a window of size k around where lbl is the type of the dependency relation be- the target word w, 2k contexts are produced: the tween the head and the modifier (e.g. nsubj, dobj, 1 k words before and the k words after w. For prep with, amod) and lbl− is used to mark the k = 2, the contexts of the target word w are inverse-relation. Relations that include a preposi- w 2, w 1, w+1, w+2. In our example, the contexts tion are “collapsed” prior to context extraction, by − − of discovers are Australian, scientist, star, with.2 directly connecting the head and the object of the Note that a context window of size 2 may miss preposition, and subsuming the preposition itself some important contexts (telescope is not a con- into the dependency label. An example of the de- text of discovers), while including some acciden- pendency context extraction is given in Figure 1. tal ones (Australian is a context discovers). More- Notice that syntactic dependencies are both over, the contexts are unmarked, resulting in dis- more inclusive and more focused than bag-of- covers being a context of both stars and scientist, words. They capture relations to words that are which may result in stars and scientists ending far apart and thus “out-of-reach” with small win- up as neighbours in the embedded space. A win- dow bag-of-words (e.g. the instrument of discover dow size of 5 is commonly used to capture broad is telescope/prep with), and also filter out “coinci- topical content, whereas smaller windows contain dental” contexts which are within the window but more focused information about the target word. not directly related to the target word (e.g. Aus- tralian is not used as the context for discovers). In Dependency-Based Contexts An alternative to addition, the contexts are typed, indicating, for ex- the bag-of-words approach is to derive contexts ample, that stars are objects of discovery and sci- based on the syntactic relations the word partic- entists are subjects. We thus expect the syntactic ipates in. This is facilitated by recent advances contexts to yield more focused embeddings, cap- in parsing technology (Goldberg and Nivre, 2012; turing more functional and less topical similarity. Goldberg and Nivre, 2013) that allow parsing to syntactic dependencies with very high speed and 4 Experiments and Evaluation near state-of-the-art accuracy. We experiment with 3 training conditions: BOW5 After parsing each sentence, we derive word (bag-of-words contexts with k = 5), BOW2 contexts as follows: for a target word w with (same, with k = 2) and DEPS (dependency-based modifiers m1, . . . , mk and a head h, we consider syntactic contexts). We modified word2vec to 1 the contexts (m1, lbl1),..., (mk, lblk), (h, lblh− ), support arbitrary contexts, and to output the con- text embeddings in addition to the word embed- 2word2vec’s implementation is slightly more compli- cated. The software defaults to prune rare words based on dings. For bag-of-words contexts we used the their frequency, and has an option for sub-sampling the fre- original word2vec implementation, and for syn- quent words. These pruning and sub-sampling happen before tactic contexts, we used our modified version. The the context extraction, leading to a dynamic window size. In addition, the window size is not fixed to k but is sampled negative-sampling parameter (how many negative uniformly in the range [1, k] for each word. contexts to sample for every correct one) was 15.

304 Target Word BOW5 BOW2 DEPS All embeddings were trained on English nightwing superman superman Wikipedia. For DEPS, the corpus was tagged aquaman superboy superboy batman catwoman aquaman supergirl with parts-of-speech using the Stanford tagger superman catwoman catwoman (Toutanova et al., 2003) and parsed into labeled manhunter batgirl aquaman dumbledore evernight sunnydale Stanford dependencies (de Marneffe and Man- hallows sunnydale collinwood ning, 2008) using an implementation of the parser hogwarts half-blood garderobe calarts malfoy blandings greendale described in (Goldberg and Nivre, 2012). All to- snape collinwood millfield nondeterministic non-deterministic pauling kens were converted to lowercase, and words and non-deterministic finite-state hotelling contexts that appeared less than 100 times were turing computability nondeterministic heting deterministic buchi lessing filtered. This resulted in a vocabulary of about finite-state primality hamming 175,000 words, with over 900,000 distinct syntac- gainesville fla texas fla alabama louisiana tic contexts. We report results for 300 dimension florida jacksonville gainesville georgia tampa tallahassee california embeddings, though similar trends were also ob- lauderdale texas carolina served with 600 dimensions. aspect-oriented aspect-oriented event-driven smalltalk event-driven domain-specific object-oriented event-driven objective-c rule-based 4.1 Qualitative Evaluation prolog dataflow data-driven domain-specific 4gl human-centered Our first evaluation is qualitative: we manually in- singing singing singing dance dance rapping spect the 5 most similar words (by cosine similar- dancing dances dances breakdancing ity) to a given set of target words (Table 1). dancers breakdancing miming tap-dancing clowning busking The first target word, Batman, results in similar sets across the different setups. This is the case for Table 1: Target words and their 5 most similar words, as in- many target words. However, other target words duced by different embeddings. show clear differences between embeddings. In Hogwarts - the school of magic from the We also tried using the subsampling option fictional Harry Potter series - it is evident that (Mikolov et al., 2013b) with BOW contexts (not BOW contexts reflect the domain aspect, whereas shown). Since word2vec removes the subsam- DEPS yield a list of famous schools, capturing pled words from the corpus before creating the the semantic type of the target word. This ob- window contexts, this option effectively increases 3 servation holds for Turing and many other nouns the window size, resulting in greater topicality. as well; BOW find words that associate with w, while DEPS find words that behave like w. Turney 4.2 Quantitative Evaluation (2012) described this distinction as domain simi- We supplement the examples in Table 1 with larity versus functional similarity. quantitative evaluation to show that the qualita- The Florida example presents an ontologi- tive differences pointed out in the previous sec- cal difference; bag-of-words contexts generate tion are indeed widespread. To that end, we use meronyms (counties or cities within Florida), the WordSim353 dataset (Finkelstein et al., 2002; while dependency-based contexts provide cohy- Agirre et al., 2009). This dataset contains pairs of ponyms (other US states). We observed the same similar words that reflect either relatedness (top- behavior with other geographical locations, partic- ical similarity) or similarity (functional similar- ularly with countries (though not all of them). ity) relations.4 We use the embeddings in a re- The next two examples demonstrate that simi- trieval/ranking setup, where the task is to rank the larities induced from DEPS share a syntactic func- similar pairs in the dataset above the related ones. tion (adjectives and gerunds), while similarities The pairs are ranked according to cosine sim- based on BOW are more diverse. Finally, we ob- ilarities between the embedded words. We then serve that while both BOW5 and BOW2 yield top- draw a recall-precision curve that describes the ical similarities, the larger window size result in embedding’s affinity towards one subset (“sim- more topicality, as expected. ilarity”) over another (“relatedness”). We ex-

3 pect DEPS’s curve to be higher than BOW2’s DEPS generated a list of scientists whose name ends with “ing”. This is may be a result of occasional POS-tagging curve, which in turn is expected to be higher than errors. Still, the embedding does a remarkable job and re- trieves scientists, despite the noisy POS. The list contains 4Some word pairs are judged to exhibit both types of sim- more mathematicians without “ing” further down. ilarity, and were ignored in this experiment.

305 batman hogwarts turing 1 1 1 superman/conj− students/prep at− machine/nn− 1 1 1 spider-man/conj− educated/prep at− test/nn− 1 1 superman/conj student/prep at− theorem/poss− 1 1 spider-man/conj stay/prep at− machines/nn− 1 1 robin/conj learned/prep at− tests/nn− florida object-oriented dancing 1 1 marlins/nn− programming/amod− dancing/conj 1 1 1 beach/appos− language/amod− dancing/conj− 1 1 1 jacksonville/appos− framework/amod− singing/conj− 1 1 tampa/appos− interface/amod− singing/conj 1 1 florida/conj− software/amod− ballroom/nn Figure 2: Recall-precision curve when attempting to rank the similar words above the related ones. (a) is based on the Table 2: Words and their top syntactic contexts. WordSim353 dataset, and (b) on the Chiarello et al. dataset. not associated with subjects or objects of verbs BOW5’s. The graph in Figure 2a shows this is in- (or their inverse), but rather with conjunctions, ap- deed the case. We repeated the experiment with a positions, noun-compounds and adjectivial modi- different dataset (Chiarello et al., 1990) that was fiers. Additionally, the collapsed preposition rela- used by Turney (2012) to distinguish between do- tion is very useful (e.g. for capturing the school main and functional similarities. The results show aspect of hogwarts). The presence of many con- a similar trend (Figure 2b). When reversing the junction contexts, such as superman/conj for task such that the goal is to rank the related terms batman and singing/conj for dancing, may above the similar ones, the results are reversed, as explain the functional similarity observed in Sec- expected (not shown).5 tion 4; conjunctions in natural language tend to en- force their conjuncts to share the same semantic 5 Model Introspection types and inflections. In the future, we hope that insights from such Neural word embeddings are often considered model introspection will allow us to develop better opaque and uninterpretable, unlike sparse vec- contexts, by focusing on conjunctions and prepo- tor space representations in which each dimen- sitions for example, or by trying to figure out why sion corresponds to a particular known context, or the subject and object relations are absent and LDA models where dimensions correspond to la- finding ways of increasing their contributions. tent topics. While this is true to a large extent, we observe that SKIPGRAM does allow a non-trivial 6 Conclusions amount of introspection. Although we cannot as- sign a meaning to any particular dimension, we We presented a generalization of the SKIP- can indeed get a glimpse at the kind of informa- GRAM embedding model in which the linear bag- tion being captured by the model, by examining of-words contexts are replaced with arbitrary ones, which contexts are “activated” by a target word. and experimented with dependency-based con- Recall that the learning procedure is attempting texts, showing that they produce markedly differ- to maximize the dot product v v for good (w, c) ent kinds of similarities. These results are ex- c · w pairs and minimize it for bad ones. If we keep the pected, and follow similar findings in the distri- context embeddings, we can query the model for butional semantics literature. We also demon- the contexts that are most activated by (have the strated how the resulting embedding model can be highest dot product with) a given target word. By queried for the discriminative contexts for a given doing so, we can see what the model learned to be word, and observed that the learning procedure a good discriminative context for the word. seems to favor relatively local syntactic contexts, To demonstrate, we list the 5 most activated as well as conjunctions and objects of preposition. contexts for our example words with DEPS em- We hope these insights will facilitate further re- beddings in Table 2. Interestingly, the most dis- search into improved context modeling and better, criminative syntactic contexts in these cases are possibly task-specific, embedded representations. Our software, allowing for experimentation with 5 Additional experiments (not presented in this paper) re- arbitrary contexts, together with the embeddings inforce our conclusion. In particular, we found that DEPS perform dramatically worse than BOW contexts on analogy described in this paper, are available for download tasks as in (Mikolov et al., 2013c; Levy and Goldberg, 2014). at the authors’ websites.

306 References Lev Finkelstein, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Ey- Eneko Agirre, Enrique Alfonseca, Keith Hall, Jana tan Ruppin. 2002. Placing search in context: The Kravalova, Marius Pasca, and Aitor Soroa. 2009. concept revisited. ACM Transactions on Informa- A study on similarity and relatedness using distribu- tion Systems, 20(1):116–131. tional and wordnet-based approaches. In Proceed- ings of Human Language Technologies: The 2009 Yoav Goldberg and Omer Levy. 2014. word2vec Annual Conference of the North American Chap- explained: deriving mikolov et al.’s negative- ter of the Association for Computational Linguistics, sampling word-embedding method. arXiv preprint pages 19–27, Boulder, Colorado, June. Association arXiv:1402.3722. for Computational Linguistics. Yoav Goldberg and Joakim Nivre. 2012. A dynamic Rami Al-Rfou, Bryan Perozzi, and Steven Skiena. oracle for the arc-eager system. In Proc. of COLING 2013. Polyglot: Distributed word representations 2012. for multilingual nlp. In Proc. of CoNLL 2013. Yoav Goldberg and Joakim Nivre. 2013. Training Marco Baroni and Alessandro Lenci. 2010. Dis- deterministic parsers with non-deterministic oracles. tributional memory: A general framework for Transactions of the association for Computational corpus-based semantics. Computational Linguis- Linguistics, 1. tics, 36(4):673–721. Zellig Harris. 1954. Distributional structure. Word, Yoshua Bengio, Rejean´ Ducharme, Pascal Vincent, and 10(23):146–162. Christian Jauvin. 2003. A neural probabilistic lan- guage model. Journal of Re- Omer Levy and Yoav Goldberg. 2014. Linguistic search, 3:1137–1155. regularities in sparse and explicit word representa- tions. In Proceedings of the Eighteenth Conference Peter F Brown, Robert L Mercer, Vincent J on Computational Natural Language Learning, Bal- Della Pietra, and Jenifer C Lai. 1992. Class-based timore, Maryland, USA, June. Association for Com- n-gram models of natural. Computational Linguis- putational Linguistics. tics, 18(4). Dekang Lin. 1998. Automatic retrieval and clustering John A Bullinaria and Joseph P Levy. 2007. Extracting of similar words. In Proceedings of the 36th Annual semantic representations from word co-occurrence Meeting of the Association for Computational Lin- statistics: A computational study. Behavior Re- guistics and 17th International Conference on Com- search Methods, 39(3):510–526. putational Linguistics - Volume 2, ACL ’98, pages 768–774, Stroudsburg, PA, USA. Association for Christine Chiarello, Curt Burgess, Lorie Richards, and Computational Linguistics. Alma Pollock. 1990. Semantic and associative priming in the cerebral hemispheres: Some words Tomas Mikolov, Stefan Kombrink, Lukas Burget, do, some words don’t... sometimes, some places. JH Cernocky, and Sanjeev Khudanpur. 2011. Brain and Language, 38(1):75–104. Extensions of . In Acoustics, Speech and Signal Processing Raphael Cohen, Yoav Goldberg, and Michael Elhadad. (ICASSP), 2011 IEEE International Conference on, 2012. Domain adaptation of a dependency parser pages 5528–5531. IEEE. with a class-class selectional preference model. In Proceedings of ACL 2012 Student Research Work- Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey shop, pages 43–48, Jeju Island, Korea, July. Associ- Dean. 2013a. Efficient estimation of word represen- ation for Computational Linguistics. tations in vector space. CoRR, abs/1301.3781. Ronan Collobert and Jason Weston. 2008. A unified Tomas Mikolov, Ilya Sutskever, Kai Chen, Gregory S. architecture for natural language processing: Deep Corrado, and Jeffrey Dean. 2013b. Distributed rep- neural networks with multitask learning. In Pro- resentations of words and phrases and their com- ceedings of the 25th International Conference on positionality. In Advances in Neural Information Machine Learning, pages 160–167. Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Pro- Ronan Collobert, Jason Weston, Leon´ Bottou, Michael ceedings of a meeting held December 5-8, 2013, Karlen, Koray Kavukcuoglu, and Pavel Kuksa. Lake Tahoe, Nevada, , pages 3111– 2011. Natural language processing (almost) from 3119. scratch. The Journal of Machine Learning Re- search, 12:2493–2537. Tomas Mikolov, Wen-tau Yih, and Geoffrey Zweig. 2013c. Linguistic regularities in continuous space Marie-Catherine de Marneffe and Christopher D. Man- word representations. In Proceedings of the 2013 ning. 2008. The Stanford typed dependencies rep- Conference of the North American Chapter of the resentation. In Coling 2008: Proceedings of the Association for Computational Linguistics: Human workshop on Cross-Framework and Cross-Domain Language Technologies, pages 746–751, Atlanta, Parser Evaluation, pages 1–8, Manchester, UK, Au- Georgia, June. Association for Computational Lin- gust. Coling 2008 Organizing Committee. guistics.

307 Andriy Mnih and Geoffrey E Hinton. 2008. A scal- able hierarchical distributed language model. In Ad- vances in Neural Information Processing Systems, pages 1081–1088. Sebastian Pado´ and Mirella Lapata. 2007. Dependency-based construction of semantic space models. Computational Linguistics, 33(2):161–199. Alan Ritter, Mausam, and Oren Etzioni. 2010. A la- tent dirichlet allocation method for selectional pref- erences. In ACL, pages 424–434.

Diarmuid OS´ eaghdha.´ 2010. Latent variable models of selectional preference. In ACL, pages 435–444. Richard Socher, Jeffrey Pennington, Eric H Huang, Andrew Y Ng, and Christopher D Manning. 2011. Semi-supervised recursive for predict- ing sentiment distributions. In Proceedings of the Conference on Empirical Methods in Natural Lan- guage Processing, pages 151–161. Association for Computational Linguistics. Kristina Toutanova, Dan Klein, Chris Manning, and Yoram Singer. 2003. Feature-rich part-of-speech tagging with a cyclic dependency network. In Pro- ceedings of NAACL. Joseph Turian, Lev Ratinov, and Yoshua Bengio. 2010. Word representations: a simple and general method for semi-. In Proceedings of the 48th Annual Meeting of the Association for Compu- tational Linguistics, pages 384–394. Association for Computational Linguistics. P.D. Turney and P. Pantel. 2010. From frequency to meaning: Vector space models of semantics. Jour- nal of Artificial Intelligence Research, 37(1):141– 188. Peter D. Turney. 2012. Domain and function: A dual- space model of semantic relations and compositions. Journal of Artificial Intelligence Research, 44:533– 585. Jakob Uszkoreit and Thorsten Brants. 2008. Dis- tributed word clustering for large scale class-based language modeling in machine translation. In Proc. of ACL, pages 755–762.

308