<<

Antonymy and Conceptual Vectors

Didier Schwab, Mathieu Lafourcade and Violaine Prince LIRMM Laboratoire d’informatique, de Robotique et de Micro´electronique de Montpellier MONTPELLIER - FRANCE. schwab,lafourca,prince @lirmm.fr http://www.lirmm.fr/{ ˜ schwab,} lafourca, prince { }

Abstract a kernel of manually indexed terms is necessary For meaning representations in NLP, we focus for bootstrapping the analysis. The transver- 1 our attention on thematic aspects and concep- sal relationships , such as synonymy (LP01), tual vectors. The learning strategy of concep- antonymy and hyperonymy, that are more or tual vectors relies on a morphosyntaxic analy- less explicitly mentioned in definitions can be sis of human usage definitions linked used as a way to globally increase the coher- to vector propagation. This analysis currently ence of vectors. In this paper, we describe a doesn’t take into account negation phenomena. vectorial function of antonymy. This can help This work aims at studying the antonymy as- to improve the learning system by dealing with pects of negation, in the larger goal of its inte- negation and antonym tags, as they are often gration into the thematic analysis. We present a present in definition texts. The antonymy func- model based on the idea of symmetry compat- tion can also help to find an thema to ible with conceptual vectors. Then, we define be used in all generative text applications: op- antonymy functions which allows the construc- posite ideas research, paraphrase (by negation tion of an antonymous vector and the enumer- of the antonym), summary, etc. ation of its potentially antinomic lexical items. Finally, we introduce a measure which evaluates 2 Conceptual Vectors how a given is an acceptable antonym for We represent thematic aspects of textual seg- a term. ments (documents, paragraph, syntagms, etc) by conceptual vectors. Vectors have been used 1 Introduction in information retrieval for long (SM83) and Research in meaning representation in NLP is for meaning representation by the LSI model an important problem still addressed through (DDL+90) from latent semantic analysis (LSA) several approaches. The NLP team at LIRMM studies in psycholinguistics. In computational currently works on thematic and lexical disam- linguistics, (Cha90) proposes a formalism for biguation text analysis (Laf01). Therefore we the projection of the linguistic notion of se- built a system, with automated learning capa- mantic field in a vectorial space, from which bilities, based on conceptual vectors for mean- our model is inspired. From a set of elemen- ing representation. Vectors are supposed to en- tary concepts, it is possible to build vectors code ‘ideas’ associated to or to expres- (conceptual vectors) and to associate them to sions. The conceptual vectors learning system lexical items2. The hypothesis3 that considers automatically defines or revises its vectors ac- a set of concepts as a generator to language cording to the following procedure. It takes, as has been long described in (Rog52). Polysemic an input, definitions in natural language con- words combine different vectors corresponding tained in electronic for human us- age. These definitions are then fed to a morpho- 1well known as lexical functions (MCP95) 2 syntactic parser that provides tagging and anal- Lexical items are words or expressions which consti- tute lexical entries. For instance, ,car- or ,white ant- are ysis trees. Trees are then used as an input lexical items. In the following we will (some what) use to a procedure that computes vectors using sometimes word or term to speak about a . tree geometry and syntactic functions. Thus, 3that we call hypothesis. to different meanings. This vector approach (45 degrees) A and B are thematically close and π is based on known mathematical properties, it share many concepts. For DA(A, B) 4 , the is thus possible to undertake well founded for- thematic proximity between A and B ≥would be π mal manipulations attached to reasonable lin- considered as loose. Around 2 , they have no guistic interpretations. Concepts are defined relation. DA is a real distance function. It ver- from a thesaurus (in our prototype applied to ifies the properties of reflexivity, symmetry and French, we have chosen (Lar92) where 873 con- triangular inequality. We have, for example, cepts are identified). To be consistent with the the following angles(values are in radian and de- thesaurus hypothesis, we consider that this set grees). constitutes a generator family for the words and DA(V(,tit-), V(,tit-))=0 (0) their meanings. This familly is probably not DA(V(,tit-), V(,bird-))=0.55 (31) DA(V(,tit-), V(,sparrow-))=0.35 (20) free (no proper vectorial base) and as such, any DA(V(,tit-), V(,train-))=1.28 (73) word would project its meaning on it according DA(V(,tit-), V(,insect-))=0.57 (32) to the following principle. Let be a finite set The first one has a straightforward interpreta- C of n concepts, a conceptual vector V is a linear tion, as a ,tit- cannot be closer to anything else combinaison of elements c of . For a meaning than itself. The second and the third are not i C A, a vector V (A) is the description (in exten- very surprising since a ,tit- is a kind of ,sparrow- sion) of activations of all concepts of . For ex- which is a kind of ,bird-. A ,tit- has not much C ample, the different meanings of ,door- could be in common with a ,train-, which explains a large projected on the following concepts (the CON- angle between them. One can wonder why there CEPT[intensity] are ordered by decreasing val- is 32 degrees angle between ,tit- and ,insect-, ues): V(,door-) = (OPENING[0.8], BARRIER[0.7], which makes them rather close. If we scruti- LIMIT[0.65], PROXIMITY [0.6], EXTERIOR[0.4], IN- nise the definition of ,tit- from which its vector TERIOR[0.39], . . . is computed (Insectivourous passerine bird with In practice, the larger is, the finer the mean- colorful feather.) perhaps the interpretation of ing descriptions are. InCreturn, the computing these values seems clearer. In effect, the the- is less easy: for dense vectors4, the enumera- matic is by no way an ontological distance. tion of activated concepts is long and difficult 2.2 Conceptual Vectors Construction. to evaluate. We prefer to select the themati- cally closest terms, i.e., the neighbourhood. For The conceptual vector construction is based on instance, the closest terms ordered by increas- definitions from different sources (dictionaries, ing distance to ,door- are: (,door-)=,portal-, lists, manual indexations, etc). Defini- ,portiere-, ,opening-, ,gate-, ,barrierV -,. . . tions are parsed and the corresponding concep- tual vector is computed. This analysis method 2.1 Angular Distance shapes, from existing conceptual vectors and Let us define Sim(A, B) as one of the similar- definitions, new vectors. It requires a bootstrap ity measures between two vectors A et B, of- with a kernel composed of pre-computed vec- ten used in information retrieval (Mor99). We tors. This reduced set of initial vectors is man- can express this function as: Sim(A, B) = ually indexed for the most frequent or difficult A B terms. It constitutes a relevant lexical items cos(A, B) = · with “ ” as the scalar A B · basis on which the learning can start and rely. product. We ksuppk×kosek here that vector com- One way to build an coherent learning system ponendts are positive or null. Then, we define is to take care of the semantic relations between an angular distance D between two vectors A A items. Then, after some fine and cyclic compu- and B as D (A, B) = arccos(Sim(A, B)). In- A tation, we obtain a relevant conceptual vector tuitively, this function constitutes an evaluation basis. At the moment of writing this article, of the thematic proximity and measures the an- our system counts more than 71000 items for gle between the two vectors. We would gener- French and more than 288000 vectors, in which ally consider that, for a distance D (A, B) π A ≤ 4 2000 items are concerned by antonymy. These 4Dense vectors are those which have very few null items are either defined through negative sen- coordinates. In practice, by construction, all vectors are tences, or because antonyms are directly in the dense. dictionnary. Example of a negative definition: ,non-existence-: property of what does not exist. of the terms implies the negation of the other. Example of a definition stating antonym: ,love-: Complementary antonymy presents two kinds antonyms: ,disgust-, ,aversion-. of symmetry, (i) a value symmetry in a boolean system, as in the examples above and (ii) a sym- 3 Definition and Characterisation of metry about the application of a property (black Antonymy is the absence of color, so it is “opposed” to all We propose a definition of antonymy compat- other colors or color combinaisons). ible with the vectorial model used. Two lexi- cal items are in antonymy relation if there is 4 Antonymy Functions a symmetry between their semantic components 4.1 Principles and Definitions. relatively to an axis. For us, antonym construc- tion depends on the type of the medium that The aim of our work is to create a function supports symmetry. For a term, either we can that would improve the learning system by sim- have several kinds of antonyms if several possi- ulating antonymy. In the following, we will be bilities for symmetry exist, or we cannot have mainly interested in antonym generation, which an obvious one if a medium for symmetry is not gives a good satisfaction clue for these functions. to be found. We can distinguish different sorts We present a function which, for a given lex- of media: (i) a property that shows scalar val- ical item, gives the n closest antonyms as the neighbourhood function provides the n clos- ues (hot and cold which are symmetrical values V of temperature), (ii) the true-false relevance or est items of a vector. In order to know which application of a property (e.g. existence/non- particular meaning of the word we want to op- existence) (iii) cultural symmetry or opposition pose, we have to assess by what context mean- (e.g. sun/moon).From the point of view of lex- ing has to be constrained. However, context is ical functions, if we compare synonymy and not always sufficient to give a symmetry axis antonymy, we can say that synonymy is the for antonymy. Let us consider the item ,father-. research of resemblance with the test of sub- In the ,family- context, it can be opposite to stitution (x is synonym of y if x may replace ,mother- or to ,children- being therefore ambigu- y), antonymy is the research of the symmetry, ous because ,mother- and ,children- are by no way that comes down to investigating the existence similar items. It should be useful, when context and nature of the symmetry medium. We have cannot be used as a symmetry axis, to refine identified three types of symmetry by relying the context with a conceptual vector which is on (Lyo77), (Pal76) and (Mue97). Each sym- considered as the referent. In our example, we metry type characterises one particular type of should take as referent ,filiation-, and thus the antonymy. In this paper, for the sake of clarity antonym would be ,children- or the specialised and precision, we expose only the complemen- similar terms (e.g. ,sons- , ,daughters-) ,marriage- tary antonymy. The same method is used for or ,masculine- and thus the antonym would be the other types of antonymy, only the list of ,mother-. antonymous concepts are different. The function AntiLexS returns the n closest antonyms of the word A in the context defined 3.1 Complementary Antonymy by C and in reference to the word R. The complementary antonyms are couples like AntiLexS (A, C, R, n) event/unevent, presence/absence. AntiLexR(A, C, n) = AntiLexS (A, C, C, n) he is present he is not absent he is absent ⇒he is not present AntiLexB (A, R, n) = AntiLexS (A, R, R, n) he is not absen⇒t he is present AntiLex (A, n) = AntiLex (A, A, A, n) he is not present⇒ he is absent A S ⇒ In logical terms, we would have: The partial function AntiLexR has been de- fined to take care of the fact that in most cases, x P (x) Q(x) x P (x) Q(x) context is enough to determine a symmetry axis. ∀x Q(x) ⇒ ¬P (x) ∀x ¬Q(x) ⇒ P (x) ∀ ⇒ ¬ ∀ ¬ ⇒ AntiLexB is defined to determine a symmetry This corresponds to the exclusive disjunction axis rather than a context. In practice, we have relation. In this frame, the assertion of one AntiLexB = AntiLexR. The last function is the absolute antonymy function. For polysemic concept, context, vector . This list is called words, its usage is delicate because only one antonymh vectors of concepti list (AVC). word defines at the same time three things: the 4.2.1 AVC construction. word we oppose, the context and the referent. The Antonym Vectors of Concepts list is manu- This increases the probability to get unsatis- ally built only for the conceptual vectors of the factory results. However, given usage habits, generating set. For any concept we can have the we should admit that, practically, this function antonym vectors such as: will be the most used. It’s sequence process is presented in picture 1. We note Anti(A,C) the AntiC(EXISTENCE, V ) = V (NON-EXISTENCE) AntiC(NON-EXISTENCE, V ) = V (EXISTENCE) CALCULATION CONCEPTUAL VECTORS AntiC(AGITATION, V ) = V (INERTIA) V (REST) OF THE Vcx, Vcr ⊕ ITEMS CORRESPONDING AntiC(PLAY, V ) = V (PLAY) X, C, R VECTORS strong contextualisation V ∀ CALCULATION AntiC(ORDER, V (order) V (disorder)) = anti OF THE ⊕ ANTONYMOUS ANTONYMOUS V (DISORDER) ITEMS VECTOR X1, X2, ..., Xn AntiC(ORDER, V (classification) V (order)) = neighbourhood V (CLASSIFICATION) ⊕ IDENTIFICATION OF THE CLOSEST CONCEPTUAL VECTOR ITEMS VAnti As items, concepts can have, according to the context, a different opposite vector even Figure 1: run of the functions AntiLex if they are not polysemic. For instance, DE- STRUCTION can have for antonyms PRESERVA- antonymy function at the vector level. Here, TION, CONSTRUCTION, REPARATION or PROTEC- A is the vector we want to oppose and C the TION. So, we have defined for each concept, one context vector. conceptual vector which allows the selection of Items without antonyms: it is the case the best antonym according to the situation. of material objects like car, bottle, boat, etc. For example, the concept EXISTENCE has the The question that raises is about the continu- vector NON-EXISTENCE for antonym for any con- ity the antonymy functions in the vector space. text. The concept DISORDER has the vector of When symmetry is at stake, then fixed points ORDER for antonym in a context constituted by or plans are always present. We consider the the vectors of ORDER DISORDER5 and has CLAS- ⊕ case of these objects, and in general, non op- SIFICATION in a context constituted by CLASSI- posable terms, as belonging to the fixed space FICATION and ORDER. of the symmetry. This allows to redirect the The function AntiC(Ci, Vcontext) returns for question of antonymy to the opposable proper- a given concept Ci and the context defined by ties of the concerned object. For instance, if we Vcontext , the complementary antonym vector in want to compute the antonym of a ,motorcycle-, the list. which is a ROAD TRANSPORT, its opposable prop- 4.3 Construction of the antonym erties being NOISY and FAST, we consider its cat- vector: the Anti Function egory (i.e. ROAD TRANSPORT) as a fixed point, and we will look for a road transport (SILEN- 4.3.1 Definitions CIOUS and SLOW ), something like a ,bicycle- or We define the relative antonymy function an ,electric car-. With this method, thanks to AntiR(A, C) which returns the opposite vec- the fixed points of symmetry, opposed “ideas” tor of A in the context C and the absolute or antonyms, not obvious to the reader, could antonymy function AntiA(A) = AntiR(A, A). be discovered. The usage of AntiA is delicate because the lexi- cal item is considered as being its own context. 4.2 Antonym vectors of concept lists We will see in 4.4.1 that this may cause real Anti functions are context-dependent and can- problems because of sense selection. We should not be free of concepts organisation. They stress now on the construction of the antonym need to identify for every concept and for ev- vector from two conceptual vectors: Vitem, for ery kind of antonymy, a vector considered as 5 xi+yi the opposite. We had to build a list of triples is the normalised sum V = A B vi = V ⊕ ⊕ | k k the item we want to oppose and the other, Vc, To catch the searched meaning of the item and, for the context (referent). if it is different from the context, to catch the 4.3.2 Construction of the Antonym selection of the meaning of the referent, we use Vector the strong contextualisation method. It com- putes, for a given item, a vector. In this vector, The method is to focus on the salient notions in some meanings are favoured against others ac- V and V . If these notions can be opposed item c cording to the context. Like this, the context then the antonym should have the inverse ideas vector is also contextualised. in the same proportion. That leads us to define this function as follows: This contextualisation shows the problem caused by the absolute antonymy function N AntiR(Vitem, Vc) = Pi AntiC(Ci, Vc) i=1 × AntiαR . In this case, the method will compute the vector of the word item in the context item. 1+CV (Vitem) with Pi = V L max(Vitem , Vc ) itemi × i i This is not a problem if item has only one defini- We crafted the definition of the weight P after tion because, in this case, the strong contextu- several experiments. We noticed that the func- alisation has no effect. Otherwise, the returned tion couldn’t be symmetric (we cannot reason- conceptual vector will stress on the main idea it ably have AntiR(V(,hot-),V(,temperature-)) = contains which one is not necessary the appro- AntiR(V(,temperature-),V(,hot-))). That is why priate one. we introduce this power, to stress more on the 4.4.2 Transition Conceptual Vectors ideas present in the vector we want to oppose. Lexical Items → We note also that the more conceptual6 the vec- This transition is easier. We just have to com- tor is, the more important this power should be. pute the neighbourhood of the antonym vector That is why the power is the variation coeffi- V to obtain the items which are in thematic cient7 which is a good clue for “conceptuality”. ant antonymy with V . With this method, we To finish, we introduce this function max be- item have, for instance: cause an idea presents in the item, even if this idea is not present in the referent, has to be op- (AnticR (death, ,death- & ,life-))=(LIFE 0.4) V posed in the antonym. For example, if we want (,killer- 0.449) (,murderer- 0.467) (,blood sucker- 0.471) (,strige- 0.471) (,to die- 0.484) (,to live- 0.486) the antonym of ,cold- in the ,temperature- con- text, the weight of ,cold- has to be important (AnticR (life, ,death- & ,life-))=(,death- 0.336) V even if it is not present in ,temperature-. (DEATH 0.357) (,murdered- 0.367) (,killer- 0.377) (C3:AGE OF LIFE 0.481) (,tyrannicide- 0.516) (,to kill- 4.4 Lexical Items and Vectors: 0.579) (,dead- 0.582) Problem and Solutions (AntiCc (LIFE))=(DEATH 0.034) (,death- 0.427) V A The goal of the functions AntiLex is to return (C3:AGE OF LIFE 0.551) (,killer- 0.568) (,mudered- antonym of a lexical item. They are defined 0.588) (,tyrannicide- 0.699) (C2:HUMAN 0.737) (,to with the Anti function. So, we have to use tools kill- 0.748) (,dead- 0.77) which allow the passage between lexical items and vectors. This transition is difficult because It is not important to contextualise the con- of polysemy, i.e. how to choose the right relation cept LIFE because we can consider that, for ev- between an item and a vector. In other words, ery context, the opposite vector is the same. how to choose the good meaning of the word. In complementary antonymy, the closest item is DEATH. This result looks satisfactory. We can 4.4.1 Transition lexical items see that the distance between the antonymy vec- Conceptual Vectors → tor and DEATH is not null. It is because our As said before, antonymy is relative to a con- method is not and cannot be an exact method. text. In some cases, this context cannot be suf- The goal of our function is to build the best ficient to select a symmetry axis for antonymy. (closest) antonymy vector it is possible to have. 6In this paragraph, conceptual means: closeness of a The construction of the generative vectors is the vector to a concept second explanation. Generative vectors are in- 7 SD(V ) The variation coefficient is µ(V ) with SD as the terdependent. Their construction is based on an standart deviation and µ as the arithmetic mean. ontology. To take care of this fact, we don’t have boolean vectors, with which, we would have ex- mistake to consider that two would be π actly the same vector. The more polysemic the at a distance of about 2 . Two lexical items at term is, the farthest the closest item is, as we π 8 2 have not much in common . We would rather can see it in the first two examples. see here the illustration that two antonyms We cannot consider, even if the potential of share some ideas, specifically those which are antonymy measure is correct, the closest lexical not opposable or those which are opposable with item from Vanti as the antonym. We have to a strong activation. Only specific activated con- consider morphological features. Simply speak- cepts would participate in the opposition. A ing, if the antonym of a verb is wanted, the re- π distance of 2 between two items should rather sult would be better if a verb is caught. be interpreted as these two items do not share much idea, a kind of anti-synonymy. This re- 4.5 Antonymy Evaluation Measure sult confirms the fact that antonymy is not the Besides computing an antonym vector, it seems exact inverse of synonymy but looks more like a relevant to assess wether two lexical items can ‘negative synonymy’ where items remains quite be antonyms. To give an answer to this ques- related. To sum up, the antonym of w is not tion, we have created a measure of antonymy a word that doesn’t share ideas with w, but a evaluation. Let A and B be two vectors. word that opposes some features of w. The question is precisely to know if they can reasonably be antonyms in the context of C. 4.5.1 Examples The antonymy measure MantiEval is the an- gle between the sum of A and B and the sum In the following examples, the context has been ommited for clarity sake. In these cases, the of AnticR (A, C) and AnticR (B, C). Thus, we have: context is the sum of the vectors of the two items.

MantiEval = DA(A B, AntiR(A, C) AntiR(B, C)) ⊕ ⊕ MantiEval(EXISTENCE,NON-EXISTENCE) = 0.03 MantiEvalC (,existence-, ,non-existence-) = 0.44 MantiEvalC (EXISTENCE, CAR) = 1.45 Anti(A,C)+Anti(B,C) MantiEvalC (,existence-, ,car-) = 1.06 CAR CAR A+B MantiEvalC ( , ) = 0.006 MantiEvalC (,car-, ,car-) = 0.407 Anti(A,C)

B The above examples confirm what pre- sented. Concepts EXISTENCE and NON- EXISTENCE are very strong antonyms in comple- A mentary antonymy. The effects of the polysemy may explain that the lexical items ,existence- and Anti(B,C) ,non-existence- are less antonyms than their re- lated concepts. In complementary antonymy, Figure 2: 2D geometric representation of the antonymy CAR is its own antonym. The antonymy mea- evaluation measure Manti Eval sure between CAR and EXISTENCE is an exam- The antonymy measure is a pseudo-distance. ple of our previous remark about vectors shar- It verifies the properties of reflexivity, symme- ing few ideas and that around π/2 this mea- try and triangular inequality only for the subset sure is close to the angular distance (we have of items which doesn’t accept antonyms. In this DA(existence, car) = 1.464.). We could con- case, notwithstanding the noise level, the mea- sider of using this function to look in a concep- sure is equal to the angular distance. In the tual for the best antonyms. However, general case, it doesn’t verify reflexivity. The the computation cost (around a minute on a P4 conceptual vector components are positive and at 1.3 GHz) would be prohibitive. we have the property: Dist [0, π ]. The anti ∈ 2 smaller the measure, the more ‘antonyms’ the 8This case is mostly theorical, as there is no language two lexical items are. However, it would be a where two lexical items are without any possible relation. 5 Action on learning and method in text generation. It also determines opposite evaluation ideas in some negation cases in analysis. Many improvements are still possible, the The function is now used in the learning process. first of them being revision of the VAC lists. We can use the evaluation measure to show the These lists have been manually constructed by increase of coherence between terms: a reduced group of persons and should widely be validated and expanded especially by linguists.

MantiEvalC new old We are currently working on possible improve- ,existence-, ,non-existence- 0.33 0.44 ments of results through learning on a corpora. ,existence-, ,car- 1.1 1.06 ,car-, ,car- 0.3 0, 407 References There is no change in concepts because they are Jacques Chauch´e. D´etermination s´emantique not learned. In the opposite, the antonymy eval- en analyse structurelle : une exp´erience bas´ee uation measure is better on items. The exemple sur une d´efinition de distance. TAL Informa- shows that ,existence- and ,non-existence- have tion, 1990. been largely modified. Now, the two items are Scott C. Deerwester, Susan T. Dumais, stronger antonyms than before and the vector Thomas K. Landauer, George W. Furnas, and basis is more coherent. Of course, we can test Richard A. Harshman. Indexing by latent se- these results on the 71000 lexical items which mantic analysis. Journal of the American So- have been modified more or less directly by the ciety of Information Science, 41(6):391–407, antonymy function. We have run the test on 1990. about 10% of the concerned items and found an Mathieu Lafourcade. Lexical sorting and lexical improvement of the angular distance through transfer by conceptual vectors. In Proceeding MantiEvalC ranking to 0.1 radian. of the First International Workshop on Mul- tiMedia Annotation, Tokyo, January 2001. 6 Conclusion Larousse. Th´esaurus Larousse - des id´ees aux mots, des mots aux id´ees. Larousse, 1992. This paper has presented a model of antonymy Mathieu Lafourcade and Violaine Prince. Syn- using the formalism of conceptual vectors. Our onymies et vecteurs conceptuels. In actes de aim was to be able: (1) to spot antonymy if TALN’2001, Tours, France, July 2001. it was not given in definition and thus provide John Lyons. . Cambridge University an antonym as a result, (2) to use antonyms Press, 1977. (discovered or given) to control or to ensure the coherence of an item vector, build by learning, Igor Mel’ˇcuk, Andr´e Clas, and Alain Polgu`ere. which could be corrupted. In NLP, antonymy is Introduction a` la lexicologie explicative et a pivotal aspect, its major applications are the- combinatoire. Duculot, 1995. matic analysis of texts, construction of large lex- Emmanuel Morin. Extraction de liens ical databases and word sense disambiguation. s´emantiques entre termes a` partir de We grounded our research on a computable lin- corpus techniques. PhD thesis, Universit´e de guisitic theory being tractable with vectors for Nantes, 1999. computational sake. This preliminary work on Victoria Lynn Muehleisen. Antonymy and se- antonymy has also been conducted under the mantic range in english. PhD thesis, North- spotlight of symmetry, and allowed us to express western university, 1997. antonymy in terms of conceptual vectors. These F.R. Palmer. Semantics : a new introduction. functions allow, from a vector and some contex- Cambridge University Press, 1976. tual information, to compute an antonym vec- P. Roget. Roget’s Thesaurus of English Words tor. Some extensions have also been proposed so and Phrases. Longman, London, 1852. that these functions may be defined and usable Gerard Salton and Michael McGill. Introduc- from lexical items. A measure has been identi- tion to Modern Information Retrieval. Mc- fied to assess the level of antonymy between two GrawHill, 1983. items. The antonym vector construction is nec- essary for the selection of opposed lexical items