<<

Data Augmentation for Hypernymy Detection Thomas Kober Julie Weeds Rasa University of Sussex Berlin, DE Brighton, UK [email protected] [email protected]

Lorenzo Bertolini David Weir University of Sussex University of Sussex Brighton, UK Brighton, UK [email protected] [email protected]

Abstract Furthermore, the recognition and discovery of hyponym-hypernym relations is a foundational part The automatic detection of hypernymy rela- tionships represents a challenging problem of constructing taxonomies, which has a range of in NLP. The successful application of state- practical applications in a variety of domains such of-the-art supervised approaches using dis- as Healthcare (Bariseviciusˇ et al., 2018) or Fash- tributed representations has generally been im- ion1. peded by the limited availability of high qual- While distributed representations of are ity training data. We have developed two commonly used to find semantically similar words, novel data augmentation techniques which they do not straightforwardly provide a way to dis- generate new training examples from existing ones. First, we combine the linguistic prin- tinguish more fine-grained semantic information, ciples of hypernym transitivity and intersec- such as hypernymy, co-hyponymy and meronymy tive modifier-noun composition to generate ad- relationships. This deficiency has attracted sub- ditional pairs of vectors, such as small dog - stantial attention in the literature, and with regard dog or small dog - animal, for which a hy- to the task of hypernymy detection, both unsuper- pernymy relationship can be assumed. Sec- vised approaches (Hearst, 1992; Weeds et al., 2004; ond, we use generative adversarial networks Kotlerman et al., 2010; Santus et al., 2014; Rimell, (GANs) to generate pairs of vectors for which the hypernymy relation can also be assumed. 2014; Nguyen et al., 2017; Chang et al., 2018) and We furthermore present two complementary supervised approaches (Weeds et al., 2014; Roller strategies for extending an existing dataset by et al., 2014; Roller and Erk, 2016; Shwartz et al., leveraging linguistic resources such as - 2016; Vulic´ and Mrksiˇ c´, 2018; Rei et al., 2018; Net. Using an evaluation across 3 different Kamath et al., 2019) have been proposed. datasets for hypernymy detection and 2 differ- Supervised methods have, however, been ent vector spaces, we demonstrate that both of the proposed automatic data augmentation severely hampered by a lack of adequate train- and dataset extension strategies substantially ing data. Not only has a paucity of labelled data improve classifier performance. been an obstacle in the adoption of deep neural net- works and other more complex supervised methods, 1 Introduction but two compounding problem-specific issues have The detection of hypernymy relationships between been identified. First, there is a need to avoid lex- terms represents a challenging commonsense infer- ical overlap between the training and test sets in ence problem and is a major component in recog- order to avoid the lexical memorisation problem nising paraphrase and textual entailment in larger (Weeds et al., 2014; Levy et al., 2015a), where a units of text. Consequently, it is important for supervised method simply learns the relationships Question-Answering, Text Simplification and Au- between lexemes rather than generalising to their tomatic Summarization. For example, distributional features. Second, the performance of There are lots of cars and vans at the port today. classifiers given just the hypernym word (at training

might be adequately summarised by 1See e.g. https://www.voguebusiness.com/ There are lots of vehicles at the port today. technology/taxonomy%2Dis%2Dthe%2Dnew% 2Dfashion%2Dtech%2Dessential%2Dthe% as car and van both lexically entail, i.e. they are 2Dyes or https://fashionbrain-project. both hyponyms of the more general term vehicle. eu/fashion%2Dtaxonomy/.

1034

Proceedings of the 16th Conference of the European Chapter of the Association for Computational , pages 1034–1048 April 19 - 23, 2021. ©2021 Association for Computational Linguistics and testing) has been shown to be almost as good as been specialised using the entirety of WordNet, we performance given both words (Weeds et al., 2014; need to guard against the danger that evaluations Shwartz et al., 2017). This suggests that classifiers are simply measuring how well WordNet has been are learning the distributional features that make encoded, rather than how well the general hyper- something a more general term or a more specific nymy relationship has been learned. In light of this, term. Our conjecture is that in order to learn the we introduce a new dataset (that we call HP4K) more complex function, more complex machinery, which does not rely on WordNet in its construction. and hence more labelled data is required. We evaluate our two data augmentation tech- In computer vision or speech recognition, it is niques against two methods for increasing the size common to use data augmentation to increase the of the training data which rely on finding or mining size of the training set (Shrivastava et al., 2017; more non-synthetic examples. First, we consider Park et al., 2019). The idea is that there are certain the extraction of additional examples from Word- transformations of the data under which the class Net. Second, we consider extracting examples au- label remains invariant. For example, rotating an tomatically from a Wikipedia corpus using Hearst image does not change whether that image contains Patterns (Hearst, 1992). This provides us with what a face or not. By providing a supervised classifier one would expect to be an upper bound on what we with rotated examples, it can better generalise. might reasonably expect to achieve with a similar In this work, we consider the use of linguistic amount of synthetic examples generated using our transformations to augment existing datasets for data augmentation techniques. hypernymy detection. The challenge is to identify Our contributions are thus threefold. First, we transformations that can be applied to the repre- have identified two novel data augmentation tech- sentations of two words that are known to be in niques for the task of hypernymy detection which a hypernym relationship, such that the entailment have the potential to generate almost limitless quan- relation still holds between the transformed repre- tities of synthetic data. Second, we show, rather sentations. We propose two ways to achieve this. surprisingly, that adding synthetic data is more ef- Our first augmentation technique is based on the fective than adding non-synthetic data in almost all hypothesis that lexical entailment is transitive and cases. Third, we release a new benchmark evalua- therefore invariant under certain compositions. For tion dataset for the lexical entailment task that is example, if A entails B and B entails C then A also not dependent on WordNet. entails C. Suitable candidates for A can be found 2 Related Work by composing common intersective adjectives with the noun B. For example, if we know that car Data augmentation has recently become a very pop- entails vehicle, then we can augment the dataset ular research topic in NLP and has successfully with fast car entails car and fast car entails vehicle. been applied in machine translation systems (Sen- Our second augmentation technique is based on nrich et al., 2016; Fadaee et al., 2017; Wang et al., the hypothesis that lexical entailment is invariant 2018; Li et al., 2019; Tong et al., 2019; Matos Veliz within a certain threshold of similarity. If A entails et al., 2019; Xia et al., 2019; Gao et al., 2019; Li B, A0 is very similar to A and B0 is very similar and Specia, 2019; Liu et al., 2019), but also for to B then A0 will also entail B0. In order to obtain tasks such as relation extraction (Can et al., 2019; vectors which are sufficiently similar to the words Yan et al., 2019), text classification (Wei and Zou, in the training data, we apply generative adver- 2019), or natural inference (Kang et al., sarial networks (GANs) to create realistic-looking 2018; Junghyun et al., 2020). Most similar to our synthetic vectors, from which we choose the most usage of GANs for data augmentation is the pro- similar to the words in the training data. posal of Kang et al.(2018) who leverage a GAN- We evaluate the proposed techniques on three based setup together with WordNet for data aug- hypernymy detection datasets. The first two are mentation for natural language inference. standard benchmark tasks in this area (Weeds et al., While to the best of our knowledge this work 2014; Baroni et al., 2012), both of which are gen- represents the first application of data augmenta- erated from WordNet (Fellbaum, 1998). However, tion for lexical entailment, a number of alternative since many of the approaches to hypernmy classi- approaches have been proposed. Most proposals fication involve vector space models which have rely on supervised methods for injecting an exter-

1035 nal source of knowledge into distributional rep- hcar, vehiclei as expressing the same semantic re- resentations. Starting with retro-fitting (Faruqui lation when the modifier-noun is com- et al., 2015), vector-specialization methods mod- posed with an intersective composition function. ify existing representations to embed desired fea- We focus on adjective-noun (AN) and noun- tures (Vulic´ and Mrksiˇ c´, 2018; Rei et al., 2018; noun (NN) compounds, extracted from our ref- Kamath et al., 2019; Glavasˇ and Vulic´, 2019). erence corpus where each AN or NN phrase oc- curred at least 50 times. We filtered pairs with 3 Data Augmentation Strategies for non-subsective adjectives using a wordlist from Hypernymy Detection Nayak et al.(2014) 2. We consider two strategies for automatically Given a labelled dataset DX of triples (i) (i) (i) (i) constructing positive hyponym-hypernym pairs: hx , x , y(i)i, where x , x ∈ X hypo hyper hypo hyper hsmall dog, dogi (i) simple positive cases such as and y ∈ {0, 1}, we define data augmentation or hfast car, cari; and gapped positive cases that as adding additional hyponym-hypernym triples mimic the transitivity of hypernym relations, where 0 (j) 0 (j) 0 (j) hxhypo, xhyper, y i coming from an automatically we pair the hypernym of an existing hyponym- generated augmentation set AX 0 , such that hypernym pair with a compound hyponym. For 0 (j) 0 (j) 0 0 (j) xhypo, xhyper ∈ X and y ∈ {0, 1}, to the example if hcar, vehiclei is in the training data, we existing training set of D. We ensure that the combine car with one of its modifiers to create the data augmentation does not introduce any lexical pair hfast car, vehiclei. overlap with the existing test set, i.e. X ∩ X 0 = ∅. We construct negative pairs from the simple pos- Data augmentation strategies in NLP can roughly itive cases using two strategies: creating compo- be divided into two categories: linguistically sitional co-hyponyms such as hfast car, red cari, grounded augmentation and artificial augmenta- where we keep the head noun fixed and pair it tion. In the former, which has been the dominant with two different modifiers; and creating perturbed paradigm in NLP, any additional instances that are simple positive examples, such as hsmall dog, cati added to a training set have an actual surface form where we select the incorrect hypernym (e.g. cat) representation, i.e. the data points correspond to from the n most similar nouns to the composed hy- actual words or sentences (Kim et al., 2019; Ku- ponym (e.g. dog). We apply the same methodology mar et al., 2019; Gao et al., 2019; Andreas, 2020; to the perturbed gapped positive examples, replac- Croce et al., 2020). The latter adds instances that ing the correct hypernym with a noun from the are fully or partly artificial, meaning they do not top n neighbours of the compositional hyponym’s correspond to any words or sentences. In this work head noun. For example, given a positive pair such we propose methods for both categories, data aug- as hdog, animali, this would result in negative ex- mentation via distributional composition adds data amples such as hsmall dog, vehiclei, where the hy- points grounded in real language to a training set, ponym dog is paired with a modifier and the hyper- and data augmentation based on GANs infers plau- nym animal is replaced with one of its neighbours, sible points in latent space, which however, do not in this case, vehicle. correspond to any real linguistic objects. In neural word embeddings, an additive com- Furthermore, we distinguish between data aug- position function approximates the intersection mentation and dataset extension, where in the for- of the corresponding feature spaces (Tian et al., mer case we only leverage knowledge from the 2017), hence by creating positive pairs such as existing dataset and in the latter case we rely on hsmall dog, dogi, we encode the distributional in- expanding the training set with additionally mined clusion hypothesis (Weeds et al., 2004; Geffet and hyponym-hypernym pairs. Below, we discuss two Dagan, 2005) in the augmentation set. ways of augmenting and two ways of extending a GAN based Augmentation. We create an aug- training set. We make use of a cleaned October mentation set using Generative Adversarial Net- 2013 Wikipedia dump (Wilson, 2015) as reference works (Goodfellow et al., 2014). GANs consist of corpus to determine word and bigram frequencies. two model components — the generator and the Distributional Composition based Augmenta- discriminator — which are typically implemented tion. We take a modified noun as being in a hy- 2In preliminary experiments we did not find that filtering pernymy relation with the unmodified noun. For non-subsective adjectives had much of an effect, but decided example, we treat the pairs hfast car, cari and to move forward with the filtered data nonetheless.

1036 as neural networks. The generator’s task is to cre- extract all hypernyms and hyponyms of a given lex- ate data that mimics the distribution of the original eme, and add it as a positive hyponym-hypernym data, while the discriminator’s task is to distinguish pair if the original lexeme and any extracted hy- between data coming from the real distribution and pernym/hyponym occurs at least 30 times in our synthetic data coming from the generator. Both reference corpus. components are trained jointly until the generator We construct negative pairs based on distribu- succeeds in creating realistic data. Using GANs tional similarity, where we calculate the pairwise for data augmentation has been shown to be a suc- cosine similarities between all lexemes in the posi- cessful strategy for a number of computer vision tive set. Subsequently we use all antecedent (LHS) tasks (Shrivastava et al., 2017; Frid-Adar et al., lexemes from the extracted positive pairs and select 2018; Neff, 2018). Our goal is to create synthetic the top n most similar words for each antecedent hyponym-hypernym pairs that are similar to real as negative examples6. examples. Unlike most other scenarios involving Pattern based Extension. Hearst Pat- GANs for NLP tasks, our generated vectors do not terns (Hearst, 1992) are textual patterns such as need to correspond to actual words. a car is-a vehicle and can be automatically mined For our model - GANDALF3 - we used a list of from text corpora in an unsupervised way. This ≈40K nouns for which we had vector representa- has recently been shown to deliver strong perfor- tions as the “real” data input to GANDALF, and mance on the hypernymy detection task (Roller sampled the synthetic vectors from a Gaussian dis- et al., 2018). In this work, we leverage Hearst Pat- tribution, optimising a binary cross-entropy error terns to mine additional hyponym-hypernym pairs criterion for the generator and the discriminator, in order to extend a training set. We treat any ex- which are both simple feedforward networks with tracted noun pairs as additional positive examples a single hidden layer. We provide GANDALF’s and create the negative pairs in the same way as for full model details in AppendixA. As an additional the WordNet-based approach above. quality check for the generated vectors, we tested whether a logistic regression classifier could distin- 4 Experiments guish the synthetic and non-synthetic vectors. Typ- We evaluate our models on the datasets ically, the accuracy of the classifier was between Weeds (Weeds et al., 2014) and LEDS (Ba- 0.55-0.65, meaning the classifier is barely able to roni et al., 2012): well-studied and frequently distinguish between “real” vectors and generated used benchmarks for the hypernymy detection ones. task (Roller et al., 2014; Vilnis and McCallum, Figure1 illustrates the training loop of GAN- 2014; Roller and Erk, 2016; Carmona and Riedel, DALF as well as the selection process for construct- 2017; Shwartz et al., 2017). Since both datasets ing an augmented training set. Essentially, once use WordNet during construction, this can give GANDALF has been trained, the generator is used rise to a bias in favour of those models that also to create a large collection4 of synthetic noun vec- make use of WordNet. To address this concern, tors. To augment a dataset, D , for each triple, X we have created a new entailment dataset, HP4K, hx , x , yi ∈ D we find the n synthetic hypo hyper X that makes use of Hearst Patterns, and is manually vectors most similar to x and the n synthetic hypo annotated, thereby avoiding the use of WordNet. vectors most similar to xhyper and for each of the 2 0 0 Weeds: The dataset is based on nouns sampled n synthetic vector pairs, hxhypo, xhyperi, we create the triple hx0 , x0 , yi. The augmented training from WordNet where each noun had to occur at hypo hyper 100 set is formed by randomly sub-sampling this set of least times in Wikipedia, and its predominant triples. sense had to account for more than 50% of the WordNet based Extension. WordNet (Fell- occurrences in SemCor (Miller et al., 1993). We baum, 1998) is a large manually curated lexical use the predefined split of Weeds et al.(2014), that resource, covering a wide range of lexical relations avoids any lexical overlap between the training and between words, where groups of semantically sim- evaluation sets. The split contains 2012 examples ilar words form “synsets”5. For each synset we in the training set, evenly balanced between pos- itive and negative hyponym-hypernym pairs, and 3GAN-based Data Augmentation for Lexical inFerence. 4We would typically create half a million synthetic nouns. 2002), using WordNet 3.0. 5We use the API provided by NLTK (Loper and Bird, 6Ensuring we don’t accidentally add any real positive pairs.

1037 Figure 1: The left panel shows the training loop for GANDALF where synthetic noun vectors are generated from Gaussian noise. The discriminator’s task is to distinguish between the generated synthetic vectors and real word representations from word2vec or HyperVec. The dashed line indicates error backpropagation, on the basis of the discriminator’s current performance, through to the generator. After a final set of synthetic vectors has been generated with GANDALF, we choose the top n most similar synthetic vectors to a real representation and add these with the same label as the original hyponym-hypernym pair from an existing training set (e.g. hcat, animali) to the augmented training set (right panel).

502 examples in the evaluation set. Negative pairs were of a more diverse nature LEDS: The dataset consists of 2770 examples, and included a range of different relations, such evenly balanced between positive and negative as co-hyponyms, meronyms or reverse hyponym- hyponym-hypernym pairs. The positive examples hypernyms. Negative pairs can also be comprised are based on direct hyponym-hypernym relation- of two random nouns or two nouns without any ships from WordNet and the negative examples are semantic relation due to some amount of noise in based on a random permutation of the hypernyms extracting candidates solely on the basis of Hearst of the positive pairs. As there is no predefined Patterns. Table1 shows positive and negative ex- training/evaluation split, we make use of the 20- amples from the dataset. fold cross-validation methodology of Roller and Pair Relationship Label Erk(2016) that avoids any lexical overlap between hdog, animali hyponymy-hypernymy (Subtype) True hNirvana, bandi hyponymy-hypernymy (Instance) True training and evaluation sets. hbeef , stewi meronymy False hpedestrian, roadi topical relatedness False HP4K: We extracted Hearst Patterns from our hchemical, adenosinei reverse hyponymy-hypernymy False reference Wikipedia corpus and randomly selected hcherry, plumi co-hyponymy False 4500 unigram pairs. Subsequently, we manually hform, situationi none False annotated each pair according to whether it con- Table 1: Examples from our proposed HP4K dataset. stitutes a correct hyponymy-hypernymy relation or not. The labelling was carried out by 4 experi- HP4K consists of 4369 pairs with a class distri- enced annotators — all domain experts, familiar bution of 45:55 (positive : negative). Subsequently with the problem of hypernymy detection. We then we split the dataset into a training and evaluation split up the annotators in two teams, with each set, ensuring that there is no lexical overlap be- team annotating one half of the dataset. The initial tween the two sets. This resulted in a training set round of annotations resulted in a Cohen’s κ score of size 3426 and an evaluation set of size 9437. of 0.714, indicating substantial agreement (Viera and Garrett, 2005). Conflicts were resolved on a 4.1 Models cross-team basis such that team A would resolve We conduct experiments with two distributional team B’s annotation conflicts and vice-versa. vector space models, word2vec (Mikolov et al., During annotation we noticed that positive pairs 2013) and HyperVec (Nguyen et al., 2017). Hyper- typically fall into one of two categories. Either they Vec is based on word2vec’s skip-gram architecture were “true” subtype-supertype relations, such as and leverages WordNet to optimise the word rep- hdog, animali, or they were individual-class rela- resentations for the hypernym detection task. Hi- tionships where the hyponym is typically a named erarchical information is encoded in the norm of entity and represents a specific instance of the more 7All resources are available from https://github. general class, as for example in hNirvana, bandi. com/tttthomasssss/le-augmentation.

1038 the learned vectors, such that lexemes higher up in of accuracy for the LR and FF model, as well as the hypernymy hierarchy have larger norms than both vector spaces, averaged across all datasets. lexemes in lower parts. While in total data augmentation as well as dataset For word2vec we use the 300-dimensional pre- extension has a positive impact, the gains are larger trained Google News vectors8 and for HyperVec we for the FF model, suggesting that a higher capac- trained 100-dimensional embeddings on a October ity model is necessary to more effectively lever- 2013 Wikipedia dump (Wilson, 2015), using the age the additional information from the augmen- recommended settings of Nguyen et al.(2017), as tation source. Furthermore, before starting our ex- our augmentation sets contained many words that periments we exptected that extending an existing were OOV in the pre-trained HyperVec vectors9. dataset with WordNet represents an upper bound In our experiments, we consider a supervised on performance, given that WordNet is manually scenario where a classifier predicts a hyponym- annotated and curated. However in our experiments hypernym relation between two given word em- we found that data augmentation by either distribu- beddings. We use two different models as clas- tional composition or by using GANDALF remark- sifier a logistic regression classifier (LR), and a ably surpassed performance of the WordNet-based 3-layer feedforward neural network (FF). In both extension technique regularly. cases, the classifier takes the aggregated hypothe- The effect of data augmentation and dataset ex- sised hyponym-hypernym pair as input and predicts tension in absolute points of accuracy on each whether the pair is in a hyponym-hypernym rela- dataset individually for the FF model is shown in tion. We report a detailed overview of the model Figure3. It highlights consistent improvements parameterisation in AppendixA. across the board with only a single performance The two models share the same procedure for ag- degradation in the case of extending the LEDS gregating the word embeddings of the hypothesised dataset with Hearst Patterns when using HyperVec- hyponym-hypernym pair. For data augmentation based word representations. The results per dataset based on distributional composition, we use vec- for the LR model are presented in AppendixA and tor averaging as composition function, which gave show that the LR model is less effective in lever- substantially better performance than addition in aging the augmented data, causing more frequent preliminary experiments. performance drops. This suggests that models with more capacity are able to make more efficient use 4.2 Results of additional data and are more robust in the pres- For the FF network, we performed 10-fold cross- ence of noise which is inevitably introduced by validation on the Weeds and HP4K training sets. automatic methods. As our evaluation for LEDS is based on a 20-fold Table2 compares our FF model using word2vec cross-valiation split, rather than a pre-defined train- embeddings with all proposed techniques for aug- ing/evaluation split as for Weeds and HP4K, the menting or extending a dataset. Our techniques same procedure for hyperparameter tuning is not are able to outperform a non-augmented model by straightforwardly applicable without exposing the 4-6 points in accuracy, representing a relative error model to some of the evaluation data. However, we reduction of 14%-26%. While the primary objec- found that the top parameterisations for Weeds and tive in this work is to improve an existing model HP4K were quite similar and therefore applied hy- setup with data augmentation, our augmented mod- perparameters to the FF model for LEDS that per- els compare favourably with previously published formed well in 10-fold cross-validation on Weeds results.10 In general, data augmentation by distri- and HP4K. For data augmentation and dataset ex- butional composition or by GANDALF overcomes tension, we consider the following amounts of ad- two key weaknesses of simply extending a dataset ditional data: {0.2K, 1K, 2K, 4K, 10K, 20K, 40K}. with more data from WordNet or Hearst Patterns. All augmentation sets are balanced between posi- First, many of the hyponym-hypernym pairs we tive and negative pairs. mined from WordNet contain low-frequency words, Figure2 shows the increase in absolute points 10We note that due to the use of different performance met- 8Available from: https://code.google.com/ rics and cross-validation splits, direct model-to-model com- archive/p/word2vec/. parisons are difficult on the LEDS and Weeds datasets. Thus 9We used the HyperVec code from www.ims. we only compare to approaches that use the same evaluation uni-stuttgart.de/data/hypervec. protocol as we do.

1039 Figure 2: Effect of data augmentation and dataset extension in absolute points of accuracy averaged across all datasets over the same model without augmentation or extension. The 2 heatmaps on the left are based on data augmentation (DC=Distributional Composition, GAN=GANDALF). The 2 heatmaps on the right use dataset extension (WN=WordNet, HP=Hearst Patterns).

Figure 3: Effect of data augmentation and dataset extension in absolute points of accuracy on all datasets for the FF model.

Model Weeds LEDS HP4K dataset optimal performance is achieved with larger No Augmentation 0.72 0.77 0.67 Distributional Composition 0.76 0.83 0.70 amounts of augmentation data. One explanation GANDALF 0.75 0.80 0.71 for the different augmentation characteristics of WordNet Extension 0.75 0.83 0.69 the HP4K dataset in comparison to the other two Hearst Patterns Extension 0.74 0.81 0.68 Weeds et al.(2014) 0.75 - - datasets is its independence of WordNet during the Carmona and Riedel(2017) 0.63 0.81 - development of the dataset.

Table 2: Accuracy scores for the data augmentation and the 5.1 Data Augmentation in Space two dataset extension strategies in comparison to the same FF model without any augmentation or extension. In order to visualise what area of the vector space the GANDALF vectors and the composed vectors inhabit, we created a t-SNE (van der Maaten and which may have poor representations in our vector Hinton, 2008) projection of the vector spaces in Fig- space models. Second, while using Hearst Pat- ure5. For the visualisation we produced the near- terns typically returned higher frequency words, est neighbours of standard word2vec embeddings the retrieved candidates frequently did not repre- and augmentation embeddings for 5 exemplary sent hyponymy-hypernymy relationships. words and project them into the same space. Fig- 5 Analysis ure5 shows that the generated augmentation points, marked with an “x”, fit in with the real neighbours The concrete amount of data augmentation, i.e. the and do not deviate much from the “natural” neigh- number of additional hyponym-hypernym pairs bourhood of a given word. GANDALF vectors that are added to the training set, represents a typically inhabit the edges of a neighbour cluster, tuneable parameter. Figure4 shows the effect of whereas composed vectors are frequently closer to varying amounts of data augmentation for the FF the cluster centroid. Table3 lists the nearest neigh- model, using word2vec representations, across all bours for the example words. For word2vec and datasets. We note that all amounts of additional the composed augmentation vectors, we simply list augmentation data share the same quality, i.e. it is the nearest neighbours of each query word. For not the case that a smaller augmentation set con- GANDALF we list the nearest neighbours of the sists of “better data” or contains less noise than a generated vector that correspond to actual words. larger set. For the Weeds and LEDS datasets, peak For example, if the vector GANDALF-234 is clos- performance is typically achieved with smaller est to the representations of sugar, GANDALF-451 amounts of additional data, whereas for the HP4K and mountain, we only list sugar and mountain as

1040 Figure 4: Impact of different amounts of data augmentation.

Figure 5: t-SNE visualisation of the data augmentation spaces. Data points marked with “x” denote the representation as coming from the data augmentation set. neighbours of GANDALF-234. in the hypernym, but is able to focus on specific The composed neighbours for each word are features in a given hyponym-hypernym pair. For typically closely related to the original query, e.g. data augmentation by distributional composition raw sugar for sugar, or zoo animal for animal. The there appears to be a correlation between the per- GANDALF neighbours on the other hand have a formance of the hypernym-only and the full model, much weaker association with the query word, but i.e. a stronger model on the whole dataset also are frequently still related to it somehow as in the results in better performance for the hypernymy- example of akpeteshie, which is a spirit on sugar only model. Hence augmentation by distributional cane basis, as a neighbour for sugar. composition might not be effective in helping the model to generalise in its current form. For aug- 5.2 Data Augmentation as Regularisation mentation with GANDALF however, performance for the full model improves, while performance In the past, a prominent criticism of distributional of the hypernym-only model slightly drops, sug- methods for hypernymy detection was that such gesting that the evoked GANDALF representations models were found to frequently identify features have a regularisation effect, while also improving of a prototypical hypernym in the distributional generalisation. Hence, a fruitful avenue for future representations, rather than being able to dynami- work will be further leveraging data augmentation cally focus on the relevant features that are indica- for regularisation. tive of a hypernymy relation for a specific pair of words (Weeds et al., 2014; Levy et al., 2015b). We 6 Conclusion therefore briefly investigate whether data augmen- tation can be used as a regularisation mechanism In NLP, in contrast to computer vision, data aug- that helps prevent models from overfitting on pro- mentation has not been applied as standard due totypical hypernym features. to the apparent lack of universal rules for label- Table4 shows the results on the Weeds dataset invariant language transformations. using a hypernym-only FF model with word2vec We have considered the problem of hypernymy representations, in comparison to the same model detection, and proposed two novel techniques for variant that makes use of the hyponym and the data augmentation. These techniques rely on se- hypernym. Ideally, we would hope to see weak per- mantic rules rather than an external knowledge formance for the hypernym-only and strong perfor- source, and have the potential to generate almost mance on the full model. This would indicate that limitless synthetic data for this task. We demon- the classifier does not rely on prototypical features strate that these techniques perform better than

1041 Word word2vec Neighbours GANDALF Neighbours Composed Neighbours sugar refined sugar, cane sugar, cmh rawalpindi, prescribed antipsychotic medication, raw sugar, white sugar, brown sugar, turbinado, cocoa, sugars sugar, mumtaz bhutto, akpeteshie sugar price, sugar industry dog dogs, puppy, pit bull, pooch, cat ellis burks, sniffing glue, microchip implants, pet cat, dog fighting, cat breed, rat terrier liz klekamp, cf rocco baldelli terrier breed ocean sea, oceans, pacific ocean, pacific ocean, heavily vegetated, alaska aleutian, ocean basin, pacific ocean, shallow sea, atlantic ocean, oceanic seagrasses, plutoid sea fish, sea mammal animal animals, animal welfare, dog, pet, animals, pet, hallway feeds, poop scooping, first animal, adult animal, zoo animal, cats panhandle animal welfare different animal, animal organization energy renewable energy, enery, radial velocity measurements, stopped, energy efficiency, solar energy, state energy, electricity, enegy, fossil fuel renewable energy, bicycle advisory, steinkuehler food energy, energy company

Table 3: Nearest neighbours for word2vec, GAN vectors and composed vectors.

Augmentation Hypernym-Only Full No Data Augmentation 0.59 0.72 until it finally converged towards acceptance. Distributional Composition (size=100) 0.60 0.74 Distributional Composition (size=500) 0.57 0.71 GANDALF (size=500) 0.58 0.75 GANDALF (size=1000) 0.60 0.73 References Table 4: Accuracy for the hypernym-only and full models on Jacob Andreas. 2020. Good-enough compositional the Weeds dataset with no, DC or GAN augmentation. data augmentation. In Proceedings of the 58th An- nual Meeting of the Association for Computational Linguistics, pages 7556–7566, Online. Association extending the training set with additional non- for Computational Linguistics. synthetic data, drawn from an external knowledge source in most cases. Our results are consistent Gintaras Barisevicius,ˇ Martin Coste, David Geleta, Damir Juric, Mohammad Khodadadi, Giorgos Stoi- across evaluation benchmarks, word vector spaces los, and Ilya Zaihrayeu. 2018. Supporting digital and classification architectures. We have also healthcare services using semantic web technologies. shown that our approach is effective even when In The Semantic Web – ISWC 2018, pages 291–306. the word vector space model has already been spe- Springer International Publishing. cialised for hypernymy detection. Marco Baroni, Raffaella Bernardi, Ngoc-Quynh Do, Since WordNet is widely used as a source of and Chung-chieh Shan. 2012. Entailment above the information about semantic relations, we have pro- word level in distributional . In Proceed- posed a new evaluation benchmark that is indepen- ings of the 13th Conference of the European Chap- dent of WordNet. Whilst results are lower across ter of the Association for Computational Linguis- tics, pages 23–32, Avignon, France. Association for the board on this dataset, suggesting that it is more Computational Linguistics. difficult than the others, we see the same pattern of increasing performance with a more complex Duy-Cat Can, Hoang-Quynh Le, Quang-Thuy Ha, and classifier and the use of data augmentation. Nigel Collier. 2019. A richer-but-smarter shortest dependency path with attentive augmentation for re- Future work includes leveraging data augmenta- lation extraction. In Proceedings of the 2019 Con- tion for more complex models and the extension ference of the North American Chapter of the Asso- of our approach to a multilingual setup as well as ciation for Computational Linguistics: Human Lan- domains with a more specialised vocabulary such guage Technologies, Volume 1 (Long and Short Pa- pers), pages 2902–2912, Minneapolis, Minnesota. as Healthcare or Fashion. Association for Computational Linguistics.

Acknowledgements Vicente Ivan Sanchez Carmona and Sebastian Riedel. 2017. How well can we predict hypernyms from Part of this research was supported by EPSRC grant word embeddings? a dataset-centric analysis. In no. 2129720: Composition and Entailment in Dis- Proceedings of the European Chapter of the Asso- tributed Word Representations. We thank Mark ciation for Computational Linguistics. Steedman for valuable feedback and insightful dis- Haw-Shiuan Chang, Ziyun Wang, Luke Vilnis, and An- cussions that had substantial impact on the quality drew McCallum. 2018. Distributional inclusion vec- the paper. We also thank Nicola De Cao, without tor embedding for unsupervised hypernymy detec- whose help GANDALF wouldn’t have been able to tion. In Proceedings of the 2018 Conference of the defeat the Balrog of mediocre performance. Lastly, North American Chapter of the Association for Com- putational Linguistics: Human Language Technolo- we thank several mini-batches of reviewers for their gies, Volume 1 (Long Papers), pages 485–495, New stochastic gradients helpful comments that itera- Orleans, Louisiana. Association for Computational tively improved the paper, rejection after rejection, Linguistics.

1042 Danilo Croce, Giuseppe Castellucci, and Roberto Marti Hearst. 1992. Automatic acquisition of hy- Basili. 2020. GAN-BERT: Generative adversarial ponyms from large text corpora. In Proceedings of learning for robust text classification with a bunch the International Conference on Computational Lin- of labeled examples. In Proceedings of the 58th An- guistics. nual Meeting of the Association for Computational Linguistics, pages 2114–2119, Online. Association Sergey Ioffe and Christian Szegedy. 2015. Batch nor- for Computational Linguistics. malization: Accelerating deep network training by reducing internal covariate shift. In Proceedings Marzieh Fadaee, Arianna Bisazza, and Christof Monz. of the 32Nd International Conference on Interna- 2017. Data augmentation for low-resource neural tional Conference on Machine Learning - Volume 37, machine translation. In Proceedings of the 55th An- ICML’15, pages 448–456. JMLR.org. nual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 567– Min Junghyun, McCoy R. Thomas, Dipanjan Das, 573, Vancouver, Canada. Association for Computa- Pitler Emily, and Linzen Tal. 2020. Syntactic tional Linguistics. data augmentation increases robustness to inference heuristics. In Proceedings of the 58th Annual Meet- Manaal Faruqui, Jesse Dodge, Sujay Kumar Jauhar, ing of the Association for Computational Linguistics, Chris Dyer, Eduard Hovy, and Noah A. Smith. 2015. Seattle, Washington. Association for Computational Retrofitting word vectors to semantic . In Linguistics. Proceedings of the 2015 Conference of the North American Chapter of the Association for Computa- Aishwarya Kamath, Jonas Pfeiffer, Edoardo Maria tional Linguistics: Human Language Technologies, Ponti, Goran Glavas,ˇ and Ivan Vulic.´ 2019. Spe- pages 1606–1615, Denver, Colorado. Association cializing distributional vectors of all words for lex- for Computational Linguistics. ical entailment. In Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP- Christiane Fellbaum. 1998. WordNet: An Electronic 2019), pages 72–83, Florence, Italy. Association for Lexical Database. Bradford Books. Computational Linguistics. M. Frid-Adar, E. Klang, M. Amitai, J. Goldberger, Dongyeop Kang, Tushar Khot, Ashish Sabharwal, and and H. Greenspan. 2018. Synthetic data augmenta- Eduard Hovy. 2018. AdvEntuRe: Adversarial train- tion using gan for improved liver lesion classifica- ing for textual entailment with knowledge-guided ex- tion. In 2018 IEEE 15th International Symposium amples. In Proceedings of the 56th Annual Meet- on Biomedical Imaging (ISBI 2018), pages 289–293. ing of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2418–2428, Mel- Fei Gao, Jinhua Zhu, Lijun Wu, Yingce Xia, Tao bourne, Australia. Association for Computational Qin, Xueqi Cheng, Wengang Zhou, and Tie-Yan Liu. Linguistics. 2019. Soft contextual data augmentation for neural machine translation. In Proceedings of the 57th An- nual Meeting of the Association for Computational Hwa-Yeon Kim, Yoon-Hyung Roh, and Young-Kil Linguistics, pages 5539–5544, Florence, Italy. Asso- Kim. 2019. Data augmentation by data noising ciation for Computational Linguistics. for open-vocabulary slots in spoken language under- standing. In Proceedings of the 2019 Conference of M. Geffet and I. Dagan. 2005. The distributional inclu- the North American Chapter of the Association for sion hypotheses and lexical entailment. In Proceed- Computational Linguistics: Student Research Work- ings of the 43rd Annual Meeting on Association for shop, pages 97–102, Minneapolis, Minnesota. Asso- Computational Linguistics, pages 107–114. Associ- ciation for Computational Linguistics. ation for Computational Linguistics. Diederik P. Kingma and Jimmy Ba. 2014. Adam: Goran Glavasˇ and Ivan Vulic.´ 2019. Generalized tun- A method for stochastic optimization. CoRR, ing of distributional word vectors for monolingual abs/1412.6980. and cross-lingual lexical entailment. In Proceedings of the 57th Annual Meeting of the Association for Lili Kotlerman, Ido Dagan, Idan Szpektor, and Maayan Computational Linguistics, pages 4824–4830, Flo- Zhitomirsky-Geffet. 2010. Directional distribu- rence, Italy. Association for Computational Linguis- tional similarity for lexical inference. Special Issue tics. of Natural Language Engineering on Distributional , 4(16):359–389. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Varun Kumar, Hadrien Glaude, Cyprien de Lichy, and Courville, and Yoshua Bengio. 2014. Generative Wlliam Campbell. 2019. A closer look at feature adversarial nets. In Z. Ghahramani, M. Welling, space data augmentation for few-shot intent classi- C. Cortes, N. D. Lawrence, and K. Q. Weinberger, fication. In Proceedings of the 2nd Workshop on editors, Advances in Neural Information Processing Deep Learning Approaches for Low-Resource NLP Systems 27, pages 2672–2680. Curran Associates, (DeepLo 2019), pages 1–10, Hong Kong, China. As- Inc. sociation for Computational Linguistics.

1043 Omer Levy, Yoav Goldberg, and Ido Dagan. 2015a. Im- George A. Miller, Claudia Leacock, Randee Tengi, and proving distributional similarity with lessons learned Ross T. Bunker. 1993. A semantic concordance. In from word embeddings. Transactions of the Associ- Proceedings of the Arpa Workshop on Human Lan- ation for Computational Linguistics, 3:211–225. guage Technology, pages 303–308. Omer Levy, Steffen Remus, Chris Biemann, and Ido Neha Nayak, Mark Kowarsky, Gabor Angeli, and Dagan. 2015b. Do supervised distributional meth- Christopher D. Manning. 2014. A of non- ods really learn lexical inference relations? In Pro- subsective adjectives. Technical Report CSTR 2014- ceedings of the 2015 Conference of the North Amer- 04, Department of Computer Science, Stanford Uni- ican Chapter of the Association for Computational versity. Linguistics: Human Language Technologies, pages 970–976, Denver, Colorado. Thomas Neff. 2018. Data augmentation in deep learn- ing using generative adversarial networks. Master’s Guanlin Li, Lemao Liu, Guoping Huang, Conghui Zhu, thesis, Graz University of Technology. and Tiejun Zhao. 2019. Understanding data aug- mentation in neural machine translation: Two per- Kim Anh Nguyen, Maximilian Koper,¨ Sabine spectives towards generalization. In Proceedings of Schulte im Walde, and Ngoc Thang Vu. 2017. the 2019 Conference on Empirical Methods in Nat- Hierarchical embeddings for hypernymy detection ural Language Processing and the 9th International and directionality. In Proceedings of the 2017 Con- Joint Conference on Natural Language Processing ference on Empirical Methods in Natural Language Processing (EMNLP-IJCNLP), pages 5689–5695, Hong Kong, , pages 233–243, Copenhagen, Denmark. China. Association for Computational Linguistics. Association for Computational Linguistics. Daniel S. Park, William Chan, Yu Zhang, Chung- Zhenhao Li and Lucia Specia. 2019. Improving neu- Cheng Chiu, Barret Zoph, Ekin D. Cubuk, and ral machine translation robustness via data augmen- Quoc V. Le. 2019. Specaugment: A simple data aug- tation: Beyond back-translation. In Proceedings of mentation method for automatic speech recognition. the 5th Workshop on Noisy User-generated Text (W- CoRR, abs/1904.08779. NUT 2019), pages 328–336, Hong Kong, China. As- sociation for Computational Linguistics. Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Angli Liu, Jingfei Du, and Veselin Stoyanov. 2019. Lin, Alban Desmaison, Luca Antiga, and Adam Knowledge-augmented language model and its ap- Lerer. 2017. Automatic differentiation in pytorch. plication to unsupervised named-entity recognition. In NIPS-W. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Compu- Fabian Pedregosa, Gael¨ Varoquaux, Alexandre Gram- tational Linguistics: Human Language Technolo- fort, Vincent Michel, Bertrand Thirion, Olivier gies, Volume 1 (Long and Short Papers), pages Grisel, Mathieu Blondel, Peter Prettenhofer, Ron 1142–1150, Minneapolis, Minnesota. Association Weiss, Vincent Dubourg, Jake Vanderplas, Alexan- for Computational Linguistics. dre Passos, David Cournapeau, Matthieu Brucher, Matthieu Perrot, and Edouard´ Duchesnay. 2011. Edward Loper and Steven Bird. 2002. Nltk: The natu- Scikit-learn: Machine learning in python. Journal ral language toolkit. In Proceedings of the Workshop of Machine Learning Research, 12:2825–2830. on Effective Tools and Methodologies for Teaching Natural Language Processing and Computational Marek Rei, Daniela Gerz, and Ivan Vulic. 2018. Scor- Linguistics - Volume 1, pages 63–70. Association for ing lexical entailment with a supervised directional Computational Linguistics. similarity network. In ACL. Laurens van der Maaten and Geoffrey Hinton. 2008. Laura Rimell. 2014. Distributional lexical entailment Visualizing data using t-SNE. Journal of Machine by topic coherence. In Proceedings of the 14th Learning Research, 9:2579–2605. Conference of the European Chapter of the Asso- ciation for Computational Linguistics, pages 511– Claudia Matos Veliz, Orphee De Clercq, and Veronique 519, Gothenburg, Sweden. Association for Compu- Hoste. 2019. Benefits of data augmentation for tational Linguistics. NMT-based text normalization of user-generated content. In Proceedings of the 5th Workshop on Stephan Roller, Katrin Erk, and Gemma Boleda. 2014. Noisy User-generated Text (W-NUT 2019), pages Inclusive yet selective: Supervised distributional hy- 275–285, Hong Kong, China. Association for Com- pernymy detection. In Proceedings of the Interna- putational Linguistics. tional Conference on Computational Linguistics. Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Cor- Stephen Roller and Katrin Erk. 2016. Relations such rado, and Jeff Dean. 2013. Distributed representa- as hypernymy: Identifying and exploiting hearst pat- tions of words and phrases and their composition- terns in distributional vectors for lexical entailment. ality. In C.J.C. Burges, L. Bottou, M. Welling, In Proceedings of the 2016 Conference on Empiri- Z. Ghahramani, and K.Q. Weinberger, editors, Ad- cal Methods in Natural Language Processing, pages vances in Neural Information Processing Systems 2163–2172, Austin, Texas. Association for Compu- 26, pages 3111–3119. Curran Associates, Inc. tational Linguistics.

1044 Stephen Roller, Douwe Kiela, and Maximilian Nickel. of the 6th Workshop on Asian Translation, pages 2018. Hearst patterns revisited: Automatic hyper- 147–151, Hong Kong, China. Association for Com- nym detection from large text corpora. In Proceed- putational Linguistics. ings of the 56th Annual Meeting of the Associa- tion for Computational Linguistics (Volume 2: Short Anthony J. Viera and Joanne M. Garrett. 2005. Under- Papers), pages 358–363. Association for Computa- standing interobserver agreement: The kappa statis- tional Linguistics. tic. Family medicine, 37:360–3. Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Luke Vilnis and Andrew McCallum. 2014. Word Vicki Cheung, Alec Radford, Xi Chen, and Xi Chen. representations via gaussian embedding. CoRR, 2016. Improved techniques for training gans. In abs/1412.6623. D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, Ivan Vulic´ and Nikola Mrksiˇ c.´ 2018. Specialising word and R. Garnett, editors, Advances in Neural Informa- vectors for lexical entailment. In Proceedings of the tion Processing Systems 29, pages 2234–2242. Cur- 2018 Conference of the North American Chapter of ran Associates, Inc. the Association for Computational Linguistics: Hu- man Language Technologies, Volume 1 (Long Pa- Enrico Santus, Alessandro Lenci, and Qin Lu. 2014. pers), pages 1134–1145. Association for Computa- Chasing hypernyms in vector space with entropy. In Proceedings of the European Chapter of the Associ- tional Linguistics. ation for Computational Linguistics. Xinyi Wang, Hieu Pham, Zihang Dai, and Graham Neubig. 2018. SwitchOut: an efficient data aug- Rico Sennrich, Barry Haddow, and Alexandra Birch. mentation algorithm for neural machine translation. 2016. Improving neural machine translation mod- In Proceedings of the 2018 Conference on Empiri- els with monolingual data. In Proceedings of the cal Methods in Natural Language Processing, pages 54th Annual Meeting of the Association for Compu- 856–861, Brussels, Belgium. Association for Com- tational Linguistics (Volume 1: Long Papers), pages putational Linguistics. 86–96, Berlin, Germany. Association for Computa- tional Linguistics. Julie Weeds, Daoud Clarke, Jeremy Reffin, David Weir, and Bill Keller. 2014. Learning to distinguish hyper- A. Shrivastava, T. Pfister, O. Tuzel, J. Susskind, nyms and co-hyponyms. In Proceedings of COL- W. Wang, and R. Webb. 2017. Learning from sim- ING 2014, the 25th International Conference on ulated and unsupervised images through adversarial Computational Linguistics: Technical Papers, pages training. In 2017 IEEE Conference on Computer Vi- 2249–2259, Dublin, Ireland. Dublin City University sion and Pattern Recognition (CVPR), pages 2242– and Association for Computational Linguistics. 2251. Julie Weeds, David Weir, and Diana McCarthy. 2004. Vered Shwartz, Yoav Goldberg, and Ido Dagan. 2016. Characterising measures of lexical distributional Improving hypernymy detection with an integrated similarity. In Proceedings of the 20th International path-based and distributional method. In Proceed- Conference on Computational Linguistics, pages ings of the 54th Annual Meeting of the Association 1015–1021, Geneva, Switzerland. for Computational Linguistics (Volume 1: Long Pa- pers), pages 2389–2398, Berlin, Germany. Jason Wei and Kai Zou. 2019. EDA: Easy data aug- mentation techniques for boosting performance on Vered Shwartz, Enrico Santus, and Dominik text classification tasks. In Proceedings of the Schlechtweg. 2017. Hypernyms under siege: 2019 Conference on Empirical Methods in Natu- Linguistically-motivated artillery for hypernymy ral Language Processing and the 9th International detection. In Proceedings of the 15th Conference Joint Conference on Natural Language Processing of the European Chapter of the Association for (EMNLP-IJCNLP), pages 6382–6388, Hong Kong, Computational Linguistics: Volume 1, Long Papers, China. Association for Computational Linguistics. pages 65–75, Valencia, Spain. Association for Computational Linguistics. Benjamin Wilson. 2015. The unknown perils of min- ing wikipedia. Https://blog.lateral.io/2015/06/the- Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, unknown-perils-of-mining-wikipedia/. Ilya Sutskever, and Ruslan Salakhutdinov. 2014. Dropout: A simple way to prevent neural networks Mengzhou Xia, Xiang Kong, Antonios Anastasopou- from overfitting. Journal of Machine Learning Re- los, and Graham Neubig. 2019. Generalized data search, 15:1929–1958. augmentation for low-resource translation. In Pro- ceedings of the 57th Annual Meeting of the Asso- Ran Tian, Naoaki Okazaki, and Kentaro Inui. 2017. ciation for Computational Linguistics, pages 5786– The mechanism of additive composition. Machine 5796, Florence, Italy. Association for Computa- Learning, 106(7):1083–1130. tional Linguistics. Yixuan Tong, Liang Liang, Boyan Liu, Shanshan Jianhao Yan, Lin He, Ruqin Huang, Jian Li, and Ying Jiang, and Bin Dong. 2019. Supervised neural ma- Liu. 2019. Relation extraction with temporal rea- chine translation based on data augmentation and im- soning based on memory augmented distant super- proved training & inference process. In Proceedings vision. In Proceedings of the 2019 Conference of

1045 the North American Chapter of the Association for Computational Linguistics: Human Language Tech- nologies, Volume 1 (Long and Short Papers), pages 1019–1030, Minneapolis, Minnesota. Association for Computational Linguistics.

1046 A Supplemental Material A.3 Results Figure6 below shows the effect of data augmentia- A.1 GANDALF Model Details tion in terms of points of Accuracy for the logistic regression classifier per vector space model and The generator and discriminator in GANDALF are dataset. Unlike for the higher-capacity feedforward single layer feedforward networks, with tanh acti- model, data augmentation frequently causes per- vations and a dropout ratio (Srivastava et al., 2014) formance to go down for the simpler linear model. of 0.3. We used ADAM (Kingma and Ba, 2014) This suggests that more complex models are re- to optimise a binary cross entropy error criterion quired to fully leverage the additional information with a learning rate of 0.0002 and β values of 0.5 from the augmentation sets. and 0.999. We found that GANDALF required Table5 below gives an overview over the com- quite a bit of wizardry to achieve strong perfor- plete results for both classifier and vector space mance and we found the website https://github. models, across all datasets. It shows the consistent com/soumith/ganhacks very helpful. For exam- positive effect of data augmentation on the more ple we applied label noise and soft labels (Sali- complex feedforward model in comparison to the mans et al., 2016) and used a batch normalisation logistic regression classifier, which is less robust to layer (Ioffe and Szegedy, 2015), which had the the small amounts of noise that is inevitably intro- largest impact on model performance. GANDALF duced by the automatic augmentation algorithm. is implemented in PyTorch (Paszke et al., 2017) and we release our code on https://github.com/ tttthomasssss/le-augmentation.

A.2 Model Details

For our linear model we use the logistic regrs- sion classifier implemented in scikit-learn (Pe- dregosa et al., 2011). Our neural network model is 3-layer feedforward model implemented in Py- Torch (Paszke et al., 2017). We tuned the parameters of the Feedforward neu- ral network by 10-fold cross-validation on the re- spective training sets, except for LEDS, where we chose the parameters on the basis of a model that performed well on the Weeds and HP4K. Our pa- rameter grid consisted of activation function: {tanh, relu}, dropout: { 0.0, 0.1, 0.3 } and hidden layer sizes, where we considered { 200-200-200, 200- 100-50, 200-50-30 } for Hypervec and { 600-600- 600, 600-400-200, 600-300-100, 600-200-50} for word2vec. We furthermore considered 3 different aggregation functions: diff (Weeds et al., 2014), which simply takes the elementwise difference of the embedding pair; asym (Roller et al., 2014) which is the concatenation of the difference and the squared difference of the embedding pair; and concat-asym (Roller and Erk, 2016), which is the concatenation of the embedding pair, their dif- ference, and their squared difference. We trained all models for 30 epochs with early stopping and used ADAM with a learning rate of 0.01 to opti- mise a cross entropy error criterion.

1047 Figure 6: Effect of data augmentation and dataset extension in absolute points of accuracy on all datasets for the LR model.

Weeds LEDS HP4K Model None DC GAN WN HP None DC GAN WN HP None DC GAN WN HP Vector Space LR 0.69 0.73 0.72 0.74 0.70 0.81 0.80 0.78 0.80 0.80 0.64 0.61 0.63 0.63 0.63 word2vec FF 0.72 0.76 0.75 0.75 0.74 0.77 0.83 0.80 0.83 0.81 0.67 0.70 0.71 0.69 0.68 LR 0.70 0.71 0.74 0.73 0.70 0.79 0.76 0.78 0.79 0.78 0.63 0.69 0.69 0.68 0.67 hypervec FF 0.71 0.74 0.75 0.75 0.74 0.79 0.80 0.81 0.80 0.75 0.66 0.70 0.72 0.70 0.70

Table 5: Accuracy scores for the data augmentation strategies (DC and GAN), and the two dataset extension strategies (WN and HP), and the baseline that neither uses augmentation nor extension (None). Boldfaced results denote top performance per vector space and dataset, underlined results denote improved performance in comparison to the baseline without data augmentation.

1048