<<

Linking Knowledge Graphs across with and Machine

John P. McCrae, Mihael Arcan,ˇ Paul Buitelaar Insight Centre for Data Analytics National University of Ireland Galway Galway, Ireland [email protected], [email protected], [email protected]

Abstract be well exploited [25]. In this paper, we propose a hybrid approach Knowledge graphs and ontologies under- that combines dataset alignment techniques and pin many natural processing ap- ontology translation techniques in order to trans- plications, and to apply these to new lan- late a dataset to a new language. In particular, guages, these knowledge graphs must be we consider three cases: firstly, we use dataset translated. Up until now, this has been alignment techniques to link the European Health achieved either by direct label transla- Records ontology1 with a highly multilingual re- tion or by cross-lingual alignment, which source, namely DBpedia to translate this ontology matches the concepts in the graph to an- from English into German. Secondly, we apply other graph in the target languages. We the OTTO machine translation system for ontolo- show that these two approaches can, in gies [1] to translate the labels of the ontology di- fact, be combined and that the combi- rectly. Finally, we combine these two approaches nation of machine translation and cross- showing a 30 point increase in BLEU score over lingual alignment can obtain improved re- the alignment model and 2 point increase over the sults for translating a biomedical ontology machine translation model. from English to German. 2 Related Work 1 Introduction The growth of semantic information published in Knowledge graphs, including large databases such recent years has been stimulated to a large extent as DBpedia [18], have found a wide range of use by the emergence of linked data, which showed to cases in many domains and in many languages. be the best way to expose, share and connect data The applications that depend on these knowledge in a language-independent fashion on the Seman- graphs frequently use and thus tic Web [15]. Although the idea of the Semantic to adapt them to new languages it is necessary to Web depends on language independent informa- add labels for concepts in this language. More- tion access, most of the ontologies published are over, sometimes it is the case that there already expressed in English, leading to a biased access to exists a suitable knowledge graph created in that the . On the other hand, the infor- language and as such it would be preferable to mation on the traditional Web can be found in dif- align the two knowledge graphs. The task of cross- ferent languages, hence is language-specific. For lingual alignment has generally required statistical these reasons, approaches have to be put in place machine translation (SMT) and for the most part to interlink and lexicalise these resources on a has simply translated the vocabulary and applied multilingual basis. Therefore, language communi- a monolingual alignment, although some authors ties started publishing data in different languages, have suggested translating into a third ‘pivot’ lan- so that no matter in which language ontologies guage [34]. In most cases there is not a full and or documents are published, access to knowledge exactly similar counterpart to a knowledge graph will be supported [5]. on the Web, but there is a large amount of relevant Nevertheless, multilingualism is one of the multilingual data already on the Web and the abil- ity to reuse this in machine translation has still not 1http://trajano.us.es/˜isabel/EHR/ biggest challenges for the Semantic Web since it task focuses on finding whether two occurrences is important to access information in the language in different languages refer to the same object. of the user’s choice. Most of the previous re- The authors evaluate the suitability of Microsoft lated work tackled this problem by using multilin- Translator to interlink RDF resources, represented gual lexical resources, like EuroWordNet or IATE as text documents in English and Chinese. The [7,8]. Their work focuses on the identification results demonstrated that the method can iden- of the lexical overlap between the ontology la- tify most of the correct matches using minimum bels and the multilingual resources, which guar- information in a resource description with preci- antees a high precision but a low recall. There- sion over 98%. [27] demonstrates the synergy of fore, external translation services like BabelFish, Wikipedia and Princeton WordNet [12] structure SDL FreeTranslation tool or Translate as a wide-coverage multilingual ontology, called were used to overcome this issue [11, 13]. Ad- BabelNet. This resource is created by linking ditionally, [11] and [23] performed ontology la- Wikipedia entries and Wordnet synsets together. bel disambiguation, where the ontology structure Since the lexical knowledge is the key for under- is used to annotate the labels with their seman- standing and decoding a continuously changing tic senses. Similarly, [24] show positive effects world, they use commercial translation systems to of different domain adaptation techniques, i.e., us- fill the lexical gaps. ing Web resources as additional bilingual knowl- There has also been research in the process edge, re-scoring with Explicit Seman- of finding translingual semantic representations, tic Analysis and adaptation for au- starting with methods that merge parallel corpora tomatic ontology translation. to obtain a translingual representation by means of latent topic modelling methods such as Latent The task on ontology alignment frequently Semantic Analysis [10] and Latent Dirichlet Allo- builds on a label translation system and thus con- cation [26]. These methods can be used to esti- verts the task of finding a cross-lingual alignment mate the latent similarity between ontology labels to that of finding a monolingual alignment among in different languages, but until recently such ap- translated labels [34]. [13] used Google Trans- proaches did not yield results that outperformed late in their work to translate the labels nested direct translation [32] in comparable tasks. In par- within the context of a larger label. The au- ticular, recent methods such as Orientated Prin- thors stress the importance of the translation qual- ciple Component Analysis [32], Kernel Canoni- ity to generate high-quality matching results for cal Correlation Analysis [36, CCA] and Orthonor- monolingual matching tools but do not provide mal Explicit Topic Analysis [22, ONETA] cal- a deeper evaluation on the translation part. In culate a translingual representation by means of their follow-up work [14], the authors illustrate estimating the correlation between term frequen- a semantic oriented cross-lingual ontology map- cies in a document-aligned corpus and this has ping, with disambiguating labels on the target been shown to outperform -based ma- side. Their strategy is based on a manual pro- chine translation [17]. Additionally, ontology la- cessing of the ontology labels and the usage of bel disambiguation was performed by [23], where the knowledge extracted from bilingual corpora. the structure of the ontology along with existing Additionally their approach also linguistically en- multilingual ontologies was used to annotate the riches the labels in the ontology. Their semantic- labels with their semantic senses. In a similar di- oriented cross-lingual ontology mapping approach rection, [6] derived cross-lingual alignments be- (SOCOM) uses off-the-shelf machine translation tween English and Chinese WordNet by means of systems, such as . They use the a distributional similarity approach. ontology relations, if the ontology is rich in struc- ture, otherwise, nesting approaches of near labels 3 Methodology were used as contextual information. Neverthe- less, they observe that resource constraints (gran- 3.1 Dataset linking ularity, size of ontologies) have an impact on the Dataset linking is the process of discovering translation selection approach. [21] describe an equivalent entries between entities in two datasets instance-based interlinking method that mostly re- and this is based on a process of identifying rel- lies upon machine translation . Their ative ‘facets’ of entities features that can then be processed into a feature that represents the simi- • Length Ratio: The ratio of the number of larity between the facets of a candidate pair of en- tokens in each sentence. For symmetry this tities. Many such features of these candidates are ratio is defined as: then combined into a single score representing the min(x, y) similarity of two entities, and finally a ‘matching’ ρ(x, y) = . is used to find the linking between the max(x, y) datasets. As an example, we may extract the label of each entity in a dataset (the ‘facets’), and then • Average Length Ratio: The average compute similarity features of the labels such as length of each word in the text are also com- edit distance, a set of such features are then com- pared as above. bined with a function learnt using a typical super- vised procedure to a similarity • Negation: One if both texts or neither text score. Finally, the optimal mapping is calculated contain a negation word (‘not’, ‘never’, etc.), according to some constraint such as the bipartite zero otherwise. assumption, which states that no entity in a dataset is connected to more than one entity in the dataset, • Number: One if all numbers (e.g, ‘6’) in and this can be efficiently computed by each text are found in the other, zero other- such as the Hungarian Algorithm [20]. This work- wise. flow is realized by a tool called NAISC (Nearly Automatic Integration of Schema)2. NAISC can • GloVe Similarity: For each word in each extract a number of ‘facets’ from a dataset includ- text we extract the GloVe vectors [30] and ing, labels of terms, descriptions of terms, and calculate the cosine similarity between these most similar label of direct or indirect superterms. ’ vectors vi and the vj using cosine sim- Then for each pair of strings extracted this way we ilarity as follows: develop a number of features as follows: 1 X g(x, y) = max cos(vi, vj), n j=1,...n • Jaccard, Dice, Containment: We consider i=1,...n the two strings both as a set of words and a set of characters and compute the following where n is the length of the first string and m functions: is the length of the second. |A ∩ B| J(A, B) = , • LSTM: We calculate a similarity using the |A ∪ B| LSTM approach described by [35]. 2|A ∩ B| D(A, B) = , Having extracted the features for each pair, we |A| + |B| learn the similarity using a regression SVM [31] |A ∩ B| and learn the optimal alignment using the Hungar- C(A, B) = . min(|A|, |B|) ian algorithm [20].

• Smoothed Jaccard: Smoothed Jaccard is 3.1.1 Alignment as Machine Translation calculated only on the word level for the con- Alignment cannot itself translate an ontology or catenated labels facet. It is calculated as fol- knowledge graph, but we can use the alignment lows: procedure to align the ontology we wish to trans- σ(|A ∩ B|) late to another resource, which has the relevant J (A, B) = , σ (σ(|A|) + σ(|B|) − σ(|A ∩ B|)) translations. In particular, we align to the DBpedia resource, which is derived from Wikipedia, and as σ(x) = 1.0 − exp(−αx). such contains a large number of translations for concepts given by the interlingual links between This is a variant of Jaccard that can be ad- pages. As such, if we can correctly select the DB- justed to distinguish matches on shorter texts; pedia entity for an element in our input ontology, it tends to Jaccard at α → 0. we can generate a translation for this term in the 2‘Naisc’ means ‘links’ in Irish and is pronounced ‘nashk’ target language. 3.2 Translating labels with Machine ity in highly specific domains. With the usage Translation of disambiguated DBpedia entries showed signifi- A similar approach to localizing ontologies is cant translation improvements of domains-specific translating domain-specific labels within the rel- expressions and ontology labels [3,4], which con- evant contextual information, which is selected firms the applicability of such techniques in a real from a pool of generic sentences based on the scenario. Therefore, we combine the linking ap- lexical3 and semantic4 overlap with the labels to proach described in Section 3.1 and automatic la- be translated. The goal is to identify sentences bel translation with SMT in Section 3.2, whereby that are domain-specific in respect of the target we give preference to the translations provided by domain and contain as much as possible relevant the linking approach and automatically translate words that can allow the SMT system to learn the ontology labels with OTTO, which were not the translations of the labels. For instance, with aligned with the DBpedia entries. the sentence selection we aim to retain only En- 4 Experimental Setting glish sentences where the word injection belongs to the medical domain, but not to the technical do- In this section, we give an overview on the dataset main. This selection process demonstrated trans- and the translation toolkit used in our experi- lation improvement of labels, since we try to trans- ment. Furthermore, we describe the SMT evalua- late them within disambiguated sentences that be- tion techniques, considering the translation direc- long to the targeted domain. For this work we used tion from English to German. our ontology translation system, called OTTO5, which has been evaluated by translating the labels 4.1 Statistical Machine Translation and of several ontologies, i.e. Organic.Lingua, DOAP, Training Dataset GeoSkills, STW, TheSoz, into different languages To align labels within knowledge graphs across [2]. different languages, we use machine translation Nonetheless, some of the domain-specific labels techniques, where we wish to find the best trans- within knowledge graphs may not be automati- lation, of a source string, given by a log-linear cally translatable with SMT, due to the fact that model combining a set of features. The translation the bilingual information is missing and cannot be that maximizes the score of the log-linear model learned from the parallel sentences. is obtained by searching all possible translations candidates. The decoder or search procedure, re- 3.3 Linking and Translating through spectively, provides the most probable translation Multilingual Linked Data based on statistical translation model learned from Since the task of translating labels in knowl- the training data. edge graphs often needs to deal with domain- For a broader domain coverage of datasets specific expressions, we require lexical knowl- necessary to train an SMT system, we merged edge of the domain. Although SMT systems several parallel corpora, e.g. JRC-Acquis, Eu- nowadays are suitable to translate very frequent roparl, DGT (translation memories generated by expressions with an acceptable quality, they fail the Directorate-General for Translation), KDE, in translating infrequent domain-specific expres- GNOME and TED talks among others, into one sions. This mostly depends on the lack of domain- parallel dataset (Table1). For the translation specific parallel data from which the SMT sys- approach, the Moses toolkit [19] and GIZA++ tems can learn. A valuable solution to handle [28] for word alignment generation were used, domain-specific terms is represented by multilin- whereby a language model was build with KenLM gual linked data resources, e.g. DBpedia or IATE, [16]. which are continuously updated and can be easily queried. For this reason, the identification, disam- DBpedia as Multilingual Alignment Resource biguation of domain-specific expressions is a cru- The DBpedia project [18] aims to extract struc- cial step towards increasing the translation qual- tured content from the knowledge added to the Wikipedia repository. DBpedia allows users to se- 3 string, substring and character overlap mantically query relationships and properties of 4measuring cosine similarity of word embeddings 5http://server1.nlp.insight-centre. Wikipedia resources, including links to other re- org/otto/ lated datasets. English German BLEU-1 METEOR Chrf # of tokens 145,757,562 135,627,679 NAISC linking 25.0 17.03 27.81 # of types 561,667 1,156,379 OTTO 53.6 29.53 52.18 # of parallel sent. 10,065,235 NAISC-then-OTTO 55.5 30.63 53.92

Table 1: Statistics on the merged English-German Table 2: Translation evaluation based ion the EHR parallel corpus. ontology labels.

Since some of the domain-specific ontology la- BLEU, the metric produces good correlation with bels may not be automatically translatable with human judgement at the sentence or segment level. SMT, due to the fact that the bilingual information chrF3 is a character n-gram metric, which has is missing and cannot be learned from the paral- shown very good correlations with human judge- lel sentences. We therefore use the information ments. contained in the DBpedia knowledge base6 to im- prove the translation of expressions which may not 5 Evaluation be known to the SMT system. The automatic evaluation on the translated EHR Evaluation dataset For the evaluation cam- ontology labels provided by Dbpedia and the paign of the proposed experiment we used OTTO translation system is based on the corre- the Electronic Health Records (EHR) ontology7, spondence between the automatically generated which is based on the work of openEHR8. The on- output and reference translations (gold standard). tology, which holds 75 labels, tries to overcome For the automatic translation evaluation of the the issue of patient information access, which is EHR labels we used the BLEU, METEOR and distributed among several independent and hetero- chrF metrics. geneous systems that may be syntactically or se- mantically incompatible. Therefore the EHR on- Table2 presents the evaluation scores of the tology focuses on the developed within the frame- translations of the EHR ontology. We observed, work of a federated information system. that the NAISC linking approach using DBpedia on the whole EHR dataset performs worse com- Machine Translation Evaluation For the auto- pared to the OTTO translation system. This is due matic evaluation we used the BLEU [29], ME- the low number of established alignments between TEOR [9] and chrF [33] metrics. the target ontology, i.e. EHR, and the Dbpedia re- BLEU (Bilingual Evaluation Understudy) is source, where only 32 (out of 75) alignments with calculated for individual translated segments (n- translations into German were found. The linking grams) by comparing them with a data set of ref- approach found label alignments and their trans- erence translations. Considering the shortness of lations like, organiser–Veranstalter, observation– the entries in the targeted ontology, we report Beobachtung or entry–Eingabe , but did not pro- scores based on the unigram overlap (BLEU-1). vide any alignments for clinical context or list Those scores, between 0 and 100 (perfect trans- folder. In the next step, we engage the OTTO lation), are then averaged over the whole evalu- translation system to translate all labels of the ation data set to reach an estimate of the transla- EHR ontology. With this approach all evalua- tion’s overall quality. METEOR (Metric for Eval- tion scores significantly increase compared to the uation of Translation with Explicit ORdering) is DBpedia linking method. This can be deducted based on the harmonic mean of precision and re- to the possibility of translating all labels stored call, whereby recall is weighted higher than pre- in the EHR ontology. Finally, we combined the cision. Along with exact word (or phrase) match- provided multilingual knowledge from both ap- ing it has additional features, i.e. , para- proaches, i.e. data linking and automatic trans- phrasing and synonymy matching. In contrast to lation, respectively. In this experiment, we first give preference to translation of a labels provided 6http://wiki.dbpedia.org/ downloads-2016-10 by NAISC and if no translation is provided, we use 7http://trajano.us.es/˜isabel/EHR/ the OTTO system to fill the lexical gap in German. 8http://www.openehr.org/ Due to the precise translation provided by NAISC and the possibility to translate the remaining EHR the Association for Computational , Bei- labels, we further improve the translation quality jing, China, July 2015. in terms of the used evaluation metrics. [5] P. Buitelaar, K.-S. Choi, P. Cimiano, and E. H. Hovy. The Multilingual Semantic Web (Dagstuhl Seminar 6 Conclusion 12362). Dagstuhl Reports, 2(9):15–94, 2013.

In this work we propose an approach to link dif- [6] M. Carpuat, G. Ngai, P. Fung, and K. Church. Cre- ferent knowledge graphs across languages. To ating a bilingual ontology: a corpus-based approach enable the usage of knowledge graphs for multi- for aligning WordNet and HowNet. In Proceedings lingual NLP applications, these resources, which of the 1st Global WordNet Conference, 2002. are mostly represented in English only, have to be [7] P. Cimiano, E. Montiel-Ponsoda, P. Buitelaar, linked to other multilingual resource. This can be M. Espinoza, and A. Gomez-P´ erez.´ A note on ontol- performed by using existing multilingual knowl- ogy localization. Appl. Ontol., 5(2):127–137, 2010. edge graphs, e.g. DBPedia, or by the usage of [8] T. Declerck, A. G. Perez,´ O. Vela, Z. Gantner, machine translation. Our ongoing work focuses and D. Manzano. Multilingual lexical semantic re- on the synergy of the linking approaches and ma- sources for ontology translation. In In Proceedings chine translation, specifically, how dataset linking of the 5th International Conference on Language can benefit from machine translation and how dis- Resources and Evaluation, Saarbrucken,¨ Germany, 2006. ambiguation and translation quality of labels can be improved from existing multilingual knowl- [9] M. Denkowski and A. Lavie. Meteor universal: edge graphs. The proposed approach provides Language specific translation evaluation for any tar- the possibility to translate highly specific vocab- get language. In Proceedings of the EACL 2014 Workshop on Statistical Machine Translation, 2014. ulary without particular in-domain parallel data. Furthermore, we observed that more than 25% of [10] S. T. Dumais, T. A. Letsche, M. L. Littman, and the identified lexical knowledge consists of multi- T. K. Landauer. Automatic cross-language retrieval word-expressions, e.g. “fatal familial in-somnia”. using latent semantic indexing. In AAAI spring sym- posium on cross-language text and speech retrieval, For this reason, we also focus on the alignment of volume 15, page 21, 1997. nested knowledge inside those expressions (com- posing partial translation). [11] M. Espinoza, E. Montiel-Ponsoda, and A. Gomez-´ Perez.´ Ontology localization. In Proceedings of the Acknowledgments Fifth International Conference on Knowledge Cap- ture, NY, USA, 2009. This work was supported by the Science Founda- [12] C. Fellbaum. Wordnet. In Theory and applications tion Ireland under Grant Number SFI/12/RC/2289 of ontology: computer applications, pages 231–243. (Insight). Springer, 2010.

[13] B. Fu, R. Brennan, and D. O’Sullivan. Cross- References lingual ontology mapping–an investigation of the impact of machine translation. In The Semantic [1] M. Arcan, K. Asooja, H. Ziad, and P. Buitelaar. Web, pages 1–15. Springer, 2009. OTTO Ontology Translation System. In ISWC 2015 Posters & Demonstrations Track, volume 1486, [14] B. Fu, R. Brennan, and D. O’Sullivan. Cross- Bethlehem (PA), USA, 2015. lingual ontology mapping and its use on the mul- Proceedings of the 1st In- [2] M. Arcan, M. Dragoni, and P. Buitelaar. Translat- tilingual semantic web. In ternational Workshop on the Multilingual Semantic ing ontologies in real-world settings. In Proceedings of the 15th International Semantic Web Conference Web, pages 13–20, 2010. (ISWC-2016) , Kobe, Japan, 2016. [15] J. Gracia, E. Montiel-Ponsoda, P. Cimiano, [3] M. Arcan, C. Giuliano, M. Turchi, and P. Buitelaar. A. Gomez-P´ erez,´ P. Buitelaar, and J. McCrae. Chal- Identification of Bilingual Terms from Monolingual lenges for the multilingual web of data. Web Se- Documents for Statistical Machine Translation. In mantics: Science, Services and Agents on the World Proceedings of the 4th International Workshop on Wide Web, 11, 2012. Computational Terminology (Computerm), Dublin, Ireland, 2014. [16] K. Heafield. KenLM: faster and smaller language model queries. In Proceedings of the EMNLP 2011 [4] M. Arcan, M. Turchi, and P. Buitelaar. Knowledge Sixth Workshop on Statistical Machine Translation, portability with semantic expansion of ontology la- pages 187–197, Edinburgh, Scotland, United King- bels. In Proceedings of the 53rd Annual Meeting of dom, July 2011. [17] J. Jagarlamudi, R. Udupa, H. Daume´ III, and [27] R. Navigli. BabelNet goes to the (Multilingual) A. Bhole. Improving bilingual projections via Semantic Web. In Proceedings of the 3rd Workshop sparse covariance matrices. In Proceedings of the on the Multilingual Semantic Web, 2012. Conference on Empirical Methods in Natural Lan- guage Processing, pages 930–940. Association for [28] F. J. Och and H. Ney. A systematic comparison of Computational Linguistics, 2011. various statistical alignment models. Computational Linguistics, 29, 2003. [18] Jens Lehmann and Robert Isele and Max Jakob and Anja Jentzsch and Dimitris Kontokostas and Pablo [29] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu. N. Mendes and Sebastian Hellmann and Mohamed BLEU: a method for automatic evaluation of ma- Morsey and Patrick van Kleef and Soren¨ Auer and chine translation. In Proceedings of the 40th Annual Christian Bizer. DBpedia a large-scale, multilin- Meeting on Association for Computational Linguis- gual knowledge base extracted from Wikipedia. Se- tics, pages 311–318, 2002. mantic Web Journal, 6(2):167–195, 2015. [30] J. Pennington, R. Socher, and C. D. Manning. [19] P. Koehn, H. Hoang, A. Birch, C. Callison-Burch, Glove: Global vectors for word representation. In M. Federico, N. Bertoldi, B. Cowan, W. Shen, EMNLP, volume 14, pages 1532–1543, 2014. C. Moran, R. Zens, C. Dyer, O. Bojar, A. Con- [31] J. Platt. Fast training of support vector ma- stantin, and E. Herbst. Moses: Open source toolkit chines using sequential minimal optimization. In for statistical machine translation. In Proceedings B. Schoelkopf, C. Burges, and A. Smola, edi- of the 45th Annual Meeting of the ACL on Inter- tors, Advances in Kernel Methods - Support Vector active Poster and Demonstration Sessions, Strouds- Learning. MIT Press, 1998. burg, PA, USA, 2007. [32] J. C. Platt, K. Toutanova, and W.-T. Yih. Translin- [20] H. W. Kuhn. The hungarian method for the assign- gual document representations from discriminative ment problem. Naval Research Logistics Quarterly, projections. In Proceedings of the 2010 Conference (2):83–97, 1955. on Empirical Methods in Natural Language Pro- [21] T. Lesnikova, J. David, and J. Euzenat. Inter- cessing, pages 251–261. Association for Computa- linking English and Chinese RDF data using Babel- tional Linguistics, 2010. Net. In Proceedings of the 2015 ACM Symposium [33] M. Popovic.´ chrF: character n-gram F-score for on Document Engineering, DocEng ’15, pages 39– automatic MT evaluation. In Proceedings of the 42, New York, NY, USA, 2015. ACM. Tenth Workshop on Statistical Machine Translation, pages 392–395, Lisbon, Portugal, September 2015. [22] J. McCrae, P. Cimiano, and R. Klinger. Orthonor- Association for Computational Linguistics. mal explicit topic analysis for cross-lingual docu- ment matching. In Proceedings of the 2013 Con- [34] D. Spohr, L. Hollink, and P. Cimiano. A machine ference on Empirical Methods in Natural Language learning approach to multilingual and cross-lingual Processing, pages 1732–1740, 2013. ontology matching. The Semantic Web–ISWC 2011, pages 665–680, 2011. [23] J. McCrae, E. Montiel-Ponsoda, G. Aguado de Cea, M. J. Espinoza Mej´ıa, and P. Cimiano. Com- [35] K. S. Tai, R. Socher, and C. D. Manning. Improved bining statistical and semantic approaches to the semantic representations from tree-structured long translation of ontologies and taxonomies. In Pro- short-term memory networks. arXiv preprint ceedings of 7th Workshop on , Structure and arXiv:1503.00075, 2015. in Statistical Translation, pages 116–125, 2011. [36] A. Vinokourov, N. Cristianini, and J. S. Shawe- taylor. Inferring a semantic representation of text via [24] J. P. McCrae, M. Arcan, K. Asooja, J. Gracia, cross-language correlation analysis. In Advances in P. Buitelaar, and P. Cimiano. Domain adaptation neural information processing systems, pages 1473– for ontology localization. Web Semantics, 36:23– 1480, 2002. 31, 2016.

[25] J. P. McCrae and P. Cimiano. Mining translations from the web of open linked data. In Proceedings of the Joint Workshop on NLP&LOD and SWAIE: Semantic Web, Linked Open Data and Infromation Extraction, pages 9–13, 2013.

[26] D. Mimno, H. M. Wallach, J. Naradowsky, D. A. Smith, and A. McCallum. Polylingual topic models. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 2, pages 880–889. Association for Computational Linguistics, 2009.