Perspectives on Ontology Learning
Total Page:16
File Type:pdf, Size:1020Kb
Perspectives on Ontology Learning Jens Lehmann and Johanna V¨olker (Eds.) Informatics Institute, University of Leipzig, Germany Data & Web Science Research Group, University of Mannheim, Germany IOS Press Foreword Philipp CIMIANO a, a Semantic Computing Group, Bielefeld University Equipping machines with commonsense as well as domain-specific knowledge in order to endow them with human-like understanding of certain problem domains has been and still is a main goal of artificial intelligence research. In this context, a crucial question is how high the cost actually is for encoding all the relevant knowledge in such a way that it can be exploited by machines for automatic reasoning, inconsistency detec- tion, etc. While there has been some recent work on developing methodologies allowing us to estimate the cost of knowledge engineering projects [12], it is legitimate to assume that not all the relevant knowledge can be encoded manually. Techniques that can extract and discover knowledge by analyzing human behaviour and data produced as a result thereof can offer an important contribution in this respect. The field of ontology learning, a term coined by Alexander Mädche and Steffen Staab in 2001 [7], is concerned with the development of methods that can induce relevant ontological knowledge from data. The field can by now look back into more than ten years of intense research. Early research in the field focused on applying shallow meth- ods to term and concept extraction as well as hierarchical and non-hierarchical relation extraction [7]. Later on, in my PhD thesis with the title “Ontology Learning and Pop- ulation from Text: Algorithms, Evaluation and Applications”, I defined ontology learn- ing as the acquisition of a domain model from data and attempted to provide a system- atic overview of ontology learning tasks by introducing the so called ontology learn- ing layer cake, which has received wide attention since then. In recent years, several re- searchers have attempted to increase the expressivity of the ontologies learned from text data, in particular by attempting to extract deeper axiomatic knowledge (e.g. see [13], [14] and [4]). Some contributions along these lines can also be found in this volume, e.g. aiming at learning OWL axioms by applying inductive techniques (cf. Lehmann et al. [5] and Lisi [6] in this volume). The problem of ontology learning has turned out to be much more difficult than ex- pected. The main reason for this is, in my view, that an ontology always reflects a way of conceptualizing the world or a given domain, while the results of an ontology learning al- gorithm that learns from a set of data essentially reflects the idiosyncrasies of the dataset in question. As such, turning the results of an ontology algorithm into an ontology that actually reflects the conceptualization one has of a domain can be more costly than actu- ally building the ontology from scratch. The problem of ontology learning has turned out to be much more difficult than expected. The main reason for this is, in my view, that an ontology always reflects a way of conceptualizing the world or a given domain, while the results of an ontology learning algorithm that learns from a set of data essentially reflects the idiosyncrasies of the dataset in question. As such, turning the results of an ontology vi Cimiano / Foreword learning algorithm into an ontology that actually reflects the conceptualization one has of a domain can be more costly than actually building the ontology from scratch. A second problem so far has been the lack of applications for automatically learned ontologies. While Hotho, Bloehdorn and myself [2] showed some positive impacts of au- tomatically learned ontologies on classification and clustering tasks, not many other con- vincing applications were at sight at that stage. Recently, ontology learning has, however, seen interesting applications for inducing the semantics of tags used in social media data and folksonomies [1]. Recently, Meilicke et al. have shown that automatically induced knowledge, disjointeness axioms in particular, can be deployed to debug ontology map- pings [9]. The fact that such applications are emerging is clearly a good sign that there is definitely progress in the field. A further interesting and very promising application potential for ontology learning lies in the field of Linked Data. Learning from Linked Data will allow us to induce schemata in a bottom-up fashion and let the schema evolve with the data. Ontology population will also continue to play a crucial role in taming and structuring the large amount of unstructured data available, e.g. in Scientific Publi- cations. In many applications domains, one needs to consider all data together in order to extract key facts and knowledge, structure this knowledge in the form of a database in order to aggregate and summarize the data and provide analytical procedures that support decision making by experts. In the mid-term, I foresee two very interesting research directions for ontology learning. When modelling knowledge, it is relatively easy to model “the obvious” and straightforward knowledge in a particular domain. However, the “not-so-obvious” and more complex relationships are harder to come up with for a knowledge engineer. This is where algorithms that induce more complex and non-trivial relationships from data can assist a human in the process of modelling more complex axioms. This is especially relevant and valuable in the Linked Data era where not many people seem to want to put effort into axiomatizing the vocabulary used in their datasets. This is tightly related to the second future direction I regard as crucial within the field of ontology learning. So far, there has not been too much focus on how humans and machines can collaborate on the task of modelling relevant knowledge for a given domain. The role of machine agents should be to derive interesting axioms and relation- ships from data, generating hypotheses induced from data and asking the human for val- idation and clarification of them. Humans would then rely on their domain knowledge to confirm the induced knowledge or reject it by providing counter-examples. Methodolo- gies that define and clarify the role of machines and humans in ontology engineering and ontology learning are urgently needed in order to exploit the capabilities of both humans and machines optimally. In this volume, for instance, there is one contribution by Simperl et al. [11] that shows how humans can be involved in the process of ontology learning through games with a purpose. Unless we have good methodologies incorporating the human in the loop, I dare to predict that major breakthroughs in ontology learning can not be expected. Concerning applications, in my view there is one very important application for on- tology learning, i.e. natural language processing (NLP). Given that language processing inherently requires world knowledge to guide the interpretation process [3], it strikes that so far – at least to my knowledge – there have not been too many convincing applica- tions of proper OWL ontologies within NLP. We may wonder why this is the case. Is the field of NLP not aware of the techniques developed in the Semantic Web community? Cimiano / Foreword vii Or are existing ontology learning techniques too noisy for researchers that are used to work with highly curated, hand-crafted resources such as WordNet, FrameNet etc.? Or is it simply the case that the knowledge contained in OWL ontologies and produced by state-of-the-art ontology learning tools is not adequate for NLP purposes? Being far from knowing the answer, let me speculate a bit about the reasons: Lack of coverage: domain ontologies – whether learned or not – typically have a • limited coverage and scope, whereas NLP researchers are typically used to work with domain-independent resources such as WordNet, FrameNet, etc. An NLP researcher would have to work with many, possibly overlapping ontologies. Limited quality: Automatically learned ontologies might be noisy, but still use- • ful as some applications have been shown. This requires also a paradigm shift towards working with non-perfect, possibly noisy, incomplete or even logically inconsistent ontologies. Limitations of crisp knowledge: In human language, the meaning of words is • often vague so that prototypes or distributions might be better suitable for NLP applications.rather than crisp knowledge representations formalisms. No need for expressive knowledge (yet!): With the statistical turn in the 80s and • 90s, NLP focused mainly on statistical approaches to natural language processing, moving away from purely symbolic approaches. However, one observes a move back into symbolic approaches as people realize more and more that inference is crucial for NLP. New models such as Markov Logic Networks [10] that combine statistic with symbolic, first-order theories have received in fact wide attention in the NLP and Semantic Web communities. They might, thus, offer a point of con- vergence for both communities. Concentrating on learning probabilistic knowl- edge will thus be an important avenue for future work. Lack of awareness: For sure, the NLP community is not particularly aware of • what is going on in the Semantic Web and ontology learning communities. This is clearly corroborated by the fact that only few NLP researchers attend Seman- tic Web conferences. The number of references to Semantic Web related work in NLP papers at major NLP conference converges practically to zero, as an in- formal empirical study by Josef van Genabith showed1. For sure, there is a lot that the Semantic Web community could do about this, i.e. organizing tutorials and workshops at NLP conferences, but also regarding NLP as a potential con- sumer of ontologies – whether learned or not. A number of recent activities in the Semantic Web field, such as the development of NIF (Hellmann et.