Probabilistic of Universal Schema

Limin Yao Sebastian Riedel Andrew McCallum Department of Computer Science University of Massachusetts, Amherst {lmyao,riedel,mccallum}@cs.umass.edu

Abstract knowledge. For example, Freebase (Bollacker et al., 2008) captures the content of Wikipedia to some In data integration we transform information extent, but has no criticized(Person,Person) from a source into a target schema. A gen- and hence cannot answer a question like “Who crit- eral problem in this task is loss of fidelity and icized George Bush?”, even though partial answers coverage: the source expresses more knowl- are expressed in Wikipedia. This makes the edge than can fit into the target schema, or knowledge that is hard to fit into any schema schema a major bottleneck in information extrac- at all. This problem is taken to an extreme in tion (IE). From a more general point of , data in- information extraction (IE) where the source tegration always suffers from schema mismatch be- is natural language. To address this issue, one tween knowledge source and knowledge target. can either automatically learn a latent schema To overcome this problem, one could attempt to emergent in text (a brittle and ill-defined task), manually extend the schema whenever needed, but or manually extend schemas. We propose in- this is a time-consuming and expensive process. Al- stead to store data in a probabilistic database of universal schema. This schema is simply ternatively, in the case of IE, we can automatically the union of all source schemas, and the proba- induce latent schemas from text, but this is a brit- bilistic database learns how to predict the cells tle, ill-defined and error-prone task. This paper pro- of each source relation in this union. For ex- poses a third alternative: sidestep the issue of in- ample, the database could store Freebase re- complete schemas altogether, by simply combining lations and relations that correspond to natu- the relations of all knowledge sources into what we ral language surface patterns. The database will refer to as a universal schema. In the case of IE would learn to predict what freebase relations this means maintaining a database with one per hold true based on what surface patterns ap- pear, and vice versa. We describe an anal- natural language surface pattern. For data integra- ogy between such databases and collaborative tion from structured sources it simply means storing filtering models, and use it to implement our the original tables as is. Crucially, the database will paradigm with probabilistic PCA, a scalable not only store what each source table does contain, and effective collaborative filtering method. it will also learn a probabilistic model about which other rows each source table should correctly con- 1 Introduction tain. Let us illustrate this approach in the context of Natural language is a highly expressive representa- IE. First we copy tables such as profession from a tion of knowledge. Yet, for many tasks databases structured source (say, DBPedia). Next we create are more suitable, as they support more effective de- one table per surface pattern, such as was-criticized- cision support, question answering and data min- by and was-attacked-by and fill these tables with the ing. But given a fixed schema, any database can entity pairs that appear with this pattern in some nat- only capture so much of the information natural lan- ural language corpus (say, the NYT Corpus). At guage can express, even if we restrict us to factual this point, our database is a simple combination of a structured and an OpenIE (Etzioni et al., 2008) r1!"#$"%"&'(')'*+"*,"- r2!"#.)"/%01+2$"- v v knowledge representation. However, while we insert r1 r2 this knowledge, we can learn a probabilistic model which is able to predict was-criticized-by pairs based x x ae e,r1 e,r2 on information from the was-attacked-by relation. e 3 4 "!"":*<=12"-**=)$" In addition, it learns that the profession relation in >'?*+"@">

Freebase can help disambiguate between physical vr1 vr2 attacks in sports and verbal attacks in politics. At 9:;# the same time, the model learns that the natural lan- x x a e,r1 e,r2 guage relation was-criticized-by can help predict the e e"!"":*<=12"-**=)$" profession information in Freebase. Moreover, often 3 56758 >'?*+"@">

4.2 Structured Data 4.4 Integrating the NYT Corpus The second set of source relations stems from Free- We investigate how gPCA can help us answer ques- base. We choose those relations that hold for entity tions based on only single data source: the NYT pairs appearing in the NYT corpus. This adds 116 Corpus. Table 1 presents, for two source relations relations to our universal schema. For each of the re- (aka surface patterns), a set of observed entity pairs lations we import only those rows which correspond (Obs.) and the most likely inferred entity pairs to entity tuples also found in the NYT corpus. In or- (New). The table shows that we can answer a ques- der to link entity mentions in the text to entities in tion like “Who owns percentages of Scania AB?” Freebase, we follow a simple string-match heuristic. even though the corpus does not explicitly contain the answer. In our case, it only contains “buy-stake- 4.3 Experimental Setup and Training in(VOLVO,SCANIA AB).” gPCA achieves about 49% recall, at about 67% In our experiments, we hold out some of the ob- precision. Interestingly, the model learns more served source rows and try to predict these based than just paraphrasing. Instead, it captures some on other observed rows. In particular, for each en- notion of entailment. This can be observed in tity pair, we traverse over all source relations. For its asymmetric beliefs. For example, the model each relation we throw an unbiased coin to deter- learned to predict “professor-at(K.BOYLE,OHIO mine whether it is observed for the given pair. Then STATE)” based on “historian-at(KEVIN BOYLE, we train a gPCA model of 50 components on the OHIO STATE)” but would not make the infer- observed rows, and use it to predict the unobserved ence “historian-at(R.FREEMAN,HARVARD)” based ones. Here a pair e is set to be in a given relation on “professor-at(R.FREEMAN,HARVARD).” r if P (r (e)) > 0.5 according to our model. Since we generally do not have observed negative informa- 4.5 Integrating Freebase tion,1 we sub-sample a set of negative rows for each relation r to create a more balanced training set. What happens if we integrate additional structured We evaluate recall of our method by measuring sources into our probabilistic database? We observe how many of the true held out rows we predict. We that by incorporating Freebase tables in addition to could use a similar approach to measure precision the NYT data we can improve recall from 49% to by considering each positive prediction to be a false 52% on surface patterns. The precision also in- positive if the observed held-out data does not con- creases by 2%. tain the corresponding fact. However, this approach Table 2 sums the results and also gives an ex- underestimates precision since our sources are gen- ample of how Freebase helps improve both preci- erally incomplete. To overcome this issue, we use sion and recall. Without Freebase, the gPCA pre- dicts that Maher Arar was arrested in Syria—primarily human annotations for the precision measure. In because he lived in Syria and the NYT often talks particular, we randomly sample a subset of entity about arrests of people in the city they live in2. After pairs and ask human annotators to assess the pre- learning placeOfBirth(ARAR,SYRIA) from Freebase, the dicted positive relations of each. gPCA model infers wasBornIn(ARAR,SYRIA) as well as grewUpIn(ARAR,SYRIA). 1Just because a particular e has not yet been seen in particu- lar relation r we cannot infer that r (e) is false. 2In fact, he was arrested in US References Table 2: Relation predictions w/o Freebase. without Freebase with Freebase Kurt Bollacker, Colin Evans, Praveen Paritosh, Tim Prec. 0.687 0.666 Sturge, and Jamie Taylor. 2008. Freebase: a collabo- Rec. 0.491 0.520 ratively created for structuring human E.g. M. Arar, Syria (Freebase: placeOfBirth) knowledge. In SIGMOD ’08: Proceedings of the 2008 ACM SIGMOD international conference on Manage- A was arrested in B A was born in B ment of data, pages 1247–1250, New York, NY, USA. Pred. A appeal to B A grow up in B ACM. A, who represent B A’s home in B Michael Collins, Sanjoy Dasgupta, and Robert E. Schapire. 2001. A generalization of principal com- ponent analysis to the exponential family. In Proceed- 5 Conclusion ings of NIPS. Oren Etzioni, Michele Banko, Stephen Soderland, and In our approach we do not design or infer new re- Daniel S. Weld. 2008. Open information extraction lations to accommodate information from different from the web. Commun. ACM, 51(12):68–74. sources. Instead we simply combine source rela- Johannes Hoffart, Fabian M. Suchanek, Klaus Berberich, tions into a universal schema, and learn a proba- and Gerhard Weikum. 2012. Yago2: A spatially and bilistic model to predict what other rows the sources temporally enhanced knowledge base from wikipedia. could contain. This simple paradigm allows us to Artificial Intelligence. perform data alignment, information extraction, and Dekang Lin and Patrick Pantel. 2001. DIRT - discovery other forms of data integration, while minimizing of inference rules from text. In Knowledge Discovery both loss of information and the need for schema and Data Mining, pages 323–328. maintenance. Mike Mintz, Steven Bills, Rion Snow, and Daniel Juraf- sky. 2009. Distant supervision for relation extrac- At the heart of our approach is the hypothesis that tion without labeled data. In Proceedings of the Joint we should concentrate on building models to predict Conference of the 47th Annual Meeting of the ACL source data—a relatively well defined task—as op- and the 4th International Joint Conference on Natu- posed to models of semantic equivalence that match ral Language Processing of the AFNLP (ACL ’09), our intuition. Our future work will therefore investi- pages 1003–1011. Association for Computational Lin- gate such predictive models in more detail, and ask guistics. how to (a) incorporate relations of different arities, Patrick Pantel, Rahul Bhagat, Bonaventura Coppola, (b) employ background knowledge, (c) optimize the Timothy Chklovski, and Eduard Hovy. 2007. ISP: Learning Inferential Selectional Preferences. In Pro- choice of negative data and (d) scale up both in terms ceedings of NAACL HLT. of rows and tables. Sebastian Riedel, Limin Yao, and Andrew McCallum. 2010. Modeling relations and their mentions without Acknowledgments labeled text. In Proceedings of the European Confer- ence on Machine Learning and Knowledge Discovery This work was supported in part by the Center for in Databases (ECML PKDD ’10). Intelligent Information Retrieval and the Univer- Evan Sandhaus, 2008. The New York Times Annotated sity of Massachusetts and in part by UPenn NSF Corpus. Linguistic Data Consortium, Philadelphia. medium IIS-0803847. We gratefully acknowledge Stefan Schoenmackers, Oren Etzioni, and Daniel S. the support of Defense Advanced Research Projects Weld. 2008. Scaling textual inference to the web. Agency (DARPA) Machine Reading Program under In EMNLP ’08: Proceedings of the Conference on Air Force Research Laboratory (AFRL) prime con- Empirical Methods in Natural Language Processing, tract no. FA8750-09-C-0181. Any opinions, find- pages 79–88, Morristown, NJ, USA. Association for Computational Linguistics. ings, and conclusion or recommendations expressed Stefan Schoenmackers, Oren Etzioni, Daniel S. Weld, in this material are those of the authors and do not and Jesse Davis. 2010. Learning first-order horn necessarily reflect the view of DARPA, AFRL, or clauses from web text. In Proceedings of the 2010 the US government. Conference on Empirical Methods in Natural Lan- guage Processing, EMNLP ’10, pages 1088–1098, Stroudsburg, PA, USA. Association for Computational Linguistics. Limin Yao, Aria Haghighi, Sebastian Riedel, and Andrew McCallum. 2011. Structured relation discovery using generative models. In Proceedings of the Conference on Empirical methods in natural language processing (EMNLP ’11), July. Alexander Yates and Oren Etzioni. 2009. Unsupervised methods for determining object and relation synonyms on the web. Journal of Artificial Intelligence Research, 34:255–296.