TTOKU Summarization Based Systems at NTCIR-9 1CLICK Task
Total Page:16
File Type:pdf, Size:1020Kb
Proceedings of NTCIR-9 Workshop Meeting, December 6-9, 2011, Tokyo, Japan TTOKU Summarization Based Systems at NTCIR-9 1CLICK task Hajime Morita Takuya Makino Tokyo Institute of Technology Tokyo Institute of Technology Kanagawa, Japan Kanagawa, Japan [email protected] [email protected] Tetsuya Sakai Hiroya Takamura Manabu Okumura Microsoft Research Asia Tokyo Institute of Technology Tokyo Institute of Technology Beijing, China Kanagawa, Japan Kanagawa, Japan [email protected] [email protected] [email protected] ABSTRACT comes important when we generate a summary from web We describe our two query-oriented summarization systems texts. implemented for the NTCIR-9 1CLICK task. We regard We also assumed that users want to know various facts a Question Answering problem as a summarization process. concerning their queries. That is to say, a generated sum- Both of the systems are based on the integer linear program- mary has to contain more diversified contents that are rele- ming technique, and consist of an abstractive summarization vant to user’s query than in the generic query-focused sum- model and a model ensuring to cover diversified aspects for marization task. We settled on aspects that are pre-defined answering user’s complex question. The first system adopts diverse types of needed information on satisfying user’s re- QSBP (Query SnowBall with word pair) for query oriented quests. The pre-defined information aspects depending on summarization, and extends the method to abstractive sum- query category help to consider covering these information marization in order to recognize and extract the parts of a explicitly in our approach. Thus we propose another method sentence that are related to the query. On the other hand, to ensure covering several aspects of information needed by The second system ensures covering pre-defined several as- users query. pects of information needed by users. We decided the types In this paper, we described our systems and the evaluation of aspects depending on the category of a query. Our first results at the NTCIR-9 1CLICK task. These systems consist and second systems achieved 0.1585 and 0.1484 S-measure(I) of two different summarization methods, both of which use score respectively in the desktop task. Furthermore, our first Integer Linear Programming. The first system adopts QSBP and second systems achieved 0.0866 and 0.0829 S-measure(I) (QuerySnowball with word pair) for query oriented summa- score respectively in the mobile task. rization, and extends the method to abstractive summariza- tion. Second system employs Max-min problem to solve sentence extraction problem for extractive summarization. Keywords When this system selects sentences, the generated summary Summarization, Integer Linear Programming covers various informations widely according to the max- min problem. Our first and second systems achieved 0.1585 1. INTRODUCTION and 0.1484 S-measure(I) score respectively in the desktop task. Furthermore, our first and second systems achieved Automatic text summarization has been studied for decades. 0.0866 and 0.0829 S-measure(I) score respectively in the mo- Although newspaper articles have been chosen as the target bile task. text of summarization studies, web text becomes increas- The rest of this paper is organized as follows. Section 2 de- ingly important as the target text, because web texts are scribes a brief introduction to Integer Linear Programming increasing at an explosive pace. First, we propose a method we employed, and Section 3 describes our systems. Section 4 that discriminate where the query relates and extract the describes our experimental settings, and Section 5 describes necessary parts of a sentence. Web texts tend to have re- official results and discussion. Section 6 concludes this pa- dundant or irrelative parts with important description. Par- per. ticularly, we should remove unnecessary parts that spoil co- herence or readability. Abstractive summarization also be- 2. INTEGER LINEAR PROBLEM Integer Linear Programming(ILP) has been studied for summarization recently. The problem is a kind of linear programming, that is constrained its variables are trapped in integer. Summarization is often expressed as a maximiza- tion of containing information about source document under length constraint. ILP can solve the maximization problem exactly. In contexts of summarization, Carbonell and Gold- stein [2] proposed the Maximal Marginal Relevance (MMR) ― 208 ― Proceedings of NTCIR-9 Workshop Meeting, December 6-9, 2011, Tokyo, Japan r r2 2 R2 Japanese N-gram [5]. Our second clue is the weight prop- r2 r2 r1 R1 agated from the center of the co-occurrence graph shown r2 r1 r1 in Figure 1. Below, we describe how to compute the word Q r2 q scores for words in R1 and then those for words in R2. root r q r1 2 r2 As Figure 1 suggests, the query relevance score for r1 ∈ R1 q r2 r1 is computed based not only on its base word score but also on r2 r q ∈ Q r r1 r the relationship between 1and . To be more specific, 2 1 r2 r2 let freq(w, w ) denote the within-sentence co-occurrence fre- r r2 2 quency for words w and w, and let distance(w, w) denote w w Figure 1: Co-occurrence Graph (Query Snowball) the minimum dependency distance between and :Ade- pendency distance is the path length between nodes w and w within a dependency parse tree; the minimum depen- criteria for non-redundant sentence selection, which con- dency distance is the shortest path length among all depen- sist of the document similarity and the redundancy penalty. dency parse trees of source-document sentences in which w Then, McDonald [7] presented an approximate dynamic pro- and w co-occur. Then, the query relevance score for r1can gramming approach to maximize the MMR criteria and a be computed as: formulation of the summarization task as an Integer Linear sb(q) freq(q, r1) Problem to provide exact solution. Yih et al. [11] formu- sr(r1) = sb(r1) (1) sumQ distance(q, r1) + 1.0 lated the document summarization problem as an MCKP q∈Q that is a class of ILP, and proposed a supervised method. where sumQ = q∈Q sb(q). It can be observed that the Whereas, our method is unsupervised. Filatova and Hatzi- query relevance score sr(r1) reflects the base word scores vassiloglou [3] also formulated summarization as an MCKP. of both q and r1, as well as the co-occurrence frequency Moreover, it is also used in several NLP tasks [4, 10] . freq(q, r1). Moreover, sr(r1) depends on distance(q, r1), the Although ILP a has great power and a wide application minimum dependency distance between q and r1, which re- possibility, there are some difficulties in using ILP in prac- flects the strength of relationship between q and r1. This tical situations. To solve ILP problems require large com- quantity is used in one of its denominators in Eq.1 as small putational resources even if ILP problems are a small prob- values of distance(q, r1) imply a strong relationship between lem. Computational time becomes exponentially longer, as q and r1. The 1.0 in the denominator avoids division by zero. a problem becomes larger. It is hard to estimate computa- r ∈ R 1 2 Similarly, the query relevance score for 2 2 is com- tional cost in advance. In general, GLPK and CPLEX puted based on the base word score of r2 and the relationship are commonly used to solve ILP problems. between r2andr1 ∈ R1: sr(r1) freq(r1,r2) 3. PROPOSED METHODS sr(r2) = sb(r2) (2) sumR1 distance(r1,r2) + 1.0 r1∈R1 sum s r 3.1 First approach where R1 = r1∈R1 r( 1). 3.1.1 Query Snowball method (QSBP) 3.1.2 Word Pairs The basic idea behind QSB(QuerySnowball) [8] is to close Having determined the query relevance score, the next the gap between the query (i.e. information need represen- step is to define the summary score. To this end, we use tation) and relevant sentences by enriching the information word pairs rather than individual words as the basic unit. need representation based on co-occurrences. To this end, This is because word pairs are more informative for discrimi- QSB computes a query relevance score for each word in the nating across different pieces of information than single com- source documents as described below. mon words. Thus, the word pair score is simply defined as: Figure 1 shows the concept of QSB. Here, Q is the set sp(w1,w2)=sr(w1)sr(w2) and the summary score is com- puted as: of query terms (each represented by q), R1 is the set of r words ( 1) that co-occur with a query term in the same sen- fQSBP (S)= sp(w1,w2)(3) R r tence, and 2 is the set of words ( 2) that co-occur with {w1,w2|w1=w2 and w1,w2∈u and u∈S} a word from R1, excluding those that are already in R1. where u is a textual unit, which in our case is a sentence. The imaginary root node at the center represents the infor- Our problem then is to select S to maximize fQSBP (S). mation need, and we assume that the need is propagated through this graph, where edges represent within-sentence co-occurrences. Thus, to compute sentence scores, we use 3.1.3 Summarization not only the query terms but also the words in R1 and R2. We derive a summary maximizing this score function using Our first clue for computing a word score is the query- ILP solver. QSB and QSBP are designed to process natural independent importance of the word.