Information-Content Based Sentence Extraction for Text Summarization

Information-Content Based Sentence Extraction for Text Summarization

Information-Content Based Sentence Extraction for Text Summarization Daniel Mallett1 James Elding2 Mario A. Nascimento1 1Department of Computing Science, University of Alberta, Canada fmallett, [email protected] 2Workers' Compensation Board, Alberta, Canada [email protected] Abstract To address this problem, most existing text summarizers, e.g. [9, 13, 15], use the framework given in Figure 1. This paper proposes the FULL-COVERAGE summarizer: an efficient, information retrieval oriented method to extract INPUT : Full text for document D non-redundant sentences from text for summarization pur- OUTPUT: Summary text of document D poses. Our method leverages existing Information Retrieval 1: Parse the document into sentences. technology by extracting key-sentences on the premise that 2: Determine the saliency (rank) of each sentence. the relevance of a sentence is proportional to its similarity 3: Generate a summary using a subset of the ranked to the whole document. We show that our method can pro- sentences (usually a few top ones). duce sentence-based summaries that are up to 78% smaller than the original text with only 3% loss in retrieval perfor- Figure 1: Text Summarization Framework mance. Summaries are generally evaluated along two orthogonal axes: information content (our focus) and readability. Our goal is to find a set of non-redundant sentences faithful to 1 Introduction the information content of the original text (steps 1 and 2 in Figure 1) from which one could generate high-quality Text summarization is the process of condensing a human readable summaries (step 3 in Figure 1). source text while preserving its information content and Most existing summarization techniques use some vari- maintaining readability. The main (and large) difference ant of the well known tf ¤ idf model [2]. Our algorithm between automatic and human-based text summarization extracts sentences that “fully covers” the concept-space of is that humans can capture and convey subtle themes that a document by iteratively measuring the similarity of each permeate documents, whereas automatic approaches have a sentence to the whole document and striking-out words that large difficulty to do the same. Nonetheless, as the amount have already been covered. As our experiments corroborate, of information available in electronic format continues to the FULL-COVERAGE approach is a simple yet efficient al- grow, research into automatic text summarization has taken ternative for steps 1 and 2 of the framework in Figure 1. on renewed interest. Indeed, A whole conference series, One of the great challenges of text summarization is the Document Understanding Conferences (DUC)1 has been summary evaluation [14]. In order to evaluate how well devoted to the topic in the recent past. our sentence-extractor performs, we propose an extrin- We model the summarization problem as follows: sic retrieval-oriented evaluation that, unlike most previous work, does not rely on human assessors. Our experimental Given a document D, composed of a set of N sen- methodology aims at measuring how well a summary cap- tences S = fS1; S2; :::; SN g, and a percentage p, tures the salient information content of a document. We ex- 0 select a subset of sentences S ⊆ S such that: (1) periment with data from SMART’s TIME Magazine Collec- 0 0 jS j · p£jSj and (2) using S as a representative tion2 as well as the TREC3 documents used for 2002’s edi- of D is as effective as using S in terms of retrieval tion of DUC. As we will discuss, the results from our eval- performance. 2ftp://ftp.cs.cornell.edu/pub/smart/time/ 1http://duc.nist.gov 3http://trec.nist.gov uation are promising. A summarized version of the TIME Maximal Marginal Relevance (MMR) is a technique ex- Magazine collection, 40% the size of the original text, loses plicitly concerned with reducing redundancy [4]. MMR re- only about 5% in terms of retrieval precision. The DUC re- ranks retrieval query results (relevant document lists) based sults are even more interesting; our FULL-COVERAGE ap- on a document’s dissimilarity to other relevant documents. proach ranks highly against other DUC competitors, pro- The method is extended to summarize single-documents by ducing summaries 22% the size of the original texts with re-ranking salient sentences instead of documents. Our re- only a 3% loss in retrieval performance. dundancy reducing mechanism is strikingly different than This paper is structured as follows. In Section 2 we re- that approach. view previous research related to the problem of text sum- marization and summary evaluation. Section 3 presents our 2.2 Summary Evaluation FULL-COVERAGE key-sentence extraction method. Sec- tion 4 provides experiments comparing our method to ten Summaries can be evaluated using intrinsic or extrinsic other summarization approaches. Finally, Section 5 con- measures [14]. While intrinsic methods attempt to measure cludes the paper. summary quality using human evaluation thereof, extrin- sic methods measure the same through a task-based perfor- 2 Related Work mance measure such the information retrieval-oriented task. Typically, the former is used, e.g., [1, 13], [15] being a no- Important seminal papers and recent advance papers on table exception of the latter. text summarization can be found in Mani and Maybury’s We propose an extrinsic IR evaluation that measures the book on text summarization [9]. The next section out- information content of a summary with respect to the orig- lines recent advances in text summarization relevant to our inal document. We posit that our evaluation measures the FULL-COVERAGE approach, followed by a section dis- ability of the summary to retain the information content of cussing related evaluation methodologies. the original text, i.e., if the original text is relevant to a cer- tain set of queries then the summary will be as well. 2.1 Summarization Techniques The TIPSTER/SUMMAC and NTCIR conferences have experimented with the retrieval-oriented task and use pre- Text summarization by extraction can employ various cision and recall in their evaluations. These works focus levels of granularity, e.g., keyword, sentence, or paragraph. on the relevance assessment task, not the information re- Most research concentrates on sentence-extraction because trieval task itself, and require considerable human involve- the readability of a list of keywords is typically low while ment. Sakai et al [15] use a performance measure simi- paragraphs are unlikely to cover the information content of lar to our evaluation. A striking difference though is that a document given summary space constraints. they chose a priori which documents are relevant to their MEAD [13], a state of the art sentence-extractor and a queries whereas we experiment with a standard collection top performer at DUC, aims to extracts sentences central to of queries and relevant documents. the overall topic of a document. The system employs (1) a centroid score representing the centrality of a sentence to the overall document, (2) a position score which is inversely 3 The FULL-COVERAGE Algorithm proportional to the position of a sentence in the document, and (3) an overlap-with-first score which is the inner prod- The intuition behind our FULL-COVERAGE summarizer uct of the tf ¤ idf with the first sentence of the document. is to consider key-sentence extraction from an information MEAD attempts to reduce summary redundancy by elimi- retrieval perspective based on the premise that the relevance nating sentences above a similarity threshold parameter. As of a sentence is proportional to its similarity to the whole we will see, our proposal is simpler than MEAD and consis- document. Like MMR [4], we aim to minimize summary tently outperformed it in the experiments we carried. redundancy, however our method is different from MMR Other approaches for sentence extraction include NLP in that we do not consider sentence dissimilarity, but in- methods [1, 3] and machine-learning techniques [11, 16]. stead focus on reducing summary redundancy by removing These approaches tend to be computationally expensive and query terms that have already been covered by other salient genre-dependent even though they are typically based on sentences. If several sentences from a document share a the more general tf ¤ idf framework. Work on genera- set of common terms, i.e., all refer to the same concept, tive algorithms includes sentence compression [6], sentence only a very small subset of those sentences, likely a single fusion [5], and sentence modification[10]. We envision one, which covers the same concept space will be consid- our FULL-COVERAGE approach as providing the input ex- ered salient using our algorithm. Next we discuss how our tracted sentences into these type of generative algorithms. proposed method fits in the framework shown in Figure 1. The first phase of the FULL-COVERAGE algorithm (Fig- from the query string Q (step 2). Querying again (step 3), ure 1 step 1) is to parse a document into sentences. We S2 is returned, and any occurrences of the words found in use a technique from [8] for this task. During this phase S2 (“E” and “A”) are struck out from Q (step 4). In step 5, we also remove stop-words and apply the Porter stemming the algorithm stops because Q is empty. The resulting full algorithm [12]. coverage set has covered all words from D. Our experimentation employs only the standard tf ¤ idf INPUT : Document D weighting scheme as defined by the vector model [2] for OUTPUT: FULL-COVERAGE of D implementing the sim() function. The key advantage to 1: Q D Ã using a single weighting scheme is that the algorithm in 2: repeat Figure 2 can be easily implemented on top of most (if not 3: S¤ S sim Q; S (1 · i · N) Ã i with max ( i), any) vector model based implementation such as that pro- 4: F C F C S¤ Ã [ vided by MG [17].

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    5 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us