Automatic Evaluation of Summaries Using N-Gram Co-Occurrence Statistics

Automatic Evaluation of Summaries Using N-Gram Co-Occurrence Statistics

Proceedings of HLT-NAACL 2003 Main Papers , pp. 71-78 Edmonton, May-June 2003 Automatic Evaluation of Summaries Using N-gram Co-Occurrence Statistics Chin-Yew Lin and Eduard Hovy Information Sciences Institute University of Southern California 4676 Admiralty Way Marina del Rey, CA 90292 {cyl,hovy}@isi.edu out that 18% of the data contained multiple judgments Abstract in the DUC 2001 single document evaluation1. To further progress in automatic summarization, in this Following the recent adoption by the machine paper we conduct an in-depth study of automatic translation community of automatic evalua- evaluation methods based on n-gram co-occurrence in tion using the BLEU/NIST scoring process, the context of DUC. Due to the setup in DUC, the we conduct an in-depth study of a similar idea evaluations we discussed here are intrinsic evaluations for evaluating summaries. The results show (Spärck Jones and Galliers 1996). Section 2 gives an that automatic evaluation using unigram co- overview of the evaluation procedure used in DUC. occurrences between summary pairs correlates Section 3 discusses the IBM BLEU (Papineni et al. surprising well with human evaluations, based 2001) and NIST (2002) n-gram co-occurrence scoring on various statistical metrics; while direct ap- procedures and the application of a similar idea in plication of the BLEU evaluation procedure evaluating summaries. Section 4 compares n-gram co- does not always give good results. occurrence scoring procedures in terms of their correla- tion to human results and on the recall and precision of statistical significance prediction. Section 5 concludes 1 Introduction this paper and discusses future directions. Automated text summarization has drawn a lot of inter- 2 Document Understanding Conference est in the natural language processing and information 2 retrieval communities in the recent years. A series of The 2002 Document Understanding Conference in- workshops on automatic text summarization (WAS cluded the follow two main tasks: 2000, 2001, 2002), special topic sessions in ACL, • Fully automatic single-document summarization: COLING, and SIGIR, and government sponsored given a document, participants were required to evaluation efforts in the United States (DUC 2002) and create a generic 100-word summary. The training Japan (Fukusima and Okumura 2001) have advanced set comprised 30 sets of approximately 10 docu- the technology and produced a couple of experimental ments each, together with their 100-word human online systems (Radev et al. 2001, McKeown et al. written summaries. The test set comprised 30 un- 2002). Despite these efforts, however, there are no seen documents. common, convenient, and repeatable evaluation meth- ods that can be easily applied to support system devel- • Fully automatic multi-document summarization: opment and just-in-time comparison among different given a set of documents about a single subject, summarization methods. participants were required to create 4 generic sum- The Document Understanding Conference (DUC 2002) maries of the entire set, containing 50, 100, 200, run by the National Institute of Standards and Technol- and 400 words respectively. The document sets ogy (NIST) sets out to address this problem by provid- were of four types: a single natural disaster event; a ing annual large scale common evaluations in text summarization. However, these evaluations involve 1 Multiple judgments occur when more than one performance human judges and hence are subject to variability (Rath score is given to the same system (or human) and human sum- et al. 1961). For example, Lin and Hovy (2002) pointed mary pairs by the same human judge. 2 DUC 2001 and DUC 2002 have similar tasks, but summaries of 10, 50, 100, and 200 words are requested in the multi- document task in DUC 2002. Figure 1. SEE in an evaluation session. single event; multiple instances of a type of event; mary of 50, 100, 200, or 400 words. Only one baseline and information about an individual. The training (baseline1) was created for the single document summa- set comprised 30 sets of approximately 10 docu- rization task. ments, each provided with their 50, 100, 200, and 400-word human written summaries. The test set 2.2 Summary Evaluation Environment comprised 30 unseen sets. To evaluate system performance NIST assessors who A total of 11 systems participated in the single- created the ‘ideal’ written summaries did pairwise com- document summarization task and 12 systems partici- parisons of their summaries to the system-generated pated in the multi-document task. summaries, other assessors’ summaries, and baseline 2.1 Evaluation Materials summaries. They used the Summary Evaluation Envi- ronment (SEE) 2.0 developed by (Lin 2001) to support For each document or document set, one human sum- the process. Using SEE, the assessors compared the mary was created as the ‘ideal’ model summary at each system’s text (the peer text) to the ideal (the model specified length. Two other human summaries were text). As shown in Figure 1, each text was decomposed also created at each length. In addition, baseline sum- into a list of units and displayed in separate windows. maries were created automatically for each length as SEE 2.0 provides interfaces for assessors to judge both reference points. For the multi-document summariza- the content and the quality of summaries. To measure tion task, one baseline, lead baseline, took the first 50, content, assessors step through each model unit, mark 100, 200, and 400 words in the last document in the all system units sharing content with the current model collection. A second baseline, coverage baseline, took unit (green/dark gray highlight in the model summary the first sentence in the first document, the first sentence window), and specify that the marked system units ex- in the second document and so on until it had a sum- press all, most, some, or hardly any of the content of the current model unit. To measure quality, assessors rate summary, the better it is. The question is: “Can we ap- grammaticality3, cohesion4, and coherence5 at five dif- ply BLEU directly without any modifications to evalu- ferent levels: all, most, some, hardly any, or none6. For ate summaries as well?”. We first ran IBM’s BLEU example, as shown in Figure 1, an assessor marked sys- evaluation script unmodified over the DUC 2001 model tem units 1.1 and 10.4 (red/dark underlines in the left and peer summary set. The resulting Spearman rank pane) as sharing some content with the current model order correlation coefficient (ρ) between BLEU and the unit 2.2 (highlighted green/dark gray in the right). human assessment for the single document task is 0.66 using one reference summary and 0.82 using three ref- 2.3 Evaluation Metrics erence summaries; while Spearman ρ for the multi- Recall at different compression ratios has been used in document task is 0.67 using one reference and 0.70 us- summarization research to measure how well an auto- ing three. These numbers indicate that they positively α 8 matic system retains important content of original correlate at = 0.01 . Therefore, BLEU seems a prom- documents (Mani et al. 1998). However, the simple sen- ising automatic scoring metric for summary evaluation. tence recall measure cannot differentiate system per- According to Papineni et al. (2001), BLEU is essentially formance appropriately, as is pointed out by Donaway a precision metric. It measures how well a machine et al. (2000). Therefore, instead of pure sentence recall translation overlaps with multiple human translations score, we use coverage score C. We define it as fol- using n-gram co-occurrence statistics. N-gram precision lows7: in BLEU is computed as follows: • − = (Number of MUs marked) E ∑ ∑Countclip (n gram) C (1) CC∈∈{Candidates} n−gram Total number of MUs in the model summary p = (2) n − E, the ratio of completeness, ranges from 1 to 0: 1 for ∑∑Count(n gram) all, 3/4 for most, 1/2 for some, 1/4 for hardly any, and 0 CC∈∈{Candidates} n−gram for none. If we ignore E (set it to 1), we obtain simple Where Countclip(n-gram) is the maximum number of n- sentence recall score. We use average coverage scores grams co-occurring in a candidate translation and a ref- derived from human judgments as the references to erence translation, and Count(n-gram) is the number of evaluate various automatic scoring methods in the fol- n-grams in the candidate translation. To prevent very lowing sections. short translations that try to maximize their precision scores, BLEU adds a brevity penalty, BP, to the for- 3 BLEU and N-gram Co-Occurrence mula: 1 if c > r To automatically evaluate machine translations the ma- BP = (3) (1−|r|/|c|) ≤ chine translation community recently adopted an n-gram e if c r co-occurrence scoring procedure BLEU (Papineni et al. Where |c| is the length of the candidate translation and 2001). The NIST (NIST 2002) scoring metric is based |r| is the length of the reference translation. The BLEU on BLEU. The main idea of BLEU is to measure the formula is then written as follows: translation closeness between a candidate translation N and a set of reference translations with a numerical met- BLEU = BP • exp w log p (4) ric. To achieve this goal, they used a weighted average ∑ n n n=1 of variable length n-gram matches between system N is set at 4 and w , the weighting factor, is set at 1/N. translations and a set of human reference translations n For summaries by analogy, we can express equation (1) and showed that a weighted average metric, i.e. BLEU, in terms of n-gram matches following equation (2): correlating highly with human assessments.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    8 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us