1998 Presentations (Communicative Events)
Summarization evaluation methods: Experiments and analysis
Two methods are used for evaluation of summarization systems: an evaluation of generated summaries against an "ideal" summary and evaluation of how well summaries help a person perform in a task such as information retrieval. We carried out two large experiments to study the two evaluation methods. Our results show that different parameters of an experiment can dramatically affect how well a system scores. For example, summary length was found to affect both types of evaluations. For the "ideal" summary based evaluation, accuracy decreases as summary length increases, while for task based evaluations summary length and accuracy on an information retrieval task appear to correlate randomly. In this paper, we show how this parameter and others can affect evaluation results and describe how parameters can be controlled to produce a sound evaluation.
Subjects
Files
- jing_al_98.pdf application/pdf 188 KB Download File
More About This Work
- Academic Units
- Computer Science
- Publisher
- AAAI Symposium on Intelligent Summarization
- Published Here
- April 29, 2013