On the Role of Summary Content Units in Text Summarization Evaluation
arxiv(2024)
摘要
At the heart of the Pyramid evaluation method for text summarization lie
human written summary content units (SCUs). These SCUs are concise sentences
that decompose a summary into small facts. Such SCUs can be used to judge the
quality of a candidate summary, possibly partially automated via natural
language inference (NLI) systems. Interestingly, with the aim to fully automate
the Pyramid evaluation, Zhang and Bansal (2021) show that SCUs can be
approximated by automatically generated semantic role triplets (STUs). However,
several questions currently lack answers, in particular: i) Are there other
ways of approximating SCUs that can offer advantages? ii) Under which
conditions are SCUs (or their approximations) offering the most value? In this
work, we examine two novel strategies to approximate SCUs: generating SCU
approximations from AMR meaning representations (SMUs) and from large language
models (SGUs), respectively. We find that while STUs and SMUs are competitive,
the best approximation quality is achieved by SGUs. We also show through a
simple sentence-decomposition baseline (SSUs) that SCUs (and their
approximations) offer the most value when ranking short summaries, but may not
help as much when ranking systems or longer summaries.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要