HyFit: Hybrid Fine-Tuning With Diverse Sampling for Abstractive Summarization

Shu Zhao, Yuanfang Cheng,Yanping Zhang,Jie Chen,Zhen Duan, Yang Sun, Xinyuan Wang

IEEE Transactions on Big Data(2024)

引用 0|浏览13
暂无评分
摘要
Abstractive summarization has made significant progress in recent years, which aims to generate a concise and coherent summary that contains the most important facts from the source document. Current fine-tuning approaches based on pre-training models typically rely on autoregressive and maximum likelihood estimation, which may result in inconsistent historical distributions generated during the training and inference stages, i.e., exposure bias problem. To alleviate this problem, we propose a hybrid fine-tuning model(HyFit), which combines contrastive learning and reinforcement learning in a diverse sampling space. Firstly, we introduce reparameterization and probability-based sampling methods to generate a set of summary candidates called candidates bank, which improves the diversity and quality of the decoding sampling space and incorporates the potential for uncertainty. Secondly, hybrid fine-tuning with sampled candidates bank, upweighting confident summaries and downweighting unconfident ones. Experiments demonstrate that HyFit significantly outperforms the state-of-the-art models on SAMSum and DialogSum. HyFit also shows good performance on low-resource summarization, on DialogSum dataset, using only approximate 8% of the examples exceed the performance of the base model trained on all examples.
更多
查看译文
关键词
Abstractive summarization,contrastive learning,hybrid fine-tuning,reinforcement learning,sampling method
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要