Are Neural Language Models Good Plagiarists? A Benchmark for Neural Paraphrase Detection

arXiv (Cornell University)(2023)

引用 0|浏览1
暂无评分
摘要
The rise of language models such as BERT allows for high-quality text paraphrasing. This is a problem to academic integrity, as it is difficult to differentiate between original and machine-generated content. We propose a benchmark consisting of paraphrased articles using recent language models relying on the Transformer architecture. Our contribution fosters future research of paraphrase detection systems as it offers a large collection of aligned original and paraphrased documents, a study regarding its structure, classification experiments with state-of-the-art systems, and we make our findings publicly available.
更多
查看译文
关键词
neural,benchmark
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络