谷歌浏览器插件
订阅小程序
在清言上使用

Integrating R-Drop and Pre-trained Language Model for Short Text Classification

2022 8th International Conference on Big Data and Information Analytics (BigDIA)(2022)

引用 0|浏览11
暂无评分
摘要
At present, the rapid development of computer technology has led to the emergence of massive short text data represented by e-commerce user comments. Due to the sparse feature and fast update speed of short text data, effective classification methods of short text have been widely concerned and applied. In this paper, a short text classification method integrating R-Drop and pre-trained language model IRDP-STC is proposed. The original data and generated data are input into the pre-trained language model Bert with Dropout, and KL divergence loss is added to approximate the two outputs. IRDP-STC makes full use of short text information, takes into account the robustness of the model on the basis of ensuring the high classification performance of the model, and is more suitable for low training sample scenarios. The experimental results on THUCNews dataset are better than the baseline model, which verifies the effectiveness and universality of the proposed method.
更多
查看译文
关键词
R-Drop,Pre-trained language model,Short text classification
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要