谷歌浏览器插件
订阅小程序
在清言上使用

Research of Weibo Text Classification Based on Knowledge Distillation and Joint Model

Zhengwu Yuan, Xiankang Peng

2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC)(2021)

引用 1|浏览5
暂无评分
摘要
Text classification is a basic task in natural language processing. In 2018, the BERT(Bidirectional Encoder Representation from Transformers) was proposed. This model greatly improves the effect of natural language processing related tasks. However, the model size of the pre-trained language model is large, and because of its huge network structure, the predict time is long. In order to solve these problems, an improved model that uses knowledge distillation and adversarial perturbation is proposed. In the training phase, RoBERTa-wwm-ext is used as the teacher model, the joint model of Text-CNN and Text-RCNN is used as a student model, combined with label smoothing and adversarial perturbation method to improve the classification accuracy of the student model, compared with Text-CNN on two datasets, is improved by 1.91% and 1.21% respectively. Using the student model to classify texts has the advantages of easy for deployment and taking less prediction time while obtaining ideal accuracy.
更多
查看译文
关键词
Training,Perturbation methods,Text categorization,Blogs,Predictive models,Natural language processing,Task analysis
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要