Anti-Distillation: Knowledge Transfer from a Simple Model to the Complex One

2022 Ivannikov Ispras Open Conference (ISPRAS)(2022)

引用 0|浏览6
暂无评分
摘要
The paper considers the problem of adapting the model to new data with a large amount of information. We propose to build a more complex model using the parameters of a simple one. We take into account not only the accuracy of the prediction on the original samples but also the adaptability to new data and the robustness of the obtained solution. The work is devoted to developing the method that allows adapting the pre-trained model to a more heterogeneous dataset. In the computational experiment, we analyse the quality of predictions and model robustness on Fashion-MNIST dataset.
更多
查看译文
关键词
knowledge transfer,weight initialization,distillation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要