Efficient Trigger Word Insertion

2023 9th International Conference on Big Data and Information Analytics (BigDIA)(2023)

引用 0|浏览5
暂无评分
摘要
With the rapid advancements in the natural language processing (NLP) domain in recent years, the emergence of backdoor attacks presents substantial threats to deep neural network models. However, prior research has often overlooked the influence of the poisoning rate. This paper aims to address this gap by prioritizing the reduction of poisoned samples while still attaining a comparable Attack Success Rate (ASR) in the context of text backdoor attacks. Our primary focus revolves around introducing an efficient strategy for trigger word insertion, encompassing both trigger word optimization and poisoned sample selection. To achieve our objectives, extensive experiments were conducted across diverse datasets and models, showcasing the significant enhancements brought forth by our proposed methodology in the realm of text classification tasks. Remarkable outcomes include an ASR surpassing 90%, utilizing a mere 10 poisoned samples in the dirty-label setting, and delivering compelling performance with only 1.5% of the training data in the clean-label setting.
更多
查看译文
关键词
text backdoor attacks,data efficiency
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要