Assessing Prompt Injection Risks in 200+ Custom GPTs.
CoRR(2023)
摘要
In the rapidly evolving landscape of artificial intelligence, ChatGPT has
been widely used in various applications. The new feature: customization of
ChatGPT models by users to cater to specific needs has opened new frontiers in
AI utility. However, this study reveals a significant security vulnerability
inherent in these user-customized GPTs: prompt injection attacks. Through
comprehensive testing of over 200 user-designed GPT models via adversarial
prompts, we demonstrate that these systems are susceptible to prompt
injections. Through prompt injection, an adversary can not only extract the
customized system prompts but also access the uploaded files. This paper
provides a first-hand analysis of the prompt injection, alongside the
evaluation of the possible mitigation of such attacks. Our findings underscore
the urgent need for robust security frameworks in the design and deployment of
customizable GPT models. The intent of this paper is to raise awareness and
prompt action in the AI community, ensuring that the benefits of GPT
customization do not come at the cost of compromised security and privacy.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要