Revisiting Demonstration Selection Strategies in In-Context Learning
CoRR(2024)
摘要
Large language models (LLMs) have shown an impressive ability to perform a
wide range of tasks using in-context learning (ICL), where a few examples are
used to describe a task to the model. However, the performance of ICL varies
significantly with the choice of demonstrations, and it is still unclear why
this happens or what factors will influence its choice. In this work, we first
revisit the factors contributing to this variance from both data and model
aspects, and find that the choice of demonstration is both data- and
model-dependent. We further proposed a data- and model-dependent demonstration
selection method, TopK + ConE, based on the assumption that
the performance of a demonstration positively correlates with its
contribution to the model's understanding of the test samples, resulting in a
simple and effective recipe for ICL. Empirically, our method yields consistent
improvements in both language understanding and generation tasks with different
model scales. Further analyses confirm that, besides the generality and
stability under different circumstances, our method provides a unified
explanation for the effectiveness of previous methods. Code will be released.
更多查看译文
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要