Transitive Vision-Language Prompt Learning for Domain Generalization
CoRR(2024)
摘要
The vision-language pre-training has enabled deep models to make a huge step
forward in generalizing across unseen domains. The recent learning method based
on the vision-language pre-training model is a great tool for domain
generalization and can solve this problem to a large extent. However, there are
still some issues that an advancement still suffers from trading-off between
domain invariance and class separability, which are crucial in current DG
problems. However, there are still some issues that an advancement still
suffers from trading-off between domain invariance and class separability,
which are crucial in current DG problems. In this paper, we introduce a novel
prompt learning strategy that leverages deep vision prompts to address domain
invariance while utilizing language prompts to ensure class separability,
coupled with adaptive weighting mechanisms to balance domain invariance and
class separability. Extensive experiments demonstrate that deep vision prompts
effectively extract domain-invariant features, significantly improving the
generalization ability of deep models and achieving state-of-the-art
performance on three datasets.
更多查看译文
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要