Adversarial Invariant Learning

2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021(2021)

引用 3|浏览75
暂无评分
摘要
Though machine learning algorithms are able to achieve pattern recognition from the correlation between data and labels, the presence of spurious features in the data decreases the robustness of these learned relationships with respect to varied testing environments. This is known as out-of-distribution (OoD) generalization problem. Recently, invariant risk minimization (IRM) attempts to tackle this issue by penalizing predictions based on the unstable spurious features in the data collected from different environments. However, similar to domain adaptation or domain generalization, a prevalent non -trivial limitation in these works is that the environment information is assigned by human specialists, i.e. a priori, or determined heuristically. However, an inappropriate group partitioning can dramatically deteriorate the OoD generalization and this process is expensive and time-consuming. To deal with this issue, we propose a novel theoretically principled min-max framework to iteratively construct a worst-case splitting, i.e. creating the most challenging environment splittings for the backbone learning paradigm (e.g. IRM) to learn the robust feature representation. We also design a differentiable training strategy to facilitate the feasible gradientbased computation. Numerical experiments show that our algorithmic framework has achieved superior and stable performance in various datasets, such as Colored MNIST and Punctuated stanford sentiment treebank (SST). Furthermore, we also find our algorithm to be robust even to a strong data poisoning attack. To the best of our knowl- edge, this is one of the first to adopt differentiable environment splitting method to enable stable predictions across environments without environment index information, which achieves the state-of-the-art performance on datasets with strong spurious correlation, such as Colored MNIST.
更多
查看译文
关键词
strong data poisoning attack,differentiable environment splitting method,stable predictions,environment index information,strong spurious correlation,Colored MNIST,adversarial invariant learning,machine learning algorithms,pattern recognition,learned relationships,varied testing environments,out-of-distribution generalization problem,invariant risk minimization,unstable spurious features,domain adaptation,nontrivial limitation,environment information,human specialists,inappropriate group partitioning,worst-case splitting,challenging environment splittings,backbone learning paradigm,IRM,robust feature representation,differentiable training strategy,feasible gradient,algorithmic framework,superior performance,theoretically principled min-max framework,punctuated Stanford sentiment treebank,OoD generalization problem
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要