Learning With Confident Examples: Rank Pruning For Robust Classification With Noisy Labels

CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017)(2017)

引用 133|浏览76
暂无评分
摘要
(P) over tilde(N) over tilde learning is the problem of binary classification when training examples may be mislabeled (flipped) uniformly with noise rate rho(1) for positive examples and rho(0) for negative examples. We propose Rank Pruning (RP) to solve (P) over tilde(N) over tilde learning and the open problem of estimating the noise rates. Unlike prior solutions, RP is efficient and general, requiring (9 (T) for any unrestricted choice of probabilistic classifier with T fitting time. We prove RP achieves consistent noise estimation and equivalent expected risk as learning with uncorrupted labels in ideal conditions, and derive closed-form solutions when conditions are non-ideal. RP achieves state-of-the-art noise estimation and F1, error, and AUC-PR for both MNIST and CIFAR datasets, regardless of the amount of noise. To highlight, RP with a CNN classifier can predict if an MNIST digit is a one or not with only 0.25% error, and 0.46% error across all digits, even when 50% of positive examples are mislabeled and 50% of observed positive labels are mislabeled negative examples.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要