Rethinking statistical learning theory: learning using statistical invariants

Machine Learning(2018)

引用 59|浏览64
暂无评分
摘要
This paper introduces a new learning paradigm, called Learning Using Statistical Invariants (LUSI), which is different from the classical one. In a classical paradigm, the learning machine constructs a classification rule that minimizes the probability of expected error; it is data-driven model of learning. In the LUSI paradigm, in order to construct the desired classification function, a learning machine computes statistical invariants that are specific for the problem, and then minimizes the expected error in a way that preserves these invariants; it is thus both data- and invariant-driven learning. From a mathematical point of view, methods of the classical paradigm employ mechanisms of strong convergence of approximations to the desired function, whereas methods of the new paradigm employ both strong and weak convergence mechanisms. This can significantly increase the rate of convergence.
更多
查看译文
关键词
Intelligent teacher,Privileged information,Support vector machine,Neural network,Classification,Learning theory,Regression,Conditional probability,Kernel function,Ill-Posed problem,Reproducing Kernel Hilbert space,Weak convergence,68Q32,68T05,68T30,83C32
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要