Rethinking statistical learning theory: learning using statistical invariants

作者:Vladimir Vapnik, Rauf Izmailov

摘要

This paper introduces a new learning paradigm, called Learning Using Statistical Invariants (LUSI), which is different from the classical one. In a classical paradigm, the learning machine constructs a classification rule that minimizes the probability of expected error; it is data-driven model of learning. In the LUSI paradigm, in order to construct the desired classification function, a learning machine computes statistical invariants that are specific for the problem, and then minimizes the expected error in a way that preserves these invariants; it is thus both data- and invariant-driven learning. From a mathematical point of view, methods of the classical paradigm employ mechanisms of strong convergence of approximations to the desired function, whereas methods of the new paradigm employ both strong and weak convergence mechanisms. This can significantly increase the rate of convergence.

论文关键词:Intelligent teacher, Privileged information, Support vector machine, Neural network, Classification, Learning theory, Regression, Conditional probability, Kernel function, Ill-Posed problem, Reproducing Kernel Hilbert space, Weak convergence

论文评审过程:

论文官网地址:https://doi.org/10.1007/s10994-018-5742-0