Learning from partially supervised data using mixture models and belief functions

作者:

Highlights:

摘要

This paper addresses classification problems in which the class membership of training data are only partially known. Each learning sample is assumed to consist of a feature vector xi∈X and an imprecise and/or uncertain “soft” label mi defined as a Dempster–Shafer basic belief assignment over the set of classes. This framework thus generalizes many kinds of learning problems including supervised, unsupervised and semi-supervised learning. Here, it is assumed that the feature vectors are generated from a mixture model. Using the generalized Bayesian theorem, an extension of Bayes’ theorem in the belief function framework, we derive a criterion generalizing the likelihood function. A variant of the expectation maximization (EM) algorithm, dedicated to the optimization of this criterion is proposed, allowing us to compute estimates of model parameters. Experimental results demonstrate the ability of this approach to exploit partial information about class labels.

论文关键词:Dempster–Shafer theory,Transferable belief model,Mixture models,EM algorithm,Classification,Clustering,Partially supervised learning,Semi-supervised learning

论文评审过程:Received 13 November 2007, Revised 10 July 2008, Accepted 15 July 2008, Available online 30 July 2008.

论文官网地址:https://doi.org/10.1016/j.patcog.2008.07.014