Discriminant feature extraction using empirical probability density estimation and a local basis library

作者:

Highlights:

摘要

The authors previously developed the so-called local discriminant basis (LDB) method for signal and image classification problems. The original LDB method relies on differences in the time–frequency energy distribution of each class: it selects the subspaces where these energy distributions are well separated by some measure such as the Kullback–Leibler divergence. Through our experience and experiments on various datasets, however, we realized that the time–frequency energy distribution is not always the best quantity to analyze for classification. In this paper, we propose to use the discrimination of coordinates based, instead, on empirical probability densities. That is, we estimate the probability density of each class in each coordinate in the wavelet packet/local trigonometric bases after expanding signals into such bases. We then evaluate a power of discrimination of each subspace by selecting the m most discriminant coordinates in terms of the “distance” among the corresponding densities (e.g., by the Kullback–Leibler divergence among the densities). This information is then used for selecting a basis for classification. We will demonstrate the capability of this algorithm using both synthetic and real datasets.

论文关键词:Local feature extraction,Pattern classification,Density estimation,Kullback–Leibler divergence,Hellinger distance

论文评审过程:Received 18 December 2000, Accepted 20 November 2001, Available online 19 February 2002.

论文官网地址:https://doi.org/10.1016/S0031-3203(02)00019-5