An unsupervised approach to feature discretization and selection

作者:

Highlights:

摘要

Many learning problems require handling high dimensional datasets with a relatively small number of instances. Learning algorithms are thus confronted with the curse of dimensionality, and need to address it in order to be effective. Examples of these types of data include the bag-of-words representation in text classification problems and gene expression data for tumor detection/classification. Usually, among the high number of features characterizing the instances, many may be irrelevant (or even detrimental) for the learning tasks. It is thus clear that there is a need for adequate techniques for feature representation, reduction, and selection, to improve both the classification accuracy and the memory requirements. In this paper, we propose combined unsupervised feature discretization and feature selection techniques, suitable for medium and high-dimensional datasets. The experimental results on several standard datasets, with both sparse and dense features, show the efficiency of the proposed techniques as well as improvements over previous related techniques.

论文关键词:Feature discretization,Feature quantization,Feature selection,Linde–Buzo–Gray algorithm,Sparse data,Support vector machines,Naïve Bayes,k-nearest neighbor

论文评审过程:Available online 23 December 2011.

论文官网地址:https://doi.org/10.1016/j.patcog.2011.12.008