Study on Unsupervised Feature Selection Method Based on Extended Entropy
DOI:
https://doi.org/10.31577/cai_2019_1_223Keywords:
Unsupervised feature selection, extended entropy, information loss, correlation valueAbstract
Feature selection techniques are designed to find the relevant feature subset of the original features that can facilitate clustering, classification and retrieval. It is an important research topic in pattern recognition and machine learning. Feature selection is mainly partitioned into two classes, i.e. supervised and unsupervised methods. Currently research mostly concentrates on supervised ones. Few efficient unsupervised feature selection methods have been developed because no label information is available. On the other hand, it is difficult to evaluate the selected features. An unsupervised feature selection method based on extended entropy is proposed here. The information loss based on extended entropy is used to measure the correlation between features. The method assures that the selected features have both big individual information and little redundancy information with the selected features. At last, the efficiency of the proposed method is illustrated with some practical datasets.Downloads
Download data is not yet available.
Downloads
Published
2019-04-26
How to Cite
Sun, Z., Li, F., & Huang, H. (2019). Study on Unsupervised Feature Selection Method Based on Extended Entropy. Computing and Informatics, 38(1), 223–239. https://doi.org/10.31577/cai_2019_1_223
Issue
Section
Articles