TY - GEN
T1 - Non-convex regularized self-representation for unsupervised feature selection
AU - Wang, Weizhi
AU - Zhang, Hongzhi
AU - Zhu, Pengfei
AU - Zhang, Dapeng
AU - Zuo, Wangmeng
PY - 2015/1/1
Y1 - 2015/1/1
N2 - Feature selection aims to select a subset of features to decrease time complexity, reduce storage burden and improve the generalization ability of classification or clustering. For the countless unlabeled high dimensional data, unsupervised feature selection is effective in alleviating the curse of dimension-ality and can find applications in various fields. In this paper, we propose a non-convex regularized self-representation (RSR) model where features can be represented by a linear combination of other features, and propose to impose L2,p norm (0 < p < 1) regularization on self-representation coefficients for unsupervised feature selection. Compared with the conventional L2, 1 norm regularization, when p < 1, much sparser solution is obtained on the self-representation coefficients, and it is also more effective in selecting salient features. To solve the non-convex RSR model, we further propose an efficient iterative reweighted least squares (IRLS) algorithm with guaranteed convergence to fixed point. Extensive experimental results on nine datasets show that our feature selection method with small p is more effective. It mostly outperforms features selected at p = 1 and other state-of-the-art unsupervised feature selection methods in terms of classification accuracy and clustering result.
AB - Feature selection aims to select a subset of features to decrease time complexity, reduce storage burden and improve the generalization ability of classification or clustering. For the countless unlabeled high dimensional data, unsupervised feature selection is effective in alleviating the curse of dimension-ality and can find applications in various fields. In this paper, we propose a non-convex regularized self-representation (RSR) model where features can be represented by a linear combination of other features, and propose to impose L2,p norm (0 < p < 1) regularization on self-representation coefficients for unsupervised feature selection. Compared with the conventional L2, 1 norm regularization, when p < 1, much sparser solution is obtained on the self-representation coefficients, and it is also more effective in selecting salient features. To solve the non-convex RSR model, we further propose an efficient iterative reweighted least squares (IRLS) algorithm with guaranteed convergence to fixed point. Extensive experimental results on nine datasets show that our feature selection method with small p is more effective. It mostly outperforms features selected at p = 1 and other state-of-the-art unsupervised feature selection methods in terms of classification accuracy and clustering result.
KW - L norm 2p
KW - Self-representation
KW - Sparse representation
KW - Unsupervised feature selection
UR - http://www.scopus.com/inward/record.url?scp=84945940939&partnerID=8YFLogxK
U2 - 10.1007/978-3-319-23862-3_6
DO - 10.1007/978-3-319-23862-3_6
M3 - Conference article published in proceeding or book
SN - 9783319238616
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 55
EP - 65
BT - Intelligence Science and Big Data Engineering
PB - Springer Verlag
T2 - 5th International Conference on Intelligence Science and Big Data Engineering, IScIDE 2015
Y2 - 14 June 2015 through 16 June 2015
ER -