Feature selection via joint embedding learning and sparse regression
View Researcher II's Other CodesDisclaimer: “The provided code links for this paper are external links. Science Nest has no responsibility for the accuracy, legality or content of these links. Also, by downloading this code(s), you agree to comply with the terms of use as set out by the author(s) of the code(s).”
Please contact us in case of a broken link from here
Authors | Chenping Hou, Feiping Nie, Dongyun Yi, Yi Wu |
Journal/Conference Name | The 22nd International Joint Conference on Artificial Intelligence (IJCAI) |
Paper Category | Computer Science |
Paper Abstract | The problem of feature selection has aroused considerable research interests in the past few years. Traditional learning based feature selection methods separate embedding learning and feature ranking. In this paper, we introduce a novel unsupervised feature selection approach via Joint Embedding Learning and Sparse Regression (JELSR). Instead of simply employing the graph laplacian for embedding learning and then regression, we use the weight via locally linear approximation to construct graph and unify embedding learning and sparse regression to perform feature selection. By adding the 2,1-norm regularization, we can learn a sparse matrix for feature ranking. We also provide an effective method to solve the proposed problem. Compared with traditional unsupervised feature selection methods, our approach could integrate the merits of embedding learning and sparse regression simultaneously. Plenty of experimental results are provided to show the validity. |
Date of publication | 2011 |
Code Programming Language | MATLAB |
Comment |