A feature selection method for multivariate performance measures

Feature selection with specific multivariate performance measures is the key to the success of many applications such as image retrieval and text classification. The existing feature selection methods are usually designed for classification error. In this paper, we propose a generalized sparse regul...

全面介紹

Saved in:
書目詳細資料
Main Authors: Mao, Qi, Tsang, Ivor Wai-Hung
其他作者: School of Computer Engineering
格式: Article
語言:English
出版: 2013
主題:
在線閱讀:https://hdl.handle.net/10356/100979
http://hdl.handle.net/10220/16693
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
實物特徵
總結:Feature selection with specific multivariate performance measures is the key to the success of many applications such as image retrieval and text classification. The existing feature selection methods are usually designed for classification error. In this paper, we propose a generalized sparse regularizer. Based on the proposed regularizer, we present a unified feature selection framework for general loss functions. In particular, we study the novel feature selection paradigm by optimizing multivariate performance measures. The resultant formulation is a challenging problem for high-dimensional data. Hence, a two-layer cutting plane algorithm is proposed to solve this problem, and the convergence is presented. In addition, we adapt the proposed method to optimize multivariate measures for multiple-instance learning problems. The analyses by comparing with the state-of-the-art feature selection methods show that the proposed method is superior to others. Extensive experiments on large-scale and high-dimensional real-world datasets show that the proposed method outperforms l1-SVM and SVM-RFE when choosing a small subset of features, and achieves significantly improved performances over SVMperl in terms of F1-score.