On feature selection with principal component analysis for one-class SVM

In this short note, we demonstrate the use of principal components analysis (PCA) for one-class support vector machine (one-class SVM) as a dimension reduction tool. However, unlike almost all other usage of PCA which extracts the eigenvectors associated with top eigenvalues as the projection direct...

Full description

Saved in:
Bibliographic Details
Main Author: Lian, Heng
Other Authors: School of Physical and Mathematical Sciences
Format: Article
Language:English
Published: 2013
Subjects:
Online Access:https://hdl.handle.net/10356/105603
http://hdl.handle.net/10220/17154
http://dx.doi.org/10.1016/j.patrec.2012.01.019
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:In this short note, we demonstrate the use of principal components analysis (PCA) for one-class support vector machine (one-class SVM) as a dimension reduction tool. However, unlike almost all other usage of PCA which extracts the eigenvectors associated with top eigenvalues as the projection directions, here it is the eigenvectors associated with small eigenvalues that are of interests, and in particular the null of the eigenspace, since the null space in fact characterizes the common features of the training samples. Image retrieval examples are used to illustrate the effectiveness of dimension reduction.