Multiview semi-supervised learning with consensus
Obtaining high-quality and up-to-date labeled data can be difficult in many real-world machine learning applications. Semi-supervised learning aims to improve the performance of a classifier trained with limited number of labeled data by utilizing the unlabeled ones. This paper demonstrates a way to...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2013
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/99262 http://hdl.handle.net/10220/13512 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-99262 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-992622020-05-28T07:18:16Z Multiview semi-supervised learning with consensus Li, Guangxia Chang, Kuiyu Hoi, Steven C. H. School of Computer Engineering DRNTU::Engineering::Computer science and engineering::Data Obtaining high-quality and up-to-date labeled data can be difficult in many real-world machine learning applications. Semi-supervised learning aims to improve the performance of a classifier trained with limited number of labeled data by utilizing the unlabeled ones. This paper demonstrates a way to improve the transductive SVM, which is an existing semi-supervised learning algorithm, by employing a multiview learning paradigm. Multiview learning is based on the fact that for some problems, there may exist multiple perspectives, so called views, of each data sample. For example, in text classification, the typical view contains a large number of raw content features such as term frequency, while a second view may contain a small but highly informative number of domain specific features. We propose a novel two-view transductive SVM that takes advantage of both the abundant amount of unlabeled data and their multiple representations to improve classification result. The idea is straightforward: train a classifier on each of the two views of both labeled and unlabeled data, and impose a global constraint requiring each classifier to assign the same class label to each labeled and unlabeled sample. We also incorporate manifold regularization, a kind of graph-based semi-supervised learning method into our framework. The proposed two-view transductive SVM was evaluated on both synthetic and real-life data sets. Experimental results show that our algorithm performs up to 10 percent better than a single-view learning approach, especially when the amount of labeled data is small. The other advantage of our two-view semi-supervised learning approach is its significantly improved stability, which is especially useful when dealing with noisy data in real-world applications. 2013-09-18T01:19:08Z 2019-12-06T20:05:11Z 2013-09-18T01:19:08Z 2019-12-06T20:05:11Z 2012 2012 Journal Article Li, G., Chang, K., & Hoi, S. C. H. (2012). Multiview semi-supervised learning with consensus. IEEE transactions on knowledge and data engineering, 24(11), 2040-2051. 1041-4347 https://hdl.handle.net/10356/99262 http://hdl.handle.net/10220/13512 10.1109/TKDE.2011.160 en IEEE transactions on knowledge and data engineering © 2012 IEEE |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering::Data |
spellingShingle |
DRNTU::Engineering::Computer science and engineering::Data Li, Guangxia Chang, Kuiyu Hoi, Steven C. H. Multiview semi-supervised learning with consensus |
description |
Obtaining high-quality and up-to-date labeled data can be difficult in many real-world machine learning applications. Semi-supervised learning aims to improve the performance of a classifier trained with limited number of labeled data by utilizing the unlabeled ones. This paper demonstrates a way to improve the transductive SVM, which is an existing semi-supervised learning algorithm, by employing a multiview learning paradigm. Multiview learning is based on the fact that for some problems, there may exist multiple perspectives, so called views, of each data sample. For example, in text classification, the typical view contains a large number of raw content features such as term frequency, while a second view may contain a small but highly informative number of domain specific features. We propose a novel two-view transductive SVM that takes advantage of both the abundant amount of unlabeled data and their multiple representations to improve classification result. The idea is straightforward: train a classifier on each of the two views of both labeled and unlabeled data, and impose a global constraint requiring each classifier to assign the same class label to each labeled and unlabeled sample. We also incorporate manifold regularization, a kind of graph-based semi-supervised learning method into our framework. The proposed two-view transductive SVM was evaluated on both synthetic and real-life data sets. Experimental results show that our algorithm performs up to 10 percent better than a single-view learning approach, especially when the amount of labeled data is small. The other advantage of our two-view semi-supervised learning approach is its significantly improved stability, which is especially useful when dealing with noisy data in real-world applications. |
author2 |
School of Computer Engineering |
author_facet |
School of Computer Engineering Li, Guangxia Chang, Kuiyu Hoi, Steven C. H. |
format |
Article |
author |
Li, Guangxia Chang, Kuiyu Hoi, Steven C. H. |
author_sort |
Li, Guangxia |
title |
Multiview semi-supervised learning with consensus |
title_short |
Multiview semi-supervised learning with consensus |
title_full |
Multiview semi-supervised learning with consensus |
title_fullStr |
Multiview semi-supervised learning with consensus |
title_full_unstemmed |
Multiview semi-supervised learning with consensus |
title_sort |
multiview semi-supervised learning with consensus |
publishDate |
2013 |
url |
https://hdl.handle.net/10356/99262 http://hdl.handle.net/10220/13512 |
_version_ |
1681056088848334848 |