Multiview semi-supervised learning with consensus

Obtaining high-quality and up-to-date labeled data can be difficult in many real-world machine learning applications. Semi-supervised learning aims to improve the performance of a classifier trained with limited number of labeled data by utilizing the unlabeled ones. This paper demonstrates a way to...

Full description

Saved in:
Bibliographic Details
Main Authors: Li, Guangxia, Chang, Kuiyu, Hoi, Steven C. H.
Other Authors: School of Computer Engineering
Format: Article
Language:English
Published: 2013
Subjects:
Online Access:https://hdl.handle.net/10356/99262
http://hdl.handle.net/10220/13512
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-99262
record_format dspace
spelling sg-ntu-dr.10356-992622020-05-28T07:18:16Z Multiview semi-supervised learning with consensus Li, Guangxia Chang, Kuiyu Hoi, Steven C. H. School of Computer Engineering DRNTU::Engineering::Computer science and engineering::Data Obtaining high-quality and up-to-date labeled data can be difficult in many real-world machine learning applications. Semi-supervised learning aims to improve the performance of a classifier trained with limited number of labeled data by utilizing the unlabeled ones. This paper demonstrates a way to improve the transductive SVM, which is an existing semi-supervised learning algorithm, by employing a multiview learning paradigm. Multiview learning is based on the fact that for some problems, there may exist multiple perspectives, so called views, of each data sample. For example, in text classification, the typical view contains a large number of raw content features such as term frequency, while a second view may contain a small but highly informative number of domain specific features. We propose a novel two-view transductive SVM that takes advantage of both the abundant amount of unlabeled data and their multiple representations to improve classification result. The idea is straightforward: train a classifier on each of the two views of both labeled and unlabeled data, and impose a global constraint requiring each classifier to assign the same class label to each labeled and unlabeled sample. We also incorporate manifold regularization, a kind of graph-based semi-supervised learning method into our framework. The proposed two-view transductive SVM was evaluated on both synthetic and real-life data sets. Experimental results show that our algorithm performs up to 10 percent better than a single-view learning approach, especially when the amount of labeled data is small. The other advantage of our two-view semi-supervised learning approach is its significantly improved stability, which is especially useful when dealing with noisy data in real-world applications. 2013-09-18T01:19:08Z 2019-12-06T20:05:11Z 2013-09-18T01:19:08Z 2019-12-06T20:05:11Z 2012 2012 Journal Article Li, G., Chang, K., & Hoi, S. C. H. (2012). Multiview semi-supervised learning with consensus. IEEE transactions on knowledge and data engineering, 24(11), 2040-2051. 1041-4347 https://hdl.handle.net/10356/99262 http://hdl.handle.net/10220/13512 10.1109/TKDE.2011.160 en IEEE transactions on knowledge and data engineering © 2012 IEEE
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering::Data
spellingShingle DRNTU::Engineering::Computer science and engineering::Data
Li, Guangxia
Chang, Kuiyu
Hoi, Steven C. H.
Multiview semi-supervised learning with consensus
description Obtaining high-quality and up-to-date labeled data can be difficult in many real-world machine learning applications. Semi-supervised learning aims to improve the performance of a classifier trained with limited number of labeled data by utilizing the unlabeled ones. This paper demonstrates a way to improve the transductive SVM, which is an existing semi-supervised learning algorithm, by employing a multiview learning paradigm. Multiview learning is based on the fact that for some problems, there may exist multiple perspectives, so called views, of each data sample. For example, in text classification, the typical view contains a large number of raw content features such as term frequency, while a second view may contain a small but highly informative number of domain specific features. We propose a novel two-view transductive SVM that takes advantage of both the abundant amount of unlabeled data and their multiple representations to improve classification result. The idea is straightforward: train a classifier on each of the two views of both labeled and unlabeled data, and impose a global constraint requiring each classifier to assign the same class label to each labeled and unlabeled sample. We also incorporate manifold regularization, a kind of graph-based semi-supervised learning method into our framework. The proposed two-view transductive SVM was evaluated on both synthetic and real-life data sets. Experimental results show that our algorithm performs up to 10 percent better than a single-view learning approach, especially when the amount of labeled data is small. The other advantage of our two-view semi-supervised learning approach is its significantly improved stability, which is especially useful when dealing with noisy data in real-world applications.
author2 School of Computer Engineering
author_facet School of Computer Engineering
Li, Guangxia
Chang, Kuiyu
Hoi, Steven C. H.
format Article
author Li, Guangxia
Chang, Kuiyu
Hoi, Steven C. H.
author_sort Li, Guangxia
title Multiview semi-supervised learning with consensus
title_short Multiview semi-supervised learning with consensus
title_full Multiview semi-supervised learning with consensus
title_fullStr Multiview semi-supervised learning with consensus
title_full_unstemmed Multiview semi-supervised learning with consensus
title_sort multiview semi-supervised learning with consensus
publishDate 2013
url https://hdl.handle.net/10356/99262
http://hdl.handle.net/10220/13512
_version_ 1681056088848334848