Multiple-instance learning from unlabeled bags with pairwise similarity

In multiple-instance learning (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL methods generally assume that training bags are fully labeled. Ho...

Full description

Saved in:
Bibliographic Details
Main Authors: Feng, Lei, Shu, Senlin, Cao, Yuzhou, Tao, Lue, Wei, Hongxin, Xiang, Tao, An, Bo, Niu, Gang
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/172864
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-172864
record_format dspace
spelling sg-ntu-dr.10356-1728642023-12-27T02:53:24Z Multiple-instance learning from unlabeled bags with pairwise similarity Feng, Lei Shu, Senlin Cao, Yuzhou Tao, Lue Wei, Hongxin Xiang, Tao An, Bo Niu, Gang School of Computer Science and Engineering Engineering::Computer science and engineering Multiple-Instance Learning Empirical Risk Minimization In multiple-instance learning (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL methods generally assume that training bags are fully labeled. However, the exact labels of training examples may not be accessible, due to security, confidentiality, and privacy concerns. Fortunately, it could be easier for us to access the pairwise similarity between two bags (indicating whether two bags share the same label or not) and unlabeled bags, as we do not need to know the underlying label of each bag. In this paper, we provide the first attempt to investigate MIL from only similar-dissimilar-unlabeled bags. To solve this new MIL problem, we first propose a strong baseline method that trains an instance-level classifier by employing an unlabeled-unlabeled learning strategy. Then, we also propose to train a bag-level classifier based on a convex formulation and theoretically derive a generalization error bound for this method. Comprehensive experimental results show that our instance-level classifier works well, while our bag-level classifier even has better performance. National Research Foundation (NRF) This work was supported in part by the National Key R&D Program of China under Grant 2022YFB3103500. The work of Bo An is supported by the National Research Foundation, Singapore under its Industry Alignment Fund – Pre-positioning (IAF-PP) Funding Initiative. The work of Lei Feng is supported by the National Natural Science Foundation of China under Grant 62106028, in part by the Chongqing Overseas Chinese Entrepreneurship and Innovation Support Program, and CAAI-Huawei MindSpore Open Fund. 2023-12-27T02:53:24Z 2023-12-27T02:53:24Z 2023 Journal Article Feng, L., Shu, S., Cao, Y., Tao, L., Wei, H., Xiang, T., An, B. & Niu, G. (2023). Multiple-instance learning from unlabeled bags with pairwise similarity. IEEE Transactions On Knowledge and Data Engineering, 35(11), 11599-11609. https://dx.doi.org/10.1109/TKDE.2022.3232141 1041-4347 https://hdl.handle.net/10356/172864 10.1109/TKDE.2022.3232141 2-s2.0-85147261041 11 35 11599 11609 en IEEE Transactions on Knowledge and Data Engineering © 2023 IEEE. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Multiple-Instance Learning
Empirical Risk Minimization
spellingShingle Engineering::Computer science and engineering
Multiple-Instance Learning
Empirical Risk Minimization
Feng, Lei
Shu, Senlin
Cao, Yuzhou
Tao, Lue
Wei, Hongxin
Xiang, Tao
An, Bo
Niu, Gang
Multiple-instance learning from unlabeled bags with pairwise similarity
description In multiple-instance learning (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL methods generally assume that training bags are fully labeled. However, the exact labels of training examples may not be accessible, due to security, confidentiality, and privacy concerns. Fortunately, it could be easier for us to access the pairwise similarity between two bags (indicating whether two bags share the same label or not) and unlabeled bags, as we do not need to know the underlying label of each bag. In this paper, we provide the first attempt to investigate MIL from only similar-dissimilar-unlabeled bags. To solve this new MIL problem, we first propose a strong baseline method that trains an instance-level classifier by employing an unlabeled-unlabeled learning strategy. Then, we also propose to train a bag-level classifier based on a convex formulation and theoretically derive a generalization error bound for this method. Comprehensive experimental results show that our instance-level classifier works well, while our bag-level classifier even has better performance.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Feng, Lei
Shu, Senlin
Cao, Yuzhou
Tao, Lue
Wei, Hongxin
Xiang, Tao
An, Bo
Niu, Gang
format Article
author Feng, Lei
Shu, Senlin
Cao, Yuzhou
Tao, Lue
Wei, Hongxin
Xiang, Tao
An, Bo
Niu, Gang
author_sort Feng, Lei
title Multiple-instance learning from unlabeled bags with pairwise similarity
title_short Multiple-instance learning from unlabeled bags with pairwise similarity
title_full Multiple-instance learning from unlabeled bags with pairwise similarity
title_fullStr Multiple-instance learning from unlabeled bags with pairwise similarity
title_full_unstemmed Multiple-instance learning from unlabeled bags with pairwise similarity
title_sort multiple-instance learning from unlabeled bags with pairwise similarity
publishDate 2023
url https://hdl.handle.net/10356/172864
_version_ 1787136791500816384