Multiple-instance learning from unlabeled bags with pairwise similarity
In multiple-instance learning (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL methods generally assume that training bags are fully labeled. Ho...
Saved in:
Main Authors: | , , , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/172864 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-172864 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1728642023-12-27T02:53:24Z Multiple-instance learning from unlabeled bags with pairwise similarity Feng, Lei Shu, Senlin Cao, Yuzhou Tao, Lue Wei, Hongxin Xiang, Tao An, Bo Niu, Gang School of Computer Science and Engineering Engineering::Computer science and engineering Multiple-Instance Learning Empirical Risk Minimization In multiple-instance learning (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL methods generally assume that training bags are fully labeled. However, the exact labels of training examples may not be accessible, due to security, confidentiality, and privacy concerns. Fortunately, it could be easier for us to access the pairwise similarity between two bags (indicating whether two bags share the same label or not) and unlabeled bags, as we do not need to know the underlying label of each bag. In this paper, we provide the first attempt to investigate MIL from only similar-dissimilar-unlabeled bags. To solve this new MIL problem, we first propose a strong baseline method that trains an instance-level classifier by employing an unlabeled-unlabeled learning strategy. Then, we also propose to train a bag-level classifier based on a convex formulation and theoretically derive a generalization error bound for this method. Comprehensive experimental results show that our instance-level classifier works well, while our bag-level classifier even has better performance. National Research Foundation (NRF) This work was supported in part by the National Key R&D Program of China under Grant 2022YFB3103500. The work of Bo An is supported by the National Research Foundation, Singapore under its Industry Alignment Fund – Pre-positioning (IAF-PP) Funding Initiative. The work of Lei Feng is supported by the National Natural Science Foundation of China under Grant 62106028, in part by the Chongqing Overseas Chinese Entrepreneurship and Innovation Support Program, and CAAI-Huawei MindSpore Open Fund. 2023-12-27T02:53:24Z 2023-12-27T02:53:24Z 2023 Journal Article Feng, L., Shu, S., Cao, Y., Tao, L., Wei, H., Xiang, T., An, B. & Niu, G. (2023). Multiple-instance learning from unlabeled bags with pairwise similarity. IEEE Transactions On Knowledge and Data Engineering, 35(11), 11599-11609. https://dx.doi.org/10.1109/TKDE.2022.3232141 1041-4347 https://hdl.handle.net/10356/172864 10.1109/TKDE.2022.3232141 2-s2.0-85147261041 11 35 11599 11609 en IEEE Transactions on Knowledge and Data Engineering © 2023 IEEE. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Multiple-Instance Learning Empirical Risk Minimization |
spellingShingle |
Engineering::Computer science and engineering Multiple-Instance Learning Empirical Risk Minimization Feng, Lei Shu, Senlin Cao, Yuzhou Tao, Lue Wei, Hongxin Xiang, Tao An, Bo Niu, Gang Multiple-instance learning from unlabeled bags with pairwise similarity |
description |
In multiple-instance learning (MIL), each training example is represented by a bag of instances. A training bag is either negative if it contains no positive instances or positive if it has at least one positive instance. Previous MIL methods generally assume that training bags are fully labeled. However, the exact labels of training examples may not be accessible, due to security, confidentiality, and privacy concerns. Fortunately, it could be easier for us to access the pairwise similarity between two bags (indicating whether two bags share the same label or not) and unlabeled bags, as we do not need to know the underlying label of each bag. In this paper, we provide the first attempt to investigate MIL from only similar-dissimilar-unlabeled bags. To solve this new MIL problem, we first propose a strong baseline method that trains an instance-level classifier by employing an unlabeled-unlabeled learning strategy. Then, we also propose to train a bag-level classifier based on a convex formulation and theoretically derive a generalization error bound for this method. Comprehensive experimental results show that our instance-level classifier works well, while our bag-level classifier even has better performance. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Feng, Lei Shu, Senlin Cao, Yuzhou Tao, Lue Wei, Hongxin Xiang, Tao An, Bo Niu, Gang |
format |
Article |
author |
Feng, Lei Shu, Senlin Cao, Yuzhou Tao, Lue Wei, Hongxin Xiang, Tao An, Bo Niu, Gang |
author_sort |
Feng, Lei |
title |
Multiple-instance learning from unlabeled bags with pairwise similarity |
title_short |
Multiple-instance learning from unlabeled bags with pairwise similarity |
title_full |
Multiple-instance learning from unlabeled bags with pairwise similarity |
title_fullStr |
Multiple-instance learning from unlabeled bags with pairwise similarity |
title_full_unstemmed |
Multiple-instance learning from unlabeled bags with pairwise similarity |
title_sort |
multiple-instance learning from unlabeled bags with pairwise similarity |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/172864 |
_version_ |
1787136791500816384 |