Active crowdsourcing for annotation

Crowdsourcing has shown great potential in obtaining large-scale and cheap labels for different tasks. However, obtaining reliable labels is challenging due to several reasons, such as noisy annotators, limited budget and so on. The state-of-the-art approaches, either suffer in some noisy scenarios,...

Full description

Saved in:
Bibliographic Details
Main Authors: HAO, Shuji, MIAO, Chunyan, HOI, Steven C. H., ZHAO, Peilin
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2015
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/3173
https://ink.library.smu.edu.sg/context/sis_research/article/4174/viewcontent/Active_Crowdsourcing_for_Annotation_accepted.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-4174
record_format dspace
spelling sg-smu-ink.sis_research-41742020-03-31T05:50:05Z Active crowdsourcing for annotation HAO, Shuji MIAO, Chunyan HOI, Steven C. H. ZHAO, Peilin Crowdsourcing has shown great potential in obtaining large-scale and cheap labels for different tasks. However, obtaining reliable labels is challenging due to several reasons, such as noisy annotators, limited budget and so on. The state-of-the-art approaches, either suffer in some noisy scenarios, or rely on unlimited resources to acquire reliable labels. In this article, we adopt the learning with expert~(AKA worker in crowdsourcing) advice framework to robustly infer accurate labels by considering the reliability of each worker. However, in order to accurately predict the reliability of each worker, traditional learning with expert advice will consult with external oracles~(AKA domain experts) on the true label of each instance. To reduce the cost of consultation, we proposed two active learning approaches, margin-based and weighted difference of advices based. Meanwhile, to address the problem of limited annotation budget, we proposed a reliability-based assigning approach which actively decides who to annotate the next instance based on each worker's cumulative performance. The experimental results both on real and simulated datasets show that our algorithms can achieve robust and promising performance both in the normal and noisy scenarios with limited budget. 2015-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/3173 info:doi/10.1109/WI-IAT.2015.34 https://ink.library.smu.edu.sg/context/sis_research/article/4174/viewcontent/Active_Crowdsourcing_for_Annotation_accepted.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Active Learning Crowdsourcing Online Learning Databases and Information Systems Numerical Analysis and Scientific Computing
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Active Learning
Crowdsourcing
Online Learning
Databases and Information Systems
Numerical Analysis and Scientific Computing
spellingShingle Active Learning
Crowdsourcing
Online Learning
Databases and Information Systems
Numerical Analysis and Scientific Computing
HAO, Shuji
MIAO, Chunyan
HOI, Steven C. H.
ZHAO, Peilin
Active crowdsourcing for annotation
description Crowdsourcing has shown great potential in obtaining large-scale and cheap labels for different tasks. However, obtaining reliable labels is challenging due to several reasons, such as noisy annotators, limited budget and so on. The state-of-the-art approaches, either suffer in some noisy scenarios, or rely on unlimited resources to acquire reliable labels. In this article, we adopt the learning with expert~(AKA worker in crowdsourcing) advice framework to robustly infer accurate labels by considering the reliability of each worker. However, in order to accurately predict the reliability of each worker, traditional learning with expert advice will consult with external oracles~(AKA domain experts) on the true label of each instance. To reduce the cost of consultation, we proposed two active learning approaches, margin-based and weighted difference of advices based. Meanwhile, to address the problem of limited annotation budget, we proposed a reliability-based assigning approach which actively decides who to annotate the next instance based on each worker's cumulative performance. The experimental results both on real and simulated datasets show that our algorithms can achieve robust and promising performance both in the normal and noisy scenarios with limited budget.
format text
author HAO, Shuji
MIAO, Chunyan
HOI, Steven C. H.
ZHAO, Peilin
author_facet HAO, Shuji
MIAO, Chunyan
HOI, Steven C. H.
ZHAO, Peilin
author_sort HAO, Shuji
title Active crowdsourcing for annotation
title_short Active crowdsourcing for annotation
title_full Active crowdsourcing for annotation
title_fullStr Active crowdsourcing for annotation
title_full_unstemmed Active crowdsourcing for annotation
title_sort active crowdsourcing for annotation
publisher Institutional Knowledge at Singapore Management University
publishDate 2015
url https://ink.library.smu.edu.sg/sis_research/3173
https://ink.library.smu.edu.sg/context/sis_research/article/4174/viewcontent/Active_Crowdsourcing_for_Annotation_accepted.pdf
_version_ 1770572969606643712