TY - JOUR
T1 - Crowdsourcing Thousands of Specialized Labels
T2 - A Bayesian Active Training Approach
AU - Servajean, Maximilien
AU - Joly, Alexis
AU - Shasha, Dennis
AU - Champ, Julien
AU - Pacitti, Esther
N1 - Funding Information:
Manuscript received July 18, 2016; revised November 24, 2016; accepted January 2, 2017. Date of publication January 16, 2017; date of current version May 13, 2017. The work of D. Shasha was supported by the INRIA International Chair. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Dong Xu.
Publisher Copyright:
© 2017 IEEE.
PY - 2017/6
Y1 - 2017/6
N2 - Large-scale annotated corpora have yielded impressive performance improvements in computer vision and multimedia content analysis. However, such datasets depend on an enormous amount of human labeling effort. When the labels correspond to well-known concepts, it is straightforward to train the annotators by giving a few examples with known answers. It is also straightforward to judge the quality of their labels. Neither is true when there are thousands of complex domain-specific labels. Training on all labels is infeasible and the quality of an annotator's judgements may be vastly different for some subsets of labels than for others. This paper proposes a set of data-driven algorithms to 1) train image annotators on how to disambiguate among automatically generated candidate labels, 2) evaluate the quality of annotators' label suggestions, and 3) weigh predictions. The algorithms adapt to the skills of each annotator both in the questions asked and the weights given to their answers. The underlying judgements are Bayesian, based on adaptive priors. We measure the benefits of these algorithms on a live user experiment related to image-based plant identification involving around 1000 people. The proposed methods are shown to enable huge gains in annotation accuracy. A standard user can correctly label around 2% of our data. This goes up to 80% with machine learning assisted training and assignment and up to almost 90% when doing a weighted combination of several annotators' labels.
AB - Large-scale annotated corpora have yielded impressive performance improvements in computer vision and multimedia content analysis. However, such datasets depend on an enormous amount of human labeling effort. When the labels correspond to well-known concepts, it is straightforward to train the annotators by giving a few examples with known answers. It is also straightforward to judge the quality of their labels. Neither is true when there are thousands of complex domain-specific labels. Training on all labels is infeasible and the quality of an annotator's judgements may be vastly different for some subsets of labels than for others. This paper proposes a set of data-driven algorithms to 1) train image annotators on how to disambiguate among automatically generated candidate labels, 2) evaluate the quality of annotators' label suggestions, and 3) weigh predictions. The algorithms adapt to the skills of each annotator both in the questions asked and the weights given to their answers. The underlying judgements are Bayesian, based on adaptive priors. We measure the benefits of these algorithms on a live user experiment related to image-based plant identification involving around 1000 people. The proposed methods are shown to enable huge gains in annotation accuracy. A standard user can correctly label around 2% of our data. This goes up to 80% with machine learning assisted training and assignment and up to almost 90% when doing a weighted combination of several annotators' labels.
KW - Bayes methods
KW - Crowdsourcing
KW - Taylor series
KW - parameter estimation
UR - http://www.scopus.com/inward/record.url?scp=85028344775&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85028344775&partnerID=8YFLogxK
U2 - 10.1109/TMM.2017.2653763
DO - 10.1109/TMM.2017.2653763
M3 - Article
AN - SCOPUS:85028344775
SN - 1520-9210
VL - 19
SP - 1376
EP - 1391
JO - IEEE Transactions on Multimedia
JF - IEEE Transactions on Multimedia
IS - 6
M1 - 7819540
ER -