Publication | Closed Access
Adaptive Task Assignment for Crowdsourced Classification
200
Citations
19
References
2013
Year
Unknown Venue
Crowdsourcing markets have gained popularity as a tool for inexpensively collecting data from diverse populations of workers. Classification tasks, in which workers provide labels (such as “offensive ” or “not offensive”) for instances (such as “websites”), are among the most common tasks posted, but due to human error and the prevalence of spam, the labels collected are often noisy. This problem is typically addressed by collecting labels for eachinstancefrommultipleworkersandcombining them in a clever way, but the question ofhowtochoosewhichtaskstoassigntoeach worker is often overlooked. We investigate the problem of task assignment and label inference for heterogeneous classification tasks. By applying online primal-dual techniques, we derive a provably near-optimal adaptive assignment algorithm. We show that adaptively assigning workers to tasks can lead to more accurate predictions at a lower cost when the available workers are diverse. 1.
| Year | Citations | |
|---|---|---|
2024 | 3.9K | |
2008 | 1.9K | |
1979 | 1.6K | |
2010 | 983 | |
2010 | 734 | |
2011 | 476 | |
2009 | 331 | |
2012 | 320 | |
2012 | 318 | |
2021 | 298 |
Page 1
Page 1