Abstract
Nowadays crowdsourcing is widely used in supervised machine learning to facilitate the collection of ratings for unlabelled training sets. In order to get good quality results it is worth rejecting results from noisy/unreliable raters, as soon as they are discovered. Many techniques for filtering unreliable raters rely on the presentation of training instances to the raters identified as most accurate to date. Early in the process, the true rater reliabilities are not known and unreliable raters may be used as a result. This paper explores improving the quality of ratings for train- ing instances by performing re-rating. The re-rating relies on the detection of such in- stances and the acquisition of additional ratings for them when the rating process is over. We compare different approaches to re-rating and compare the improvements in labeling accuracy and the labeling costs of these approaches.
| Original language | English |
|---|---|
| DOIs | |
| Publication status | Published - 2013 |
| Event | International Conference on Machine Learning - Atlanta, United States Duration: 1 Jan 2013 → … |
Conference
| Conference | International Conference on Machine Learning |
|---|---|
| Country/Territory | United States |
| City | Atlanta |
| Period | 1/01/13 → … |
Keywords
- crowdsourcing
- supervised machine learning
- ratings
- unreliable raters
- training instances
- re-rating
- labeling accuracy
- labeling costs