Improving Performance by Re-Rating in the Dynamic Estimation of Rater Reliability

Alexey Tarasov, Sarah Jane Delany, Brian MacNamee

Research output: Contribution to conferencePaperpeer-review

Abstract

Nowadays crowdsourcing is widely used in supervised machine learning to facilitate the collection of ratings for unlabelled training sets. In order to get good quality results it is worth rejecting results from noisy/unreliable raters, as soon as they are discovered. Many techniques for filtering unreliable raters rely on the presentation of training instances to the raters identified as most accurate to date. Early in the process, the true rater reliabilities are not known and unreliable raters may be used as a result. This paper explores improving the quality of ratings for train- ing instances by performing re-rating. The re-rating relies on the detection of such in- stances and the acquisition of additional ratings for them when the rating process is over. We compare different approaches to re-rating and compare the improvements in labeling accuracy and the labeling costs of these approaches.
Original languageEnglish
DOIs
Publication statusPublished - 2013
EventInternational Conference on Machine Learning - Atlanta, United States
Duration: 1 Jan 2013 → …

Conference

ConferenceInternational Conference on Machine Learning
Country/TerritoryUnited States
CityAtlanta
Period1/01/13 → …

Keywords

  • crowdsourcing
  • supervised machine learning
  • ratings
  • unreliable raters
  • training instances
  • re-rating
  • labeling accuracy
  • labeling costs

Fingerprint

Dive into the research topics of 'Improving Performance by Re-Rating in the Dynamic Estimation of Rater Reliability'. Together they form a unique fingerprint.

Cite this