Systematic Labeling Bias: De-biasing where Everyone is Wrong

Cabrera, GF; Miller, CJ; Schneider, J

Abstract

Many real world classification problems use ground truth labels created by human annotators. However, observed data is never perfect, and even labels assigned by perfect annotators can be systematically biased due to poor quality of the data they are labeling. This bias is not created by the annotators from measurement error, but is intrinsic to the observational data. We present a method for de-biasing labels which simultaneously learns a classification model, estimates the intrinsic biases in the ground truth, and provides new de-biased labels. We test our algorithm on simulated and real data and show that it is superior to standard de-noising algorithms, like instance weighted logistic regression.

Más información

Título según WOS: Systematic Labeling Bias: De-biasing where Everyone is Wrong
Título según SCOPUS: Systematic labeling bias: De-biasing where everyone is wrong
Título de la Revista: 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR)
Editorial: IEEE
Fecha de publicación: 2014
Página de inicio: 4417
Página final: 4422
Idioma: English
DOI:

10.1109/ICPR.2014.756

Notas: ISI, SCOPUS