Systematic Labeling Bias: De-biasing where Everyone is Wrong
Abstract
Many real world classification problems use ground truth labels created by human annotators. However, observed data is never perfect, and even labels assigned by perfect annotators can be systematically biased due to poor quality of the data they are labeling. This bias is not created by the annotators from measurement error, but is intrinsic to the observational data. We present a method for de-biasing labels which simultaneously learns a classification model, estimates the intrinsic biases in the ground truth, and provides new de-biased labels. We test our algorithm on simulated and real data and show that it is superior to standard de-noising algorithms, like instance weighted logistic regression.
Más información
Título según WOS: | Systematic Labeling Bias: De-biasing where Everyone is Wrong |
Título según SCOPUS: | Systematic labeling bias: De-biasing where everyone is wrong |
Título de la Revista: | 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) |
Editorial: | IEEE |
Fecha de publicación: | 2014 |
Página de inicio: | 4417 |
Página final: | 4422 |
Idioma: | English |
DOI: |
10.1109/ICPR.2014.756 |
Notas: | ISI, SCOPUS |