Robust Active Label Correction

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Standard

Robust Active Label Correction. / Kremer, Jan; Sha, Fei; Igel, Christian.

Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics. Bind 84 PMLR, 2018. s. 308-316 (Proceedings of Machine Learning Research). (Proceedings of Machine Learning Research, Bind 84).

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Harvard

Kremer, J, Sha, F & Igel, C 2018, Robust Active Label Correction. i Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics. bind 84, PMLR, Proceedings of Machine Learning Research, Proceedings of Machine Learning Research, bind 84, s. 308-316, 21st International Conference on Artificial Intelligence and Statistics, Playa Blanca, Lanzarote, Canary Islands, Spanien, 09/04/2018.

APA

Kremer, J., Sha, F., & Igel, C. (2018). Robust Active Label Correction. I Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics (Bind 84, s. 308-316). PMLR. Proceedings of Machine Learning Research, Proceedings of Machine Learning Research, Bind. 84

Vancouver

Kremer J, Sha F, Igel C. Robust Active Label Correction. I Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics. Bind 84. PMLR. 2018. s. 308-316. (Proceedings of Machine Learning Research). (Proceedings of Machine Learning Research, Bind 84).

Author

Kremer, Jan ; Sha, Fei ; Igel, Christian. / Robust Active Label Correction. Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics. Bind 84 PMLR, 2018. s. 308-316 (Proceedings of Machine Learning Research). (Proceedings of Machine Learning Research, Bind 84).

Bibtex

@inproceedings{ddbb5305e84844b1bfc751cf9941897a,
title = "Robust Active Label Correction",
abstract = "Active label correction addresses the problem of learning from input data for which noisy labels are available (e.g., from imprecise measurements or crowd-sourcing) and each true label can be obtained at a significant cost (e.g., through additional measurements or human experts). To minimize these costs, we are interested in identifying training patterns for which knowing the true labels maximally improves the learning performance. We approximate the true label noise by a model that learns the aspects of the noise that are class-conditional (i.e., independent of the input given the observed label). To select labels for correction, we adopt the active learning strategy of maximizing the expected model change. We consider the change in regularized empirical risk functionals that use different pointwise loss functions for patterns with noisy and true labels, respectively. Different loss functions for the noisy data lead to different active label correction algorithms. If loss functions consider the label noise rates, these rates are estimated during learning, where importance weighting compensates for the sampling bias. We show empirically that viewing the true label as a latent variable and computing the maximum likelihood estimate of the model parameters performs well across all considered problems. A maximum a posteriori estimate of the model parameters was beneficial in most test cases. An image classification experiment using convolutional neural networks demonstrates that the class-conditional noise model, which can be learned efficiently, can guide re-labeling in real-world applications.",
author = "Jan Kremer and Fei Sha and Christian Igel",
year = "2018",
language = "English",
volume = "84",
series = "Proceedings of Machine Learning Research",
publisher = "PMLR",
pages = "308--316",
booktitle = "Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics",
note = "21st International Conference on Artificial Intelligence and Statistics ; Conference date: 09-04-2018 Through 11-04-2018",

}

RIS

TY - GEN

T1 - Robust Active Label Correction

AU - Kremer, Jan

AU - Sha, Fei

AU - Igel, Christian

PY - 2018

Y1 - 2018

N2 - Active label correction addresses the problem of learning from input data for which noisy labels are available (e.g., from imprecise measurements or crowd-sourcing) and each true label can be obtained at a significant cost (e.g., through additional measurements or human experts). To minimize these costs, we are interested in identifying training patterns for which knowing the true labels maximally improves the learning performance. We approximate the true label noise by a model that learns the aspects of the noise that are class-conditional (i.e., independent of the input given the observed label). To select labels for correction, we adopt the active learning strategy of maximizing the expected model change. We consider the change in regularized empirical risk functionals that use different pointwise loss functions for patterns with noisy and true labels, respectively. Different loss functions for the noisy data lead to different active label correction algorithms. If loss functions consider the label noise rates, these rates are estimated during learning, where importance weighting compensates for the sampling bias. We show empirically that viewing the true label as a latent variable and computing the maximum likelihood estimate of the model parameters performs well across all considered problems. A maximum a posteriori estimate of the model parameters was beneficial in most test cases. An image classification experiment using convolutional neural networks demonstrates that the class-conditional noise model, which can be learned efficiently, can guide re-labeling in real-world applications.

AB - Active label correction addresses the problem of learning from input data for which noisy labels are available (e.g., from imprecise measurements or crowd-sourcing) and each true label can be obtained at a significant cost (e.g., through additional measurements or human experts). To minimize these costs, we are interested in identifying training patterns for which knowing the true labels maximally improves the learning performance. We approximate the true label noise by a model that learns the aspects of the noise that are class-conditional (i.e., independent of the input given the observed label). To select labels for correction, we adopt the active learning strategy of maximizing the expected model change. We consider the change in regularized empirical risk functionals that use different pointwise loss functions for patterns with noisy and true labels, respectively. Different loss functions for the noisy data lead to different active label correction algorithms. If loss functions consider the label noise rates, these rates are estimated during learning, where importance weighting compensates for the sampling bias. We show empirically that viewing the true label as a latent variable and computing the maximum likelihood estimate of the model parameters performs well across all considered problems. A maximum a posteriori estimate of the model parameters was beneficial in most test cases. An image classification experiment using convolutional neural networks demonstrates that the class-conditional noise model, which can be learned efficiently, can guide re-labeling in real-world applications.

M3 - Article in proceedings

VL - 84

T3 - Proceedings of Machine Learning Research

SP - 308

EP - 316

BT - Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics

PB - PMLR

T2 - 21st International Conference on Artificial Intelligence and Statistics

Y2 - 9 April 2018 through 11 April 2018

ER -

ID: 216873000