Learning from noisy large-scale datasets with minimal supervision

Publikation: Bidrag til tidsskriftKonferenceartikelForskningfagfællebedømt

Standard

Learning from noisy large-scale datasets with minimal supervision. / Veit, Andreas; Alldrin, Neil; Chechik, Gal; Krasin, Ivan; Gupta, Abhinav; Belongie, Serge.

I: Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, 06.11.2017, s. 6575-6583.

Publikation: Bidrag til tidsskriftKonferenceartikelForskningfagfællebedømt

Harvard

Veit, A, Alldrin, N, Chechik, G, Krasin, I, Gupta, A & Belongie, S 2017, 'Learning from noisy large-scale datasets with minimal supervision', Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, s. 6575-6583. https://doi.org/10.1109/CVPR.2017.696

APA

Veit, A., Alldrin, N., Chechik, G., Krasin, I., Gupta, A., & Belongie, S. (2017). Learning from noisy large-scale datasets with minimal supervision. Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, 6575-6583. https://doi.org/10.1109/CVPR.2017.696

Vancouver

Veit A, Alldrin N, Chechik G, Krasin I, Gupta A, Belongie S. Learning from noisy large-scale datasets with minimal supervision. Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017. 2017 nov. 6;6575-6583. https://doi.org/10.1109/CVPR.2017.696

Author

Veit, Andreas ; Alldrin, Neil ; Chechik, Gal ; Krasin, Ivan ; Gupta, Abhinav ; Belongie, Serge. / Learning from noisy large-scale datasets with minimal supervision. I: Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017. 2017 ; s. 6575-6583.

Bibtex

@inproceedings{2b9d91fe87c0441bb6f2b5bcac75a098,
title = "Learning from noisy large-scale datasets with minimal supervision",
abstract = "We present an approach to effectively use millions of images with noisy annotations in conjunction with a small subset of cleanly-annotated images to learn powerful image representations. One common approach to combine clean and noisy data is to first pre-train a network using the large noisy dataset and then fine-tune with the clean dataset. We show this approach does not fully leverage the information contained in the clean set. Thus, we demonstrate how to use the clean annotations to reduce the noise in the large dataset before fine-tuning the network using both the clean set and the full set with reduced noise. The approach comprises a multi-task network that jointly learns to clean noisy annotations and to accurately classify images. We evaluate our approach on the recently released Open Images dataset, containing ∼9 million images, multiple annotations per image and over 6000 unique classes. For the small clean set of annotations we use a quarter of the validation set with ∼40k images. Our results demonstrate that the proposed approach clearly outperforms direct fine-tuning across all major categories of classes in the Open Image dataset. Further, our approach is particularly effective for a large number of classes with wide range of noise in annotations (20-80% false positive annotations).",
author = "Andreas Veit and Neil Alldrin and Gal Chechik and Ivan Krasin and Abhinav Gupta and Serge Belongie",
note = "Funding Information: We would like to thank Ramakrishna Vedantam for insightful feedback as well as the AOL Connected Experiences Laboratory at Cornell Tech. This work was funded in part by a Google Focused Research Award. Publisher Copyright: {\textcopyright} 2017 IEEE.; 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017 ; Conference date: 21-07-2017 Through 26-07-2017",
year = "2017",
month = nov,
day = "6",
doi = "10.1109/CVPR.2017.696",
language = "English",
pages = "6575--6583",
journal = "Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017",

}

RIS

TY - GEN

T1 - Learning from noisy large-scale datasets with minimal supervision

AU - Veit, Andreas

AU - Alldrin, Neil

AU - Chechik, Gal

AU - Krasin, Ivan

AU - Gupta, Abhinav

AU - Belongie, Serge

N1 - Funding Information: We would like to thank Ramakrishna Vedantam for insightful feedback as well as the AOL Connected Experiences Laboratory at Cornell Tech. This work was funded in part by a Google Focused Research Award. Publisher Copyright: © 2017 IEEE.

PY - 2017/11/6

Y1 - 2017/11/6

N2 - We present an approach to effectively use millions of images with noisy annotations in conjunction with a small subset of cleanly-annotated images to learn powerful image representations. One common approach to combine clean and noisy data is to first pre-train a network using the large noisy dataset and then fine-tune with the clean dataset. We show this approach does not fully leverage the information contained in the clean set. Thus, we demonstrate how to use the clean annotations to reduce the noise in the large dataset before fine-tuning the network using both the clean set and the full set with reduced noise. The approach comprises a multi-task network that jointly learns to clean noisy annotations and to accurately classify images. We evaluate our approach on the recently released Open Images dataset, containing ∼9 million images, multiple annotations per image and over 6000 unique classes. For the small clean set of annotations we use a quarter of the validation set with ∼40k images. Our results demonstrate that the proposed approach clearly outperforms direct fine-tuning across all major categories of classes in the Open Image dataset. Further, our approach is particularly effective for a large number of classes with wide range of noise in annotations (20-80% false positive annotations).

AB - We present an approach to effectively use millions of images with noisy annotations in conjunction with a small subset of cleanly-annotated images to learn powerful image representations. One common approach to combine clean and noisy data is to first pre-train a network using the large noisy dataset and then fine-tune with the clean dataset. We show this approach does not fully leverage the information contained in the clean set. Thus, we demonstrate how to use the clean annotations to reduce the noise in the large dataset before fine-tuning the network using both the clean set and the full set with reduced noise. The approach comprises a multi-task network that jointly learns to clean noisy annotations and to accurately classify images. We evaluate our approach on the recently released Open Images dataset, containing ∼9 million images, multiple annotations per image and over 6000 unique classes. For the small clean set of annotations we use a quarter of the validation set with ∼40k images. Our results demonstrate that the proposed approach clearly outperforms direct fine-tuning across all major categories of classes in the Open Image dataset. Further, our approach is particularly effective for a large number of classes with wide range of noise in annotations (20-80% false positive annotations).

UR - http://www.scopus.com/inward/record.url?scp=85041911514&partnerID=8YFLogxK

U2 - 10.1109/CVPR.2017.696

DO - 10.1109/CVPR.2017.696

M3 - Conference article

AN - SCOPUS:85041911514

SP - 6575

EP - 6583

JO - Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017

JF - Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017

T2 - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017

Y2 - 21 July 2017 through 26 July 2017

ER -

ID: 301826772