Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation

Publikation: Working paperPreprintForskning

Standard

Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation. / Musgrave, Kevin; Belongie, Serge; Lim, Ser Nam.

arXiv.org, 2023.

Publikation: Working paperPreprintForskning

Harvard

Musgrave, K, Belongie, S & Lim, SN 2023 'Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation' arXiv.org. <https://arxiv.org/abs/2208.07360>

APA

Musgrave, K., Belongie, S., & Lim, S. N. (2023). Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation. arXiv.org. https://arxiv.org/abs/2208.07360

Vancouver

Musgrave K, Belongie S, Lim SN. Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation. arXiv.org. 2023.

Author

Musgrave, Kevin ; Belongie, Serge ; Lim, Ser Nam. / Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation. arXiv.org, 2023.

Bibtex

@techreport{0dc0cfdeca484c96ac05123f56905c23,
title = "Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation",
abstract = "Changes to hyperparameters can have a dramatic effect on model accuracy. Thus, the tuning of hyperparameters plays an important role in optimizing machine-learning models. An integral part of the hyperparameter-tuning process is the evaluation of model checkpoints, which is done through the use of {"}validators{"}. In a supervised setting, these validators evaluate checkpoints by computing accuracy on a validation set that has labels. In contrast, in an unsupervised setting, the validation set has no such labels. Without any labels, it is impossible to compute accuracy, so validators must estimate accuracy instead. But what is the best approach to estimating accuracy? In this paper, we consider this question in the context of unsupervised domain adaptation (UDA). Specifically, we propose three new validators, and we compare and rank them against five other existing validators, on a large dataset of 1,000,000 checkpoints. Extensive experimental results show that two of our proposed validators achieve state-of-the-art performance in various settings. Finally, we find that in many cases, the state-of-the-art is obtained by a simple baseline method. To the best of our knowledge, this is the largest empirical study of UDA validators to date. Code is available at this https URL.",
author = "Kevin Musgrave and Serge Belongie and Lim, {Ser Nam}",
year = "2023",
language = "English",
publisher = "arXiv.org",
type = "WorkingPaper",
institution = "arXiv.org",

}

RIS

TY - UNPB

T1 - Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation

AU - Musgrave, Kevin

AU - Belongie, Serge

AU - Lim, Ser Nam

PY - 2023

Y1 - 2023

N2 - Changes to hyperparameters can have a dramatic effect on model accuracy. Thus, the tuning of hyperparameters plays an important role in optimizing machine-learning models. An integral part of the hyperparameter-tuning process is the evaluation of model checkpoints, which is done through the use of "validators". In a supervised setting, these validators evaluate checkpoints by computing accuracy on a validation set that has labels. In contrast, in an unsupervised setting, the validation set has no such labels. Without any labels, it is impossible to compute accuracy, so validators must estimate accuracy instead. But what is the best approach to estimating accuracy? In this paper, we consider this question in the context of unsupervised domain adaptation (UDA). Specifically, we propose three new validators, and we compare and rank them against five other existing validators, on a large dataset of 1,000,000 checkpoints. Extensive experimental results show that two of our proposed validators achieve state-of-the-art performance in various settings. Finally, we find that in many cases, the state-of-the-art is obtained by a simple baseline method. To the best of our knowledge, this is the largest empirical study of UDA validators to date. Code is available at this https URL.

AB - Changes to hyperparameters can have a dramatic effect on model accuracy. Thus, the tuning of hyperparameters plays an important role in optimizing machine-learning models. An integral part of the hyperparameter-tuning process is the evaluation of model checkpoints, which is done through the use of "validators". In a supervised setting, these validators evaluate checkpoints by computing accuracy on a validation set that has labels. In contrast, in an unsupervised setting, the validation set has no such labels. Without any labels, it is impossible to compute accuracy, so validators must estimate accuracy instead. But what is the best approach to estimating accuracy? In this paper, we consider this question in the context of unsupervised domain adaptation (UDA). Specifically, we propose three new validators, and we compare and rank them against five other existing validators, on a large dataset of 1,000,000 checkpoints. Extensive experimental results show that two of our proposed validators achieve state-of-the-art performance in various settings. Finally, we find that in many cases, the state-of-the-art is obtained by a simple baseline method. To the best of our knowledge, this is the largest empirical study of UDA validators to date. Code is available at this https URL.

M3 - Preprint

BT - Three New Validators and a Large-Scale Benchmark Ranking for Unsupervised Domain Adaptation

PB - arXiv.org

ER -

ID: 384616250