Why is unsupervised alignment of English embeddings from different algorithms so hard?

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Standard

Why is unsupervised alignment of English embeddings from different algorithms so hard? / Hartmann, Mareike; Kementchedjhieva, Yova Radoslavova; Søgaard, Anders.

Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2018. s. 582–586.

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Harvard

Hartmann, M, Kementchedjhieva, YR & Søgaard, A 2018, Why is unsupervised alignment of English embeddings from different algorithms so hard? i Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, s. 582–586, 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgien, 31/10/2018.

APA

Hartmann, M., Kementchedjhieva, Y. R., & Søgaard, A. (2018). Why is unsupervised alignment of English embeddings from different algorithms so hard? I Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (s. 582–586). Association for Computational Linguistics.

Vancouver

Hartmann M, Kementchedjhieva YR, Søgaard A. Why is unsupervised alignment of English embeddings from different algorithms so hard? I Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics. 2018. s. 582–586

Author

Hartmann, Mareike ; Kementchedjhieva, Yova Radoslavova ; Søgaard, Anders. / Why is unsupervised alignment of English embeddings from different algorithms so hard?. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2018. s. 582–586

Bibtex

@inproceedings{b1f6822e7c0342f78a081c20a676ecf3,
title = "Why is unsupervised alignment of English embeddings from different algorithms so hard?",
abstract = "This paper presents a challenge to the community:Generative adversarial networks (GANs)can perfectly align independent English wordembeddings induced using the same algorithm,based on distributional informationalone; but fails to do so, for two different embeddingsalgorithms. Why is that? We believeunderstanding why, is key to understand bothmodern word embedding algorithms and thelimitations and instability dynamics of GANs.This paper shows that (a) in all these cases,where alignment fails, there exists a lineartransform between the two embeddings (so algorithmbiases do not lead to non-linear differences),and (b) similar effects can not easilybe obtained by varying hyper-parameters. Oneplausible suggestion based on our initial experimentsis that the differences in the inductivebiases of the embedding algorithms lead toan optimization landscape that is riddled withlocal optima, leading to a very small basin ofconvergence, but we present this more as achallenge paper than a technical contribution.",
author = "Mareike Hartmann and Kementchedjhieva, {Yova Radoslavova} and Anders S{\o}gaard",
year = "2018",
language = "English",
pages = "582–586",
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
publisher = "Association for Computational Linguistics",
note = "null ; Conference date: 31-10-2018 Through 04-11-2018",

}

RIS

TY - GEN

T1 - Why is unsupervised alignment of English embeddings from different algorithms so hard?

AU - Hartmann, Mareike

AU - Kementchedjhieva, Yova Radoslavova

AU - Søgaard, Anders

PY - 2018

Y1 - 2018

N2 - This paper presents a challenge to the community:Generative adversarial networks (GANs)can perfectly align independent English wordembeddings induced using the same algorithm,based on distributional informationalone; but fails to do so, for two different embeddingsalgorithms. Why is that? We believeunderstanding why, is key to understand bothmodern word embedding algorithms and thelimitations and instability dynamics of GANs.This paper shows that (a) in all these cases,where alignment fails, there exists a lineartransform between the two embeddings (so algorithmbiases do not lead to non-linear differences),and (b) similar effects can not easilybe obtained by varying hyper-parameters. Oneplausible suggestion based on our initial experimentsis that the differences in the inductivebiases of the embedding algorithms lead toan optimization landscape that is riddled withlocal optima, leading to a very small basin ofconvergence, but we present this more as achallenge paper than a technical contribution.

AB - This paper presents a challenge to the community:Generative adversarial networks (GANs)can perfectly align independent English wordembeddings induced using the same algorithm,based on distributional informationalone; but fails to do so, for two different embeddingsalgorithms. Why is that? We believeunderstanding why, is key to understand bothmodern word embedding algorithms and thelimitations and instability dynamics of GANs.This paper shows that (a) in all these cases,where alignment fails, there exists a lineartransform between the two embeddings (so algorithmbiases do not lead to non-linear differences),and (b) similar effects can not easilybe obtained by varying hyper-parameters. Oneplausible suggestion based on our initial experimentsis that the differences in the inductivebiases of the embedding algorithms lead toan optimization landscape that is riddled withlocal optima, leading to a very small basin ofconvergence, but we present this more as achallenge paper than a technical contribution.

M3 - Article in proceedings

SP - 582

EP - 586

BT - Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing

PB - Association for Computational Linguistics

Y2 - 31 October 2018 through 4 November 2018

ER -

ID: 214760789