Clustering Monolingual Vocabularies to Improve Cross-Lingual Generalization

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Dokumenter

  • Fulltext

    Forlagets udgivne version, 828 KB, PDF-dokument

Multilingual language models exhibit better performance for some languages than for others (Singh et al., 2019), and many languages do not seem to benefit from multilingual sharing at all, presumably as a result of poor multilingual segmentation (Pyysal o et al., 2020). This work explores the idea of learning multilingual language models based on clustering of monolingual segments. We show significant improvements over standard multilingual segmentation and training across nine languages on a question answering task, both in a small model regime and for a model of the size of BERT-base.
OriginalsprogEngelsk
TitelProceedings of the 1st Workshop on Multilingual Representation Learning
ForlagAssociation for Computational Linguistics
Publikationsdato2021
Sider32–40
DOI
StatusUdgivet - 2021
Begivenhed1st Workshop on Multilingual Representation Learning - Online
Varighed: 11 nov. 202111 nov. 2021

Konference

Konference1st Workshop on Multilingual Representation Learning
ByOnline
Periode11/11/202111/11/2021

Antal downloads er baseret på statistik fra Google Scholar og www.ku.dk


Ingen data tilgængelig

ID: 300080332