Historical Text Normalization with Delayed Rewards

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Dokumenter

Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words.
OriginalsprogEngelsk
TitelProceedings of the 57th Annual Meeting of the Association for Computational Linguistics
ForlagAssociation for Computational Linguistics
Publikationsdato2019
Sider1614-1619
DOI
StatusUdgivet - 2019
Begivenhed57th Annual Meeting of the Association for Computational Linguistics - Florence, Italien
Varighed: 1 jul. 20191 jul. 2019

Konference

Konference57th Annual Meeting of the Association for Computational Linguistics
LandItalien
ByFlorence,
Periode01/07/201901/07/2019

Antal downloads er baseret på statistik fra Google Scholar og www.ku.dk


Ingen data tilgængelig

ID: 239617712