Historical Text Normalization with Delayed Rewards

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Standard

Historical Text Normalization with Delayed Rewards. / Flachs, Simon; Bollmann, Marcel; Søgaard, Anders.

Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 2019. s. 1614-1619.

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Harvard

Flachs, S, Bollmann, M & Søgaard, A 2019, Historical Text Normalization with Delayed Rewards. i Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, s. 1614-1619, 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italien, 01/07/2019. https://doi.org/10.18653/v1/P19-1157

APA

Flachs, S., Bollmann, M., & Søgaard, A. (2019). Historical Text Normalization with Delayed Rewards. I Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (s. 1614-1619). Association for Computational Linguistics. https://doi.org/10.18653/v1/P19-1157

Vancouver

Flachs S, Bollmann M, Søgaard A. Historical Text Normalization with Delayed Rewards. I Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics. 2019. s. 1614-1619 https://doi.org/10.18653/v1/P19-1157

Author

Flachs, Simon ; Bollmann, Marcel ; Søgaard, Anders. / Historical Text Normalization with Delayed Rewards. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 2019. s. 1614-1619

Bibtex

@inproceedings{f56b041d68144f9a9d7a246e533680a5,
title = "Historical Text Normalization with Delayed Rewards",
abstract = "Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words",
author = "Simon Flachs and Marcel Bollmann and Anders S{\o}gaard",
year = "2019",
doi = "10.18653/v1/P19-1157",
language = "English",
pages = "1614--1619",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
publisher = "Association for Computational Linguistics",
note = "57th Annual Meeting of the Association for Computational Linguistics ; Conference date: 01-07-2019 Through 01-07-2019",

}

RIS

TY - GEN

T1 - Historical Text Normalization with Delayed Rewards

AU - Flachs, Simon

AU - Bollmann, Marcel

AU - Søgaard, Anders

PY - 2019

Y1 - 2019

N2 - Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words

AB - Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words

U2 - 10.18653/v1/P19-1157

DO - 10.18653/v1/P19-1157

M3 - Article in proceedings

SP - 1614

EP - 1619

BT - Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics

PB - Association for Computational Linguistics

T2 - 57th Annual Meeting of the Association for Computational Linguistics

Y2 - 1 July 2019 through 1 July 2019

ER -

ID: 239617712