Locke’s Holiday: Belief Bias in Machine Reading

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Standard

Locke’s Holiday : Belief Bias in Machine Reading. / Søgaard, Anders.

Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2021. s. 8240–8245.

Publikation: Bidrag til bog/antologi/rapportKonferencebidrag i proceedingsForskningfagfællebedømt

Harvard

Søgaard, A 2021, Locke’s Holiday: Belief Bias in Machine Reading. i Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, s. 8240–8245, 2021 Conference on Empirical Methods in Natural Language Processing, 07/11/2021. https://doi.org/10.18653/v1/2021.emnlp-main.649

APA

Søgaard, A. (2021). Locke’s Holiday: Belief Bias in Machine Reading. I Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (s. 8240–8245). Association for Computational Linguistics. https://doi.org/10.18653/v1/2021.emnlp-main.649

Vancouver

Søgaard A. Locke’s Holiday: Belief Bias in Machine Reading. I Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics. 2021. s. 8240–8245 https://doi.org/10.18653/v1/2021.emnlp-main.649

Author

Søgaard, Anders. / Locke’s Holiday : Belief Bias in Machine Reading. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2021. s. 8240–8245

Bibtex

@inproceedings{18dbf9a48bfa4f6487af069ebd3453a6,
title = "Locke{\textquoteright}s Holiday: Belief Bias in Machine Reading",
abstract = "I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. For example, machine reading systems fail to answer What did Elizabeth want? correctly in the context of {\textquoteleft}My kingdom for a cough drop, cried Queen Elizabeth.{\textquoteright} Biased by co-occurrence statistics in the training data of pretrained language models, systems predict my kingdom, rather than a cough drop. I argue such biases are analogous to human belief biases and present a carefully designed challenge dataset for English machine reading, called Auto-Locke, to quantify such effects. Evaluations of machine reading systems on Auto-Locke show the pervasiveness of belief bias in machine reading.",
author = "Anders S{\o}gaard",
year = "2021",
doi = "10.18653/v1/2021.emnlp-main.649",
language = "English",
pages = "8240–8245",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
publisher = "Association for Computational Linguistics",
note = "2021 Conference on Empirical Methods in Natural Language Processing ; Conference date: 07-11-2021 Through 11-11-2021",

}

RIS

TY - GEN

T1 - Locke’s Holiday

T2 - 2021 Conference on Empirical Methods in Natural Language Processing

AU - Søgaard, Anders

PY - 2021

Y1 - 2021

N2 - I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. For example, machine reading systems fail to answer What did Elizabeth want? correctly in the context of ‘My kingdom for a cough drop, cried Queen Elizabeth.’ Biased by co-occurrence statistics in the training data of pretrained language models, systems predict my kingdom, rather than a cough drop. I argue such biases are analogous to human belief biases and present a carefully designed challenge dataset for English machine reading, called Auto-Locke, to quantify such effects. Evaluations of machine reading systems on Auto-Locke show the pervasiveness of belief bias in machine reading.

AB - I highlight a simple failure mode of state-of-the-art machine reading systems: when contexts do not align with commonly shared beliefs. For example, machine reading systems fail to answer What did Elizabeth want? correctly in the context of ‘My kingdom for a cough drop, cried Queen Elizabeth.’ Biased by co-occurrence statistics in the training data of pretrained language models, systems predict my kingdom, rather than a cough drop. I argue such biases are analogous to human belief biases and present a carefully designed challenge dataset for English machine reading, called Auto-Locke, to quantify such effects. Evaluations of machine reading systems on Auto-Locke show the pervasiveness of belief bias in machine reading.

U2 - 10.18653/v1/2021.emnlp-main.649

DO - 10.18653/v1/2021.emnlp-main.649

M3 - Article in proceedings

SP - 8240

EP - 8245

BT - Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing

PB - Association for Computational Linguistics

Y2 - 7 November 2021 through 11 November 2021

ER -

ID: 299822827