A Latent-Variable Model for Intrinsic Probing

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Standard

A Latent-Variable Model for Intrinsic Probing. / Stańczak, Karolina; Torroba Hennigen, Lucas; Williams, Adina; Cotterell, Ryan; Augenstein, Isabelle.

I: AAAI Conference on Artificial Intelligence, Bind 37, Nr. 11, 2023, s. 13591-13599.

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Harvard

Stańczak, K, Torroba Hennigen, L, Williams, A, Cotterell, R & Augenstein, I 2023, 'A Latent-Variable Model for Intrinsic Probing', AAAI Conference on Artificial Intelligence, bind 37, nr. 11, s. 13591-13599. https://doi.org/10.1609/aaai.v37i11.26593

APA

Stańczak, K., Torroba Hennigen, L., Williams, A., Cotterell, R., & Augenstein, I. (2023). A Latent-Variable Model for Intrinsic Probing. AAAI Conference on Artificial Intelligence, 37(11), 13591-13599. https://doi.org/10.1609/aaai.v37i11.26593

Vancouver

Stańczak K, Torroba Hennigen L, Williams A, Cotterell R, Augenstein I. A Latent-Variable Model for Intrinsic Probing. AAAI Conference on Artificial Intelligence. 2023;37(11):13591-13599. https://doi.org/10.1609/aaai.v37i11.26593

Author

Stańczak, Karolina ; Torroba Hennigen, Lucas ; Williams, Adina ; Cotterell, Ryan ; Augenstein, Isabelle. / A Latent-Variable Model for Intrinsic Probing. I: AAAI Conference on Artificial Intelligence. 2023 ; Bind 37, Nr. 11. s. 13591-13599.

Bibtex

@article{d6288d97e2574a7bbf0063557bb38f2a,
title = "A Latent-Variable Model for Intrinsic Probing",
abstract = "The success of pre-trained contextualized representations has prompted researchers to analyze them for the presence of linguistic information. Indeed, it is natural to assume that these pre-trained representations do encode some level of linguistic knowledge as they have brought about large empirical improvements on a wide variety of NLP tasks, which suggests they are learning true linguistic generalization. In this work, we focus on intrinsic probing, an analysis technique where the goal is not only to identify whether a representation encodes a linguistic attribute but also to pinpoint where this attribute is encoded. We propose a novel latent-variable formulation for constructing intrinsic probes and derive a tractable variational approximation to the log-likelihood. Our results show that our model is versatile and yields tighter mutual information estimates than two intrinsic probes previously proposed in the literature. Finally, we find empirical evidence that pre-trained representations develop a cross-lingually entangled notion of morphosyntax.",
author = "Karolina Sta{\'n}czak and {Torroba Hennigen}, Lucas and Adina Williams and Ryan Cotterell and Isabelle Augenstein",
year = "2023",
doi = "10.1609/aaai.v37i11.26593",
language = "English",
volume = "37",
pages = "13591--13599",
journal = "AAAI Conference on Artificial Intelligence",
issn = "2159-5399",
number = "11",

}

RIS

TY - JOUR

T1 - A Latent-Variable Model for Intrinsic Probing

AU - Stańczak, Karolina

AU - Torroba Hennigen, Lucas

AU - Williams, Adina

AU - Cotterell, Ryan

AU - Augenstein, Isabelle

PY - 2023

Y1 - 2023

N2 - The success of pre-trained contextualized representations has prompted researchers to analyze them for the presence of linguistic information. Indeed, it is natural to assume that these pre-trained representations do encode some level of linguistic knowledge as they have brought about large empirical improvements on a wide variety of NLP tasks, which suggests they are learning true linguistic generalization. In this work, we focus on intrinsic probing, an analysis technique where the goal is not only to identify whether a representation encodes a linguistic attribute but also to pinpoint where this attribute is encoded. We propose a novel latent-variable formulation for constructing intrinsic probes and derive a tractable variational approximation to the log-likelihood. Our results show that our model is versatile and yields tighter mutual information estimates than two intrinsic probes previously proposed in the literature. Finally, we find empirical evidence that pre-trained representations develop a cross-lingually entangled notion of morphosyntax.

AB - The success of pre-trained contextualized representations has prompted researchers to analyze them for the presence of linguistic information. Indeed, it is natural to assume that these pre-trained representations do encode some level of linguistic knowledge as they have brought about large empirical improvements on a wide variety of NLP tasks, which suggests they are learning true linguistic generalization. In this work, we focus on intrinsic probing, an analysis technique where the goal is not only to identify whether a representation encodes a linguistic attribute but also to pinpoint where this attribute is encoded. We propose a novel latent-variable formulation for constructing intrinsic probes and derive a tractable variational approximation to the log-likelihood. Our results show that our model is versatile and yields tighter mutual information estimates than two intrinsic probes previously proposed in the literature. Finally, we find empirical evidence that pre-trained representations develop a cross-lingually entangled notion of morphosyntax.

U2 - 10.1609/aaai.v37i11.26593

DO - 10.1609/aaai.v37i11.26593

M3 - Journal article

VL - 37

SP - 13591

EP - 13599

JO - AAAI Conference on Artificial Intelligence

JF - AAAI Conference on Artificial Intelligence

SN - 2159-5399

IS - 11

ER -

ID: 381157186