TY - GEN
T1 - Verb Conjugation in Transformers Is Determined by Linear Encodings of Subject Number
AU - Hao, Sophie
AU - Linzen, Tal
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics.
PY - 2023
Y1 - 2023
N2 - Deep architectures such as Transformers are sometimes criticized for having uninterpretable “black-box” representations. We use causal intervention analysis to show that, in fact, some linguistic features are represented in a linear, interpretable format. Specifically, we show that BERT's ability to conjugate verbs relies on a linear encoding of subject number that can be manipulated with predictable effects on conjugation accuracy. This encoding is found in the subject position at the first layer and the verb position at the last layer, but is distributed across positions at middle layers, particularly when there are multiple cues to subject number.
AB - Deep architectures such as Transformers are sometimes criticized for having uninterpretable “black-box” representations. We use causal intervention analysis to show that, in fact, some linguistic features are represented in a linear, interpretable format. Specifically, we show that BERT's ability to conjugate verbs relies on a linear encoding of subject number that can be manipulated with predictable effects on conjugation accuracy. This encoding is found in the subject position at the first layer and the verb position at the last layer, but is distributed across positions at middle layers, particularly when there are multiple cues to subject number.
UR - http://www.scopus.com/inward/record.url?scp=85183303460&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85183303460&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85183303460
T3 - Findings of the Association for Computational Linguistics: EMNLP 2023
SP - 4531
EP - 4539
BT - Findings of the Association for Computational Linguistics
PB - Association for Computational Linguistics (ACL)
T2 - 2023 Findings of the Association for Computational Linguistics: EMNLP 2023
Y2 - 6 December 2023 through 10 December 2023
ER -