Verb Conjugation in Transformers Is Determined by Linear Encodings of Subject Number

Sophie Hao, Tal Linzen

    Research output: Chapter in Book/Report/Conference proceedingConference contribution

    Abstract

    Deep architectures such as Transformers are sometimes criticized for having uninterpretable “black-box” representations. We use causal intervention analysis to show that, in fact, some linguistic features are represented in a linear, interpretable format. Specifically, we show that BERT's ability to conjugate verbs relies on a linear encoding of subject number that can be manipulated with predictable effects on conjugation accuracy. This encoding is found in the subject position at the first layer and the verb position at the last layer, but is distributed across positions at middle layers, particularly when there are multiple cues to subject number.

    Original languageEnglish (US)
    Title of host publicationFindings of the Association for Computational Linguistics
    Subtitle of host publicationEMNLP 2023
    PublisherAssociation for Computational Linguistics (ACL)
    Pages4531-4539
    Number of pages9
    ISBN (Electronic)9798891760615
    StatePublished - 2023
    Event2023 Findings of the Association for Computational Linguistics: EMNLP 2023 - Singapore, Singapore
    Duration: Dec 6 2023Dec 10 2023

    Publication series

    NameFindings of the Association for Computational Linguistics: EMNLP 2023

    Conference

    Conference2023 Findings of the Association for Computational Linguistics: EMNLP 2023
    Country/TerritorySingapore
    CitySingapore
    Period12/6/2312/10/23

    ASJC Scopus subject areas

    • Computational Theory and Mathematics
    • Computer Science Applications
    • Information Systems
    • Language and Linguistics
    • Linguistics and Language

    Fingerprint

    Dive into the research topics of 'Verb Conjugation in Transformers Is Determined by Linear Encodings of Subject Number'. Together they form a unique fingerprint.

    Cite this