Syntactic Surprisal from Neural Models Predicts, but Underestimates, Human Processing Difficulty from Syntactic Ambiguities

Suhas Arehalli, Brian Dillon, Tal Linzen

    Research output: Chapter in Book/Report/Conference proceedingConference contribution

    Abstract

    Humans exhibit garden path effects: When reading sentences that are temporarily structurally ambiguous, they slow down when the structure is disambiguated in favor of the less preferred alternative. Surprisal theory (Hale, 2001; Levy, 2008), a prominent explanation of this finding, proposes that these slowdowns are due to the unpredictability of each of the words that occur in these sentences. Challenging this hypothesis, van Schijndel and Linzen (2021) find that estimates of the cost of word predictability derived from language models severely underestimate the magnitude of human garden path effects. In this work, we consider whether this underestimation is due to the fact that humans weight syntactic factors in their predictions more highly than language models do. We propose a method for estimating syntactic predictability from a language model, allowing us to weigh the cost of lexical and syntactic predictability independently. We find that treating syntactic predictability independently from lexical predictability indeed results in larger estimates of garden path. At the same time, even when syntactic predictability is independently weighted, surprisal still greatly underestimate the magnitude of human garden path effects. Our results support the hypothesis that predictability is not the only factor responsible for the processing cost associated with garden path sentences.

    Original languageEnglish (US)
    Title of host publicationCoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference
    PublisherAssociation for Computational Linguistics (ACL)
    Pages301-313
    Number of pages13
    ISBN (Electronic)9781959429074
    StatePublished - 2022
    Event26th Conference on Computational Natural Language Learning, CoNLL 2022 collocated and co-organized with EMNLP 2022 - Abu Dhabi, United Arab Emirates
    Duration: Dec 7 2022Dec 8 2022

    Publication series

    NameCoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference

    Conference

    Conference26th Conference on Computational Natural Language Learning, CoNLL 2022 collocated and co-organized with EMNLP 2022
    Country/TerritoryUnited Arab Emirates
    CityAbu Dhabi
    Period12/7/2212/8/22

    ASJC Scopus subject areas

    • Artificial Intelligence
    • Human-Computer Interaction
    • Linguistics and Language

    Fingerprint

    Dive into the research topics of 'Syntactic Surprisal from Neural Models Predicts, but Underestimates, Human Processing Difficulty from Syntactic Ambiguities'. Together they form a unique fingerprint.

    Cite this