TY - GEN
T1 - Syntactic Surprisal from Neural Models Predicts, but Underestimates, Human Processing Difficulty from Syntactic Ambiguities
AU - Arehalli, Suhas
AU - Dillon, Brian
AU - Linzen, Tal
N1 - Funding Information:
This work was supported by the National Science Foundation, grant nos. BCS-2020945 and BCS-2020914, by the United States–Israel Binational Science Foundation (award no. 2018284), and in part through the NYU IT High Performance Computing resources, services, and staff expertise. We would also like to thank members of the NYU Computation and Psycholinguistics lab, as well as members of the Society for Human Sentence Processing community, for insightful discussion around this work.
Publisher Copyright:
©2022 Association for Computational Linguistics.
PY - 2022
Y1 - 2022
N2 - Humans exhibit garden path effects: When reading sentences that are temporarily structurally ambiguous, they slow down when the structure is disambiguated in favor of the less preferred alternative. Surprisal theory (Hale, 2001; Levy, 2008), a prominent explanation of this finding, proposes that these slowdowns are due to the unpredictability of each of the words that occur in these sentences. Challenging this hypothesis, van Schijndel and Linzen (2021) find that estimates of the cost of word predictability derived from language models severely underestimate the magnitude of human garden path effects. In this work, we consider whether this underestimation is due to the fact that humans weight syntactic factors in their predictions more highly than language models do. We propose a method for estimating syntactic predictability from a language model, allowing us to weigh the cost of lexical and syntactic predictability independently. We find that treating syntactic predictability independently from lexical predictability indeed results in larger estimates of garden path. At the same time, even when syntactic predictability is independently weighted, surprisal still greatly underestimate the magnitude of human garden path effects. Our results support the hypothesis that predictability is not the only factor responsible for the processing cost associated with garden path sentences.
AB - Humans exhibit garden path effects: When reading sentences that are temporarily structurally ambiguous, they slow down when the structure is disambiguated in favor of the less preferred alternative. Surprisal theory (Hale, 2001; Levy, 2008), a prominent explanation of this finding, proposes that these slowdowns are due to the unpredictability of each of the words that occur in these sentences. Challenging this hypothesis, van Schijndel and Linzen (2021) find that estimates of the cost of word predictability derived from language models severely underestimate the magnitude of human garden path effects. In this work, we consider whether this underestimation is due to the fact that humans weight syntactic factors in their predictions more highly than language models do. We propose a method for estimating syntactic predictability from a language model, allowing us to weigh the cost of lexical and syntactic predictability independently. We find that treating syntactic predictability independently from lexical predictability indeed results in larger estimates of garden path. At the same time, even when syntactic predictability is independently weighted, surprisal still greatly underestimate the magnitude of human garden path effects. Our results support the hypothesis that predictability is not the only factor responsible for the processing cost associated with garden path sentences.
UR - http://www.scopus.com/inward/record.url?scp=85145277658&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85145277658&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85145277658
T3 - CoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference
SP - 301
EP - 313
BT - CoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference
PB - Association for Computational Linguistics (ACL)
T2 - 26th Conference on Computational Natural Language Learning, CoNLL 2022 collocated and co-organized with EMNLP 2022
Y2 - 7 December 2022 through 8 December 2022
ER -