Cargando…

The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing

Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal—those that involve repeatedly sampling fr...

Descripción completa

Detalles Bibliográficos
Autores principales: Hoover, Jacob Louis, Sonderegger, Morgan, Piantadosi, Steven T., O’Donnell, Timothy J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MIT Press 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449406/
https://www.ncbi.nlm.nih.gov/pubmed/37637302
http://dx.doi.org/10.1162/opmi_a_00086
_version_ 1785094944392740864
author Hoover, Jacob Louis
Sonderegger, Morgan
Piantadosi, Steven T.
O’Donnell, Timothy J.
author_facet Hoover, Jacob Louis
Sonderegger, Morgan
Piantadosi, Steven T.
O’Donnell, Timothy J.
author_sort Hoover, Jacob Louis
collection PubMed
description Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal—those that involve repeatedly sampling from the prior. Our first contribution is to show that simple examples of such algorithms predict runtime to increase superlinearly with surprisal, and also predict variance in runtime to increase. These two predictions stand in contrast with literature on surprisal theory (Hale, 2001; Levy, 2008a) which assumes that the expected processing cost increases linearly with surprisal, and makes no prediction about variance. In the second part of this paper, we conduct an empirical study of the relationship between surprisal and reading time, using a collection of modern language models to estimate surprisal. We find that with better language models, reading time increases superlinearly in surprisal, and also that variance increases. These results are consistent with the predictions of sampling-based algorithms.
format Online
Article
Text
id pubmed-10449406
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MIT Press
record_format MEDLINE/PubMed
spelling pubmed-104494062023-08-25 The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing Hoover, Jacob Louis Sonderegger, Morgan Piantadosi, Steven T. O’Donnell, Timothy J. Open Mind (Camb) Research Article Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal—those that involve repeatedly sampling from the prior. Our first contribution is to show that simple examples of such algorithms predict runtime to increase superlinearly with surprisal, and also predict variance in runtime to increase. These two predictions stand in contrast with literature on surprisal theory (Hale, 2001; Levy, 2008a) which assumes that the expected processing cost increases linearly with surprisal, and makes no prediction about variance. In the second part of this paper, we conduct an empirical study of the relationship between surprisal and reading time, using a collection of modern language models to estimate surprisal. We find that with better language models, reading time increases superlinearly in surprisal, and also that variance increases. These results are consistent with the predictions of sampling-based algorithms. MIT Press 2023-07-21 /pmc/articles/PMC10449406/ /pubmed/37637302 http://dx.doi.org/10.1162/opmi_a_00086 Text en © 2023 Massachusetts Institute of Technology https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/.
spellingShingle Research Article
Hoover, Jacob Louis
Sonderegger, Morgan
Piantadosi, Steven T.
O’Donnell, Timothy J.
The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
title The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
title_full The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
title_fullStr The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
title_full_unstemmed The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
title_short The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
title_sort plausibility of sampling as an algorithmic theory of sentence processing
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449406/
https://www.ncbi.nlm.nih.gov/pubmed/37637302
http://dx.doi.org/10.1162/opmi_a_00086
work_keys_str_mv AT hooverjacoblouis theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT sondereggermorgan theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT piantadosistevent theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT odonnelltimothyj theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT hooverjacoblouis plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT sondereggermorgan plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT piantadosistevent plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing
AT odonnelltimothyj plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing