Cargando…
The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing
Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal—those that involve repeatedly sampling fr...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MIT Press
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449406/ https://www.ncbi.nlm.nih.gov/pubmed/37637302 http://dx.doi.org/10.1162/opmi_a_00086 |
_version_ | 1785094944392740864 |
---|---|
author | Hoover, Jacob Louis Sonderegger, Morgan Piantadosi, Steven T. O’Donnell, Timothy J. |
author_facet | Hoover, Jacob Louis Sonderegger, Morgan Piantadosi, Steven T. O’Donnell, Timothy J. |
author_sort | Hoover, Jacob Louis |
collection | PubMed |
description | Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal—those that involve repeatedly sampling from the prior. Our first contribution is to show that simple examples of such algorithms predict runtime to increase superlinearly with surprisal, and also predict variance in runtime to increase. These two predictions stand in contrast with literature on surprisal theory (Hale, 2001; Levy, 2008a) which assumes that the expected processing cost increases linearly with surprisal, and makes no prediction about variance. In the second part of this paper, we conduct an empirical study of the relationship between surprisal and reading time, using a collection of modern language models to estimate surprisal. We find that with better language models, reading time increases superlinearly in surprisal, and also that variance increases. These results are consistent with the predictions of sampling-based algorithms. |
format | Online Article Text |
id | pubmed-10449406 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MIT Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-104494062023-08-25 The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing Hoover, Jacob Louis Sonderegger, Morgan Piantadosi, Steven T. O’Donnell, Timothy J. Open Mind (Camb) Research Article Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal—those that involve repeatedly sampling from the prior. Our first contribution is to show that simple examples of such algorithms predict runtime to increase superlinearly with surprisal, and also predict variance in runtime to increase. These two predictions stand in contrast with literature on surprisal theory (Hale, 2001; Levy, 2008a) which assumes that the expected processing cost increases linearly with surprisal, and makes no prediction about variance. In the second part of this paper, we conduct an empirical study of the relationship between surprisal and reading time, using a collection of modern language models to estimate surprisal. We find that with better language models, reading time increases superlinearly in surprisal, and also that variance increases. These results are consistent with the predictions of sampling-based algorithms. MIT Press 2023-07-21 /pmc/articles/PMC10449406/ /pubmed/37637302 http://dx.doi.org/10.1162/opmi_a_00086 Text en © 2023 Massachusetts Institute of Technology https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/. |
spellingShingle | Research Article Hoover, Jacob Louis Sonderegger, Morgan Piantadosi, Steven T. O’Donnell, Timothy J. The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing |
title | The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing |
title_full | The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing |
title_fullStr | The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing |
title_full_unstemmed | The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing |
title_short | The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing |
title_sort | plausibility of sampling as an algorithmic theory of sentence processing |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449406/ https://www.ncbi.nlm.nih.gov/pubmed/37637302 http://dx.doi.org/10.1162/opmi_a_00086 |
work_keys_str_mv | AT hooverjacoblouis theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT sondereggermorgan theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT piantadosistevent theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT odonnelltimothyj theplausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT hooverjacoblouis plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT sondereggermorgan plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT piantadosistevent plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing AT odonnelltimothyj plausibilityofsamplingasanalgorithmictheoryofsentenceprocessing |