Cargando…

Information-Theoretic Neural Decoding Reproduces Several Laws of Human Behavior

Human response times conform to several regularities including the Hick-Hyman law, the power law of practice, speed-accuracy trade-offs, and the Stroop effect. Each of these has been thoroughly modeled in isolation, but no account describes these phenomena as predictions of a unified framework. We p...

Descripción completa

Detalles Bibliográficos
Autores principales: Christie, S. Thomas, Johnson, Hayden R., Schrater, Paul R.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MIT Press 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10575563/
https://www.ncbi.nlm.nih.gov/pubmed/37840757
http://dx.doi.org/10.1162/opmi_a_00101
Descripción
Sumario:Human response times conform to several regularities including the Hick-Hyman law, the power law of practice, speed-accuracy trade-offs, and the Stroop effect. Each of these has been thoroughly modeled in isolation, but no account describes these phenomena as predictions of a unified framework. We provide such a framework and show that the phenomena arise as decoding times in a simple neural rate code with an entropy stopping threshold. Whereas traditional information-theoretic encoding systems exploit task statistics to optimize encoding strategies, we move this optimization to the decoder, treating it as a Bayesian ideal observer that can track transmission statistics as prior information during decoding. Our approach allays prominent concerns that applying information-theoretic perspectives to modeling brain and behavior requires complex encoding schemes that are incommensurate with neural encoding.