Cargando…

Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework

This paper introduces a time- and state-dependent measure of integrated information, φ, which captures the repertoire of causal states available to a system as a whole. Specifically, φ quantifies how much information is generated (uncertainty is reduced) when a system enters a particular state throu...

Descripción completa

Detalles Bibliográficos
Autores principales: Balduzzi, David, Tononi, Giulio
Formato: Texto
Lenguaje:English
Publicado: Public Library of Science 2008
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2386970/
https://www.ncbi.nlm.nih.gov/pubmed/18551165
http://dx.doi.org/10.1371/journal.pcbi.1000091
_version_ 1782155283782434816
author Balduzzi, David
Tononi, Giulio
author_facet Balduzzi, David
Tononi, Giulio
author_sort Balduzzi, David
collection PubMed
description This paper introduces a time- and state-dependent measure of integrated information, φ, which captures the repertoire of causal states available to a system as a whole. Specifically, φ quantifies how much information is generated (uncertainty is reduced) when a system enters a particular state through causal interactions among its elements, above and beyond the information generated independently by its parts. Such mathematical characterization is motivated by the observation that integrated information captures two key phenomenological properties of consciousness: (i) there is a large repertoire of conscious experiences so that, when one particular experience occurs, it generates a large amount of information by ruling out all the others; and (ii) this information is integrated, in that each experience appears as a whole that cannot be decomposed into independent parts. This paper extends previous work on stationary systems and applies integrated information to discrete networks as a function of their dynamics and causal architecture. An analysis of basic examples indicates the following: (i) φ varies depending on the state entered by a network, being higher if active and inactive elements are balanced and lower if the network is inactive or hyperactive. (ii) φ varies for systems with identical or similar surface dynamics depending on the underlying causal architecture, being low for systems that merely copy or replay activity states. (iii) φ varies as a function of network architecture. High φ values can be obtained by architectures that conjoin functional specialization with functional integration. Strictly modular and homogeneous systems cannot generate high φ because the former lack integration, whereas the latter lack information. Feedforward and lattice architectures are capable of generating high φ but are inefficient. (iv) In Hopfield networks, φ is low for attractor states and neutral states, but increases if the networks are optimized to achieve tension between local and global interactions. These basic examples appear to match well against neurobiological evidence concerning the neural substrates of consciousness. More generally, φ appears to be a useful metric to characterize the capacity of any physical system to integrate information.
format Text
id pubmed-2386970
institution National Center for Biotechnology Information
language English
publishDate 2008
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-23869702008-06-13 Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework Balduzzi, David Tononi, Giulio PLoS Comput Biol Research Article This paper introduces a time- and state-dependent measure of integrated information, φ, which captures the repertoire of causal states available to a system as a whole. Specifically, φ quantifies how much information is generated (uncertainty is reduced) when a system enters a particular state through causal interactions among its elements, above and beyond the information generated independently by its parts. Such mathematical characterization is motivated by the observation that integrated information captures two key phenomenological properties of consciousness: (i) there is a large repertoire of conscious experiences so that, when one particular experience occurs, it generates a large amount of information by ruling out all the others; and (ii) this information is integrated, in that each experience appears as a whole that cannot be decomposed into independent parts. This paper extends previous work on stationary systems and applies integrated information to discrete networks as a function of their dynamics and causal architecture. An analysis of basic examples indicates the following: (i) φ varies depending on the state entered by a network, being higher if active and inactive elements are balanced and lower if the network is inactive or hyperactive. (ii) φ varies for systems with identical or similar surface dynamics depending on the underlying causal architecture, being low for systems that merely copy or replay activity states. (iii) φ varies as a function of network architecture. High φ values can be obtained by architectures that conjoin functional specialization with functional integration. Strictly modular and homogeneous systems cannot generate high φ because the former lack integration, whereas the latter lack information. Feedforward and lattice architectures are capable of generating high φ but are inefficient. (iv) In Hopfield networks, φ is low for attractor states and neutral states, but increases if the networks are optimized to achieve tension between local and global interactions. These basic examples appear to match well against neurobiological evidence concerning the neural substrates of consciousness. More generally, φ appears to be a useful metric to characterize the capacity of any physical system to integrate information. Public Library of Science 2008-06-13 /pmc/articles/PMC2386970/ /pubmed/18551165 http://dx.doi.org/10.1371/journal.pcbi.1000091 Text en Balduzzi, Tononi. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Balduzzi, David
Tononi, Giulio
Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework
title Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework
title_full Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework
title_fullStr Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework
title_full_unstemmed Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework
title_short Integrated Information in Discrete Dynamical Systems: Motivation and Theoretical Framework
title_sort integrated information in discrete dynamical systems: motivation and theoretical framework
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2386970/
https://www.ncbi.nlm.nih.gov/pubmed/18551165
http://dx.doi.org/10.1371/journal.pcbi.1000091
work_keys_str_mv AT balduzzidavid integratedinformationindiscretedynamicalsystemsmotivationandtheoreticalframework
AT tononigiulio integratedinformationindiscretedynamicalsystemsmotivationandtheoreticalframework