Cargando…
Agent-based representations of objects and actions in the monkey pre-supplementary motor area
Information about objects around us is essential for planning actions and for predicting those of others. Here, we studied pre-supplementary motor area F6 neurons with a task in which monkeys viewed and grasped (or refrained from grasping) objects, and then observed a human doing the same task. We f...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
National Academy of Sciences
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6377463/ https://www.ncbi.nlm.nih.gov/pubmed/30696759 http://dx.doi.org/10.1073/pnas.1810890116 |
_version_ | 1783395745138737152 |
---|---|
author | Livi, Alessandro Lanzilotto, Marco Maranesi, Monica Fogassi, Leonardo Rizzolatti, Giacomo Bonini, Luca |
author_facet | Livi, Alessandro Lanzilotto, Marco Maranesi, Monica Fogassi, Leonardo Rizzolatti, Giacomo Bonini, Luca |
author_sort | Livi, Alessandro |
collection | PubMed |
description | Information about objects around us is essential for planning actions and for predicting those of others. Here, we studied pre-supplementary motor area F6 neurons with a task in which monkeys viewed and grasped (or refrained from grasping) objects, and then observed a human doing the same task. We found “action-related neurons” encoding selectively monkey’s own action [self-type (ST)], another agent’s action [other-type (OT)], or both [self- and other-type (SOT)]. Interestingly, we found “object-related neurons” exhibiting the same type of selectivity before action onset: Indeed, distinct sets of neurons discharged when visually presented objects were targeted by the monkey’s own action (ST), another agent’s action (OT), or both (SOT). Notably, object-related neurons appear to signal self and other’s intention to grasp and the most likely grip type that will be performed, whereas action-related neurons encode a general goal attainment signal devoid of any specificity for the observed grip type. Time-resolved cross-modal population decoding revealed that F6 neurons first integrate information about object and context to generate an agent-shared signal specifying whether and how the object will be grasped, which progressively turns into a broader agent-based goal attainment signal during action unfolding. Importantly, shared representation of objects critically depends upon their location in the observer’s peripersonal space, suggesting an “object-mirroring” mechanism through which observers could accurately predict others’ impending action by recruiting the same motor representation they would activate if they were to act upon the same object in the same context. |
format | Online Article Text |
id | pubmed-6377463 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | National Academy of Sciences |
record_format | MEDLINE/PubMed |
spelling | pubmed-63774632019-02-19 Agent-based representations of objects and actions in the monkey pre-supplementary motor area Livi, Alessandro Lanzilotto, Marco Maranesi, Monica Fogassi, Leonardo Rizzolatti, Giacomo Bonini, Luca Proc Natl Acad Sci U S A PNAS Plus Information about objects around us is essential for planning actions and for predicting those of others. Here, we studied pre-supplementary motor area F6 neurons with a task in which monkeys viewed and grasped (or refrained from grasping) objects, and then observed a human doing the same task. We found “action-related neurons” encoding selectively monkey’s own action [self-type (ST)], another agent’s action [other-type (OT)], or both [self- and other-type (SOT)]. Interestingly, we found “object-related neurons” exhibiting the same type of selectivity before action onset: Indeed, distinct sets of neurons discharged when visually presented objects were targeted by the monkey’s own action (ST), another agent’s action (OT), or both (SOT). Notably, object-related neurons appear to signal self and other’s intention to grasp and the most likely grip type that will be performed, whereas action-related neurons encode a general goal attainment signal devoid of any specificity for the observed grip type. Time-resolved cross-modal population decoding revealed that F6 neurons first integrate information about object and context to generate an agent-shared signal specifying whether and how the object will be grasped, which progressively turns into a broader agent-based goal attainment signal during action unfolding. Importantly, shared representation of objects critically depends upon their location in the observer’s peripersonal space, suggesting an “object-mirroring” mechanism through which observers could accurately predict others’ impending action by recruiting the same motor representation they would activate if they were to act upon the same object in the same context. National Academy of Sciences 2019-02-12 2019-01-29 /pmc/articles/PMC6377463/ /pubmed/30696759 http://dx.doi.org/10.1073/pnas.1810890116 Text en Copyright © 2019 the Author(s). Published by PNAS. http://creativecommons.org/licenses/by/4.0/ This open access article is distributed under Creative Commons Attribution License 4.0 (CC BY) (http://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | PNAS Plus Livi, Alessandro Lanzilotto, Marco Maranesi, Monica Fogassi, Leonardo Rizzolatti, Giacomo Bonini, Luca Agent-based representations of objects and actions in the monkey pre-supplementary motor area |
title | Agent-based representations of objects and actions in the monkey pre-supplementary motor area |
title_full | Agent-based representations of objects and actions in the monkey pre-supplementary motor area |
title_fullStr | Agent-based representations of objects and actions in the monkey pre-supplementary motor area |
title_full_unstemmed | Agent-based representations of objects and actions in the monkey pre-supplementary motor area |
title_short | Agent-based representations of objects and actions in the monkey pre-supplementary motor area |
title_sort | agent-based representations of objects and actions in the monkey pre-supplementary motor area |
topic | PNAS Plus |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6377463/ https://www.ncbi.nlm.nih.gov/pubmed/30696759 http://dx.doi.org/10.1073/pnas.1810890116 |
work_keys_str_mv | AT livialessandro agentbasedrepresentationsofobjectsandactionsinthemonkeypresupplementarymotorarea AT lanzilottomarco agentbasedrepresentationsofobjectsandactionsinthemonkeypresupplementarymotorarea AT maranesimonica agentbasedrepresentationsofobjectsandactionsinthemonkeypresupplementarymotorarea AT fogassileonardo agentbasedrepresentationsofobjectsandactionsinthemonkeypresupplementarymotorarea AT rizzolattigiacomo agentbasedrepresentationsofobjectsandactionsinthemonkeypresupplementarymotorarea AT boniniluca agentbasedrepresentationsofobjectsandactionsinthemonkeypresupplementarymotorarea |