Cargando…
Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing
This meta-analysis compares the brain structures and mechanisms involved in facial and vocal emotion recognition. Neuroimaging studies contrasting emotional with neutral (face: N = 76, voice: N = 34) and explicit with implicit emotion processing (face: N = 27, voice: N = 20) were collected to shed l...
Autor principal: | |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5793823/ https://www.ncbi.nlm.nih.gov/pubmed/29186621 http://dx.doi.org/10.1093/scan/nsx142 |
_version_ | 1783297028635230208 |
---|---|
author | Schirmer, Annett |
author_facet | Schirmer, Annett |
author_sort | Schirmer, Annett |
collection | PubMed |
description | This meta-analysis compares the brain structures and mechanisms involved in facial and vocal emotion recognition. Neuroimaging studies contrasting emotional with neutral (face: N = 76, voice: N = 34) and explicit with implicit emotion processing (face: N = 27, voice: N = 20) were collected to shed light on stimulus and goal-driven mechanisms, respectively. Activation likelihood estimations were conducted on the full data sets for the separate modalities and on reduced, modality-matched data sets for modality comparison. Stimulus-driven emotion processing engaged large networks with significant modality differences in the superior temporal (voice-specific) and the medial temporal (face-specific) cortex. Goal-driven processing was associated with only a small cluster in the dorsomedial prefrontal cortex for voices but not faces. Neither stimulus- nor goal-driven processing showed significant modality overlap. Together, these findings suggest that stimulus-driven processes shape activity in the social brain more powerfully than goal-driven processes in both the visual and the auditory domains. Yet, whereas faces emphasize subcortical emotional and mnemonic mechanisms, voices emphasize cortical mechanisms associated with perception and effortful stimulus evaluation (e.g. via subvocalization). These differences may be due to sensory stimulus properties and highlight the need for a modality-specific perspective when modeling emotion processing in the brain. |
format | Online Article Text |
id | pubmed-5793823 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-57938232018-02-06 Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing Schirmer, Annett Soc Cogn Affect Neurosci Original Articles This meta-analysis compares the brain structures and mechanisms involved in facial and vocal emotion recognition. Neuroimaging studies contrasting emotional with neutral (face: N = 76, voice: N = 34) and explicit with implicit emotion processing (face: N = 27, voice: N = 20) were collected to shed light on stimulus and goal-driven mechanisms, respectively. Activation likelihood estimations were conducted on the full data sets for the separate modalities and on reduced, modality-matched data sets for modality comparison. Stimulus-driven emotion processing engaged large networks with significant modality differences in the superior temporal (voice-specific) and the medial temporal (face-specific) cortex. Goal-driven processing was associated with only a small cluster in the dorsomedial prefrontal cortex for voices but not faces. Neither stimulus- nor goal-driven processing showed significant modality overlap. Together, these findings suggest that stimulus-driven processes shape activity in the social brain more powerfully than goal-driven processes in both the visual and the auditory domains. Yet, whereas faces emphasize subcortical emotional and mnemonic mechanisms, voices emphasize cortical mechanisms associated with perception and effortful stimulus evaluation (e.g. via subvocalization). These differences may be due to sensory stimulus properties and highlight the need for a modality-specific perspective when modeling emotion processing in the brain. Oxford University Press 2018-01 2017-11-23 /pmc/articles/PMC5793823/ /pubmed/29186621 http://dx.doi.org/10.1093/scan/nsx142 Text en © The Author (2017). Published by Oxford University Press. http://creativecommons.org/licenses/by-nc/4.0/ This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com |
spellingShingle | Original Articles Schirmer, Annett Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing |
title | Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing |
title_full | Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing |
title_fullStr | Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing |
title_full_unstemmed | Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing |
title_short | Is the voice an auditory face? An ALE meta-analysis comparing vocal and facial emotion processing |
title_sort | is the voice an auditory face? an ale meta-analysis comparing vocal and facial emotion processing |
topic | Original Articles |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5793823/ https://www.ncbi.nlm.nih.gov/pubmed/29186621 http://dx.doi.org/10.1093/scan/nsx142 |
work_keys_str_mv | AT schirmerannett isthevoiceanauditoryfaceanalemetaanalysiscomparingvocalandfacialemotionprocessing |