Back to Search Start Over

Adaptation to Vocal Expressions Reveals Multistep Perception of Auditory Emotion

Authors :
Bestelmeyer, Patricia E.G.
Maurage, Pierre
Rouger, Julien
Latinus, Marianne
Belin, Pascal
School of Psychology [Gwynedd, UK] (Bangor University)
Bangor University
Université Catholique de Louvain = Catholic University of Louvain (UCL)
Centre de recherche cerveau et cognition (CERCO)
Institut des sciences du cerveau de Toulouse. (ISCT)
Université Toulouse - Jean Jaurès (UT2J)
Université de Toulouse (UT)-Université de Toulouse (UT)-Université Toulouse III - Paul Sabatier (UT3)
Université de Toulouse (UT)-Centre Hospitalier Universitaire de Toulouse (CHU Toulouse)-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre National de la Recherche Scientifique (CNRS)-Université Toulouse - Jean Jaurès (UT2J)
Université de Toulouse (UT)-Centre Hospitalier Universitaire de Toulouse (CHU Toulouse)-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre National de la Recherche Scientifique (CNRS)-Centre National de la Recherche Scientifique (CNRS)
Institut de Neurosciences de la Timone (INT)
Aix Marseille Université (AMU)-Centre National de la Recherche Scientifique (CNRS)
International Laboratory for Brain, Music and Sound Research [Montréal, Canada] (BRAMS)
McGill University = Université McGill [Montréal, Canada]-Université de Montréal (UdeM)
Institute of Neurosciences and Psychology [Glasgow]
University of Glasgow
ANR-16-CONV-0002,ILCB,ILCB: Institute of Language Communication and the Brain(2016)
UCL - SSS/IONS/COSY - Systems & cognitive Neuroscience
UCL - SSS/IONS/NEUR - Clinical Neuroscience
Université Toulouse - Jean Jaurès (UT2J)-Université Toulouse III - Paul Sabatier (UT3)
Université Fédérale Toulouse Midi-Pyrénées-Université Fédérale Toulouse Midi-Pyrénées-CHU Toulouse [Toulouse]-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre National de la Recherche Scientifique (CNRS)-Université Toulouse - Jean Jaurès (UT2J)-Université Toulouse III - Paul Sabatier (UT3)
Université Fédérale Toulouse Midi-Pyrénées-Université Fédérale Toulouse Midi-Pyrénées-CHU Toulouse [Toulouse]-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre National de la Recherche Scientifique (CNRS)-Centre National de la Recherche Scientifique (CNRS)
Centre National de la Recherche Scientifique (CNRS)-Aix Marseille Université (AMU)
Université Fédérale Toulouse Midi-Pyrénées-Université Fédérale Toulouse Midi-Pyrénées-Centre Hospitalier Universitaire de Toulouse (CHU Toulouse)-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre National de la Recherche Scientifique (CNRS)-Université Toulouse - Jean Jaurès (UT2J)-Université Toulouse III - Paul Sabatier (UT3)
Université Fédérale Toulouse Midi-Pyrénées-Université Fédérale Toulouse Midi-Pyrénées-Centre Hospitalier Universitaire de Toulouse (CHU Toulouse)-Institut National de la Santé et de la Recherche Médicale (INSERM)-Centre National de la Recherche Scientifique (CNRS)-Centre National de la Recherche Scientifique (CNRS)
Bourhis, Morgane
ILCB: Institute of Language Communication and the Brain - - ILCB2016 - ANR-16-CONV-0002 - CONV - VALID
Source :
Journal of Neuroscience, Journal of Neuroscience, 2014, 34 (24), pp.8098-8105. ⟨10.1523/jneurosci.4820-13.2014⟩, The Journal of Neuroscience, Vol. 34, no. 24, p. 8098-8105 (2014), Journal of Neuroscience, Society for Neuroscience, 2014, 34 (24), pp.8098-8105. ⟨10.1523/jneurosci.4820-13.2014⟩
Publication Year :
2014
Publisher :
HAL CCSD, 2014.

Abstract

The human voice carries speech as well as important nonlinguistic signals that influence our social interactions. Among these cues that impact our behavior and communication with other people is the perceived emotional state of the speaker. A theoretical framework for the neural processing stages of emotional prosody has suggested that auditory emotion is perceived in multiple steps (Schirmer and Kotz, 2006) involving low-level auditory analysis and integration of the acoustic information followed by higher-level cognition. Empirical evidence for this multistep processing chain, however, is still sparse. We examined this question using functional magnetic resonance imaging and a continuous carry-over design (Aguirre, 2007) to measure brain activity while volunteers listened to non-speech-affective vocalizations morphed on a continuum between anger and fear. Analyses dissociated neuronal adaptation effects induced by similarity in perceived emotional content between consecutive stimuli from those induced by their acoustic similarity. We found that bilateral voice-sensitive auditory regions as well as right amygdala coded the physical difference between consecutive stimuli. In contrast, activity in bilateral anterior insulae, medial superior frontal cortex, precuneus, and subcortical regions such as bilateral hippocampi depended predominantly on the perceptual difference between morphs. Our results suggest that the processing of vocal affect recognition is a multistep process involving largely distinct neural networks. Amygdala and auditory areas predominantly code emotion-related acoustic information while more anterior insular and prefrontal regions respond to the abstract, cognitive representation of vocal affec. © 2014 Bestelmeyer et al.

Details

Language :
English
ISSN :
02706474 and 15292401
Database :
OpenAIRE
Journal :
Journal of Neuroscience, Journal of Neuroscience, 2014, 34 (24), pp.8098-8105. ⟨10.1523/jneurosci.4820-13.2014⟩, The Journal of Neuroscience, Vol. 34, no. 24, p. 8098-8105 (2014), Journal of Neuroscience, Society for Neuroscience, 2014, 34 (24), pp.8098-8105. ⟨10.1523/jneurosci.4820-13.2014⟩
Accession number :
edsair.dedup.wf.001..9e9bcf3a6fc08e67b453376df9ac5093
Full Text :
https://doi.org/10.1523/jneurosci.4820-13.2014⟩