Research output

A network analysis of audiovisual affective speech perception

Research output: Contribution to journalArticleAcademicpeer-review

Associated researcher

Associated organisations

Abstract

In this study we were interested in the neural system supporting the audiovisual integration of emotional expression and emotional prosody. To this end normal participants were exposed to short videos of a computer-animated face voicing emotionally positive or negative words with the appropriate prosody. Facial expression of the face was either neutral or emotionally appropriate. To reveal the neural network involved in affective audio-visual (AV) integration, standard univariate analysis of fMRI data was followed by a Random-Effects Granger Causality Mapping (RFX-GCM). The regions that distinguished emotional from neutral facial expressions in the univariate analysis were taken as seed regions. In trials showing emotional expressions compared to neutral trials univariate analysis showed activation primarily in bilateral amygdala, Fusiform Gyrus, Middle Temporal Gyrus / Superior Temporal Sulcus and Inferior Occipital Gyrus. When employing either the left amygdala or the right amygdala as a seed region in RFX-GCM we found connectivity with the right hemispheric Fusiform Gyrus, with the indication that the Fusiform Gyrus sends information to the Amygdala. These results led to a working model for face perception in general and for audio-visual-affective integration in particular which is an elaborated adaptation of existing models.

    Research areas

  • audiovisual speech, emotion, facial affect perception, amygdala, Granger causality, SUPERIOR TEMPORAL SULCUS, EVENT-RELATED FMRI, TIME-RESOLVED FMRI, FUSIFORM FACE AREA, FACIAL EXPRESSIONS, HUMAN BRAIN, NEURAL RESPONSES, HUMAN AMYGDALA, EFFECTIVE CONNECTIVITY, CROSSMODAL BINDING
View graph of relations

Details

Original languageEnglish
Pages (from-to)230-241
Number of pages12
JournalNeuroscience
Volume256
DOIs
Publication statusPublished - 3 Jan 2014