社交焦虑对跨模态情绪信息加工的影响
The Impact of Social Anxiety on Cross-Modal Emotional Processing
DOI: 10.12677/ap.2026.163148, PDF,    科研立项经费支持
作者: 王亚丽*:浙江财经大学马克思主义学院,浙江 杭州;石 茹:包头医学院党委学生工作部(处),内蒙古 包头
关键词: 社交焦虑跨通道情绪加工整合优势冲突效应Social Anxiety Cross-Modal Emotional Processing Integration Advantage Conflict Effect
摘要: 跨模态情绪信息的准确识别是社会交往的重要基础。既有研究表明,社交焦虑个体在情绪识别与判断中可能存在偏差,但其跨模态情绪加工特征仍有待进一步探讨。本研究筛选高、低社交焦虑大学生,考察其在视觉、听觉及视听情绪信息加工中的表现。结果显示,无论在正确率还是反应时指标上,视听一致条件均显著优于单视觉、单听觉及视听不一致条件,呈现典型的双通道加工优势;视听不一致条件正确率最低、反应时最长,表现出显著的跨模态冲突效应。进一步分析发现,在愤怒情绪条件下,高、低社交焦虑组均表现出双通道优势,其中高社交焦虑组的整合效应更为突出;在中性情绪条件下,两组均出现跨模态不一致效应,但低社交焦虑组的不一致干扰更为明显。研究结果揭示了社交焦虑水平对跨模态情绪加工的调节作用,为理解社交焦虑相关的情绪感知偏差提供了行为学证据。
Abstract: Accurate recognition of cross-modal emotional information is a fundamental prerequisite for successful social interaction. Previous studies have suggested that individuals with social anxiety may exhibit biases in emotion recognition and judgment; however, their characteristics of cross-modal emotional processing remain insufficiently understood. The present study recruited university students with high and low levels of social anxiety to examine their performance in processing visual, auditory, and audiovisual emotional information. The results showed that, in terms of both accuracy and reaction time in emotion recognition, the audiovisual congruent condition significantly outperformed the unimodal visual, unimodal auditory, and audiovisual incongruent conditions, demonstrating a typical dual-channel processing advantage. In contrast, the audiovisual incongruent condition yielded the lowest accuracy and the longest reaction times, indicating a robust cross-modal conflict effect. Further analyses revealed that under angry emotion conditions, both high and low social anxiety groups exhibited a dual-channel advantage, with a more pronounced integration effect observed in the high social anxiety group. Under neutral emotion conditions, both groups showed significant cross-modal incongruency effects; however, the interference caused by incongruent information was more pronounced in the low social anxiety group. These findings indicate that social anxiety level modulates cross-modal emotional processing and provide behavioral evidence for understanding emotion perception biases associated with social anxiety.
文章引用:王亚丽, 石茹 (2026). 社交焦虑对跨模态情绪信息加工的影响. 心理学进展, 16(3), 338-350. https://doi.org/10.12677/ap.2026.163148

参考文献

[1] 李萍, 张明明, 李帅霞, 张火垠, 罗文波(2019). 面孔表情和声音情绪信息整合加工的脑机制. 心理科学进展, 27(7), 1205-1214.
[2] 彭纯子, 龚耀先, 朱熊兆(2004). 交往焦虑量表的信效度及其在中国大学生中的适用性. 中国心理卫生杂志, 18(1), 39-41.
[3] 宋素涛, 李爽, 赵诗梦, 肖观来, 张锦秀, 郑元杰(2023). 社交焦虑者对面孔表情存在持续性注意偏向: 来自N2pc的证据. 中国临床心理学杂志, 31(2), 267-273.
[4] 张亮, 孙向红, 张侃(2009). 情绪信息的多通道整合. 心理科学进展, 17(6), 1133-1138.
[5] 张明, 王婷婷, 吴晓刚, 张月娥, 王爱君(2022). 面孔表情和声音情绪信息整合对返回抑制的影响. 心理学报, 54(4), 331-342.
[6] Belin, P., Fillion-Bilodeau, S., & Gosselin, F. (2008). The Montreal Affective Voices: A Validated Set of Nonverbal Affect Bursts for Research on Auditory Affective Processing. Behavior Research Methods, 40, 531-539.[CrossRef] [PubMed]
[7] Carleton, R. N. (2016). Fear of the Unknown: One Fear to Rule Them All? Journal of Anxiety Disorders, 41, 5-21.[CrossRef] [PubMed]
[8] Chen, J., Short, M., & Kemps, E. (2020). Interpretation Bias in Social Anxiety: A Systematic Review and Meta-Analysis. Journal of Affective Disorders, 276, 1119-1130.[CrossRef] [PubMed]
[9] Chen, M., Zhao, S., Yu, J., Leng, X., Zhai, M., Feng, C. et al. (2022). Audiovisual Emotional Congruency Modulates the Stimulus-Driven Cross-Modal Spread of Attention. Brain Sciences, 12, Article 1229.[CrossRef] [PubMed]
[10] Collignon, O., Girard, S., Gosselin, F., Roy, S., Saint-Amour, D., Lassonde, M. et al. (2008). Audio-Visual Integration of Emotion Expression. Brain Research, 1242, 126-135.[CrossRef] [PubMed]
[11] Diaconescu, A. O., Alain, C., & McIntosh, A. R. (2011). The Co-Occurrence of Multisensory Facilitation and Cross-Modal Conflict in the Human Brain. Journal of Neurophysiology, 106, 2896-2909.[CrossRef] [PubMed]
[12] Dolan, R. J., Morris, J. S., & de Gelder, B. (2001). Crossmodal Binding of Fear in Voice and Face. Proceedings of the National Academy of Sciences, 98, 10006-10010.[CrossRef] [PubMed]
[13] Donohue, S. E., Appelbaum, L. G., Park, C. J., Roberts, K. C., & Woldorff, M. G. (2013). Cross-Modal Stimulus Conflict: The Behavioral Effects of Stimulus Input Timing in a Visual-Auditory Stroop Task. PLOS ONE, 8, e62802.[CrossRef] [PubMed]
[14] Eysenck, M. W., Derakshan, N., Santos, R., & Calvo, M. G. (2007). Anxiety and Cognitive Performance: Attentional Control Theory. Emotion, 7, 336-353.[CrossRef] [PubMed]
[15] Filippi, P. (2016). Emotional and Interactional Prosody across Animal Communication Systems: A Comparative Approach to the Emergence of Language. Frontiers in Psychology, 7, Article 1393.[CrossRef] [PubMed]
[16] Gan, S., & Li, W. (2023). Aberrant Neural Correlates of Multisensory Processing of Audiovisual Social Cues Related to Social Anxiety: An Electrophysiological Study. Frontiers in Psychiatry, 14, Article 1020812.[CrossRef] [PubMed]
[17] Gao, C., Wedell, D. H., Green, J. J., Jia, X., Mao, X., Guo, C. et al. (2018). Temporal Dynamics of Audiovisual Affective Processing. Biological Psychology, 139, 59-72.[CrossRef] [PubMed]
[18] Grupe, D. W., & Nitschke, J. B. (2013). Uncertainty and Anticipation in Anxiety: An Integrated Neurobiological and Psychological Perspective. Nature Reviews Neuroscience, 14, 488-501.[CrossRef] [PubMed]
[19] Heeren, A., & McNally, R. J. (2018). Social Anxiety Disorder as a Densely Interconnected Network of Fear and Avoidance for Social Situations. Cognitive Therapy and Research, 42, 103-113.[CrossRef
[20] Heffer, N., Gradidge, M., Karl, A., Ashwin, C., & Petrini, K. (2022). High Trait Anxiety Enhances Optimal Integration of Auditory and Visual Threat Cues. Journal of Behavior Therapy and Experimental Psychiatry, 74, Article ID: 101693.[CrossRef] [PubMed]
[21] Hu, Z., Zhang, R., Zhang, Q., Liu, Q., & Li, H. (2012). Neural Correlates of Audiovisual Integration of Semantic Category Information. Brain and Language, 121, 70-75.[CrossRef] [PubMed]
[22] Joassin, F., Maurage, P., Bruyer, R., Crommelinck, M., & Campanella, S. (2004). When Audition Alters Vision: An Event-Related Potential Study of the Cross-Modal Interactions between Faces and Voices. Neuroscience Letters, 369, 132-137.[CrossRef] [PubMed]
[23] Klasen, M., Chen, Y., & Mathiak, K. (2012). Multisensory Emotions: Perception, Combination and Underlying Neural Processes. Reviews in the Neurosciences, 23, 381-392.[CrossRef] [PubMed]
[24] Klasen, M., Kenworthy, C. A., Mathiak, K. A., Kircher, T. T. J., & Mathiak, K. (2011). Supramodal Representation of Emotions. The Journal of Neuroscience, 31, 13635-13643.[CrossRef] [PubMed]
[25] Koizumi, A., Tanaka, A., Imai, H., Hiramatsu, S., Hiramoto, E., Sato, T. et al. (2011). The Effects of Anxiety on the Interpretation of Emotion in the Face-Voice Pairs. Experimental Brain Research, 213, 275-282.[CrossRef] [PubMed]
[26] Kokinous, J., Kotz, S. A., Tavano, A., & Schröger, E. (2015). The Role of Emotion in Dynamic Audiovisual Integration of Faces and Voices. Social Cognitive and Affective Neuroscience, 10, 713-720.[CrossRef] [PubMed]
[27] Kreifelts, B., Ethofer, T., Wiegand, A., Brück, C., Wächter, S., Erb, M. et al. (2020). The Neural Correlates of Face-Voice-Integration in Social Anxiety Disorder. Frontiers in Psychiatry, 11, Article 657.[CrossRef] [PubMed]
[28] Laukka, P., Månsson, K. N. T., Cortes, D. S., Manzouri, A., Frick, A., Fredborg, W. et al. (2024). Neural Correlates of Individual Differences in Multimodal Emotion Recognition Ability. Cortex, 175, 1-11.[CrossRef] [PubMed]
[29] Leary, M. R. (1983). Social Anxiousness: The Construct and Its Measurement. Journal of Personality Assessment, 47, 66-75.[CrossRef] [PubMed]
[30] Li, S., Li, Y., Li, R., Li, Y., Wu, J., Yang, W. et al. (2024). The Modulatory Effect of Exogenous Orienting on Audiovisual Emotional Integration: An ERP Study. Journal of Integrative Neuroscience, 23, Article No. 210.[CrossRef] [PubMed]
[31] Liu, P., Sutherland, M., & Pollick, F. E. (2021). Incongruence Effects in Cross-Modal Emotional Processing in Autistic Traits: An fMRI Study. Neuropsychologia, 161, Article 107997.[CrossRef] [PubMed]
[32] Müller, V. I., Habel, U., Derntl, B., Schneider, F., Zilles, K., Turetsky, B. I. et al. (2011). Incongruence Effects in Crossmodal Emotional Integration. NeuroImage, 54, 2257-2266.[CrossRef] [PubMed]
[33] Pan, F., Zhang, L., Ou, Y., & Zhang, X. (2019). The Audio-Visual Integration Effect on Music Emotion: Behavioral and Physiological Evidence. PLOS ONE, 14, e0217040.[CrossRef] [PubMed]
[34] Pan, Z., Liu, X., Luo, Y., & Chen, X. (2017). Emotional Intensity Modulates the Integration of Bimodal Angry Expressions: ERP Evidence. Frontiers in Neuroscience, 11, Article 349.[CrossRef] [PubMed]
[35] Peschard, V., & Philippot, P. (2017). Overestimation of Threat from Neutral Faces and Voices in Social Anxiety. Journal of Behavior Therapy and Experimental Psychiatry, 57, 206-211.[CrossRef] [PubMed]
[36] Proverbio, A. M., & De Benedetto, F. (2018). Auditory Enhancement of Visual Memory Encoding Is Driven by Emotional Content of the Auditory Material and Mediated by Superior Frontal Cortex. Biological Psychology, 132, 164-175.[CrossRef] [PubMed]
[37] Proverbio, A. M., Camporeale, E., & Brusa, A. (2020). Multimodal Recognition of Emotions in Music and Facial Expressions. Frontiers in Human Neuroscience, 14, Article 32.[CrossRef] [PubMed]
[38] Robins, D. L., Hunyadi, E., & Schultz, R. T. (2009). Superior Temporal Activation in Response to Dynamic Audio-Visual Emotional Cues. Brain and Cognition, 69, 269-278.[CrossRef] [PubMed]
[39] Schelenz, P. D., Klasen, M., Reese, B., Regenbogen, C., Wolf, D., Kato, h. Y. et al. (2013). Multisensory Integration of Dynamic Emotional Faces and Voices: Method for Simultaneous EEG-fMRI Measurements. Frontiers in Human Neuroscience, 7, Article 729.[CrossRef] [PubMed]
[40] Schirmer, A., & Adolphs, R. (2017). Emotion Perception from Face, Voice, and Touch: Comparisons and Convergence. Trends in Cognitive Sciences, 21, 216-228.[CrossRef] [PubMed]
[41] Schofield, C. A., Johnson, A. L., Inhoff, A. W., & Coles, M. E. (2012). Social Anxiety and Difficulty Disengaging Threat: Evidence from Eye-Tracking. Cognition & Emotion, 26, 300-311.[CrossRef] [PubMed]
[42] Tottenham, N., Tanaka, J. W., Leon, A. C., McCarry, T., Nurse, M., Hare, T. A. et al. (2009). The NimStim Set of Facial Expressions: Judgments from Untrained Research Participants. Psychiatry Research, 168, 242-249.[CrossRef] [PubMed]
[43] Watson, R., Latinus, M., Noguchi, T., Garrod, O., Crabbe, F., & Belin, P. (2013). Dissociating Task Difficulty from Incongruence in Face-Voice Emotion Integration. Frontiers in Human Neuroscience, 7, Article 744.[CrossRef] [PubMed]
[44] Yang, J., Si, F., & Cao, J. (2025). Electrophysiological Evidence of Processing Social Threat Words in Social Anxiety Participants. BMC Psychiatry, 25, Article No. 821.[CrossRef] [PubMed]
[45] Yeh, P., Geangu, E., & Reid, V. (2016). Coherent Emotional Perception from Body Expressions and the Voice. Neuropsychologia, 91, 99-108.[CrossRef] [PubMed]
[46] Yuan, J., Liu, X., Chen, X., Li, M., Liu, Z., & Li, B. (2026). Audiovisual Integration of Faces and Voices in Chinese Individuals with Social Anxiety. Behavioural Brain Research, 497, Article ID: 115905.[CrossRef