Auditory signal dominates visual in the perception of emotional social interactions

L. Piwek, K. Petrini, F. Pollick
{"title":"Auditory signal dominates visual in the perception of emotional social interactions","authors":"L. Piwek, K. Petrini, F. Pollick","doi":"10.1163/187847612X647450","DOIUrl":null,"url":null,"abstract":"Multimodal perception of emotions has been typically examined using displays of a solitary character (e.g., the face–voice and/or body–sound of one actor). We extend investigation to more complex, dyadic point-light displays combined with speech. A motion and voice capture system was used to record twenty actors interacting in couples with happy, angry and neutral emotional expressions. The obtained stimuli were validated in a pilot study and used in the present study to investigate multimodal perception of emotional social interactions. Participants were required to categorize happy and angry expressions displayed visually, auditorily, or using emotionally congruent and incongruent bimodal displays. In a series of cross-validation experiments we found that sound dominated the visual signal in the perception of emotional social interaction. Although participants’ judgments were faster in the bimodal condition, the accuracy of judgments was similar for both bimodal and auditory-only conditions. When participants watched emotionally mismatched bimodal displays, they predominantly oriented their judgments towards the auditory rather than the visual signal. This auditory dominance persisted even when the reliability of auditory signal was decreased with noise, although visual information had some effect on judgments of emotions when it was combined with a noisy auditory signal. Our results suggest that when judging emotions from observed social interaction, we rely primarily on vocal cues from the conversation, rather then visual cues from their body movement.","PeriodicalId":49553,"journal":{"name":"Seeing and Perceiving","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2012-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1163/187847612X647450","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Seeing and Perceiving","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1163/187847612X647450","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Multimodal perception of emotions has been typically examined using displays of a solitary character (e.g., the face–voice and/or body–sound of one actor). We extend investigation to more complex, dyadic point-light displays combined with speech. A motion and voice capture system was used to record twenty actors interacting in couples with happy, angry and neutral emotional expressions. The obtained stimuli were validated in a pilot study and used in the present study to investigate multimodal perception of emotional social interactions. Participants were required to categorize happy and angry expressions displayed visually, auditorily, or using emotionally congruent and incongruent bimodal displays. In a series of cross-validation experiments we found that sound dominated the visual signal in the perception of emotional social interaction. Although participants’ judgments were faster in the bimodal condition, the accuracy of judgments was similar for both bimodal and auditory-only conditions. When participants watched emotionally mismatched bimodal displays, they predominantly oriented their judgments towards the auditory rather than the visual signal. This auditory dominance persisted even when the reliability of auditory signal was decreased with noise, although visual information had some effect on judgments of emotions when it was combined with a noisy auditory signal. Our results suggest that when judging emotions from observed social interaction, we rely primarily on vocal cues from the conversation, rather then visual cues from their body movement.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
在情感社会互动的感知中,听觉信号占主导地位
情感的多模态感知通常是通过单个角色的表现(例如,一个演员的脸-声音和/或肢体-声音)来检验的。我们将研究扩展到更复杂的,结合语音的二元点光显示。一个动作和声音捕捉系统记录了20位演员在情侣中以快乐、愤怒和中性的情绪表情互动。获得的刺激在一项初步研究中得到验证,并在本研究中用于调查情感社会互动的多模态感知。参与者被要求对视觉上、听觉上或情感上一致和不一致的双峰显示的快乐和愤怒的表情进行分类。在一系列的交叉验证实验中,我们发现声音在情感社会互动的感知中占主导地位。虽然参与者的判断在双峰条件下更快,但判断的准确性在双峰条件和听觉条件下是相似的。当参与者观看情感不匹配的双峰显示时,他们主要将判断导向听觉信号而不是视觉信号。即使当听觉信号的可靠性因噪音而降低时,这种听觉优势仍然存在,尽管视觉信息与嘈杂的听觉信号结合在一起时对情绪的判断有一定影响。我们的研究结果表明,当我们从观察到的社交互动中判断情绪时,我们主要依赖于对话中的声音线索,而不是来自他们身体动作的视觉线索。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Seeing and Perceiving
Seeing and Perceiving BIOPHYSICS-PSYCHOLOGY
自引率
0.00%
发文量
0
审稿时长
>12 weeks
期刊最新文献
Chapter ten. Color Vision Chapter six. Brightness Constancy Chapter One. Our Idea of the Physical World Chapter nine. Optometrists, Ophthalmologists, Opticians: What They Do Chapter seven. Why the Rate of Unbleaching is Important
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1