Musical training generalises across modalities and reveals efficient and adaptive mechanisms for judging temporal intervals

David Aagten-Murphy, G. Cappagli, D. Burr
{"title":"Musical training generalises across modalities and reveals efficient and adaptive mechanisms for judging temporal intervals","authors":"David Aagten-Murphy, G. Cappagli, D. Burr","doi":"10.1163/187847612X646361","DOIUrl":null,"url":null,"abstract":"Expert musicians are able to accurately and consistently time their actions during a musical performance. We investigated how musical expertise influences the ability to reproduce auditory intervals and how this generalises to vision in a ‘ready-set-go’ paradigm. Subjects reproduced time intervals drawn from distributions varying in total length (176, 352 or 704 ms) or in the number of discrete intervals within the total length (3, 5, 11 or 21 discrete intervals). Overall musicians performed more veridically than non-musicians, and all subjects reproduced auditory-defined intervals more accurately than visually-defined intervals. However non-musicians, particularly with visual intervals, consistently exhibited a substantial and systematic regression towards the mean of the interval. When subjects judged intervals from distributions of longer total length they tended to exhibit more regression towards the mean, while the ability to discriminate between discrete intervals within the distribution had little influence on subject error. These results are consistent with a Bayesian model which minimizes reproduction errors by incorporating a central tendency prior weighted by the subject’s own temporal precision relative to the current intervals distribution (Cicchini et al., 2012; Jazayeri and Shadlen, 2010). Finally a strong correlation was observed between all durations of formal musical training and total reproduction errors in both modalities (accounting for 30% of the variance). Taken together these results demonstrate that formal musical training improves temporal reproduction, and that this improvement transfers from audition to vision. They further demonstrate the flexibility of sensorimotor mechanisms in adapting to different task conditions to minimise temporal estimation errors.","PeriodicalId":49553,"journal":{"name":"Seeing and Perceiving","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2012-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1163/187847612X646361","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Seeing and Perceiving","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1163/187847612X646361","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Expert musicians are able to accurately and consistently time their actions during a musical performance. We investigated how musical expertise influences the ability to reproduce auditory intervals and how this generalises to vision in a ‘ready-set-go’ paradigm. Subjects reproduced time intervals drawn from distributions varying in total length (176, 352 or 704 ms) or in the number of discrete intervals within the total length (3, 5, 11 or 21 discrete intervals). Overall musicians performed more veridically than non-musicians, and all subjects reproduced auditory-defined intervals more accurately than visually-defined intervals. However non-musicians, particularly with visual intervals, consistently exhibited a substantial and systematic regression towards the mean of the interval. When subjects judged intervals from distributions of longer total length they tended to exhibit more regression towards the mean, while the ability to discriminate between discrete intervals within the distribution had little influence on subject error. These results are consistent with a Bayesian model which minimizes reproduction errors by incorporating a central tendency prior weighted by the subject’s own temporal precision relative to the current intervals distribution (Cicchini et al., 2012; Jazayeri and Shadlen, 2010). Finally a strong correlation was observed between all durations of formal musical training and total reproduction errors in both modalities (accounting for 30% of the variance). Taken together these results demonstrate that formal musical training improves temporal reproduction, and that this improvement transfers from audition to vision. They further demonstrate the flexibility of sensorimotor mechanisms in adapting to different task conditions to minimise temporal estimation errors.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
音乐训练概括了各种模态,揭示了判断时间间隔的有效和适应性机制
专业的音乐家能够在音乐表演中准确而一致地计时他们的动作。我们调查了音乐专业知识如何影响再现听觉间隔的能力,以及这种能力如何在“准备-开始”范式中推广到视觉。受试者再现从总长度(176、352或704毫秒)或总长度内的离散间隔数(3、5、11或21个离散间隔)的分布中提取的时间间隔。总的来说,音乐家比非音乐家表现得更真实,所有的受试者都比视觉定义的音程更准确地再现了听觉定义的音程。然而,非音乐家,特别是在视觉音程方面,始终表现出对音程平均值的实质性和系统性的回归。当受试者从较长的总长度分布中判断区间时,他们倾向于向均值回归,而在分布中区分离散区间的能力对受试者误差的影响很小。这些结果与贝叶斯模型一致,该模型通过纳入由受试者自身相对于当前间隔分布的时间精度加权的集中趋势先验来最大限度地减少再现误差(Cicchini et al., 2012;Jazayeri和Shadlen, 2010)。最后,在正式音乐训练的所有持续时间和两种模式的总再现误差之间观察到很强的相关性(占方差的30%)。综上所述,这些结果表明,正规的音乐训练改善了时间再现,并且这种改善从听力转移到视觉。他们进一步证明了感觉运动机制在适应不同任务条件以最小化时间估计误差方面的灵活性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Seeing and Perceiving
Seeing and Perceiving BIOPHYSICS-PSYCHOLOGY
自引率
0.00%
发文量
0
审稿时长
>12 weeks
期刊最新文献
Chapter ten. Color Vision Chapter six. Brightness Constancy Chapter One. Our Idea of the Physical World Chapter nine. Optometrists, Ophthalmologists, Opticians: What They Do Chapter seven. Why the Rate of Unbleaching is Important
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1