Guilhem Marion , Fei Gao , Benjamin P. Gold , Giovanni M. Di Liberto , Shihab Shamma
{"title":"IDyOMpy:一个新的基于python的音乐期望值统计分析模型。","authors":"Guilhem Marion , Fei Gao , Benjamin P. Gold , Giovanni M. Di Liberto , Shihab Shamma","doi":"10.1016/j.jneumeth.2024.110347","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>: IDyOM (Information Dynamics of Music) is the statistical model of music the most used in the community of neuroscience of music. It has been shown to allow for significant correlations with EEG (Marion, 2021), ECoG (Di Liberto, 2020) and fMRI (Cheung, 2019) recordings of human music listening. The language used for IDyOM -Lisp- is not very familiar to the neuroscience community and makes this model hard to use and more importantly to modify.</div></div><div><h3>New method</h3><div>: IDyOMpy is a new Python re-implementation and extension of IDyOM. This new model allows for computing the information content and entropy for each melody note after training on a corpus of melodies. In addition to those features, two new features are presented: probability estimation of silences and enculturation modeling.</div></div><div><h3>Results</h3><div>: We first describe the mathematical details of the implementation. We extensively compare the two models and show that they generate very similar outputs. We also support the validity of IDyOMpy by using its output to replicate previous EEG and behavioral results that relied on the original Lisp version (Gold, 2019; Di Liberto, 2020; Marion, 2021). Finally, it reproduced the computation of cultural distances between two different datasets as described in previous studies (Pearce, 2018).</div></div><div><h3>Comparison with existing methods and Conclusions</h3><div>: Our model replicates the previous behaviors of IDyOM in a modern and easy-to-use language -Python. In addition, more features are presented. We deeply think this new version will be of great use to the community of neuroscience of music.</div></div>","PeriodicalId":16415,"journal":{"name":"Journal of Neuroscience Methods","volume":"415 ","pages":"Article 110347"},"PeriodicalIF":2.7000,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"IDyOMpy: A new Python-based model for the statistical analysis of musical expectations\",\"authors\":\"Guilhem Marion , Fei Gao , Benjamin P. Gold , Giovanni M. Di Liberto , Shihab Shamma\",\"doi\":\"10.1016/j.jneumeth.2024.110347\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Background</h3><div>: IDyOM (Information Dynamics of Music) is the statistical model of music the most used in the community of neuroscience of music. It has been shown to allow for significant correlations with EEG (Marion, 2021), ECoG (Di Liberto, 2020) and fMRI (Cheung, 2019) recordings of human music listening. The language used for IDyOM -Lisp- is not very familiar to the neuroscience community and makes this model hard to use and more importantly to modify.</div></div><div><h3>New method</h3><div>: IDyOMpy is a new Python re-implementation and extension of IDyOM. This new model allows for computing the information content and entropy for each melody note after training on a corpus of melodies. In addition to those features, two new features are presented: probability estimation of silences and enculturation modeling.</div></div><div><h3>Results</h3><div>: We first describe the mathematical details of the implementation. We extensively compare the two models and show that they generate very similar outputs. We also support the validity of IDyOMpy by using its output to replicate previous EEG and behavioral results that relied on the original Lisp version (Gold, 2019; Di Liberto, 2020; Marion, 2021). Finally, it reproduced the computation of cultural distances between two different datasets as described in previous studies (Pearce, 2018).</div></div><div><h3>Comparison with existing methods and Conclusions</h3><div>: Our model replicates the previous behaviors of IDyOM in a modern and easy-to-use language -Python. In addition, more features are presented. We deeply think this new version will be of great use to the community of neuroscience of music.</div></div>\",\"PeriodicalId\":16415,\"journal\":{\"name\":\"Journal of Neuroscience Methods\",\"volume\":\"415 \",\"pages\":\"Article 110347\"},\"PeriodicalIF\":2.7000,\"publicationDate\":\"2024-12-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Neuroscience Methods\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0165027024002929\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"BIOCHEMICAL RESEARCH METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Neuroscience Methods","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0165027024002929","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"BIOCHEMICAL RESEARCH METHODS","Score":null,"Total":0}
引用次数: 0
摘要
背景:IDyOM (Information Dynamics of Music)是音乐神经科学领域使用最多的音乐统计模型。研究表明,它与脑电图(Marion, 2021)、脑电图(Di Liberto, 2020)和功能磁共振成像(b张,2019)记录的人类音乐听力存在显著相关性。用于IDyOM的语言——lisp——对于神经科学社区来说不是很熟悉,这使得这个模型很难使用,更重要的是难以修改。新方法:IDyOMpy是对IDyOM的一个新的Python重新实现和扩展。这个新模型允许在旋律语料库上训练后计算每个旋律音符的信息内容和熵。在此基础上,提出了两个新的特征:沉默概率估计和适应建模。结果:我们首先描述了实现的数学细节。我们广泛地比较了这两个模型,并表明它们产生非常相似的输出。我们还通过使用IDyOMpy的输出来复制依赖于原始Lisp版本的先前EEG和行为结果来支持IDyOMpy的有效性(Gold, 2019;《自由》,2020;马里昂,2021)。最后,它再现了先前研究中描述的两个不同数据集之间文化距离的计算(即Pearce, 2018)。与现有方法和结论的比较:我们的模型在现代和易于使用的语言-Python中复制了IDyOM的先前行为。此外,还提出了更多的特性。我们深信这个新版本将对音乐的神经科学社区有很大的用处。
IDyOMpy: A new Python-based model for the statistical analysis of musical expectations
Background
: IDyOM (Information Dynamics of Music) is the statistical model of music the most used in the community of neuroscience of music. It has been shown to allow for significant correlations with EEG (Marion, 2021), ECoG (Di Liberto, 2020) and fMRI (Cheung, 2019) recordings of human music listening. The language used for IDyOM -Lisp- is not very familiar to the neuroscience community and makes this model hard to use and more importantly to modify.
New method
: IDyOMpy is a new Python re-implementation and extension of IDyOM. This new model allows for computing the information content and entropy for each melody note after training on a corpus of melodies. In addition to those features, two new features are presented: probability estimation of silences and enculturation modeling.
Results
: We first describe the mathematical details of the implementation. We extensively compare the two models and show that they generate very similar outputs. We also support the validity of IDyOMpy by using its output to replicate previous EEG and behavioral results that relied on the original Lisp version (Gold, 2019; Di Liberto, 2020; Marion, 2021). Finally, it reproduced the computation of cultural distances between two different datasets as described in previous studies (Pearce, 2018).
Comparison with existing methods and Conclusions
: Our model replicates the previous behaviors of IDyOM in a modern and easy-to-use language -Python. In addition, more features are presented. We deeply think this new version will be of great use to the community of neuroscience of music.
期刊介绍:
The Journal of Neuroscience Methods publishes papers that describe new methods that are specifically for neuroscience research conducted in invertebrates, vertebrates or in man. Major methodological improvements or important refinements of established neuroscience methods are also considered for publication. The Journal''s Scope includes all aspects of contemporary neuroscience research, including anatomical, behavioural, biochemical, cellular, computational, molecular, invasive and non-invasive imaging, optogenetic, and physiological research investigations.