Music visualization based on the MIDI specifications for multidimensional polyphonic expression

Z. Xie, Yu Chen
{"title":"Music visualization based on the MIDI specifications for multidimensional polyphonic expression","authors":"Z. Xie, Yu Chen","doi":"10.1109/ICSAI.2017.8248478","DOIUrl":null,"url":null,"abstract":"Music visualization is a branch of information visualization. First, the advantages of MIDI specifications for multidimensional polyphonic expression in “channel per note” were analyzed. Second, the new music visualization mapping model was proposed, which was based on the structural similarity of human and computer recognition in sound and image. The model corresponded to pitch bend, control changes, velocity to hue, saturation, and value respectively. Third, the scheme of music visualization was proposed. The base image was obtained by improving the brightness of original image by 20%–30%; then the influence image was divided into three parts as the left, the middle and the right solid color area, which corresponded to the bass voice, the alto voice and the treble voice respectively; next, music information from different areas were mapped to calculate the HSV values of solid color and then the HSV values were converted to the RGB values; finally, the RGB values of these three parts were glued together and combined with the base image in multiply mode to obtain the final image. Fourth, taking advantage of Cycling'74 Max/Jitter 7.3.4 and optimizing image combination calculation, five experiments were completed under the circumstances of bass voice, alto voice, treble voice, two notes shown in the same area and two notes shown in two different areas. Finally, by calculation of Jit.fpsgui, the final frame rate of the image was around 80fps if RGB values were glued every 16 milliseconds. The research shows that this method can meet real-time needs of music visualization and it can be used for live shows.","PeriodicalId":285726,"journal":{"name":"2017 4th International Conference on Systems and Informatics (ICSAI)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 4th International Conference on Systems and Informatics (ICSAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSAI.2017.8248478","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Music visualization is a branch of information visualization. First, the advantages of MIDI specifications for multidimensional polyphonic expression in “channel per note” were analyzed. Second, the new music visualization mapping model was proposed, which was based on the structural similarity of human and computer recognition in sound and image. The model corresponded to pitch bend, control changes, velocity to hue, saturation, and value respectively. Third, the scheme of music visualization was proposed. The base image was obtained by improving the brightness of original image by 20%–30%; then the influence image was divided into three parts as the left, the middle and the right solid color area, which corresponded to the bass voice, the alto voice and the treble voice respectively; next, music information from different areas were mapped to calculate the HSV values of solid color and then the HSV values were converted to the RGB values; finally, the RGB values of these three parts were glued together and combined with the base image in multiply mode to obtain the final image. Fourth, taking advantage of Cycling'74 Max/Jitter 7.3.4 and optimizing image combination calculation, five experiments were completed under the circumstances of bass voice, alto voice, treble voice, two notes shown in the same area and two notes shown in two different areas. Finally, by calculation of Jit.fpsgui, the final frame rate of the image was around 80fps if RGB values were glued every 16 milliseconds. The research shows that this method can meet real-time needs of music visualization and it can be used for live shows.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于MIDI规范的多维复调表达音乐可视化
音乐可视化是信息可视化的一个分支。首先,分析了MIDI规范在“每音通道”多维复调表达中的优势。其次,提出了基于人与计算机识别在声音和图像上的结构相似性的音乐可视化映射模型。该模型分别对应于俯仰弯曲、控制变化、速度、色调、饱和度和值。第三,提出了音乐可视化方案。将原始图像的亮度提高20% ~ 30%得到基础图像;然后将影响图像分为左、中、右三部分纯色区,分别对应低音、中音和高音;然后,对不同区域的音乐信息进行映射,计算纯色的HSV值,并将HSV值转换为RGB值;最后,将这三部分的RGB值粘合在一起,与底图进行乘法组合,得到最终的图像。第四,利用Cycling’74 Max/Jitter 7.3.4,优化图像组合计算,在低音、中音、高音、两个音符显示在同一区域和两个音符显示在两个不同区域的情况下完成5个实验。最后,通过计算Jit。在fpsgui中,如果RGB值每16毫秒粘合一次,那么图像的最终帧速率约为80fps。研究表明,该方法能够满足音乐可视化的实时性需求,可用于现场演出。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Adaptive coverage control with Guaranteed Power Voronoi Diagrams Gray relativity analysis used to track association in passive sonar system Music visualization based on the MIDI specifications for multidimensional polyphonic expression Modeling of a data modification cyber-attack in an IEC 61850 scenario using stochastic colored Petri Nets Four nonlinear multi-input multi-output ADHDP constructions and algorithms based on topology principle
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1