基于改进复合词的多乐器音乐生成模型

Yi-Jr Liao, Wang Yue, Yuqing Jian, Zijun Wang, Yuchong Gao, Chenhao Lu
{"title":"基于改进复合词的多乐器音乐生成模型","authors":"Yi-Jr Liao, Wang Yue, Yuqing Jian, Zijun Wang, Yuchong Gao, Chenhao Lu","doi":"10.1109/ICMEW56448.2022.9859531","DOIUrl":null,"url":null,"abstract":"In this work, we address the task of multi-instrument music generation. Notably, along with the development of artificial neural networks, deep learning has become a leading technique to accelerate the automatic music generation and is featured in many previous papers like MuseGan[1], MusicBert[2], and PopMAG[3]. However, seldom of them implement a well-designed representation of multi-instrumental music, and no model perfectly introduces a prior knowledge of music theory. In this paper, we leverage the Compound Word[4] and R-drop[5] method to work on multi-instrument music generation tasks. Objective and subjective evaluations show that the generated music has cost less training time, and achieved prominent music quality.","PeriodicalId":106759,"journal":{"name":"2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MICW: A Multi-Instrument Music Generation Model Based on the Improved Compound Word\",\"authors\":\"Yi-Jr Liao, Wang Yue, Yuqing Jian, Zijun Wang, Yuchong Gao, Chenhao Lu\",\"doi\":\"10.1109/ICMEW56448.2022.9859531\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this work, we address the task of multi-instrument music generation. Notably, along with the development of artificial neural networks, deep learning has become a leading technique to accelerate the automatic music generation and is featured in many previous papers like MuseGan[1], MusicBert[2], and PopMAG[3]. However, seldom of them implement a well-designed representation of multi-instrumental music, and no model perfectly introduces a prior knowledge of music theory. In this paper, we leverage the Compound Word[4] and R-drop[5] method to work on multi-instrument music generation tasks. Objective and subjective evaluations show that the generated music has cost less training time, and achieved prominent music quality.\",\"PeriodicalId\":106759,\"journal\":{\"name\":\"2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)\",\"volume\":\"98 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-07-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMEW56448.2022.9859531\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMEW56448.2022.9859531","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在这项工作中,我们解决了多乐器音乐生成的任务。值得注意的是,随着人工神经网络的发展,深度学习已经成为加速自动音乐生成的主要技术,并且在许多先前的论文中都有介绍,如MuseGan[1], MusicBert[2]和PopMAG[3]。然而,它们很少实现多乐器音乐的精心设计的表示,也没有一个模型完美地引入了音乐理论的先验知识。在本文中,我们利用Compound Word[4]和R-drop[5]方法来完成多乐器音乐生成任务。客观评价和主观评价表明,生成的音乐训练时间少,音乐质量突出。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
MICW: A Multi-Instrument Music Generation Model Based on the Improved Compound Word
In this work, we address the task of multi-instrument music generation. Notably, along with the development of artificial neural networks, deep learning has become a leading technique to accelerate the automatic music generation and is featured in many previous papers like MuseGan[1], MusicBert[2], and PopMAG[3]. However, seldom of them implement a well-designed representation of multi-instrumental music, and no model perfectly introduces a prior knowledge of music theory. In this paper, we leverage the Compound Word[4] and R-drop[5] method to work on multi-instrument music generation tasks. Objective and subjective evaluations show that the generated music has cost less training time, and achieved prominent music quality.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Emotional Quality Evaluation for Generated Music Based on Emotion Recognition Model Bottleneck Detection in Crowded Video Scenes Utilizing Lagrangian Motion Analysis Via Density and Arc Length Measures Efficient Topology Coding and Payload Partitioning Techniques for Neural Network Compression (NNC) Standard Exploring Multisensory Feedback for Virtual Reality Relaxation A Unified Video Summarization for Video Anomalies Through Deep Learning
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1