Zhenye Luo, Min Ren, Xuecai Hu, Yongzhen Huang, Li Yao
{"title":"POPDG: Popular 3D Dance Generation with PopDanceSet","authors":"Zhenye Luo, Min Ren, Xuecai Hu, Yongzhen Huang, Li Yao","doi":"arxiv-2405.03178","DOIUrl":null,"url":null,"abstract":"Generating dances that are both lifelike and well-aligned with music\ncontinues to be a challenging task in the cross-modal domain. This paper\nintroduces PopDanceSet, the first dataset tailored to the preferences of young\naudiences, enabling the generation of aesthetically oriented dances. And it\nsurpasses the AIST++ dataset in music genre diversity and the intricacy and\ndepth of dance movements. Moreover, the proposed POPDG model within the iDDPM\nframework enhances dance diversity and, through the Space Augmentation\nAlgorithm, strengthens spatial physical connections between human body joints,\nensuring that increased diversity does not compromise generation quality. A\nstreamlined Alignment Module is also designed to improve the temporal alignment\nbetween dance and music. Extensive experiments show that POPDG achieves SOTA\nresults on two datasets. Furthermore, the paper also expands on current\nevaluation metrics. The dataset and code are available at\nhttps://github.com/Luke-Luo1/POPDG.","PeriodicalId":501178,"journal":{"name":"arXiv - CS - Sound","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Sound","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.03178","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Generating dances that are both lifelike and well-aligned with music
continues to be a challenging task in the cross-modal domain. This paper
introduces PopDanceSet, the first dataset tailored to the preferences of young
audiences, enabling the generation of aesthetically oriented dances. And it
surpasses the AIST++ dataset in music genre diversity and the intricacy and
depth of dance movements. Moreover, the proposed POPDG model within the iDDPM
framework enhances dance diversity and, through the Space Augmentation
Algorithm, strengthens spatial physical connections between human body joints,
ensuring that increased diversity does not compromise generation quality. A
streamlined Alignment Module is also designed to improve the temporal alignment
between dance and music. Extensive experiments show that POPDG achieves SOTA
results on two datasets. Furthermore, the paper also expands on current
evaluation metrics. The dataset and code are available at
https://github.com/Luke-Luo1/POPDG.