Xuanchen Wang, Heng Wang, Dongnan Liu, Weidong Cai
TL;DR通过音乐作为条件输入,直接从静态图像中生成舞蹈视频的Dance Any Beat Diffusion模型引入了图像到视频生成原则,采用了音乐作为图像到视频生成的条件因素。
Abstract
The task of generating dance from music is crucial, yet current methods, which mainly produce joint sequences, lead to outputs that lack intuitiveness and complicate data collection due to the necessity for precise joint annotations. We introduce a Dance Any Beat Diffusion model, namely DabFusion, that employs music as a conditional input to directly create