Open Access
Eurasip Journal on Audio, Speech, and Music Processing, volume 2025, issue 1, publication number 8
AI-based Chinese-style music generation from video content: a study on cross-modal analysis and generation methods
Moxi Cao
1
,
Jiaxiang Zheng
1
,
Chongbin Zhang
2
1
Department of Global Cultural Convergence, Kangwon National University, Chuncheon, South Korea
2
School of Modern Music and Technology, Nanjing University of the Arts, Nanjing, China
Publication type: Journal Article
Publication date: 2025-02-12
scimago Q2
SJR: 0.414
CiteScore: 4.1
Impact factor: 1.7
ISSN: 16874714, 16874722
Abstract
In recent years, Artificial Intelligence Generated Content (AIGC) technologies have advanced rapidly, with models such as Stable Diffusion and GPT garnering significant attention across various domains. Against this backdrop, AI-driven music composition techniques have also produced significant progress. However, no existing model has yet demonstrated the capability to generate Chinese-style music corresponding to Chinese-style videos. To address this gap, this study proposes a novel Chinese-style video music generation model based on the Latent Diffusion Model (LDM) and Diffusion Transformers (DiT). Experimental results demonstrate that the proposed model generates Chinese-style music from Chinese-style videos and achieves performance comparable to the baseline models in audio quality, distribution fitting, musicality, rhythmic stability, and audio-visual synchronization. These findings indicate that the model captures the stylistic features of Chinese music. This research not only demonstrates the feasibility applications of artificial intelligence in music creation but also provides a new technological approach to preserve and innovate the traditional Chinese music culture in the digital era. Furthermore, it explores new possibilities for the dissemination and innovation of Chinese cultural arts in the digital age.
Found
Are you a researcher?
Create a profile to get free access to personal recommendations for colleagues and new articles.