Heterogeneous AI Music Generation Technology Integrating Fine-Grained Control

As artificial intelligence algorithms continue to advance, researchers have increasingly harnessed their capabilities to generate music that resonates with human emotions, offering a novel means of alleviating the escalating pressures of contemporary life. To tackle the persistent issue of low accur...

Full description

Saved in:
Bibliographic Details
Main Authors: Hongtao Wang, Li Gong
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/11096601/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:As artificial intelligence algorithms continue to advance, researchers have increasingly harnessed their capabilities to generate music that resonates with human emotions, offering a novel means of alleviating the escalating pressures of contemporary life. To tackle the persistent issue of low accuracy in current emotion recognition and music generation systems, an innovative approach was proposed that fused a graph convolutional neural network with a channel attention mechanism for emotion recognition. This integrated model was subsequently paired with a Transformer architecture, creating a sophisticated framework capable of fine-grained control and heterogeneous music generation. In comparing the performance of the emotion recognition model against other leading models, the results underscored its exceptional accuracy across five distinct electroencephalogram signal bands: 97.3%, 95.8%, 96.9%, 98.4%, and 97.6%, respectively. Crucially, all these accuracy metrics exceeded the 95% benchmark, clearly demonstrating superiority over the comparative models. Additionally, a rigorous performance assessment was conducted to evaluate the music generation model’s capabilities against alternative approaches. The findings revealed that the suggested model achieved an average mean square error of 0.27 and an average root mean square error of 0.24. These error rates were notably lower than those of the competing models, highlighting the enhanced precision and fidelity of the music generated. Together, these results validated the effectiveness of both the emotion recognition and music generation models developed in this research. This research not only propelled the existing frontiers of emotion detection and musical composition forward, but also laid a robust theoretical framework to facilitate subsequent investigations into the emerging field of emotion-aware music generation.
ISSN:2169-3536