Abstract
The classification of musical emotions is crucial for the indexing, structuring, searching, and recommending of tracks and albums across various music platforms. Consequently, the automated categorization of musical emotions has become a vital element in nearly all music applications. Recent studies have mainly concentrated on utilizing textual, audio, or multimodal data for genre classification, frequently neglecting the impact of singers, composers, and listener preferences. In practice, composers possess unique compositional styles, listeners have varied musical preferences, and singers focus on particular music genres. These different viewpoints offer significant insights into the classification of musical emotions, greatly enhancing the effectiveness of classification performance. In this paper, we introduce a novel heterogeneous graph neural network (HGN) that models the relationships of music emotion preferences among singers, composers, and listeners, in order to generate accurate node feature representations for downstream tasks. The experimental results show that our model significantly outperforms current state-of-the-art (SOTA) methods on two datasets for music emotion classification.
| Original language | English |
|---|---|
| Pages (from-to) | 76473-76480 |
| Number of pages | 8 |
| Journal | IEEE Access |
| Volume | 13 |
| Early online date | 18 Apr 2025 |
| DOIs | |
| Publication status | E-pub ahead of print - 18 Apr 2025 |
Bibliographical note
Publisher Copyright:© 2013 IEEE. 2025 The Authors. This work is licensed under a Creative Commons Attribution 4.0 License.
For more information, see https://creativecommons.org/licenses/by/4.0/
Keywords
- Music emotion
- classification algorithm
- heterogeneous graph neural networks
ASJC Scopus subject areas
- General Computer Science
- General Materials Science
- General Engineering