Cited 0 time in
Type-based mixture of experts and semi-supervised multi-task pre-training for symbolic music
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Li, Shuyu | - |
| dc.contributor.author | Sung, Yunsick | - |
| dc.date.accessioned | 2025-07-07T07:30:15Z | - |
| dc.date.available | 2025-07-07T07:30:15Z | - |
| dc.date.issued | 2025-11 | - |
| dc.identifier.issn | 0957-4174 | - |
| dc.identifier.issn | 1873-6793 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/58628 | - |
| dc.description.abstract | In the rapidly evolving field of AI-driven music applications, there is a growing interest in the understanding and generation of symbolic music (e.g., MIDI). Symbolic music, unlike audio waveforms, contains discrete representations of musical elements, making it both a detailed and challenging domain for AI models to process. While pre-training techniques from natural language processing have been adapted for music-related tasks, these pre-trained models often struggle with the hierarchical and polyphonic characteristics of symbolic music. To overcome these problems, a method is proposed comprising two components, a foundational model named type-based mixture of experts (TypeMoE) and a semi-supervised multi-task pre-training (SS-MTP) strategy. Type-MoE captures fine-grained musical features more effectively by dynamically activating specialized experts for different event types, while SS-MTP covers tasks including key-signature recognition, time-signature recognition, and causal language modeling. Unlike purely self-supervised approaches, SS-MTP utilizes a small amount of labeled data alongside extensive unlabeled data, enabling structural representation learning and promoting efficient knowledge sharing across tasks. Experimental results showed that TypeMoE, when pre-trained with the SS-MTP strategy, outperformed baseline models in both music understanding and generation tasks. Specifically, it achieved 71.80% accuracy in genre classification and 76.79% in emotion classification. For music generation, it outperformed baselines with 54.24% Hits@1 and 0.7521 BLEU-2 in continue generation, and 75.79% Hits@1 and 0.8757 BLEU-2 in conditional generation. Additionally, it obtained a CLAP-based semantic alignment score of 0.24. | - |
| dc.format.extent | 14 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | Elsevier Ltd. | - |
| dc.title | Type-based mixture of experts and semi-supervised multi-task pre-training for symbolic music | - |
| dc.type | Article | - |
| dc.publisher.location | 네델란드 | - |
| dc.identifier.doi | 10.1016/j.eswa.2025.128613 | - |
| dc.identifier.scopusid | 2-s2.0-105008489168 | - |
| dc.identifier.wosid | 001519870200008 | - |
| dc.identifier.bibliographicCitation | Expert Systems with Applications, v.292, pp 1 - 14 | - |
| dc.citation.title | Expert Systems with Applications | - |
| dc.citation.volume | 292 | - |
| dc.citation.startPage | 1 | - |
| dc.citation.endPage | 14 | - |
| dc.type.docType | Article | - |
| dc.description.isOpenAccess | Y | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.relation.journalResearchArea | Computer Science | - |
| dc.relation.journalResearchArea | Engineering | - |
| dc.relation.journalResearchArea | Operations Research & Management Science | - |
| dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
| dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
| dc.relation.journalWebOfScienceCategory | Operations Research & Management Science | - |
| dc.subject.keywordPlus | MODEL | - |
| dc.subject.keywordAuthor | Symbolic music | - |
| dc.subject.keywordAuthor | Semi-supervised learning | - |
| dc.subject.keywordAuthor | Multi-task | - |
| dc.subject.keywordAuthor | Mixture of experts | - |
| dc.subject.keywordAuthor | Pre-training | - |
| dc.subject.keywordAuthor | Fine-tuning | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
