Detailed Information

Cited 8 time in webofscience Cited 10 time in scopus
Metadata Downloads

MelodyDiffusion: Chord-Conditioned Melody Generation Using a Transformer-Based Diffusion Model

Full metadata record
DC Field Value Language
dc.contributor.authorLi, Shuyu-
dc.contributor.authorSung, Yunsick-
dc.date.accessioned2024-08-08T07:31:38Z-
dc.date.available2024-08-08T07:31:38Z-
dc.date.issued2023-04-
dc.identifier.issn2227-7390-
dc.identifier.issn2227-7390-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/19857-
dc.description.abstractArtificial intelligence, particularly machine learning, has begun to permeate various real-world applications and is continually being explored in automatic music generation. The approaches to music generation can be broadly divided into two categories: rule-based and data-driven methods. Rule-based approaches rely on substantial prior knowledge and may struggle to handle large datasets, whereas data-driven approaches can solve these problems and have become increasingly popular. However, data-driven approaches still face challenges such as the difficulty of considering long-distance dependencies when handling discrete-sequence data and convergence during model training. Although the diffusion model has been introduced as a generative model to solve the convergence problem in generative adversarial networks, it has not yet been applied to discrete-sequence data. This paper proposes a transformer-based diffusion model known as MelodyDiffusion to handle discrete musical data and realize chord-conditioned melody generation. MelodyDiffusion replaces the U-nets used in traditional diffusion models with transformers to consider the long-distance dependencies using attention and parallel mechanisms. Moreover, a transformer-based encoder is designed to extract contextual information from chords as a condition to guide melody generation. MelodyDiffusion can automatically generate diverse melodies based on the provided chords in practical applications. The evaluation experiments, in which Hits@k was used as a metric to evaluate the restored melodies, demonstrate that the large-scale version of MelodyDiffusion achieves an accuracy of 72.41% (k = 1).-
dc.format.extent15-
dc.language영어-
dc.language.isoENG-
dc.publisherMDPI-
dc.titleMelodyDiffusion: Chord-Conditioned Melody Generation Using a Transformer-Based Diffusion Model-
dc.typeArticle-
dc.publisher.location스위스-
dc.identifier.doi10.3390/math11081915-
dc.identifier.scopusid2-s2.0-85153701007-
dc.identifier.wosid000976445800001-
dc.identifier.bibliographicCitationMathematics, v.11, no.8, pp 1 - 15-
dc.citation.titleMathematics-
dc.citation.volume11-
dc.citation.number8-
dc.citation.startPage1-
dc.citation.endPage15-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaMathematics-
dc.relation.journalWebOfScienceCategoryMathematics-
dc.subject.keywordAuthormelody generation-
dc.subject.keywordAuthorconditional generation-
dc.subject.keywordAuthordiffusion model-
dc.subject.keywordAuthortransformer-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Sung, Yunsick photo

Sung, Yunsick
College of Advanced Convergence Engineering (Department of Computer Science and Artificial Intelligence)
Read more

Altmetrics

Total Views & Downloads

BROWSE