Cited 0 time in
A Survey of Generative Models for Image and Video with Diffusion Model
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Koh, Byoung Soo | - |
| dc.contributor.author | Park, Hyeong Cheol | - |
| dc.contributor.author | Park, Jin Ho | - |
| dc.date.accessioned | 2024-12-09T08:00:07Z | - |
| dc.date.available | 2024-12-09T08:00:07Z | - |
| dc.date.issued | 2024-12 | - |
| dc.identifier.issn | 2192-1962 | - |
| dc.identifier.issn | 2192-1962 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/56328 | - |
| dc.description.abstract | With recent advances in deep learning-based generative models, it is now possible to synthesize realistic data in a diverse domain. One notable method in the generative model is a diffusion-based generative model that generates realistic and high-quality images and videos. Diffusion-based generative model leverages a diffusion process to transform a Gaussian noise distribution into a complex, realistic data distribution. To illustrate the diffusion-based generative models, we give an overview diffusion probabilistic models and denoising diffusion probabilistic models. Especially, we review research that presents new methodologies for image, video, and multimedia contents generation, aiming to understand how those models efficiently learn complex data distribution using various techniques. In the meantime, using multimodal data for training generative models helps them learn more about various representations of complex data distribution, which enhances the generation of diverse images and videos. For the main contribution of this paper, we present several effective methods for synthesizing various types of data using diffusion models and multimodal data, along with their applications. In this context, we believe that presenting how diffusion models have expanded into multimedia generation along with the progression of technological advancements will provide knowledge and inspiration to many researchers. | - |
| dc.format.extent | 20 | - |
| dc.language | 영어 | - |
| dc.language.iso | ENG | - |
| dc.publisher | 한국컴퓨터산업협회 | - |
| dc.title | A Survey of Generative Models for Image and Video with Diffusion Model | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.22967/HCIS.2024.14.069 | - |
| dc.identifier.scopusid | 2-s2.0-85210031386 | - |
| dc.identifier.wosid | 001363019800001 | - |
| dc.identifier.bibliographicCitation | Human-centric Computing and Information Sciences, v.14, pp 1 - 20 | - |
| dc.citation.title | Human-centric Computing and Information Sciences | - |
| dc.citation.volume | 14 | - |
| dc.citation.startPage | 1 | - |
| dc.citation.endPage | 20 | - |
| dc.type.docType | Article | - |
| dc.identifier.kciid | ART003220704 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | scie | - |
| dc.description.journalRegisteredClass | scopus | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.relation.journalResearchArea | Computer Science | - |
| dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
| dc.subject.keywordAuthor | Deep Learning | - |
| dc.subject.keywordAuthor | Generative Model | - |
| dc.subject.keywordAuthor | Diffusion Model | - |
| dc.subject.keywordAuthor | Multimodal Learning | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
