Predicting Post-Liposuction Body Shape Using RGB Image-to-Image Translationopen access
- Authors
- Kim, Minji; Byeon, Jiseong; Chang, Jihun; Youm, Sekyoung
- Issue Date
- Apr-2025
- Publisher
- MDPI
- Keywords
- prediction of liposuction outcome; image-to-image translation; Pix2Pix; deep learning; GAN
- Citation
- Applied Sciences, v.15, no.9, pp 1 - 22
- Pages
- 22
- Indexed
- SCIE
SCOPUS
- Journal Title
- Applied Sciences
- Volume
- 15
- Number
- 9
- Start Page
- 1
- End Page
- 22
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/58414
- DOI
- 10.3390/app15094787
- ISSN
- 2076-3417
2076-3417
- Abstract
- The growing interest in weight management has elevated the popularity of liposuction. Individuals deciding whether to undergo liposuction must rely on a doctor's subjective projections or surgical outcomes for other people to gauge how their own body shape will change. However, such predictions may not be accurate. Although deep learning technology has recently achieved breakthroughs in analyzing medical images and rendering diagnoses, predicting surgical outcomes based on medical images outside clinical settings remains challenging. Hence, this study aimed to develop a method for predicting body shape changes after liposuction using only images of the subject's own body. To achieve this, we utilize data augmentation based on a conditional continuous Generative Adversarial Network (CcGAN), which generates realistic synthetic data conditioned on continuous variables. Additionally, we modify the loss function of Pix2Pix-a supervised image-to-image translation technique based on Generative Adversarial Networks (GANs)-to enhance prediction quality. Our approach quantitatively and qualitatively demonstrates that accurate, intuitive predictions before liposuction are possible.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Department of Industrial and Systems Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.