SlimDeblurGAN-Based Motion Deblurring and Marker Detection for Autonomous Drone Landingopen access
- Authors
- Noi Quang Truong; Lee, Young Won; Owais, Muhammad; Dat Tien Nguyen; Batchuluun, Ganbayar; Tuyen Danh Pham; Park, Kang Ryoung
- Issue Date
- Jul-2020
- Publisher
- MDPI
- Keywords
- unmanned aerial vehicle; autonomous landing; deep-learning-based motion deblurring and marker detection; network slimming; pruning model
- Citation
- SENSORS, v.20, no.14, pp 1 - 33
- Pages
- 33
- Indexed
- SCIE
SCOPUS
- Journal Title
- SENSORS
- Volume
- 20
- Number
- 14
- Start Page
- 1
- End Page
- 33
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/18727
- DOI
- 10.3390/s20143918
- ISSN
- 1424-8220
1424-3210
- Abstract
- Deep learning-based marker detection for autonomous drone landing is widely studied, due to its superior detection performance. However, no study was reported to address non-uniform motion-blurred input images, and most of the previous handcrafted and deep learning-based methods failed to operate with these challenging inputs. To solve this problem, we propose a deep learning-based marker detection method for autonomous drone landing, by (1) introducing a two-phase framework of deblurring and object detection, by adopting a slimmed version of deblur generative adversarial network (DeblurGAN) model and a You only look once version 2 (YOLOv2) detector, respectively, and (2) considering the balance between the processing time and accuracy of the system. To this end, we propose a channel-pruning framework for slimming the DeblurGAN model called SlimDeblurGAN, without significant accuracy degradation. The experimental results on the two datasets showed that our proposed method exhibited higher performance and greater robustness than the previous methods, in both deburring and marker detection.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Department of Electronics and Electrical Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.