Hybrid deep convolutional networks for the autonomous damage diagnosis of laminated composite structures
- Authors
- Azad, Muhammad Muzammil; Kim, Heung Soo
- Issue Date
- Feb-2024
- Publisher
- Elsevier BV
- Keywords
- Damage detection; Hybrid method; Laminated composites; SVM; Convolutional neural network; Convolutional auto -encoder
- Citation
- Composite Structures, v.329, pp 1 - 14
- Pages
- 14
- Indexed
- SCIE
SCOPUS
- Journal Title
- Composite Structures
- Volume
- 329
- Start Page
- 1
- End Page
- 14
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/20983
- DOI
- 10.1016/j.compstruct.2023.117792
- ISSN
- 0263-8223
1879-1085
- Abstract
- This article presents a robust autonomous damage diagnosis method using hybrid deep convolutional networks for the damage diagnosis of laminated composite structures. Inspired by the potential of deep learning models to autonomously extract deep discriminative features and machine learning models that provide better diagnosis on limited data, the current research integrates deep convolutional networks, namely convolutional neural networks (CNN) and convolutional autoencoder (CAE), with support vector machines (SVM) to build hybrid damage detection models. The proposed hybrid models incorporate the advantages of both convolutional operations to extract deep features, and SVM to diagnose using limited feature data. The proposed hybrid models are validated using random vibrational signals for one healthy and two delamination states of laminated composites. The results showed improved damage detection performance compared to the conventional methods, with lower computational costs. Additionally, the hybrid methods autonomously extracted deep discriminative features, eliminating the need for manual damage-sensitive feature extraction.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > Department of Mechanical, Robotics and Energy Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.