Detailed Information

Cited 6 time in webofscience Cited 7 time in scopus
Metadata Downloads

MTGEA: A Multimodal Two-Stream GNN Framework for Efficient Point Cloud and Skeleton Data Alignment

Full metadata record
DC Field Value Language
dc.contributor.authorLee, Gawon-
dc.contributor.authorKim, Jihie-
dc.date.accessioned2024-08-08T05:30:38Z-
dc.date.available2024-08-08T05:30:38Z-
dc.date.issued2023-03-
dc.identifier.issn1424-8220-
dc.identifier.issn1424-8220-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/18594-
dc.description.abstractBecause of societal changes, human activity recognition, part of home care systems, has become increasingly important. Camera-based recognition is mainstream but has privacy concerns and is less accurate under dim lighting. In contrast, radar sensors do not record sensitive information, avoid the invasion of privacy, and work in poor lighting. However, the collected data are often sparse. To address this issue, we propose a novel Multimodal Two-stream GNN Framework for Efficient Point Cloud and Skeleton Data Alignment (MTGEA), which improves recognition accuracy through accurate skeletal features from Kinect models. We first collected two datasets using the mmWave radar and Kinect v4 sensors. Then, we used zero-padding, Gaussian Noise (GN), and Agglomerative Hierarchical Clustering (AHC) to increase the number of collected point clouds to 25 per frame to match the skeleton data. Second, we used Spatial Temporal Graph Convolutional Network (ST-GCN) architecture to acquire multimodal representations in the spatio-temporal domain focusing on skeletal features. Finally, we implemented an attention mechanism aligning the two multimodal features to capture the correlation between point clouds and skeleton data. The resulting model was evaluated empirically on human activity data and shown to improve human activity recognition with radar data only. All datasets and codes are available in our GitHub.-
dc.format.extent13-
dc.language영어-
dc.language.isoENG-
dc.publisherMDPI-
dc.titleMTGEA: A Multimodal Two-Stream GNN Framework for Efficient Point Cloud and Skeleton Data Alignment-
dc.typeArticle-
dc.publisher.location스위스-
dc.identifier.doi10.3390/s23052787-
dc.identifier.scopusid2-s2.0-85149804005-
dc.identifier.wosid000948089600001-
dc.identifier.bibliographicCitationSensors, v.23, no.5, pp 1 - 13-
dc.citation.titleSensors-
dc.citation.volume23-
dc.citation.number5-
dc.citation.startPage1-
dc.citation.endPage13-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaChemistry-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaInstruments & Instrumentation-
dc.relation.journalWebOfScienceCategoryChemistry, Analytical-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.relation.journalWebOfScienceCategoryInstruments & Instrumentation-
dc.subject.keywordAuthorhuman activity recognition-
dc.subject.keywordAuthormmWave radar-
dc.subject.keywordAuthorKinect V4 sensor-
dc.subject.keywordAuthorpoint clouds-
dc.subject.keywordAuthorskeleton data-
dc.subject.keywordAuthormultimodal-
dc.subject.keywordAuthortwo stream-
dc.subject.keywordAuthorattention mechanism-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Kim, Ji Hie photo

Kim, Ji Hie
College of Advanced Convergence Engineering (Department of Computer Science and Artificial Intelligence)
Read more

Altmetrics

Total Views & Downloads

BROWSE