Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Semantic-Guided Spatial and Temporal Fusion Framework for Enhancing Monocular Video Depth Estimation

Full metadata record
DC Field Value Language
dc.contributor.authorKim, Hyunsu-
dc.contributor.authorLee, Yeongseop-
dc.contributor.authorKo, Hyunseong-
dc.contributor.authorJeong, Junho-
dc.contributor.authorSon, Yunsik-
dc.date.accessioned2026-01-20T01:30:16Z-
dc.date.available2026-01-20T01:30:16Z-
dc.date.issued2026-01-
dc.identifier.issn2076-3417-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/63469-
dc.description.abstractDespite advancements in deep learning-based Monocular Depth Estimation (MDE), applying these models to video sequences remains challenging due to geometric ambiguities in texture-less regions and temporal instability caused by independent per-frame inference. To address these limitations, we propose STF-Depth, a novel post-processing framework that enhances depth quality by logically fusing heterogeneous information-geometric, semantic, and panoptic-without requiring additional retraining. Our approach introduces a robust RANSAC-based Vanishing Point Estimation to guide Dynamic Depth Gradient Correction for background separation, alongside Adaptive Instance Re-ordering to clarify occlusion relationships. Experimental results on the KITTI, NYU Depth V2, and TartanAir datasets demonstrate that STF-Depth functions as a universal plug-and-play module. Notably, it achieved a 25.7% reduction in Absolute Relative error (AbsRel) and significantly enhanced temporal consistency compared to state-of-the-art backbone models. These findings confirm the framework's practicality for real-world applications requiring geometric precision and video stability, such as autonomous driving, robotics, and augmented reality (AR).-
dc.format.extent26-
dc.language영어-
dc.language.isoENG-
dc.publisherMDPI-
dc.titleSemantic-Guided Spatial and Temporal Fusion Framework for Enhancing Monocular Video Depth Estimation-
dc.typeArticle-
dc.publisher.location스위스-
dc.identifier.doi10.3390/app16010212-
dc.identifier.scopusid2-s2.0-105027319326-
dc.identifier.wosid001657163400001-
dc.identifier.bibliographicCitationApplied Sciences, v.16, no.1, pp 1 - 26-
dc.citation.titleApplied Sciences-
dc.citation.volume16-
dc.citation.number1-
dc.citation.startPage1-
dc.citation.endPage26-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaChemistry-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaMaterials Science-
dc.relation.journalResearchAreaPhysics-
dc.relation.journalWebOfScienceCategoryChemistry, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryEngineering, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryMaterials Science, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryPhysics, Applied-
dc.subject.keywordAuthormonocular video depth estimation-
dc.subject.keywordAuthorheterogeneous information fusion-
dc.subject.keywordAuthortemporal consistency-
dc.subject.keywordAuthorsemantic and panoptic segmentation-
dc.subject.keywordAuthorvanishing point estimation-
Files in This Item
There are no files associated with this item.
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Jeong, Jun Ho photo

Jeong, Jun Ho
College of Advanced Convergence Engineering (Department of Computer Science and Artificial Intelligence)
Read more

Altmetrics

Total Views & Downloads

BROWSE