Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

BDA: Bi-directional attention for zero-shot learning

Full metadata record
DC Field Value Language
dc.contributor.authorLee, Junseok-
dc.contributor.authorCao, Jinming-
dc.contributor.authorYin, Yifang-
dc.contributor.authorKim, Jihie-
dc.contributor.authorZimmermann, Roger-
dc.contributor.authorPark, Seongsik-
dc.date.accessioned2025-11-28T07:30:51Z-
dc.date.available2025-11-28T07:30:51Z-
dc.date.issued2025-10-
dc.identifier.issn2096-0433-
dc.identifier.issn2096-0662-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/62159-
dc.description.abstractZero-shot learning (ZSL) is an important and rapidly growing area of machine learning that aims to recognize new classes without prior training data. Despite its significance, ZSL has faced challenges with overfitting in embedding-based methods and limitations in traditional one-directional attention (ODA) based approaches. To bridge these gaps, this paper proposes the use of bi-directional attention (BDA) to integrate insights from both embedding and attention-based approaches. The proposed BDA system consists of a bi-directional attention network (BDAN) and a synthesized visual embedding network (SVEN) that facilitates visual-semantic interaction for ZSL classification. More specifically, the BDAN employs region self-attention (RSA), semantic synthesis attention (SSA), and visual synthesis attention (VSA) to overcome the overfitting issue in embedding methods and enhance transferability, to associate visual features with semantic property information, and to learn locally improved visual features. Extensive testing on CUB, SUN, and AWA2 datasets confirm the superiority of our proposed method over traditional approaches. © 2025 Elsevier B.V., All rights reserved.-
dc.format.extent21-
dc.language영어-
dc.language.isoENG-
dc.publisherTsinghua University Press-
dc.titleBDA: Bi-directional attention for zero-shot learning-
dc.typeArticle-
dc.publisher.location중국-
dc.identifier.doi10.26599/CVM.2025.9450401-
dc.identifier.scopusid2-s2.0-105021021968-
dc.identifier.wosid001616145400011-
dc.identifier.bibliographicCitationComputational Visual Media, v.11, no.5, pp 983 - 1003-
dc.citation.titleComputational Visual Media-
dc.citation.volume11-
dc.citation.number5-
dc.citation.startPage983-
dc.citation.endPage1003-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalWebOfScienceCategoryComputer Science, Software Engineering-
dc.subject.keywordAuthorbi-directional attention (BDA)-
dc.subject.keywordAuthorinteraction-
dc.subject.keywordAuthortransferability-
dc.subject.keywordAuthorzero-shot learning (ZSL)-
Files in This Item
There are no files associated with this item.
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Kim, Ji Hie photo

Kim, Ji Hie
College of Advanced Convergence Engineering (Department of Computer Science and Artificial Intelligence)
Read more

Altmetrics

Total Views & Downloads

BROWSE