상세 보기
- Gull, Sahar;
- Kim, Juntae
WEB OF SCIENCE
0SCOPUS
0초록
Cross-domain few-shot learning (CD-FSL) remains challenging in medical imaging, where labeled data are scarce and source-target domain gaps are often large due to modality differences. In particular, existing few-shot learning methods rely on source-target domain similarity, which limits their effectiveness in cross-modality settings such as MRI-to-CT transfer. To address this problem, this paper proposes an adapter-based Vision Transformer framework for cross-domain few-shot brain tumor classification. Lightweight adapter modules are inserted into a pretrained Vision Transformer to enable parameter-efficient domain adaptation without fine-tuning the entire backbone. In addition, a Prototypical Network is employed to construct class prototypes from limited labeled samples, while a prototype-level Maximum Mean Discrepancy (MMD) loss is introduced to align feature distributions across domains. Unlike prior approaches, the proposed framework introduces a unified prototype-level alignment strategy within an episodic learning paradigm, enabling direct class-wise cross-modal alignment. This design improves generalization under large modality gaps and limited labeled data by jointly optimizing representation learning and domain adaptation. The proposed framework is evaluated on MRI-to-CT brain tumor classification as well as several heterogeneous cross-domain benchmarks, including Chest X-ray, ISIC, CropDisease, and EuroSAT. Experimental results demonstrate that the proposed method achieves competitive performance compared to existing few-shot learning baselines, showing strong robustness under significant domain shifts.
키워드
- 제목
- Adapter-Based Vision Transformer for Cross Domain Few-Shot Classification Using Prototypical Networks
- 저자
- Gull, Sahar; Kim, Juntae
- 발행일
- 2026-04
- 유형
- Article
- 저널명
- Applied Sciences
- 권
- 16
- 호
- 8
- 페이지
- 1 ~ 21