Knowledge Distillation based Online Learning Methodology using Unlabeled Data Stream
- Authors
- Seo, Sanghyun; Park, Seongchul; Jeong, Changhoon; Kim, Juntae
- Issue Date
- 28-Sep-2018
- Publisher
- ASSOC COMPUTING MACHINERY
- Keywords
- Online Learning; Knowledge Distillation; Knowledge Transfer; Concept Drift
- Citation
- PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND MACHINE INTELLIGENCE (MLMI 2018), pp 68 - 71
- Pages
- 4
- Indexed
- SCOPUS
- Journal Title
- PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND MACHINE INTELLIGENCE (MLMI 2018)
- Start Page
- 68
- End Page
- 71
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/10016
- DOI
- 10.1145/3278312.3278319
- Abstract
- In supervised learning, the performance of the learning model decreases with the change of time step due to concept drift caused by overfitting of the training data. As a methodology to mitigate such concept drift, an online learning methodology has been proposed that trains the learning model on continuously input data stream. In this paper, we proposed an online learning methodology in which teacher model continuously trains student model based on knowledge distillation theory. The teacher model generates the output distribution called soft label to make a label for the unlabeled data stream and the student model trained by the unlabeled data stream with the soft label from teacher model. Experimental results show that the proposed method has better performances such as classification accuracy than that of the batch learning model trained by labeled data stream only.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.