Hybrid Approach for Efficient Quantization of Weights in Convolutional Neural Networks
- Authors
- Seo, Sanghyun; Kim, Juntae
- Issue Date
- 25-May-2018
- Publisher
- IEEE
- Keywords
- Neural Networks Compression; Convolutional Neural Networks; Weights Quantization; Hybrid Quantizer
- Citation
- 2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP), pp 638 - 641
- Pages
- 4
- Indexed
- SCOPUS
- Journal Title
- 2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP)
- Start Page
- 638
- End Page
- 641
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/10029
- DOI
- 10.1109/BigComp.2018.00114
- ISSN
- 2375-933X
- Abstract
- Convolutional neural networks(CNN) have achieved outstanding results in the fields of image recognition which classifies objects in the input images. In the deep neural networks such as CNN, the number of layers and the number of neurons in each layer are large. In other words, the deep neural networks requires relatively large storage space and calculation process. However, in embedded devices for object recognition in autonomous vehicles, large storage space and high computational complexity are constraints. For this reasons, various methodologies have been proposed to apply CNN to small embedded hardware such as mobile devices, FPGA and ASIC efficiently. In this paper, we quantize the weights of AlexNet without a large drop in accuracy by using a hybrid quantizer using uniform quantizer and k-means clustering.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.