Cited 0 time in
Initialization by using truncated distributions in artificial neural network
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Kim, MinJong | - |
| dc.contributor.author | Cho, Sungchul | - |
| dc.contributor.author | Jeong, Hyerin | - |
| dc.contributor.author | Lee, YungSeop | - |
| dc.contributor.author | Lim, Changwon | - |
| dc.date.accessioned | 2023-04-28T02:40:50Z | - |
| dc.date.available | 2023-04-28T02:40:50Z | - |
| dc.date.issued | 2019-10 | - |
| dc.identifier.issn | 1225-066X | - |
| dc.identifier.issn | 2383-5818 | - |
| dc.identifier.uri | https://scholarworks.dongguk.edu/handle/sw.dongguk/7609 | - |
| dc.description.abstract | Deep learning has gained popularity for the classification and prediction task. Neural network layers become deeper as more data becomes available. Saturation is the phenomenon that the gradient of an activation function gets closer to 0 and can happen when the value of weight is too big. Increased importance has been placed on the issue of saturation which limits the ability of weight to learn. To resolve this problem, Glorot and Bengio (Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, 249-256, 2010) claimed that efficient neural network training is possible when data flows variously between layers. They argued that variance over the output of each layer and variance over input of each layer are equal. They proposed a method of initialization that the variance of the output of each layer and the variance of the input should be the same. In this paper, we propose a new method of establishing initialization by adopting truncated normal distribution and truncated cauchy distribution. We decide where to truncate the distribution while adapting the initialization method by Glorot and Bengio (2010). Variances are made over output and input equal that are then accomplished by setting variances equal to the variance of truncated distribution. It manipulates the distribution so that the initial values of weights would not grow so large and with values that simultaneously get close to zero. To compare the performance of our proposed method with existing methods, we conducted experiments on MNIST and CIFAR-10 data using DNN and CNN. Our proposed method outperformed existing methods in terms of accuracy. | - |
| dc.format.extent | 10 | - |
| dc.language | 한국어 | - |
| dc.language.iso | KOR | - |
| dc.publisher | KOREAN STATISTICAL SOC | - |
| dc.title | Initialization by using truncated distributions in artificial neural network | - |
| dc.type | Article | - |
| dc.publisher.location | 대한민국 | - |
| dc.identifier.doi | 10.5351/KJAS.2019.32.5.693 | - |
| dc.identifier.bibliographicCitation | KOREAN JOURNAL OF APPLIED STATISTICS, v.32, no.5, pp 693 - 702 | - |
| dc.citation.title | KOREAN JOURNAL OF APPLIED STATISTICS | - |
| dc.citation.volume | 32 | - |
| dc.citation.number | 5 | - |
| dc.citation.startPage | 693 | - |
| dc.citation.endPage | 702 | - |
| dc.type.docType | Article | - |
| dc.identifier.kciid | ART002520837 | - |
| dc.description.isOpenAccess | N | - |
| dc.description.journalRegisteredClass | esci | - |
| dc.description.journalRegisteredClass | kci | - |
| dc.relation.journalResearchArea | Mathematics | - |
| dc.relation.journalWebOfScienceCategory | Statistics & Probability | - |
| dc.subject.keywordAuthor | initialization | - |
| dc.subject.keywordAuthor | saturation | - |
| dc.subject.keywordAuthor | Xavier initialization | - |
| dc.subject.keywordAuthor | truncated distribution | - |
| dc.subject.keywordAuthor | deep learning | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
30, Pildong-ro 1-gil, Jung-gu, Seoul, 04620, Republic of Korea+82-2-2260-3114
Copyright(c) 2023 DONGGUK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.
