Detailed Information

Cited 12 time in webofscience Cited 13 time in scopus
Metadata Downloads

Memristive Architectures Exploiting Self-Compliance Multilevel Implementation on 1 kb Crossbar Arrays for Online and Offline Learning Neuromorphic Applications

Authors
Kim, SungjoonJi, HyeonseungPark, KyungchulSo, HyojinKim, HyungjinKim, SungjunChoi, Woo Young
Issue Date
Sep-2024
Publisher
American Chemical Society
Keywords
crossbar array; memristor; self-compliance; online/offline learning; neuromorphic system
Citation
ACS Nano, v.18, no.36, pp 25128 - 25143
Pages
16
Indexed
SCIE
SCOPUS
Journal Title
ACS Nano
Volume
18
Number
36
Start Page
25128
End Page
25143
URI
https://scholarworks.dongguk.edu/handle/sw.dongguk/22971
DOI
10.1021/acsnano.4c06942
ISSN
1936-0851
1936-086X
Abstract
This paper suggests the practical implications of utilizing a high-density crossbar array with self-compliance (SC) at the conductive filament (CF) formation stage. By limiting the excessive growth of CF, SC functions enable the operation of a crossbar array without access transistors. An AlOx/TiOy, internal overshoot limitation structure, allows the SC to have resistive random-access memory. In addition, an overshoot-limited memristor crossbar array makes it possible to implement vector-matrix multiplication (VMM) capability in neuromorphic systems. Furthermore, AlOx/TiOy structure optimization was conducted to reduce overshoot and operation current, verifying uniform bipolar resistive switching behavior and analog switching properties. Additionally, extensive electric pulse stimuli are confirmed, evaluating long-term potentiation (LTP), long-term depression (LTD), and other forms of synaptic plasticity. We found that LTP and LTD characteristics for training an online learning neural network enable MNIST classification accuracies of 92.36%. The SC mode quantized multilevel in offline learning neural networks achieved 95.87%. Finally, the 32 x 32 crossbar array demonstrated spiking neural network-based VMM operations to classify the MNIST image. Consequently, weight programming errors make only a 1.2% point of accuracy drop to software-based neural networks.
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > Department of Electronics and Electrical Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Kim, Sung Jun photo

Kim, Sung Jun
College of Engineering (Department of Electronics and Electrical Engineering)
Read more

Altmetrics

Total Views & Downloads

BROWSE