TY - JOUR
T1 - Efficient Training of the Memristive Deep Belief Net Immune to Non-Idealities of the Synaptic Devices
AU - Wang, Wei
AU - Hoffer, Barak
AU - Greenberg-Toledo, Tzofnat
AU - Li, Yang
AU - Zou, Minhui
AU - Herbelin, Eric
AU - Ronen, Ronny
AU - Xu, Xiaoxin
AU - Zhao, Yulin
AU - Yang, Jianguo
AU - Kvatinsky, Shahar
PY - 2022
Y1 - 2022
N2 - The tunability of conductance states of various emerging nonvolatile memristive devices emulates the plasticity of biological synapses, making it promising in the hardware realization of large-scale neuromorphic systems. The inference of the neural network can be greatly accelerated by the vector-matrix multiplication (VMM) performed within a crossbar array of memristive devices in one step. Nevertheless, the implementation of the VMM needs complex peripheral circuits, and the complexity further increases as non-idealities of memristive devices prevent precise conductance tuning (especially for the online training) and largely degrade the performance of the deep neural networks (DNNs). Herein, an efficient online training method of the memristive deep belief net (DBN) is presented. The proposed memristive DBN uses stochastically binarized activations, reducing the complexity of peripheral circuits, and uses the contrastive divergence (CD)-based gradient descent learning algorithm. The analog VMM and digital CD are performed separately in a mixed-signal hardware arrangement, making the memristive DBN highly immune to non-idealities of synaptic devices. The number of write operations on memristive devices is reduced by two orders of magnitude. The recognition accuracy of 95–97% can be achieved for the MNIST dataset using pulsed synaptic behaviors of various memristive synaptic devices.
AB - The tunability of conductance states of various emerging nonvolatile memristive devices emulates the plasticity of biological synapses, making it promising in the hardware realization of large-scale neuromorphic systems. The inference of the neural network can be greatly accelerated by the vector-matrix multiplication (VMM) performed within a crossbar array of memristive devices in one step. Nevertheless, the implementation of the VMM needs complex peripheral circuits, and the complexity further increases as non-idealities of memristive devices prevent precise conductance tuning (especially for the online training) and largely degrade the performance of the deep neural networks (DNNs). Herein, an efficient online training method of the memristive deep belief net (DBN) is presented. The proposed memristive DBN uses stochastically binarized activations, reducing the complexity of peripheral circuits, and uses the contrastive divergence (CD)-based gradient descent learning algorithm. The analog VMM and digital CD are performed separately in a mixed-signal hardware arrangement, making the memristive DBN highly immune to non-idealities of synaptic devices. The number of write operations on memristive devices is reduced by two orders of magnitude. The recognition accuracy of 95–97% can be achieved for the MNIST dataset using pulsed synaptic behaviors of various memristive synaptic devices.
KW - contrastive divergence
KW - deep belief net
KW - memristive synapse
KW - non-ideality
KW - restricted Boltzmann machine
U2 - https://doi.org/10.1002/aisy.202100249
DO - https://doi.org/10.1002/aisy.202100249
M3 - مقالة
VL - 4
SP - 2100249
JO - Advanced Intelligent Systems
JF - Advanced Intelligent Systems
IS - 5
ER -