4.7 Article

Training Spiking Neural Networks Using Lessons From Deep Learning

期刊

PROCEEDINGS OF THE IEEE
卷 -, 期 -, 页码 -

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JPROC.2023.3308088

关键词

| Deep learning; neural code; neuromorphic; online learning; spiking neural networks (SNNs)

向作者/读者索取更多资源

This article discusses how to apply decades of research in deep learning, gradient descent, backpropagation, and neuroscience to biologically plausible spiking neural networks (SNNs) and how deep learning can move towards biologically plausible online learning. The article also explores the interplay between encoding data as spikes and the learning process, the challenges and solutions of applying gradient-based learning to SNNs, the subtle link between temporal backpropagation and spike timing-dependent plasticity. Some ideas are well accepted and commonly used among the neuromorphic engineering community, while others are presented or justified for the first time here.
The brain is the perfect place to look for inspiration to develop more efficient neural networks. The inner workings of our synapses and neurons provide a glimpse at what the future of deep learning might look like. This article serves as a tutorial and perspective showing how to apply the lessons learned from several decades of research in deep learning, gradient descent, backpropagation, and neuroscience to biologically plausible spiking neural networks (SNNs). We also explore the delicate interplay between encoding data as spikes and the learning process; the challenges and solutions of applying gradient-based learning to SNNs; the subtle link between temporal backpropagation and spike timing-dependent plasticity; and how deep learning might move toward biologically plausible online learning. Some ideas are well accepted and commonly used among the neuromorphic engineering community, while others are presented or justified for the first time here. A series of companion interactive tutorials complementary to this article using our Python package, snnTorch, are also made available: https://snntorch.readthedocs.io/en/latest/tutorials/index.html.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

Article Engineering, Electrical & Electronic

TAICHI: A Tiled Architecture for In-Memory Computing and Heterogeneous Integration

Xinxin Wang, Reid Pinkham, Mohammed A. Zidan, Fan-Hsuan Meng, Michael P. Flynn, Zhengya Zhang, Wei D. Lu

Summary: TAICHI is a general in-memory computing deep neural network accelerator design that uses RRAM crossbar arrays and integrates local arithmetic units and global co-processors. It efficiently maps different models while maintaining high energy efficiency and throughput. A hierarchical mesh network-on-chip is implemented to balance reconfigurability and efficiency. The system performance is estimated at several technology nodes and the heterogeneous design allows for larger models than the on-chip storage capacity.

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS (2022)

Article Engineering, Electrical & Electronic

Dynamic resistive switching devices for neuromorphic computing

Yuting Wu, Xinxin Wang, Wei D. Lu

Summary: Memristive devices have rich dynamics due to internal state variables determining conductance, which may serve as building blocks for biofaithful neuromorphic systems; these devices can be utilized in compute-in-memory architectures to process temporal data efficiently, implementing synaptic and neuronal functions effectively.

SEMICONDUCTOR SCIENCE AND TECHNOLOGY (2022)

Article Nanoscience & Nanotechnology

Tuning Resistive Switching Behavior by Controlling Internal Ionic Dynamics for Biorealistic Implementation of Synaptic Plasticity

Sangmin Yoo, Yuting Wu, Yongmo Park, Wei D. Lu

Summary: Memristive devices exhibit rich switching behaviors similar to synaptic functions, making them ideal for constructing efficient neuromorphic systems. The internal temperature serves as a second state variable to regulate ion motion and allows for the native implementation of timing- and rate-based learning rules. This study demonstrates that by engineering the internal temperature in a Ta2O5-based memristor, it is possible to tune its spike timing dependent plasticity (STDP) characteristics. When combined with an artificial post-synaptic neuron, the second-order memristor synapses can capture the temporal correlation in input streaming events.

ADVANCED ELECTRONIC MATERIALS (2022)

Article Nanoscience & Nanotechnology

Memristor-Based Binarized Spiking Neural Networks

Jason K. Eshraghian, Xinxin Wang, Wei D. Lu

IEEE NANOTECHNOLOGY MAGAZINE (2022)

Article Engineering, Electrical & Electronic

Physical Unclonable Function Systems Based on Pattern Transfer of Fingerprint-Like Patterns

Ziyu Wang, Xiaojian Zhu, Supreet Jeloka, Brian Cline, Wei D. Lu

Summary: In this letter, a physical unclonable function (PUF) system based on fingerprint-like random planar structures is demonstrated through pattern transfer of self-assembled binary polymer mixtures. The PUF achieves different types of conductance distributions with large variations, allowing it to operate in differential or on/off modes. The uniqueness of the PUF is verified using inter Hamming distance and entropy at different temperatures. The proposed fingerprint PUF is resistant to machine learning attacks based on Fourier extrapolation and is compatible with back-end-of-line (BEOL) process, offering potential for hardware security in the IoT industry.

IEEE ELECTRON DEVICE LETTERS (2022)

Review Multidisciplinary Sciences

Memristive technologies for data storage, computation, encryption, and radio-frequency communication

Mario Lanza, Abu Sebastian, Wei D. Lu, Manuel Le Gallo, Meng-Fan Chang, Deji Akinwande, Francesco M. Puglisi, Husam N. Alshareef, Ming Liu, Juan B. Roldan

Summary: Memristive devices, which can change their resistance and memory state, have potential applications in various fields. However, there are still challenges to be addressed, including performance and reliability issues.

SCIENCE (2022)

Review Nanoscience & Nanotechnology

Dynamical memristors for higher-complexity neuromorphic computing

Suhas Kumar, Xinxin Wang, John Paul Strachan, Yuchao Yang, Wei D. Lu

Summary: Research on electronic devices and materials is driven by the slowdown of transistor scaling and the growth of computing needs. Using devices like memristors to achieve complex dynamics enables new computing architectures with high energy efficiency and computing capacity.

NATURE REVIEWS MATERIALS (2022)

Article Engineering, Electrical & Electronic

Gradient-Based Neuromorphic Learning on Dynamical RRAM Arrays

Peng Zhou, Dong-Uk Choi, Wei D. Lu, Sung-Mo Kang, Jason K. Eshraghian

Summary: We present MEMprop, a gradient-based learning method for training fully memristive spiking neural networks. By harnessing the device dynamics to trigger voltage spikes, MEMprop eliminates the need for surrogate gradient methods. The implementation is fully memristive, without the need for additional circuits to implement spiking dynamics, and achieves competitive accuracy on several benchmarks.

IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS (2022)

Article Engineering, Electrical & Electronic

Exploring Compute-in-Memory Architecture Granularity for Structured Pruning of Neural Networks

Fan-Hsuan Meng, Xinxin Wang, Ziyu Wang, Eric Yeu-Jer Lee, Wei D. Lu

Summary: This paper presents how structured pruning can be efficiently implemented in CIM systems based on RRAM crossbars, improving accuracy and reducing hardware cost.

IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS (2022)

Article Computer Science, Hardware & Architecture

RM-NTT: An RRAM-Based Compute-in-Memory Number Theoretic Transform Accelerator

Yongmo Park, Ziyu Wang, Sangmin Yoo, Wei D. D. Lu

Summary: As cloud computing resources are increasingly used in machine learning, privacy-preserving techniques like homomorphic encryption (HE) have attracted attention. However, implementing deep neural networks (DNN) using HE is significantly slower than plaintext implementations.

IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS (2022)

Article Computer Science, Hardware & Architecture

Research Challenges for Energy-Efficient Computing in Automated Vehicles

Robert P. Dick, Rob Aitken, Jace Mogill, John Paul Strachan, Kirk Bresniker, Wei Lu, Yorie Nakahira, Zhiyong Li, Matthew J. Marinella, William Severa, A. Alec Talin, Craig M. Vineyard, Suhas Kumar, Christian Mailhiot, Lennie Klebanoff

Summary: Fully automated retail vehicles will require stringent real-time operational safety, sensing, communication, inference, planning, and control capabilities. However, existing technologies may not meet these requirements due to excessive energy consumption and thermal management issues. This article summarizes the research challenges in designing computationally energy-efficient, cost-effective, safe, and reliable automated retail vehicles.

COMPUTER (2023)

Article Chemistry, Multidisciplinary

Bulk-Switching Memristor-Based Compute-In-Memory Module for Deep Neural Network Training

Yuting Wu, Qiwen Wang, Ziyu Wang, Xinxin Wang, Buvna Ayyagari, Siddarth Krishnan, Michael Chudzik, Wei D. Lu

Summary: This paper proposes a mixed-precision training scheme using bulk-switching memristor-based compute-in-memory (CIM) modules to accelerate deep neural network (DNN) training. Low-precision CIM modules are used for fast computation while high-precision weight updates are accumulated in digital units. Experimental results show that this scheme enables efficient mixed-precision DNN training with comparable accuracy to software-trained models.

ADVANCED MATERIALS (2023)

Article Automation & Control Systems

PowerGAN: A Machine Learning Approach for Power Side-Channel Attack on Compute-in-Memory Accelerators

Ziyu Wang, Yuting Wu, Yongmo Park, Sangmin Yoo, Xinxin Wang, Jason K. Eshraghian, Wei D. Lu

Summary: Analog compute-in-memory (CIM) systems have the potential to accelerate deep neural network (DNN) inference. However, this study identifies a security vulnerability wherein an attacker can reconstruct the user's private input data from a power side-channel attack, even without knowledge of the stored DNN model. The study proposes an attack approach using a generative adversarial network that achieves high-quality data reconstruction from power leakage measurements. The results demonstrate the effectiveness of the attack methodology in reconstructing user input data from power leakage of the analog CIM accelerator, even at high noise levels and after countermeasures.

ADVANCED INTELLIGENT SYSTEMS (2023)

Proceedings Paper Engineering, Electrical & Electronic

Demonstration of a Multi-Level mu A-Range Bulk Switching ReRAM and its Application for Keyword Spotting

Y. Wu, F. Cai, L. Thomas, T. Liu, A. Nourbakhsh, J. Hebding, E. Smith, R. Quon, R. Smith, A. Kumar, A. Pang, J. Holt, R. Someshwar, F. Nardi, J. Anthis, S-H. Yen, C. Chevallier, A. Uppala, X. Chen, N. Breil, T. Sherwood, K. Wong, W. Cho, D. Thompson, J. Hsu, B. Ayyagari, S. Krishnan, Wei. D. Lu, M. Chudzik

Summary: This paper presents a forming-free bulk ReRAM cell with high programming accuracy and yield, showing superior analog behavior for memory computing applications.

2022 INTERNATIONAL ELECTRON DEVICES MEETING, IEDM (2022)

暂无数据