Journal of South China University of Technology(Natural Science Edition) ›› 2023, Vol. 51 ›› Issue (5): 104-113.doi: 10.12141/j.issn.1000-565X.220623

Special Issue: 2023年电子、通信与自动控制

• Electronics, Communication & Automation Technology • Previous Articles     Next Articles

Design and Implementation of Hardware Structure for Online Learning of Spiking Neural Networks Based on FPGA Parallel Acceleration

LIU Yijun1 CAO Yu2 YE Wujian1 LIN Ziqi2   

  1. 1.School of Integrated Circuits, Guangdong University of Technology, Guangzhou 510006, Guangdong, China
    2.School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, Guangdong, China
  • Received:2022-09-27 Online:2023-05-25 Published:2022-12-05
  • Contact: 叶武剑(1987-),男,博士,讲师,主要从事类脑计算机、深度学习应用研究。 E-mail:yewjian@gdut.edu.cn
  • About author:刘怡俊(1977-),男,博士,教授,博士生导师,主要从事集成电路设计、类脑计算机、深度学习研究。E-mail:yjliu@gdut.edu.cn
  • Supported by:
    the Key-Area R&D Program of Guangdong Province(2018B030338001)

Abstract:

Currently, the hardware design of spiking neural networks based on digital circuits has a low synaptic parallel nature in terms of learning function, leading to a large overall hardware delay, which limits the speed of online learning of spiking neural network models to some extent. To address the above problems, this paper proposed an efficient spiking neural network online learning hardware architecture based on FPGA parallel acceleration, which accelerates the training and inference process of the model through the dual parallel design of neurons and synapses. Firstly, a synaptic structure with parallel spike delivery function and parallel spike time-dependent plasticity learning function was designed; then, the learning layers of input encoding layer and winner-take-all structure were built, and the implementation of lateral inhibition of the winner-take-all network was optimized, forming an impulsive neural network model with a scale of 784~400. The experiments show, the hardware has a training speed of 1.61 ms/image and an energy consumption of about 3.18 mJ/image for the SNN model and an inference speed of 1.19 ms/image and an energy consumption of about 2.37 mJ/image on the MNIST dataset, with an accuracy rate of 87.51%. Based on the hardware framework designed in this paper, the synaptic parallel structure can improve the training speed by more than 38%, and reduce the hardware energy consumption by about 24.1%, which can help to promote the development of edge intelligent computing devices and technologies.

Key words: neural network, learning algorithm, acceleration, parallel architecture

CLC Number: