FireFly: A High-Throughput Hardware Accelerator for Spiking Neural Networks With Efficient DSP and Memory Optimization
文献类型:期刊论文
作者 | Li, Jindong3,4; Shen, Guobin2,4; Zhao, Dongcheng4![]() ![]() ![]() |
刊名 | IEEE Transactions on Very Large Scale Integration (VLSI) Systems
![]() |
出版日期 | 2023 |
页码 | 1178 - 1191 |
DOI | 10.1109/TVLSI.2023.3279349 |
英文摘要 | Spiking neural networks (SNNs) have been widely used due to their strong biological interpretability and high-energy efficiency. With the introduction of the backpropagation algorithm and surrogate gradient, the structure of SNNs has become more complex, and the performance gap with artificial neural networks (ANNs) has gradually decreased. However, most SNN hardware implementations for field-programmable gate arrays (FPGAs) cannot meet arithmetic or memory efficiency requirements, which significantly restricts the development of SNNs. They do not delve into the arithmetic operations between the binary spikes and synaptic weights or assume unlimited on-chip RAM resources using overly expensive devices on small tasks. To improve arithmetic efficiency, we analyze the neural dynamics of spiking neurons, generalize the SNN arithmetic operation to the multiplex-accumulate operation, and propose a high-performance implementation of such operation by utilizing the DSP48E2 hard block in Xilinx Ultrascale FPGAs. To improve memory efficiency, we design a memory system to enable efficient synaptic weights and membrane voltage memory access with reasonable on-chip RAM consumption. Combining the above two improvements, we propose an FPGA accelerator that can process spikes generated by the firing neurons on-the-fly (FireFly). FireFly is the first SNN accelerator that incorporates DSP optimization techniques into SNN synaptic operations. FireFly is implemented on several FPGA edge devices with limited resources but still guarantees a peak performance of 5.53 TOP/s at 300 MHz. As a lightweight accelerator, FireFly achieves the highest computational density efficiency compared with existing research using large FPGA devices. |
URL标识 | 查看原文 |
源URL | [http://ir.ia.ac.cn/handle/173211/57243] ![]() |
专题 | 类脑智能研究中心_类脑认知计算 |
通讯作者 | Zhang, Qian; Zeng, Yi |
作者单位 | 1.Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, China 2.School of Future Technology, University of Chinese Academy of Sciences, Beijing, China 3.School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China 4.Brain-Inspired Cognitive Intelligence Laboratory, Institute of Automation, Chinese Academy of Sciences, Beijing, China |
推荐引用方式 GB/T 7714 | Li, Jindong,Shen, Guobin,Zhao, Dongcheng,et al. FireFly: A High-Throughput Hardware Accelerator for Spiking Neural Networks With Efficient DSP and Memory Optimization[J]. IEEE Transactions on Very Large Scale Integration (VLSI) Systems,2023:1178 - 1191. |
APA | Li, Jindong,Shen, Guobin,Zhao, Dongcheng,Zhang, Qian,&Zeng, Yi.(2023).FireFly: A High-Throughput Hardware Accelerator for Spiking Neural Networks With Efficient DSP and Memory Optimization.IEEE Transactions on Very Large Scale Integration (VLSI) Systems,1178 - 1191. |
MLA | Li, Jindong,et al."FireFly: A High-Throughput Hardware Accelerator for Spiking Neural Networks With Efficient DSP and Memory Optimization".IEEE Transactions on Very Large Scale Integration (VLSI) Systems (2023):1178 - 1191. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。