Toward Accurate Binarized Neural Networks With Sparsity for Mobile Application
文献类型:期刊论文
| 作者 | Wang, Peisong ; He, Xiangyu ; Cheng, Jian
|
| 刊名 | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
![]() |
| 出版日期 | 2022-05-27 |
| 页码 | 13 |
| 关键词 | Quantization (signal) Deep learning Convolution Training Biological neural networks Optimization Neurons Acceleration binarized neural networks (BNNs) compression fixed-point quantization |
| ISSN号 | 2162-237X |
| DOI | 10.1109/TNNLS.2022.3173498 |
| 通讯作者 | Cheng, Jian(jcheng@nlpria.ac.cn) |
| 英文摘要 | While binarized neural networks (BNNs) have attracted great interest, popular approaches proposed so far mainly exploit the symmetric sign function for feature binarization, i.e., to binarize activations into -1 and +1 with a fixed threshold of 0. However, whether this option is optimal has been largely overlooked. In this work, we propose the Sparsity-inducing BNN (Si-BNN) to quantize the activations to be either 0 or +1, which better approximates ReLU using 1-bit. We further introduce trainable thresholds into the backward function of binarization to guide the gradient propagation. Our method dramatically outperforms the current state-of-the-art, lowering the performance gap between full-precision networks and BNNs on mainstream architectures, achieving the new state-of-the-art on binarized AlexNet (Top-1 50.5%), ResNet-18 (Top-1 62.2%), and ResNet-50 (Top-1 68.3%). At inference time, Si-BNN still enjoys the high efficiency of bit-wise operations. In our implementation, the running time of binary AlexNet on the CPU can be competitive with the popular GPU-based deep learning framework. |
| 资助项目 | National Key Research and Development Program of China[2021ZD0201504] ; National Natural Science Foundation of China[61906193] ; Strategic Priority Research Program of Chinese Academy of Sciences[XDA27040300] |
| WOS研究方向 | Computer Science ; Engineering |
| 语种 | 英语 |
| WOS记录号 | WOS:000805801000001 |
| 出版者 | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC |
| 资助机构 | National Key Research and Development Program of China ; National Natural Science Foundation of China ; Strategic Priority Research Program of Chinese Academy of Sciences |
| 源URL | [http://ir.ia.ac.cn/handle/173211/49509] ![]() |
| 专题 | 类脑芯片与系统研究 |
| 通讯作者 | Cheng, Jian |
| 作者单位 | Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China |
| 推荐引用方式 GB/T 7714 | Wang, Peisong,He, Xiangyu,Cheng, Jian. Toward Accurate Binarized Neural Networks With Sparsity for Mobile Application[J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,2022:13. |
| APA | Wang, Peisong,He, Xiangyu,&Cheng, Jian.(2022).Toward Accurate Binarized Neural Networks With Sparsity for Mobile Application.IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,13. |
| MLA | Wang, Peisong,et al."Toward Accurate Binarized Neural Networks With Sparsity for Mobile Application".IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS (2022):13. |
入库方式: OAI收割
来源:自动化研究所
浏览0
下载0
收藏0
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


