Fixed-point Factorized Networks
文献类型:会议论文
作者 | Wang, Peisong1,2![]() ![]() |
出版日期 | 2017 |
会议日期 | 2017.7.21-7.26 |
会议地点 | Hawaii,USA |
关键词 | Convolutional Neural Networks Ternary Quantization Network Acceleration Network Compression |
页码 | 4012-4020 |
英文摘要 |
In recent years, Deep Neural Networks (DNN) based methods have achieved remarkable performance in a wide range of tasks and have been among the most powerful and widely used techniques in computer vision. However, DNN-based methods are both computational-intensive and resource-consuming, which hinders the application of these methods on embedded systems like smart phones. To alleviate this problem, we introduce a novel Fixed-point Factorized Networks (FFN) for pretrained models to reduce the computational complexity as well as the storage requirement of networks. The resulting networks have only weights of -1, 0 and 1, which significantly eliminates the most resource-consuming multiply-accumulate operations (MACs). Extensive experiments on large-scale ImageNet classification task show the proposed FFN only requires one-thousandth of multiply operations with comparable accuracy. |
源URL | [http://ir.ia.ac.cn/handle/173211/20114] ![]() |
专题 | 自动化研究所_模式识别国家重点实验室_图像与视频分析团队 |
作者单位 | 1.Institute of Automation, Chinese Academy of Sciences 2.University of Chinese Academy of Sciences 3.Center for Excellence in Brain Science and Intelligence Technology |
推荐引用方式 GB/T 7714 | Wang, Peisong,Cheng, Jian. Fixed-point Factorized Networks[C]. 见:. Hawaii,USA. 2017.7.21-7.26. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。