中国科学院机构知识库网格
Chinese Academy of Sciences Institutional Repositories Grid
BlockQNN: Efficient Block-Wise Neural Network Architecture Generation

文献类型:期刊论文

作者Zhong, Zhao1; Yang, Zichen2; Deng, Boyang2; Yan, Junjie2; Wu, Wei2; Shao, Jing2; Liu, Cheng-Lin3,4
刊名IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE
出版日期2021-07-01
卷号43期号:7页码:2314-2328
关键词Computer architecture Task analysis Neural networks Network architecture Graphics processing units Acceleration Indexes Convolutional neural network neural architecture search AutoML reinforcement learning Q-learning
ISSN号0162-8828
DOI10.1109/TPAMI.2020.2969193
通讯作者Liu, Cheng-Lin(liucl@nlpr.ia.ac.cn)
英文摘要Convolutional neural networks have gained a remarkable success in computer vision. However, most popular network architectures are hand-crafted and usually require expertise and elaborate design. In this paper, we provide a block-wise network generation pipeline called BlockQNN which automatically builds high-performance networks using the Q-Learning paradigm with epsilon-greedy exploration strategy. The optimal network block is constructed by the learning agent which is trained to choose component layers sequentially. We stack the block to construct the whole auto-generated network. To accelerate the generation process, we also propose a distributed asynchronous framework and an early stop strategy. The block-wise generation brings unique advantages: (1) it yields state-of-the-art results in comparison to the hand-crafted networks on image classification, particularly, the best network generated by BlockQNN achieves 2.35 percent top-1 error rate on CIFAR-10. (2) it offers tremendous reduction of the search space in designing networks, spending only 3 days with 32 GPUs. A faster version can yield a comparable result with only 1 GPU in 20 hours. (3) it has strong generalizability in that the network built on CIFAR also performs well on the larger-scale dataset. The best network achieves very competitive accuracy of 82.0 percent top-1 and 96.0 percent top-5 on ImageNet.
资助项目Major Project for New Generation of AI[2018AAA0100400] ; National Natural Science Foundation of China (NSFC)[61721004] ; National Natural Science Foundation of China (NSFC)[61633021]
WOS研究方向Computer Science ; Engineering
语种英语
WOS记录号WOS:000692540900011
出版者IEEE COMPUTER SOC
资助机构Major Project for New Generation of AI ; National Natural Science Foundation of China (NSFC)
源URL[http://ir.ia.ac.cn/handle/173211/45738]  
专题自动化研究所_模式识别国家重点实验室_模式分析与学习团队
通讯作者Liu, Cheng-Lin
作者单位1.Univ Chinese Acad Sci, Inst Automat, Chinese Acad Sci, NLPR, Beijing 100190, Peoples R China
2.Sensetime Res Inst, SenseTime Grp Ltd, Beijing, Peoples R China
3.Chinese Acad Sci, Inst Automat, NLPR, Beijing, Peoples R China
4.Univ Chinese Acad Sci, CAS Ctr Excellence Brain Sci & Intelligence, Beijing 100190, Peoples R China
推荐引用方式
GB/T 7714
Zhong, Zhao,Yang, Zichen,Deng, Boyang,et al. BlockQNN: Efficient Block-Wise Neural Network Architecture Generation[J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,2021,43(7):2314-2328.
APA Zhong, Zhao.,Yang, Zichen.,Deng, Boyang.,Yan, Junjie.,Wu, Wei.,...&Liu, Cheng-Lin.(2021).BlockQNN: Efficient Block-Wise Neural Network Architecture Generation.IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,43(7),2314-2328.
MLA Zhong, Zhao,et al."BlockQNN: Efficient Block-Wise Neural Network Architecture Generation".IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 43.7(2021):2314-2328.

入库方式: OAI收割

来源:自动化研究所

浏览0
下载0
收藏0
其他版本

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。