M-NAS: Meta Neural Architecture Search
文献类型:会议论文
作者 | Wang JX(王家兴)1,5![]() ![]() ![]() |
出版日期 | 2020 |
会议日期 | 2020.02.07-2020.02.12 |
会议地点 | New York |
英文摘要 | Neural Architecture Search (NAS) has recently outperformed hand-crafted networks in various areas. However, most prevalent NAS methods only focus on a pre-defined task. For a previously unseen task, the architecture is either searched from scratch, which is inefficient, or transferred from the one obtained on some other task, which might be sub-optimal. In this paper, we investigate a previously unexplored problem: whether a universal NAS method exists, such that task-aware architectures can be effectively generated? Towards this prob- lem, we propose Meta Neural Architecture Search (M-NAS). To obtain task-specific architectures, M-NAS adopts a task-aware architecture controller for child model generation. Since optimal weights for different tasks and architectures span diversely, we resort to meta-learning, and learn meta-weights that efficiently adapt to a new task on the corresponding architecture with only several gradient descent steps. Experimental results demonstrate the superiority of M-NAS against a number of competitive baselines on both toy regression and few shot classification problems. |
会议录出版者 | AAAI Press |
语种 | 英语 |
源URL | [http://ir.ia.ac.cn/handle/173211/44761] ![]() |
专题 | 类脑芯片与系统研究 |
通讯作者 | Cheng J(程健) |
作者单位 | 1.Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100190, Peoples R China 2.Center for Excellence in Brain Science and Intelligence Technology, CAS 3.The Chinese University of Hong Kong 4.Tencent AI Lab 5.Chinese Acad Sci, Inst Automat, Beijing 100190, Peoples R China |
推荐引用方式 GB/T 7714 | Wang JX,Wu JX,Bo HL,et al. M-NAS: Meta Neural Architecture Search[C]. 见:. New York. 2020.02.07-2020.02.12. |
入库方式: OAI收割
来源:自动化研究所
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。