热门
Activated Gradients for Deep Neural Networks
文献类型:期刊论文
| 作者 | Liu, Mei1,2; Chen, Liangming1,2; Du, Xiaohao3; Jin, Long1; Shang, Mingsheng1,2,3
|
| 刊名 | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
![]() |
| 出版日期 | 2021-08-31 |
| 页码 | 13 |
| 关键词 | Training Deep learning Neural networks Optimization Visualization Newton method Eigenvalues and eigenfunctions Exploding gradient problems gradient activation function (GAF) ill-conditioned problems saddle point problems vanishing gradient problems |
| ISSN号 | 2162-237X |
| DOI | 10.1109/TNNLS.2021.3106044 |
| 通讯作者 | Jin, Long(longjin@ieee.org) |
| 英文摘要 | Deep neural networks often suffer from poor performance or even training failure due to the ill-conditioned problem, the vanishing/exploding gradient problem, and the saddle point problem. In this article, a novel method by acting the gradient activation function (GAF) on the gradient is proposed to handle these challenges. Intuitively, the GAF enlarges the tiny gradients and restricts the large gradient. Theoretically, this article gives conditions that the GAF needs to meet and, on this basis, proves that the GAF alleviates the problems mentioned above. In addition, this article proves that the convergence rate of SGD with the GAF is faster than that without the GAF under some assumptions. Furthermore, experiments on CIFAR, ImageNet, and PASCAL visual object classes confirm the GAF's effectiveness. The experimental results also demonstrate that the proposed method is able to be adopted in various deep neural networks to improve their performance. The source code is publicly available at https://github.com/LongJin-lab/Activated-Gradients-for-Deep-Neural-Networks. |
| 资助项目 | National Natural Science Foundation of China[62176109] ; National Natural Science Foundation of China[62072429] ; Natural Science Foundation of Chongqing (China)[cstc2020jcyj-zdxmX0028] ; CAS Light of West China Program ; CAAI-Huawei MindSpore Open Fund[CAAIXSJLJJ-2020-009A] ; Gansu Provincial Youth Doctoral Fund of Colleges and Universities[2021QB-003] ; Natural Science Foundation of Gansu Province (China)[20JR10RA639] |
| WOS研究方向 | Computer Science ; Engineering |
| 语种 | 英语 |
| WOS记录号 | WOS:000732099500001 |
| 出版者 | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC |
| 源URL | [http://119.78.100.138/handle/2HOD01W0/14752] ![]() |
| 专题 | 中国科学院重庆绿色智能技术研究院 |
| 通讯作者 | Jin, Long |
| 作者单位 | 1.Chinese Acad Sci, Chongqing Inst Green & Intelligent Technol, Chongqing Key Lab Big Data & Intelligent Comp, Chongqing 400714, Peoples R China 2.Univ Chinese Acad Sci, Chongqing Sch, Chongqing 400714, Peoples R China 3.Lanzhou Univ, Sch Informat Sci & Engn, Lanzhou 730000, Peoples R China |
| 推荐引用方式 GB/T 7714 | Liu, Mei,Chen, Liangming,Du, Xiaohao,et al. Activated Gradients for Deep Neural Networks[J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,2021:13. |
| APA | Liu, Mei,Chen, Liangming,Du, Xiaohao,Jin, Long,&Shang, Mingsheng.(2021).Activated Gradients for Deep Neural Networks.IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,13. |
| MLA | Liu, Mei,et al."Activated Gradients for Deep Neural Networks".IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS (2021):13. |
入库方式: OAI收割
来源:重庆绿色智能技术研究院
浏览0
下载0
收藏0
其他版本
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


