Journal of the Operations Research Society of China ›› 2020, Vol. 8 ›› Issue (2): 249-294.doi: 10.1007/s40305-020-00309-6

• • 上一篇    下一篇

  

  • 收稿日期:2019-12-18 修回日期:2020-04-13 出版日期:2020-06-30 发布日期:2020-07-07
  • 通讯作者: Ruo-Yu Sun E-mail:ruoyus@illinois.edu

Optimization for Deep Learning: An Overview

Ruo-Yu Sun1,2   

  1. 1 Department of Industrial and Enterprise Systems Engineering, and affiliated to Coordinated Science Laboratory, University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA;
    2 Department of Electrical and Computer Engineering, University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA
  • Received:2019-12-18 Revised:2020-04-13 Online:2020-06-30 Published:2020-07-07
  • Contact: Ruo-Yu Sun E-mail:ruoyus@illinois.edu

Abstract: Optimization is a critical component in deep learning. We think optimization for neural networks is an interesting topic for theoretical research due to various reasons. First, its tractability despite non-convexity is an intriguing question and may greatly expand our understanding of tractable problems. Second, classical optimization theory is far from enough to explain many phenomena. Therefore, we would like to understand the challenges and opportunities from a theoretical perspective and review the existing research in this field. First, we discuss the issue of gradient explosion/vanishing and the more general issue of undesirable spectrum and then discuss practical solutions including careful initialization, normalization methods and skip connections. Second, we review generic optimization methods used in training neural networks, such as stochastic gradient descent and adaptive gradient methods, and existing theoretical results. Third, we review existing research on the global issues of neural network training, including results on global landscape, mode connectivity, lottery ticket hypothesis and neural tangent kernel.

Key words: Deep learning, Non-convex optimization, Neural networks, Convergence, Landscape

中图分类号: