학술논문

A Stochastic Variance Reduced Nesterov's Accelerated Quasi-Newton Method
Document Type
Conference
Source
2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA) Machine Learning And Applications (ICMLA), 2019 18th IEEE International Conference On. :1874-1879 Dec, 2019
Subject
Computing and Processing
Engineering Profession
Robotics and Control Systems
Signal Processing and Analysis
Acceleration
Training
Neural networks
Computational efficiency
Approximation algorithms
Convergence
Linear programming
training algorithms
stochastic variance reduction
quasi-Newton method
momentum term
limited memory
Language
Abstract
Recently algorithms incorporating second order curvature information have become popular in training neural networks. The Nesterov's Accelerated Quasi-Newton (NAQ) method has shown to effectively accelerate the BFGS quasi-Newton method by incorporating the momentum term and Nesterov's accelerated gradient vector. A stochastic version of NAQ method was proposed for training of large-scale problems. However, this method incurs high stochastic variance noise. This paper proposes a stochastic variance reduced Nesterov's Accelerated Quasi-Newton method in full (SVR-NAQ) and limited (SVRLNAQ) memory forms. The performance of the proposed method is evaluated in Tensorflow on four benchmark problems - two regression and two classification problems respectively. The results show improved performance compared to conventional methods.