AdaS: Adaptive Scheduling of Stochastic Gradients
From MaRDI portal
Publication:6342660
arXiv2006.06587MaRDI QIDQ6342660
Author name not available (Why is that?)
Publication date: 11 June 2020
Abstract: The choice of step-size used in Stochastic Gradient Descent (SGD) optimization is empirically selected in most training procedures. Moreover, the use of scheduled learning techniques such as Step-Decaying, Cyclical-Learning, and Warmup to tune the step-size requires extensive practical experience--offering limited insight into how the parameters update--and is not consistent across applications. This work attempts to answer a question of interest to both researchers and practitioners, namely extit{"how much knowledge is gained in iterative training of deep neural networks?"} Answering this question introduces two useful metrics derived from the singular values of the low-rank factorization of convolution layers in deep neural networks. We introduce the notions of extit{"knowledge gain"} and extit{"mapping condition"} and propose a new algorithm called Adaptive Scheduling (AdaS) that utilizes these derived metrics to adapt the SGD learning rate proportionally to the rate of change in knowledge gain over successive iterations. Experimentation reveals that, using the derived metrics, AdaS exhibits: (a) faster convergence and superior generalization over existing adaptive learning methods; and (b) lack of dependence on a validation set to determine when to stop training. Code is available at url{https://github.com/mahdihosseini/AdaS}.
Has companion code repository: https://github.com/mahdihosseini/histokt
This page was built for publication: AdaS: Adaptive Scheduling of Stochastic Gradients
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6342660)