Coupling Adaptive Batch Sizes with Learning Rates

2017

Conference Paper

ps

pn


Mini-batch stochastic gradient descent and variants thereof have become standard for large-scale empirical risk minimization like the training of neural networks. These methods are usually used with a constant batch size chosen by simple empirical inspection. The batch size significantly influences the behavior of the stochastic optimization algorithm, though, since it determines the variance of the gradient estimates. This variance also changes over the optimization process; when using a constant batch size, stability and convergence is thus often enforced by means of a (manually tuned) decreasing learning rate schedule. We propose a practical method for dynamic batch size adaptation. It estimates the variance of the stochastic gradients and adapts the batch size to decrease the variance proportionally to the value of the objective function, removing the need for the aforementioned learning rate decrease. In contrast to recent related work, our algorithm couples the batch size to the learning rate, directly reflecting the known relationship between the two. On three image classification benchmarks, our batch size adaptation yields faster optimization convergence, while simultaneously simplifying learning rate tuning. A TensorFlow implementation is available.

Author(s): Lukas Balles and Javier Romero and Philipp Hennig
Book Title: Thirty-Third Conference on Uncertainty in Artificial Intelligence (UAI)
Year: 2017
Editors: Gal Elidan and Kristian Kersting

Department(s): Perceiving Systems, Probabilistic Numerics
Research Project(s): Probabilistic Methods for Nonlinear Optimization
Bibtex Type: Conference Paper (inproceedings)
Paper Type: Conference

State: Accepted
URL: https://arxiv.org/abs/1612.05086

Links: Code

BibTex

@inproceedings{balles2017coupling,
  title = {Coupling Adaptive Batch Sizes with Learning Rates},
  author = {Balles, Lukas and Romero, Javier and Hennig, Philipp},
  booktitle = {Thirty-Third Conference on Uncertainty in Artificial Intelligence (UAI)},
  editors = {Gal Elidan and Kristian Kersting},
  year = {2017},
  url = {https://arxiv.org/abs/1612.05086}
}