Sign in

Beating SGD Saturation with Tail-Averaging and Minibatching

By Nicole Mücke and others
While stochastic gradient descent (SGD) is one of the major workhorses in machine learning, the learning properties of many practically used variants are poorly understood. In this paper, we consider least squares learning in a nonparametric setting and contribute to filling this gap by focusing on the effect and interplay... Show more
May 26, 2019
=
0
Loading PDF…
Loading full text...
Similar articles
Loading recommendations...
=
0
x1
Beating SGD Saturation with Tail-Averaging and Minibatching
Click on play to start listening