Sign in

Select without Fear: Almost All Mini-Batch Schedules Generalize Optimally

By Konstantinos Nikolakakis and others
We establish matching upper and lower generalization error bounds for mini-batch Gradient Descent (GD) training with either deterministic or stochastic, data-independent, but otherwise arbitrary batch selection rules. We consider smooth Lipschitz-convex/nonconvex/strongly-convex loss functions, and show that classical upper bounds for Stochastic GD (SGD) also hold verbatim for such arbitrary nonadaptive... Show more
October 23, 2023
=
0
Loading PDF…
Loading full text...
Similar articles
Loading recommendations...
=
0
x1
Select without Fear: Almost All Mini-Batch Schedules Generalize Optimally
Click on play to start listening