Sign in

Information-Theoretic Generalization Bounds for Stochastic Gradient Descent

By Gergely Neu
We study the generalization properties of the popular stochastic optimization method known as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our main contribution is providing upper bounds on the generalization error that depend on local statistics of the stochastic gradients evaluated along the path of iterates calculated... Show more
July 20, 2021
=
0
Loading PDF…
Loading full text...
Similar articles
Loading recommendations...
=
0
x1
Information-Theoretic Generalization Bounds for Stochastic Gradient Descent
Click on play to start listening