scispace - formally typeset
B

Blake Woodworth

Researcher at Toyota Technological Institute at Chicago

Publications -  46
Citations -  2031

Blake Woodworth is an academic researcher from Toyota Technological Institute at Chicago. The author has contributed to research in topics: Upper and lower bounds & Gradient descent. The author has an hindex of 20, co-authored 42 publications receiving 1384 citations. Previous affiliations of Blake Woodworth include Toyota Technological Institute.

Papers
More filters
Proceedings Article

Implicit Regularization in Matrix Factorization

TL;DR: In this article, the authors studied implicit regularization when optimizing an underdetermined quadratic objective over a matrix $X$ with gradient descent on a factorization of X, and provided empirical and theoretical evidence that with small enough step sizes and initialization close enough to the origin, gradient descent converges to the minimum nuclear norm solution.
Posted Content

Lower Bounds for Non-Convex Stochastic Optimization.

TL;DR: It is proved that (in the worst case) any algorithm requires at least $\epsilon^{-4}$ queries to find an stationary point, and establishes that stochastic gradient descent is minimax optimal in this model.
Posted Content

Kernel and Rich Regimes in Overparametrized Models

TL;DR: This work shows how the scale of the initialization controls the transition between the "kernel" and "rich" regimes and affects generalization properties in multilayer homogeneous models and highlights an interesting role for the width of a model in the case that the predictor is not identically zero at initialization.
Proceedings ArticleDOI

Implicit Regularization in Matrix Factorization

TL;DR: It is conjecture and theoretical evidence that with small enough step sizes and initialization close enough to the origin, gradient descent on a full dimensional factorization converges to the minimum nuclear norm solution.
Posted Content

Is Local SGD Better than Minibatch SGD

TL;DR: For quadratic objectives it is proved that local SGD strictly dominates minibatch SGD and that accelerated localSGD is minimax optimal for quadratics and for general convex objectives the first guarantee that at least sometimes improves over minibatches SGD is provided.