scispace - formally typeset
F

Fangyu Zou

Researcher at Stony Brook University

Publications -  9
Citations -  379

Fangyu Zou is an academic researcher from Stony Brook University. The author has contributed to research in topics: Momentum (technical analysis) & Adaptive learning. The author has an hindex of 4, co-authored 9 publications receiving 211 citations.

Papers
More filters
Proceedings ArticleDOI

A Sufficient Condition for Convergences of Adam and RMSProp

TL;DR: In this paper, an alternative easy-to-check sufficient condition, which merely depends on the parameters of the base learning rate and combinations of historical second-order moments, was proposed to guarantee the global convergence of generic Adam/RMSProp for solving large-scale non-convex stochastic optimization.
Posted Content

A Sufficient Condition for Convergences of Adam and RMSProp.

TL;DR: An alternative easy-to-check sufficient condition is introduced, which merely depends on the parameters of the base learning rate and combinations of historical second-order moments, to guarantee the global convergence of generic Adam/RMSProp for solving large-scale non-convex stochastic optimization.
Posted Content

Weighted AdaGrad with Unified Momentum

TL;DR: AdaUSM is proposed, which has the main characteristics that it incorporates a unified momentum scheme which covers both the heavy ball momentum and the Nesterov accelerated gradient momentum, and adopts a novel weighted adaptive learning rate that can unify the learning rates of AdaGrad, AccAdaGrad, Adam, and RMSProp.
Posted Content

On the Convergence of AdaGrad with Momentum for Training Deep Neural Networks

Fangyu Zou, +1 more
TL;DR: Two new adaptive stochastic gradient methods called AdaHB and AdaNAG are proposed which integrate coordinate-wise AdaGrad with heavy ball momentum and Nesterov accelerated gradient momentum, respectively and are jointly characterized by leveraging a newly developed unified formulation of these two momentum mechanisms.
Posted Content

On the Convergence of Weighted AdaGrad with Momentum for Training Deep Neural Networks

Fangyu Zou, +1 more
- 10 Aug 2018 - 
TL;DR: Two new adaptive stochastic gradient methods are proposed called AdaHB and AdaNAG which integrate a novel weighted coordinate-wise AdaGrad with heavy ball momentum and Nesterov accelerated gradient momentum, respectively.