scispace - formally typeset
Z

Zhaoxian Wu

Researcher at Sun Yat-sen University

Publications -  11
Citations -  174

Zhaoxian Wu is an academic researcher from Sun Yat-sen University. The author has contributed to research in topics: Stochastic gradient descent & Computer science. The author has an hindex of 4, co-authored 9 publications receiving 57 citations.

Papers
More filters
Journal ArticleDOI

Federated Variance-Reduced Stochastic Gradient Descent With Robustness to Byzantine Attacks

TL;DR: A novel Byzantine attack resilient distributed (Byrd-) SAGA approach is introduced for federated learning tasks involving multiple workers that corroborate the robustness to various Byzantine attacks, as well as the merits of Byrd-SAGA over ByzantineAttack resilient distributed SGD.
Posted Content

Federated Variance-Reduced Stochastic Gradient Descent with Robustness to Byzantine Attacks

TL;DR: In this article, a Byzantine attack resilient distributed (Byrd-) SAGA approach was proposed for learning tasks involving finite-sum optimization over networks. But the robustness of the Byrd-SAGA to Byzantine attacks was not evaluated.
Journal ArticleDOI

Byzantine-Resilient Decentralized TD Learning with Linear Function Approximation

TL;DR: This paper proposes a trimmed-mean based decentralized TD(λ) algorithm to perform policy evaluation in this setting and establishes the finite-time convergence rate, as well as the asymptotic learning error that depends on the number of Byzantine agents.
Posted Content

Byzantine-Robust Variance-Reduced Federated Learning over Distributed Non-i.i.d. Data

TL;DR: A Byzantine-robust variance-reduced stochastic gradient descent method to solve the distributed finite-sum minimization problem when the data on the workers are not independent and identically distributed (i.i.d.), and it is proved that the proposed method reaches a neighborhood of the optimal solution with linear convergence rate.
Journal ArticleDOI

Communication-Censored Distributed Stochastic Gradient Descent.

TL;DR: In this article, a communication-censored distributed stochastic gradient descent (CSGD) algorithm is proposed to reduce the burden of data transmission in distributed machine learning applications by increasing the batch size.