scispace - formally typeset
Open AccessPosted Content

FedSplit: An algorithmic framework for fast federated optimization

Reads0
Chats0
TLDR
FedSplit is introduced, a class of algorithms based on operator splitting procedures for solving distributed convex minimization with additive structure and theory shows that these methods are provably robust to inexact computation of intermediate local quantities.
Abstract
Motivated by federated learning, we consider the hub-and-spoke model of distributed optimization in which a central authority coordinates the computation of a solution among many agents while limiting communication. We first study some past procedures for federated optimization, and show that their fixed points need not correspond to stationary points of the original optimization problem, even in simple convex settings with deterministic updates. In order to remedy these issues, we introduce FedSplit, a class of algorithms based on operator splitting procedures for solving distributed convex minimization with additive structure. We prove that these procedures have the correct fixed points, corresponding to optima of the original optimization problem, and we characterize their convergence rates under different settings. Our theory shows that these methods are provably robust to inexact computation of intermediate local quantities. We complement our theory with some simple experiments that demonstrate the benefits of our methods in practice.

read more

Citations
More filters
Proceedings Article

Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization

TL;DR: This paper provides the first principled understanding of the solution bias and the convergence slowdown due to objective inconsistency and proposes FedNova, a normalized averaging method that eliminates objective inconsistency while preserving fast error convergence.
Journal Article

Client Selection in Federated Learning: Convergence Analysis and Power-of-Choice Selection Strategies

TL;DR: This paper presents the first convergence analysis of federated optimization for biased client selection strategies, and quantifies how the selection bias affects convergence speed, and proposes Power-of-Choice, a communication- and computation-efficient client selection framework that can flexibly span the trade-off between convergence speed and solution bias.
Posted Content

Lower Bounds and Optimal Algorithms for Personalized Federated Learning

TL;DR: This work establishes the first lower bounds for this formulation of personalized federated learning, for both the communication complexity and the local oracle complexity, and designs several optimal methods matching these lower bounds in almost all regimes.
Journal ArticleDOI

FedPD: A Federated Learning Framework With Adaptivity to Non-IID Data

TL;DR: In this article, a primal-dual optimization strategy was proposed to design federated learning algorithms that are provably fast and require as few assumptions as possible, which can deal with nonconvex objective functions, achieves the best possible optimization and communication complexity.
Posted Content

Local SGD: Unified Theory and New Efficient Methods

TL;DR: This work was supported by the KAUST baseline research grant of P. Richt´arik and the research of E. Gorbunov was also partially funded by the Ministry of Science and Higher Education of the Russian Federation and RFBR, project number 19-31-51001.
References
More filters
Book

Convex Optimization

TL;DR: In this article, the focus is on recognizing convex optimization problems and then finding the most appropriate technique for solving them, and a comprehensive introduction to the subject is given. But the focus of this book is not on the optimization problem itself, but on the problem of finding the appropriate technique to solve it.
Book

Distributed Optimization and Statistical Learning Via the Alternating Direction Method of Multipliers

TL;DR: It is argued that the alternating direction method of multipliers is well suited to distributed convex optimization, and in particular to large-scale problems arising in statistics, machine learning, and related areas.
Posted Content

Communication-Efficient Learning of Deep Networks from Decentralized Data

TL;DR: This work presents a practical method for the federated learning of deep networks based on iterative model averaging, and conducts an extensive empirical evaluation, considering five different model architectures and four datasets.
Book

Parallel and Distributed Computation: Numerical Methods

TL;DR: This work discusses parallel and distributed architectures, complexity measures, and communication and synchronization issues, and it presents both Jacobi and Gauss-Seidel iterations, which serve as algorithms of reference for many of the computational approaches addressed later.
Related Papers (5)