scispace - formally typeset
Open AccessPosted Content

A Theorem of the Alternative for Personalized Federated Learning.

TLDR
This paper shows how the excess risks of personalized federated learning with a smooth, strongly convex loss depend on data heterogeneity from a minimax point of view, and reveals a surprising theorem of the alternative for personalized federation learning.
Abstract
A widely recognized difficulty in federated learning arises from the statistical heterogeneity among clients: local datasets often come from different but not entirely unrelated distributions, and personalization is, therefore, necessary to achieve optimal results from each individual's perspective. In this paper, we show how the excess risks of personalized federated learning with a smooth, strongly convex loss depend on data heterogeneity from a minimax point of view. Our analysis reveals a surprising theorem of the alternative for personalized federated learning: there exists a threshold such that (a) if a certain measure of data heterogeneity is below this threshold, the FedAvg algorithm [McMahan et al., 2017] is minimax optimal; (b) when the measure of heterogeneity is above this threshold, then doing pure local training (i.e., clients solve empirical risk minimization problems on their local datasets without any communication) is minimax optimal. As an implication, our results show that the presumably difficult (infinite-dimensional) problem of adapting to client-wise heterogeneity can be reduced to a simple binary decision problem of choosing between the two baseline algorithms. Our analysis relies on a new notion of algorithmic stability that takes into account the nature of federated learning.

read more

Citations
More filters
Proceedings ArticleDOI

FedAvg with Fine Tuning: Local Updates Lead to Representation Learning

TL;DR: The reason behind generalizability of the FedAvg’s output is its power in learning the common data representation among the clients’ tasks, by leveraging the diversity among client data distributions via local updates, in the multi-task linear representation setting.
Journal Article

Adaptive and Robust Multi-task Learning

Yaqiong Duan, +1 more
- 10 Feb 2022 - 
TL;DR: In this article , a family of adaptive methods that automatically utilize possible similarities among those tasks while carefully handling their differences is proposed. But their robustness against outlier tasks is questionable.
Proceedings ArticleDOI

Privacy-Preserving Federated Multi-Task Linear Regression: A One-Shot Linear Mixing Approach Inspired By Graph Regularization

TL;DR: This work focuses on the federated multi-task linear regression setting, where each machine possesses its own data for individual tasks and sharing the full local data between machines is prohibited, and proposes a novel fusion framework that only requires a one-shot communication of local estimates.
Posted Content

Personalized Federated Learning with Gaussian Processes

TL;DR: In this paper, a solution to PFL that is based on Gaussian processes (GPs) with deep kernel learning is presented, where a shared kernel function across all clients, parameterized by a neural network, with a personal GP classifier for each client.
Journal ArticleDOI

Personalized Federated Learning with Multiple Known Clusters

TL;DR: This work develops an algorithm that allows each cluster to communicate independently and derive the convergence results, and studies a hierarchical linear model to theoretically demonstrate that this approach outperforms agents learning independently and agents learning a single shared weight.
References
More filters
Proceedings ArticleDOI

Survey of Personalization Techniques for Federated Learning

TL;DR: The need for personalization is highlighted and recent research on this topic is surveyed and several techniques have been proposed to personalize global models to work better for individual clients.
Proceedings Article

Learning-to-Learn Stochastic Gradient Descent with Biased Regularization

TL;DR: A key feature of the results is that, when the number of tasks grows and their variance is relatively small, the learning-to-learn approach has a significant advantage over learning each task in isolation by Stochastic Gradient Descent without a bias term.
Journal Article

Algorithmic Stability and Meta-Learning

TL;DR: A general method is given to prove generalisation error bounds for meta-algorithms searching spaces of uniformly stable algorithms and an application to regularized least squares regression is presented.
Posted Content

Provable Guarantees for Gradient-Based Meta-Learning

TL;DR: In this article, the authors study the problem of meta-learning through the lens of online convex optimization and develop a meta-algorithm bridging the gap between popular gradient-based meta learning and classical regularization-based multi-task transfer methods.
Proceedings Article

Learning To Learn Around A Common Mean

TL;DR: It is shown that the LTL problem can be reformulated as a Least Squares (LS) problem and a novel meta- algorithm is exploited to efficiently solve it, and a bound for the generalization error of the meta-algorithm is presented, which suggests the right splitting parameter to choose.