scispace - formally typeset
Open AccessJournal ArticleDOI

Network Newton Distributed Optimization Methods

TLDR
This paper proposes the network Newton (NN) method as a distributed algorithm that incorporates second-order information via distributed implementation of approximations of a suitably chosen Newton step and proves convergence to a point close to the optimal argument at a rate that is at least linear.
Abstract
We study the problem of minimizing a sum of convex objective functions, where the components of the objective are available at different nodes of a network and nodes are allowed to only communicate with their neighbors. The use of distributed gradient methods is a common approach to solve this problem. Their popularity notwithstanding, these methods exhibit slow convergence and a consequent large number of communications between nodes to approach the optimal argument because they rely on first-order information only. This paper proposes the network Newton (NN) method as a distributed algorithm that incorporates second-order information. This is done via distributed implementation of approximations of a suitably chosen Newton step. The approximations are obtained by truncation of the Newton step's Taylor expansion. This leads to a family of methods defined by the number $K$ of Taylor series terms kept in the approximation. When keeping $K$ terms of the Taylor series, the method is called NN- $K$ and can be implemented through the aggregation of information in $K$ -hop neighborhoods. Convergence to a point close to the optimal argument at a rate that is at least linear is proven and the existence of a tradeoff between convergence time and the distance to the optimal argument is shown. The numerical experiments corroborate reductions in the number of iterations and the communication cost that are necessary to achieve convergence relative to first-order alternatives.

read more

Citations
More filters
Journal ArticleDOI

Decentralized Quasi-Newton Methods

TL;DR: The decentralized Broyden–Fletcher–Goldfarb–Shanno (D-BFGS) method is introduced as a variation of the BFGS quasi-Newton method for solving decentralized optimization problems.
Proceedings ArticleDOI

A Push-Pull Gradient Method for Distributed Optimization in Networks

TL;DR: In this article, a push-pull gradient method is proposed to minimize the sum of the cost functions while obeying the network connectivity structure, and the algorithm converges linearly for strongly convex and smooth objective functions over a directed static network.
Journal ArticleDOI

An Exact Quantized Decentralized Gradient Descent Algorithm

TL;DR: The Quantized Decentralized Gradient Descent algorithm is proposed, in which nodes update their local decision variables by combining the quantized information received from their neighbors with their local information, and it is proved that under standard strong convexity and smoothness assumptions for the objective function, QDGD achieves a vanishing mean solution error under customary conditions for quantizers.
Journal ArticleDOI

ByRDiE: Byzantine-Resilient Distributed Coordinate Descent for Decentralized Learning

TL;DR: This paper focuses on the problem of Byzantine failures, which are the hardest to safeguard against in distributed algorithms, and develops and analyzes an algorithm termed Byzantine-resilient distributed coordinate descent that enables distributed learning in the presence of Byzantine fails.
Journal ArticleDOI

Balancing Communication and Computation in Distributed Optimization

TL;DR: In this article, the authors propose an adaptive cost framework that adjusts the cost measure depending on the features of various applications, where communication and computation steps are explicitly decomposed to enable algorithm customization for various applications.
References
More filters
Book

Convex Optimization

TL;DR: In this article, the focus is on recognizing convex optimization problems and then finding the most appropriate technique for solving them, and a comprehensive introduction to the subject is given. But the focus of this book is not on the optimization problem itself, but on the problem of finding the appropriate technique to solve it.
Book

Distributed Optimization and Statistical Learning Via the Alternating Direction Method of Multipliers

TL;DR: It is argued that the alternating direction method of multipliers is well suited to distributed convex optimization, and in particular to large-scale problems arising in statistics, machine learning, and related areas.
Book

Introductory Lectures on Convex Optimization: A Basic Course

TL;DR: A polynomial-time interior-point method for linear optimization was proposed in this paper, where the complexity bound was not only in its complexity, but also in the theoretical pre- diction of its high efficiency was supported by excellent computational results.
Journal ArticleDOI

Distributed Subgradient Methods for Multi-Agent Optimization

TL;DR: The authors' convergence rate results explicitly characterize the tradeoff between a desired accuracy of the generated approximate optimal solutions and the number of iterations needed to achieve the accuracy.
Journal ArticleDOI

An Overview of Recent Progress in the Study of Distributed Multi-Agent Coordination

TL;DR: In this article, the authors reviewed some main results and progress in distributed multi-agent coordination, focusing on papers published in major control systems and robotics journals since 2006 and proposed several promising research directions along with some open problems that are deemed important for further investigations.
Related Papers (5)