scispace - formally typeset
Search or ask a question
Topic

Rate of convergence

About: Rate of convergence is a research topic. Over the lifetime, 31257 publications have been published within this topic receiving 795334 citations. The topic is also known as: convergence rate.


Papers
More filters
Journal ArticleDOI
TL;DR: In this paper, the authors study a projected multi-agent subgradient algorithm under state-dependent communication and show that the algorithm converges to the same optimal solution with probability one under different assumptions on the local constraint sets and the stepsize sequence.
Abstract: We study distributed algorithms for solving global optimization problems in which the objective function is the sum of local objective functions of agents and the constraint set is given by the intersection of local constraint sets of agents. We assume that each agent knows only his own local objective function and constraint set, and exchanges information with the other agents over a randomly varying network topology to update his information state. We assume a state-dependent communication model over this topology: communication is Markovian with respect to the states of the agents and the probability with which the links are available depends on the states of the agents. We study a projected multi-agent subgradient algorithm under state-dependent communication. The state-dependence of the communication introduces significant challenges and couples the study of information exchange with the analysis of subgradient steps and projection errors. We first show that the multi-agent subgradient algorithm when used with a constant stepsize may result in the agent estimates to diverge with probability one. Under some assumptions on the stepsize sequence, we provide convergence rate bounds on a “disagreement metric” between the agent estimates. Our bounds are time-nonhomogeneous in the sense that they depend on the initial starting time. Despite this, we show that agent estimates reach an almost sure consensus and converge to the same optimal solution of the global optimization problem with probability one under different assumptions on the local constraint sets and the stepsize sequence.

139 citations

Journal ArticleDOI
01 Sep 1996
TL;DR: In this article, three types of optimal, continuously time-varying sliding modes for robust control of second-order uncertain dynamic systems subject to input constraint are presented, two of them incorporate straight sliding lines, and the third uses the so-called terminal slider, that is a curve that guarantees system error convergence to zero in finite time.
Abstract: Three types of optimal, continuously time-varying sliding mode for robust control of second-order uncertain dynamic systems subject to input constraint are presented. Two of the modes incorporate straight sliding lines, and the third uses the so-called terminal slider, that is a curve that guarantees system error convergence to zero in finite time. At first, all three lines adapt themselves to the initial conditions of the system, and afterwards they move in such a way that, for each of them, the integral of the absolute value of the systems error is minimised over the whole period of the control action. By this means, insensitivity of the system to external disturbances and parameter uncertainties is guaranteed from the very beginning of the proposed control action, and the system error convergence rate can be increased. Performance of the three control algorithms is compared, and the Lyapunov theory is used to prove the existence of a sliding mode on the lines.

139 citations

Proceedings Article
01 Dec 1997
TL;DR: It is shown that for discounted MDPs with discount factor γ > 1/2 the asymptotic rate of convergence of Q-learning is O(1/tR(1-γ)) if R(1 - γ) 0, where pmin and pmax now become the minimum and maximum state-action occupation frequencies corresponding to the stationary distribution.
Abstract: In this paper we show that for discounted MDPs with discount factor γ > 1/2 the asymptotic rate of convergence of Q-learning is O(1/tR(1-γ)) if R(1 - γ) 0, where pmin and pmax now become the minimum and maximum state-action occupation frequencies corresponding to the stationary distribution.

139 citations

Journal ArticleDOI
TL;DR: Weak and strong convergence for some generalized proximal point algorithms are proved, which include the Eckstein and Bertsekas generalized proxiesimal point algorithm, a contraction-proximal points algorithm, and inexact proximal points algorithms.
Abstract: Weak and strong convergence for some generalized proximal point algorithms are proved. These algorithms include the Eckstein and Bertsekas generalized proximal point algorithm, a contraction-proximal point algorithm, and inexact proximal point algorithms. Convergence rate is also considered.

139 citations


Network Information
Related Topics (5)
Partial differential equation
70.8K papers, 1.6M citations
89% related
Markov chain
51.9K papers, 1.3M citations
88% related
Optimization problem
96.4K papers, 2.1M citations
88% related
Differential equation
88K papers, 2M citations
88% related
Nonlinear system
208.1K papers, 4M citations
88% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20241
2023693
20221,530
20212,129
20202,036
20191,995