Type: Article
Publication Date: 2022-11-29
Citations: 5
DOI: https://doi.org/10.1109/tac.2022.3201034
In this article, we study a decentralized variant of stochastic approximation (SA), a data-driven approach for finding the root of an operator under noisy measurements. A network of agents, each with its own operator and data observations, cooperatively find the fixed point of the aggregate operator over a decentralized communication graph. Our main contribution is to provide a finite-time analysis of this decentralized SA method when the data observed at each agent are sampled from a Markov process; this lack of independence makes the iterates biased and (potentially) unbounded. Under fairly standard assumptions, we show that the convergence rate of the proposed method is essentially the same as if the samples were independent, differing only by a log factor that accounts for the mixing time of the Markov processes. The key idea in our analysis is to introduce a novel Lyapunov–Razumikhin function, motivated by the one used in analyzing the stability of delayed ordinary differential equations. We also discuss applications of the proposed method on a number of interesting learning problems in multiagent systems.
Action | Title | Year | Authors |
---|---|---|---|
+ PDF Chat | A Two-Time-Scale Stochastic Optimization Framework with Applications in Control and Reinforcement Learning | 2024 |
Sihan Zeng Thinh T. Doan Justin Romberg |