The ODE Method for Stochastic Approximation and Reinforcement Learning with Markovian Noise
Authors
Paper Information
-
Journal:
Journal of Machine Learning Research -
Added to Tracker:
Jul 30, 2025
Abstract
Stochastic approximation is a class of algorithms that update a vector iteratively, incrementally, and stochastically, including, e.g., stochastic gradient descent and temporal difference learning. One fundamental challenge in analyzing a stochastic approximation algorithm is to establish its stability, i.e., to show that the stochastic vector iterates are bounded almost surely. In this paper, we extend the celebrated Borkar-Meyn theorem for stability from the Martingale difference noise setting to the Markovian noise setting, which greatly improves its applicability in reinforcement learning, especially in those off-policy reinforcement learning algorithms with linear function approximation and eligibility traces. Central to our analysis is the diminishing asymptotic rate of change of a few functions, which is implied by both a form of the strong law of large numbers and a form of the law of the iterated logarithm.
Author Details
Shuze Daniel Liu
AuthorShuhang Chen
AuthorShangtong Zhang
AuthorCitation Information
APA Format
Shuze Daniel Liu
,
Shuhang Chen
&
Shangtong Zhang
.
The ODE Method for Stochastic Approximation and Reinforcement Learning with Markovian Noise.
Journal of Machine Learning Research
.
BibTeX Format
@article{paper286,
title = { The ODE Method for Stochastic Approximation and Reinforcement Learning with Markovian Noise },
author = {
Shuze Daniel Liu
and Shuhang Chen
and Shangtong Zhang
},
journal = { Journal of Machine Learning Research },
url = { https://www.jmlr.org/papers/v26/24-0100.html }
}