JMLR

The ODE Method for Stochastic Approximation and Reinforcement Learning with Markovian Noise

Authors
Shuze Daniel Liu Shuhang Chen Shangtong Zhang
Paper Information
  • Journal:
    Journal of Machine Learning Research
  • Added to Tracker:
    Jul 15, 2025
Abstract

Stochastic approximation is a class of algorithms that update a vector iteratively, incrementally, and stochastically, including, e.g., stochastic gradient descent and temporal difference learning. One fundamental challenge in analyzing a stochastic approximation algorithm is to establish its stability, i.e., to show that the stochastic vector iterates are bounded almost surely. In this paper, we extend the celebrated Borkar-Meyn theorem for stability from the Martingale difference noise setting to the Markovian noise setting, which greatly improves its applicability in reinforcement learning, especially in those off-policy reinforcement learning algorithms with linear function approximation and eligibility traces. Central to our analysis is the diminishing asymptotic rate of change of a few functions, which is implied by both a form of the strong law of large numbers and a form of the law of the iterated logarithm.

Author Details
Shuze Daniel Liu
Author
Shuhang Chen
Author
Shangtong Zhang
Author
Citation Information
APA Format
Shuze Daniel Liu , Shuhang Chen & Shangtong Zhang . The ODE Method for Stochastic Approximation and Reinforcement Learning with Markovian Noise. Journal of Machine Learning Research .
BibTeX Format
@article{JMLR:v26:24-0100,
  author  = {Shuze Daniel Liu and Shuhang Chen and Shangtong Zhang},
  title   = {The ODE Method for Stochastic Approximation and Reinforcement Learning with Markovian Noise},
  journal = {Journal of Machine Learning Research},
  year    = {2025},
  volume  = {26},
  number  = {24},
  pages   = {1--76},
  url     = {http://jmlr.org/papers/v26/24-0100.html}
}
Related Papers