Reinforcement Learning for Infinite-Dimensional Systems
Authors
Paper Information
-
Journal:
Journal of Machine Learning Research -
Added to Tracker:
Dec 30, 2025
Abstract
Interest in reinforcement learning (RL) for large-scale systems, comprising extensive populations of intelligent agents interacting with heterogeneous environments, has surged significantly across diverse scientific domains in recent years. However, the large-scale nature of these systems often leads to high computational costs or reduced performance for most state-of-the-art RL techniques. To address these challenges, we propose a novel RL architecture and derive effective algorithms to learn optimal policies for arbitrarily large systems of agents. In our formulation, we model such systems as parameterized control systems defined on an infinite-dimensional function space. We then develop a moment kernel transform that maps the parameterized system and the value function into a reproducing kernel Hilbert space. This transformation generates a sequence of finite-dimensional moment representations for the RL problem, organized into a filtrated structure. Leveraging this RL filtration, we develop a hierarchical algorithm for learning optimal policies for the infinite-dimensional parameterized system. To enhance the algorithm's efficiency, we incorporate early stopping at each hierarchy, demonstrating the fast convergence property of the algorithm through the construction of a convergent spectral sequence. The performance and efficiency of the proposed algorithm are validated using practical examples in engineering and quantum systems.
Author Details
Wei Zhang
AuthorJr-Shin Li
AuthorCitation Information
APA Format
Wei Zhang
&
Jr-Shin Li
.
Reinforcement Learning for Infinite-Dimensional Systems.
Journal of Machine Learning Research
.
BibTeX Format
@article{paper716,
title = { Reinforcement Learning for Infinite-Dimensional Systems },
author = {
Wei Zhang
and Jr-Shin Li
},
journal = { Journal of Machine Learning Research },
url = { https://www.jmlr.org/papers/v26/24-1575.html }
}