JMLR

Imprecise Multi-Armed Bandits: Representing Irreducible Uncertainty as a Zero-Sum Game

Authors
Vanessa Kosoy
Research Topics
Machine Learning
Paper Information
  • Journal:
    Journal of Machine Learning Research
  • Added to Tracker:
    Sep 08, 2025
Abstract

We introduce a novel multi-armed bandit framework, where each arm is associated with a fixed unknown credal set over the space of outcomes (which can be richer than just the reward). The arm-to-credal-set correspondence comes from a known class of hypotheses. We then define a notion of regret corresponding to the lower prevision defined by these credal sets. Equivalently, the setting can be regarded as a two-player zero-sum game, where, on each round, the agent chooses an arm and the adversary chooses the distribution over outcomes from a set of options associated with this arm. The regret is defined with respect to the value of game. For certain natural hypothesis classes, loosely analogous to stochastic linear bandits (which are a special case of the resulting setting), we propose an algorithm and prove a corresponding upper bound on regret.

Author Details
Vanessa Kosoy
Author
Research Topics & Keywords
Machine Learning
Research Area
Citation Information
APA Format
Vanessa Kosoy . Imprecise Multi-Armed Bandits: Representing Irreducible Uncertainty as a Zero-Sum Game. Journal of Machine Learning Research .
BibTeX Format
@article{paper471,
  title = { Imprecise Multi-Armed Bandits: Representing Irreducible Uncertainty as a Zero-Sum Game },
  author = { Vanessa Kosoy },
  journal = { Journal of Machine Learning Research },
  url = { https://www.jmlr.org/papers/v26/24-2001.html }
}