Published February 19, 2024 | Version v1
Conference paper Open

Best-of-Both-Worlds Algorithms for Linear Contextual Bandits

  • 1. CENTAI Institute
  • 2. ROR icon University of Milan
  • 3. ROR icon The University of Tokyo
  • 4. ROR icon Politecnico di Milano

Description

We study best-of-both-worlds algorithms for K-armed linear contextual bandits. Our algorithms deliver near-optimal regret bounds in both the adversarial and stochastic regimes, without prior knowledge about the environment. In the stochastic regime, we achieve the polylogarithmic rate (dK)2polyln(dKT ) / ∆min where ∆min is the minimum suboptimality gap over the d-dimensional context space. In the adversarial regime, we obtain either the first-order O(dK√L∗) bound, or the second- order O(dK√Λ∗) bound, where L∗ is the cumulative loss of the best action and Λ∗ is a notion of the cumulative second moment for the losses incurred by the algorithm. Moreover, we develop an algorithm based on FTRL with Shannon entropy regularizer that does not require the knowledge of the inverse of the covariance matrix, and achieves a polylogarithmic regeret in the stochastic regime while obtaining O(dK √T) regret bounds in the adversarial regime.

Files

2312.15433v2.pdf

Files (654.2 kB)

Name Size Download all
md5:4522fd67702ffb6c398da5bf4cffd742
654.2 kB Preview Download

Additional details

Funding

European Commission
ELIAS – European Lighthouse of AI for Sustainability 101120237