Download citation

Decomposing and Optimizing Regret in Classical Multi-Armed Bandit Algorithms: ETC, UCB, and Thompson Sampling

ITM Web Conf., 80 (2025) 02003
DOI: https://doi.org/10.1051/itmconf/20258002003