Decomposing and Optimizing Regret in Classical Multi-Armed Bandit Algorithms: ETC, UCB, and Thompson Sampling
ITM Web Conf., 80 (2025) 02003
Published online: 16 December 2025
DOI: 10.1051/itmconf/20258002003

