Open Access
Issue |
ITM Web Conf.
Volume 73, 2025
International Workshop on Advanced Applications of Deep Learning in Image Processing (IWADI 2024)
|
|
---|---|---|
Article Number | 01012 | |
Number of page(s) | 10 | |
Section | Reinforcement Learning and Optimization Techniques | |
DOI | https://doi.org/10.1051/itmconf/20257301012 | |
Published online | 17 February 2025 |
- D. Chakrabarti, R. Kumar, F. Radlinski, and E. Upfal, Mortal multi-armed bandits, Adv. Neural Inf. Process. Syst., 21, (2008). [Google Scholar]
- Y. Shao and Z. Fang, On multi-armed bandit with impatient arms, In Forty-first International Conference on Machine Learning. [Google Scholar]
- S. Zhang, Utilizing reinforcement learning bandit algorithms in advertising optimization, Highlights in Science, Engineering and Technology, 94, 195-200 (2024). [CrossRef] [Google Scholar]
- Y. Nishimura, Ad recommender system analysis by the multi-armed bandit problem. [Google Scholar]
- A. Dimitrova, Ad campaign optimization—a targeted contextual bandit approach, Master's thesis, ETH Zurich, Department of Computer Science (2018). [Google Scholar]
- K. Abhishek, Design and analysis of algorithms for combinatorial multi-arm bandit problems under complex environments, Doctoral dissertation, International Institute of Information Technology Hyderabad (2021). [Google Scholar]
- K. Hu, Comprehensive introduction and analysis of the UCB algorithm in multi-armed bandit problems, Highlights in Science, Engineering and Technology, 94, 247-250 (2024). [CrossRef] [Google Scholar]
- O. A. Hanna, L. Yang, and C. Fragouli, Solving multi-arm bandit using a few bits of communication, In International Conference on Artificial Intelligence and Statistics, pp. 11215-11236, PMLR (2022). [Google Scholar]
- C. Zeng, Q. Wang, S. Mokhtari, and T. Li, Online context-aware recommendation with time varying multi-armed bandit, In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 2025-2034 (2016). [Google Scholar]
- Y. Zhao, N. Behari, E. Hughes, E. Zhang, D. Nagaraj, K. Tuyls, and M. Tambe, Towards a pretrained model for restless bandits via multi-arm generalization, In Proceedings of IJCAI, pp. 321-329 (2024). [Google Scholar]
- D. N. Hill, H. Nassif, Y. Liu, A. Iyer, and S. V. N. Vishwanathan, An efficient bandit algorithm for realtime multivariate optimization, In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1813-1821 (2017). [Google Scholar]
- L. Tang, R. Rosales, A. Singh, and D. Agarwal, Automatic ad format selection via contextual bandits, In Proceedings of the 22nd ACM International Conference on Information & Knowledge Management, pp. 1587-1594 (2013). [Google Scholar]
- D. Bouneffouf and I. Rish, A survey on practical applications of multi-armed and contextual bandits, arXiv preprint arXiv:1904.10040 (2019). [Google Scholar]
- J. Zuo and C. Joe-Wong, Combinatorial multi-armed bandits for resource allocation, In 2021 55th Annual Conference on Information Sciences and Systems (CISS), pp. 1-4, IEEE (2021). [Google Scholar]
- A. Sankararaman, A. Ganesh, and S. Shakkottai, Social learning in multi agent multi armed bandits, Proceedings of the ACM on Measurement and Analysis of Computing Systems, 3(3), 1-35 (2019). [CrossRef] [Google Scholar]
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.