Open Access
| Issue |
ITM Web Conf.
Volume 84, 2026
2026 International Conference on Advent Trends in Computational Intelligence and Data Science (ATCIDS 2026)
|
|
|---|---|---|
| Article Number | 03003 | |
| Number of page(s) | 12 | |
| Section | Large Language Models, Generative AI, and Multimodal Learning | |
| DOI | https://doi.org/10.1051/itmconf/20268403003 | |
| Published online | 06 April 2026 | |
- T. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal et al., Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 33, 1877–1901 (2020) [Google Scholar]
- S. Xie, A. Raghunathan, P. Liang, T. Ma, An explanation of in-context learning as implicit Bayesian inference. Int. Conf. Learn. Represent. (ICLR 2022) [Google Scholar]
- S. Garg, P. Nakkiran, D. Tsipras, What can transformers learn in-context? a case study of simple function classes. Adv. Neural Inf. Process. Syst. 35 (2022) [Google Scholar]
- O. Rubin, D. Herzig, A. Berant, Learning to retrieve prompts for in-context learning. Conf. North Amer. Ch. Assoc. Comput. Linguist. (NAACL 2022) [Google Scholar]
- O. Press, N. A. Smith, M. Lewis, Train short, test long: Attention with linear biases (ALiBi) enables length extrapolation. arXiv preprint arXiv:2108.12409 (2021) [Google Scholar]
- J. von Oswald, V. V. Ramasesh, W. Czarnecki, J. Kirkpatrick, C. Clopath, Transformers learn in-context by gradient descent. Nat. Mach. Intell. 5, 850–862 (2023) [Google Scholar]
- R. Agarwal, S. Kundu, H. Shi, J. Zhao, Many-shot in-context learning. Adv. Neural Inf. Process. Syst. (NeurIPS 2024 Spotlight) [Google Scholar]
- M. Hahn, N. Goyal, A theory of emergent in-context learning as implicit structure induction. arXiv preprint arXiv:2303.07971 (2023) [Google Scholar]
- H. Liu, S. Liu, Z. Liang, Y. Ruan, J. Wei, D. Zhou, Lost in the middle: how language models use long contexts. Trans. Assoc. Comput. Linguist. 12, 157–173 (2024) [Google Scholar]
- R. Schaeffer, B. Miranda, S. Koyejo, Are emergent abilities of large language models a mirage. arXiv preprint arXiv:2304.15004 (2023) [Google Scholar]
- K. Greshake, S. Abdelnabi, S. Mishra, C. Endres, T. Holz, M. Fritz, Not what you’ve signed up for: Compromising real-world LLM-integrated applications with indirect prompt injection. arXiv:2302.12173 (2023) [Google Scholar]
- Y. Li, G. Zhang, Q. D. Do, X. Yue, W. Chen, LongICLBench: Benchmarking long-context in-context learning. arXiv preprint arXiv:2404.02060 (2024) [Google Scholar]
- F. Falck, Z. Wang, C.C. Holmes, Is in-context learning in large language models Bayesian? A martingale perspective. Proc. 41st Int. Conf. Mach. Learn. (ICML 2024), PMLR 235, 12784–12805 (2024) [Google Scholar]
- A. Srivastava, A. Rastogi, A. Misra, et al., Beyond the Imitation Game: Quantifying and extrapolating behaviors of language models. arXiv preprint arXiv:2206.04615 (2022). [Google Scholar]
- J.-B. Alayrac, J. Donahue, P. Luc et al., Flamingo: a visual language model for few-shot learning. Adv. Neural Inf. Process. Syst. (2022) [Google Scholar]
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.

