skip to main content
research-article

A Comprehensive Overview of Large Language Models

Authors:
Humza Naveed
The University of Sydney, Sydney, Australia
,
Asad Ullah Khan
University of Engineering and Technology, Lahore, Pakistan
,
Shi Qiu
The Chinese University of Hong Kong, Hong Kong, Hong Kong
,
Muhammad Saqib
University of Technology Sydney, Sydney, Australia and Commonwealth Scientific and Industrial Research Organisation, Canberra, Australia
,
Saeed Anwar
King Fahd University of Petroleum and Minerals, Dhahran, Saudi Arabia and SDAIAKFUPM Joint Research Center for Artificial Intelligence, Dhahran, Saudi Arabia
,
Muhammad Usman
University of Ontario Institute of Technology, Oshawa, Ontario, Canada
,
Naveed Akhtar
University of Melbourne VCCC, Parkville, Australia and The University of Western Australia, Perth, Australia
,
Nick Barnes
Australian National University, Canberra, Australia
,
Ajmal Mian
The University of Western Australia - Perth Campus, Perth, Australia
Authors Info & Claims
Published: 18 August 2025 Publication History

Abstract

Large Language Models (LLMs) have recently demonstrated remarkable capabilities in natural language processing tasks and beyond. This success of LLMs has led to a large influx of research contributions in this direction. These works encompass diverse topics such as architectural innovations, better training strategies, context length improvements, fine-tuning, multimodal LLMs, robotics, datasets, benchmarking, efficiency, and more. With the rapid development of techniques and regular breakthroughs in LLM research, it has become considerably challenging to perceive the bigger picture of the advances in this direction. Considering the rapidly emerging plethora of literature on LLMs, it is imperative that the research community is able to benefit from a concise yet comprehensive overview of the recent developments in this field. This article provides an overview of the literature on a broad range of LLM-related concepts. Our self-contained comprehensive overview of LLMs discusses relevant background concepts along with covering the advanced topics at the frontier of research in LLMs. This review article is intended to provide not only a systematic survey but also a quick, comprehensive reference for the researchers and practitioners to draw insights from extensive, informative summaries of the existing works to advance the LLM research.

References

[1]
A. Chernyavskiy, D. Ilvovsky, and P. Nakov. 2021. Transformers: “The end of history” for natural language processing? In Proceedings of the Machine Learning and Knowledge Discovery in Databases. Research Track: European Conference (ECML PKDD ’21). Springer, 677–693.
[2]
A. Wang, Y. Pruksachatkun, N. Nangia, A. Singh, J. Michael, F. Hill, O. Levy, and S. Bowman. 2019. SuperGLUE: A stickier benchmark for general-purpose language understanding systems. In Proceedings of the Advances in Neural Information Processing Systems, Vol. 32
[3]
D. Adiwardana, M.-T. Luong, D. R. So, J. Hall, N. Fiedel, R. Thoppilan, Z. Yang, A. Kulshreshtha, G. Nemade, Y. Lu, et al. 2020. Towards a human-like open-domain chatbot. arXiv:2001.09977. Retrieved from https://arxiv.org/abs/2001.09977
[4]
B. A. Y. Arcas. 2022. Do large language models understand us? Daedalus 151, 2 (2022), 183–197.

Cited By

View all
  • (2026)Fusion of equipment reliability data through knowledge graphsReliability Engineering & System Safety10.1016/j.ress.2025.111785267(111785)Online publication date: Mar-2026
  • (2026)Automated carbon-aware assessment of openBIM-based ductwork design using knowledge graph–augmented LLM multi-agent frameworkAutomation in Construction10.1016/j.autcon.2025.106611181(106611)Online publication date: Jan-2026
  • (2025)Semi-Supervised Prefix Tuning of Large Language Models for Industrial Fault Diagnosis with Big DataBig Data Mining and Analytics10.26599/BDMA.2025.90200388:6(1353-1368)Online publication date: Dec-2025
  • Show More Cited By

Index Terms

  1. A Comprehensive Overview of Large Language Models

    Recommendations

    Comments