A comprehensive overview of large language models

H Naveed, AU Khan, S Qiu, M Saqib, S Anwar… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) have recently demonstrated remarkable capabilities in
natural language processing tasks and beyond. This success of LLMs has led to a large …

[HTML][HTML] A survey of GPT-3 family large language models including ChatGPT and GPT-4

KS Kalyan - Natural Language Processing Journal, 2024 - Elsevier
Large language models (LLMs) are a special class of pretrained language models (PLMs)
obtained by scaling model size, pretraining corpus and computation. LLMs, because of their …

Reflexion: Language agents with verbal reinforcement learning

N Shinn, F Cassano, A Gopinath… - Advances in …, 2024 - proceedings.neurips.cc
Large language models (LLMs) have been increasingly used to interact with external
environments (eg, games, compilers, APIs) as goal-driven agents. However, it remains …

C-pack: Packed resources for general chinese embeddings

S **ao, Z Liu, P Zhang, N Muennighoff, D Lian… - Proceedings of the 47th …, 2024 - dl.acm.org
We introduce C-Pack, a package of resources that significantly advances the field of general
text embeddings for Chinese. C-Pack includes three critical resources. 1) C-MTP is a …

Gorilla: Large language model connected with massive apis

SG Patil, T Zhang, X Wang… - Advances in Neural …, 2025 - proceedings.neurips.cc
Abstract Large Language Models (LLMs) have seen an impressive wave of advances,
withmodels now excelling in a variety of tasks, such as mathematical reasoning andprogram …

Mathematical discoveries from program search with large language models

B Romera-Paredes, M Barekatain, A Novikov, M Balog… - Nature, 2024 - nature.com
Large language models (LLMs) have demonstrated tremendous capabilities in solving
complex tasks, from quantitative reasoning to understanding natural language. However …

Leandojo: Theorem proving with retrieval-augmented language models

K Yang, A Swope, A Gu, R Chalamala… - Advances in …, 2023 - proceedings.neurips.cc
Large language models (LLMs) have shown promise in proving formal theorems using proof
assistants such as Lean. However, existing methods are difficult to reproduce or build on …

Wizardcoder: Empowering code large language models with evol-instruct

Z Luo, C Xu, P Zhao, Q Sun, X Geng, W Hu… - arxiv preprint arxiv …, 2023 - arxiv.org
Code Large Language Models (Code LLMs), such as StarCoder, have demonstrated
exceptional performance in code-related tasks. However, most existing models are solely …

Pissa: Principal singular values and singular vectors adaptation of large language models

F Meng, Z Wang, M Zhang - Advances in Neural …, 2025 - proceedings.neurips.cc
To parameter-efficiently fine-tune (PEFT) large language models (LLMs), the low-rank
adaptation (LoRA) method approximates the model changes $\Delta W\in\mathbb …

Textbooks are all you need

S Gunasekar, Y Zhang, J Aneja, CCT Mendes… - arxiv preprint arxiv …, 2023 - arxiv.org
We introduce phi-1, a new large language model for code, with significantly smaller size
than competing models: phi-1 is a Transformer-based model with 1.3 B parameters, trained …