[HTML][HTML] From turing to transformers: A comprehensive review and tutorial on the evolution and applications of generative transformer models

EY Zhang, AD Cheok, Z Pan, J Cai, Y Yan - Sci, 2023 - mdpi.com
In recent years, generative transformers have become increasingly prevalent in the field of
artificial intelligence, especially within the scope of natural language processing. This paper …

[PDF][PDF] A survey of large language models

WX Zhao, K Zhou, J Li, T Tang… - arxiv preprint arxiv …, 2023 - paper-notes.zhjwpku.com
Ever since the Turing Test was proposed in the 1950s, humans have explored the mastering
of language intelligence by machine. Language is essentially a complex, intricate system of …

Medusa: Simple llm inference acceleration framework with multiple decoding heads

T Cai, Y Li, Z Geng, H Peng, JD Lee, D Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) employ auto-regressive decoding that requires sequential
computation, with each step reliant on the previous one's output. This creates a bottleneck …

[HTML][HTML] Applying large language models and chain-of-thought for automatic scoring

GG Lee, E Latif, X Wu, N Liu, X Zhai - Computers and Education: Artificial …, 2024 - Elsevier
This study investigates the application of large language models (LLMs), specifically GPT-
3.5 and GPT-4, with Chain-of-Though (CoT) in the automatic scoring of student-written …

Evaluating the world model implicit in a generative model

K Vafa, J Chen, A Rambachan… - Advances in …, 2025 - proceedings.neurips.cc
Recent work suggests that large language models may implicitly learn world models. How
should we assess this possibility? We formalize this question for the case where the …

Evaluating large language models on controlled generation tasks

J Sun, Y Tian, W Zhou, N Xu, Q Hu, R Gupta… - arxiv preprint arxiv …, 2023 - arxiv.org
While recent studies have looked into the abilities of large language models in various
benchmark tasks, including question generation, reading comprehension, multilingual and …

A thorough examination of decoding methods in the era of llms

C Shi, H Yang, D Cai, Z Zhang, Y Wang, Y Yang… - arxiv preprint arxiv …, 2024 - arxiv.org
Decoding methods play an indispensable role in converting language models from next-
token predictors into practical task solvers. Prior research on decoding methods, primarily …

From decoding to meta-generation: Inference-time algorithms for large language models

S Welleck, A Bertsch, M Finlayson… - arxiv preprint arxiv …, 2024 - arxiv.org
One of the most striking findings in modern research on large language models (LLMs) is
that scaling up compute during training leads to better results. However, less attention has …

Beyond extractive: advancing abstractive automatic text summarization in norwegian with transformers

JJ Navjord, JMR Korsvik - 2023 - nmbu.brage.unit.no
Automatic summarization is a key area in natural language processing (NLP) and machine
learning which attempts to generate informative summaries of articles and documents …

Uncertainty in natural language generation: From theory to applications

J Baan, N Daheim, E Ilia, D Ulmer, HS Li… - arxiv preprint arxiv …, 2023 - arxiv.org
Recent advances of powerful Language Models have allowed Natural Language
Generation (NLG) to emerge as an important technology that can not only perform traditional …