[PDF][PDF] A survey of large language models

WX Zhao, K Zhou, J Li, T Tang… - arxiv preprint arxiv …, 2023 - paper-notes.zhjwpku.com
Ever since the Turing Test was proposed in the 1950s, humans have explored the mastering
of language intelligence by machine. Language is essentially a complex, intricate system of …

Metacognitive capabilities of llms: An exploration in mathematical problem solving

A Didolkar, A Goyal, NR Ke, S Guo, M Valko… - arxiv preprint arxiv …, 2024 - arxiv.org
Metacognitive knowledge refers to humans' intuitive knowledge of their own thinking and
reasoning processes. Today's best LLMs clearly possess some reasoning processes. The …

Data curation via joint example selection further accelerates multimodal learning

T Evans, N Parthasarathy, H Merzic… - arxiv preprint arxiv …, 2024 - arxiv.org
Data curation is an essential component of large-scale pretraining. In this work, we
demonstrate that jointly selecting batches of data is more effective for learning than selecting …

International Scientific Report on the Safety of Advanced AI (Interim Report)

Y Bengio, S Mindermann, D Privitera… - arxiv preprint arxiv …, 2024 - arxiv.org
This is the interim publication of the first International Scientific Report on the Safety of
Advanced AI. The report synthesises the scientific understanding of general-purpose AI--AI …

Mates: Model-aware data selection for efficient pretraining with data influence models

Z Yu, S Das, C **ong - Advances in Neural Information …, 2025 - proceedings.neurips.cc
Pretraining data selection has the potential to improve language model pretraining efficiency
by utilizing higher-quality data from massive web data corpora. Current data selection …

A systematic assessment of openai o1-preview for higher order thinking in education

E Latif, Y Zhou, S Guo, Y Gao, L Shi… - arxiv preprint arxiv …, 2024 - arxiv.org
As artificial intelligence (AI) continues to advance, it demonstrates capabilities comparable
to human intelligence, with significant potential to transform education and workforce …

Balancing cost and effectiveness of synthetic data generation strategies for llms

YC Chan, G Pu, A Shanker, P Suresh, P Jenks… - arxiv preprint arxiv …, 2024 - arxiv.org
As large language models (LLMs) are applied to more use cases, creating high quality, task-
specific datasets for fine-tuning becomes a bottleneck for model improvement. Using high …

A Noise-Oriented and Redundancy-Aware Instance Selection Framework

W Cunha, A Moreo Fernández, A Esuli… - ACM Transactions on …, 2025 - dl.acm.org
Fine-tuning transformer-based deep-learning models are currently at the forefront of natural
language processing (NLP) and information retrieval (IR) tasks. However, fine-tuning these …

Large language model-guided document selection

X Kong, T Gunter, R Pang - arxiv preprint arxiv:2406.04638, 2024 - arxiv.org
Large Language Model (LLM) pre-training exhausts an ever growing compute budget, yet
recent research has demonstrated that careful document selection enables comparable …

Data-centric ai in the age of large language models

X Xu, Z Wu, R Qiao, A Verma, Y Shu, J Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
This position paper proposes a data-centric viewpoint of AI research, focusing on large
language models (LLMs). We start by making the key observation that data is instrumental in …