History, development, and principles of large language models: an introductory survey

Z Wang, Z Chu, TV Doan, S Ni, M Yang, W Zhang - AI and Ethics, 2024 - Springer
Abstract Language models serve as a cornerstone in natural language processing, utilizing
mathematical methods to generalize language laws and knowledge for prediction and …

Llms4ol 2024 overview: The 1st large language models for ontology learning challenge

HB Giglou, J D'Souza, S Auer - arxiv preprint arxiv:2409.10146, 2024 - arxiv.org
This paper outlines the LLMs4OL 2024, the first edition of the Large Language Models for
Ontology Learning Challenge. LLMs4OL is a community development initiative collocated …

Assessment of fine-tuned large language models for real-world chemistry and material science applications

J Van Herck, MV Gil, KM Jablonka, A Abrudan… - Chemical …, 2025 - pubs.rsc.org
The current generation of large language models (LLMs) has limited chemical knowledge.
Recently, it has been shown that these LLMs can learn and predict chemical properties …

Aligning llms to be robust against prompt injection

S Chen, A Zharmagambetov, S Mahloujifar… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are becoming increasingly prevalent in modern software
systems, interfacing between the user and the internet to assist with tasks that require …

xgen-mm-vid (blip-3-video): You only need 32 tokens to represent a video even in vlms

MS Ryoo, H Zhou, S Kendre, C Qin, L Xue… - arxiv preprint arxiv …, 2024 - arxiv.org
We present xGen-MM-Vid (BLIP-3-Video): a multimodal language model for videos,
particularly designed to efficiently capture temporal information over multiple frames. BLIP-3 …

Re-task: Revisiting llm tasks from capability, skill, and knowledge perspectives

Z Wang, S Zhao, Y Wang, H Huang, S **e… - arxiv preprint arxiv …, 2024 - arxiv.org
The Chain-of-Thought (CoT) paradigm has become a pivotal method for solving complex
problems. However, its application to intricate, domain-specific tasks remains challenging …

A scalable communication protocol for networks of large language models

S Marro, E La Malfa, J Wright, G Li, N Shadbolt… - arxiv preprint arxiv …, 2024 - arxiv.org
Communication is a prerequisite for collaboration. When scaling networks of AI-powered
agents, communication must be versatile, efficient, and portable. These requisites, which we …

Critic-cot: Boosting the reasoning abilities of large language model via chain-of-thoughts critic

X Zheng, J Lou, B Cao, X Wen, Y Ji, H Lin, Y Lu… - arxiv preprint arxiv …, 2024 - arxiv.org
Self-critic has become a crucial mechanism for enhancing the reasoning performance of
LLMs. However, current approaches mainly involve basic prompts for intuitive instance-level …

Memory layers at scale

VP Berges, B Oğuz, D Haziza, W Yih… - arxiv preprint arxiv …, 2024 - arxiv.org
Memory layers use a trainable key-value lookup mechanism to add extra parameters to a
model without increasing FLOPs. Conceptually, sparsely activated memory layers …

Fine-tuning large language models for domain adaptation: Exploration of training strategies, scaling, model merging and synergistic capabilities

W Lu, RK Luu, MJ Buehler - arxiv preprint arxiv:2409.03444, 2024 - arxiv.org
The advancement of Large Language Models (LLMs) for domain applications in fields such
as materials science and engineering depends on the development of fine-tuning strategies …