Mobile edge intelligence for large language models: A contemporary survey

G Qu, Q Chen, W Wei, Z Lin, X Chen… - … Surveys & Tutorials, 2025 - ieeexplore.ieee.org
On-device large language models (LLMs), referring to running LLMs on edge devices, have
raised considerable interest since they are more cost-effective, latency-efficient, and privacy …

A survey on model compression for large language models

X Zhu, J Li, Y Liu, C Ma, W Wang - Transactions of the Association for …, 2024 - direct.mit.edu
Abstract Large Language Models (LLMs) have transformed natural language processing
tasks successfully. Yet, their large size and high computational needs pose challenges for …

Aligning large language models with human: A survey

Y Wang, W Zhong, L Li, F Mi, X Zeng, W Huang… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) trained on extensive textual corpora have emerged as
leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite …

A survey of safety and trustworthiness of large language models through the lens of verification and validation

X Huang, W Ruan, W Huang, G **, Y Dong… - Artificial Intelligence …, 2024 - Springer
Large language models (LLMs) have exploded a new heatwave of AI for their ability to
engage end-users in human-level conversations with detailed and articulate answers across …

Cmmlu: Measuring massive multitask language understanding in chinese

H Li, Y Zhang, F Koto, Y Yang, H Zhao, Y Gong… - arxiv preprint arxiv …, 2023 - arxiv.org
As the capabilities of large language models (LLMs) continue to advance, evaluating their
performance becomes increasingly crucial and challenging. This paper aims to bridge this …

Knowledge distillation of large language models

Y Gu, L Dong, F Wei, M Huang - arxiv preprint arxiv:2306.08543, 2023 - arxiv.org
Knowledge Distillation (KD) is a promising technique for reducing the high computational
demand of large language models (LLMs). However, previous KD methods are primarily …

Datasets for large language models: A comprehensive survey

Y Liu, J Cao, C Liu, K Ding, L ** - arxiv preprint arxiv:2402.18041, 2024 - arxiv.org
This paper embarks on an exploration into the Large Language Model (LLM) datasets,
which play a crucial role in the remarkable advancements of LLMs. The datasets serve as …

MiniLLM: Knowledge distillation of large language models

Y Gu, L Dong, F Wei, M Huang - The Twelfth International …, 2024 - openreview.net
Knowledge Distillation (KD) is a promising technique for reducing the high computational
demand of large language models (LLMs). However, previous KD methods are primarily …

Platypus: Quick, cheap, and powerful refinement of llms

AN Lee, CJ Hunter, N Ruiz - arxiv preprint arxiv:2308.07317, 2023 - arxiv.org
We present $\textbf {Platypus} $, a family of fine-tuned and merged Large Language Models
(LLMs) that achieves the strongest performance and currently stands at first place in …

Bactrian-x: Multilingual replicable instruction-following models with low-rank adaptation

H Li, F Koto, M Wu, AF Aji, T Baldwin - arxiv preprint arxiv:2305.15011, 2023 - arxiv.org
Instruction tuning has shown great promise in improving the performance of large language
models. However, research on multilingual instruction tuning has been limited due to the …