Llm-based edge intelligence: A comprehensive survey on architectures, applications, security and trustworthiness

O Friha, MA Ferrag, B Kantarci… - IEEE Open Journal …, 2024 - ieeexplore.ieee.org
The integration of Large Language Models (LLMs) and Edge Intelligence (EI) introduces a
groundbreaking paradigm for intelligent edge devices. With their capacity for human-like …

A tutorial on open-source large language models for behavioral science

Z Hussain, M Binz, R Mata, DU Wulff - Behavior Research Methods, 2024 - Springer
Large language models (LLMs) have the potential to revolutionize behavioral science by
accelerating and improving the research cycle, from conceptualization to data analysis …

Ladder: Enabling Efficient {Low-Precision} Deep Learning Computing through Hardware-aware Tensor Transformation

L Wang, L Ma, S Cao, Q Zhang, J Xue, Y Shi… - … USENIX Symposium on …, 2024 - usenix.org
The increasing demand for improving deep learning model performance has led to a
paradigm shift in supporting low-precision computation to harness the robustness of deep …

Efficientqat: Efficient quantization-aware training for large language models

M Chen, W Shao, P Xu, J Wang, P Gao… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are crucial in modern natural language processing and
artificial intelligence. However, they face challenges in managing their significant memory …

Llm as a system service on mobile devices

W Yin, M Xu, Y Li, X Liu - arxiv preprint arxiv:2403.11805, 2024 - arxiv.org
Being more powerful and intrusive into user-device interactions, LLMs are eager for on-
device execution to better preserve user privacy. In this work, we propose a new paradigm of …

Recurrent neural networks: vanishing and exploding gradients are not the end of the story

N Zucchet, A Orvieto - Advances in Neural Information …, 2025 - proceedings.neurips.cc
Recurrent neural networks (RNNs) notoriously struggle to learn long-term memories,
primarily due to vanishing and exploding gradients. The recent success of state-space …

On-device language models: A comprehensive review

J Xu, Z Li, W Chen, Q Wang, X Gao, Q Cai… - arxiv preprint arxiv …, 2024 - arxiv.org
The advent of large language models (LLMs) revolutionized natural language processing
applications, and running LLMs on edge devices has become increasingly attractive for …

A robust governance for the AI act: AI office, AI Board, scientific panel, and national authorities

C Novelli, P Hacker, J Morley, J Trondal… - European Journal of …, 2024 - cambridge.org
Regulation is nothing without enforcement. This particularly holds for the dynamic field of
emerging technologies. Hence, this article has two ambitions. First, it explains how the EU's …

Scaling laws for precision

T Kumar, Z Ankner, BF Spector, B Bordelon… - arxiv preprint arxiv …, 2024 - arxiv.org
Low precision training and inference affect both the quality and cost of language models, but
current scaling laws do not account for this. In this work, we devise" precision-aware" scaling …

[PDF][PDF] Scalable matmul-free language modeling

RJ Zhu, Y Zhang, E Sifferman, T Sheaves… - arxiv preprint arxiv …, 2024 - openreview.net
Matrix multiplication (MatMul) typically dominates the overall computational cost of large
language models (LLMs). This cost only grows as LLMs scale to larger embedding …