[HTML][HTML] A survey on large language model (llm) security and privacy: The good, the bad, and the ugly

Y Yao, J Duan, K Xu, Y Cai, Z Sun, Y Zhang - High-Confidence Computing, 2024 - Elsevier
Abstract Large Language Models (LLMs), such as ChatGPT and Bard, have revolutionized
natural language understanding and generation. They possess deep language …

Large language models and games: A survey and roadmap

R Gallotta, G Todd, M Zammit, S Earle… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
Recent years have seen an explosive increase in research on large language models
(LLMs), and accompanying public engagement on the topic. While starting as a niche area …

Siren's song in the AI ocean: a survey on hallucination in large language models

Y Zhang, Y Li, L Cui, D Cai, L Liu, T Fu… - arxiv preprint arxiv …, 2023 - arxiv.org
While large language models (LLMs) have demonstrated remarkable capabilities across a
range of downstream tasks, a significant concern revolves around their propensity to exhibit …

Explainability for large language models: A survey

H Zhao, H Chen, F Yang, N Liu, H Deng, H Cai… - ACM Transactions on …, 2024 - dl.acm.org
Large language models (LLMs) have demonstrated impressive capabilities in natural
language processing. However, their internal mechanisms are still unclear and this lack of …

[PDF][PDF] Trustllm: Trustworthiness in large language models

L Sun, Y Huang, H Wang, S Wu, Q Zhang… - arxiv preprint arxiv …, 2024 - mosis.eecs.utk.edu
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …

[HTML][HTML] Position: TrustLLM: Trustworthiness in large language models

Y Huang, L Sun, H Wang, S Wu… - International …, 2024 - proceedings.mlr.press
Large language models (LLMs) have gained considerable attention for their excellent
natural language processing capabilities. Nonetheless, these LLMs present many …

INSIDE: LLMs' internal states retain the power of hallucination detection

C Chen, K Liu, Z Chen, Y Gu, Y Wu, M Tao… - arxiv preprint arxiv …, 2024 - arxiv.org
Knowledge hallucination have raised widespread concerns for the security and reliability of
deployed LLMs. Previous efforts in detecting hallucinations have been employed at logit …

A survey of confidence estimation and calibration in large language models

J Geng, F Cai, Y Wang, H Koeppl, P Nakov… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable capabilities across a wide
range of tasks in various domains. Despite their impressive performance, they can be …

Estimating the hallucination rate of generative ai

A Jesson, N Beltran Velez, Q Chu… - Advances in …, 2025 - proceedings.neurips.cc
This paper presents a method for estimating the hallucination rate for in-context learning
(ICL) with generative AI. In ICL, a conditional generative model (CGM) is prompted with a …

Luq: Long-text uncertainty quantification for llms

C Zhang, F Liu, M Basaldella, N Collier - arxiv preprint arxiv:2403.20279, 2024 - arxiv.org
Large Language Models (LLMs) have demonstrated remarkable capability in a variety of
NLP tasks. However, LLMs are also prone to generate nonfactual content. Uncertainty …