[HTML][HTML] Trustworthy clinical AI solutions: a unified review of uncertainty quantification in deep learning models for medical image analysis

B Lambert, F Forbes, S Doyle, H Dehaene… - Artificial Intelligence in …, 2024 - Elsevier
The full acceptance of Deep Learning (DL) models in the clinical field is rather low with
respect to the quantity of high-performing solutions reported in the literature. End users are …

Scaling vision transformers to 22 billion parameters

M Dehghani, J Djolonga, B Mustafa… - International …, 2023 - proceedings.mlr.press
The scaling of Transformers has driven breakthrough capabilities for language models. At
present, the largest large language models (LLMs) contain upwards of 100B parameters …

Critic: Large language models can self-correct with tool-interactive critiquing

Z Gou, Z Shao, Y Gong, Y Shen, Y Yang… - arxiv preprint arxiv …, 2023 - arxiv.org
Recent developments in large language models (LLMs) have been impressive. However,
these models sometimes show inconsistencies and problematic behavior, such as …

Can llms express their uncertainty? an empirical evaluation of confidence elicitation in llms

M **ong, Z Hu, X Lu, Y Li, J Fu, J He, B Hooi - arxiv preprint arxiv …, 2023 - arxiv.org
Empowering large language models to accurately express confidence in their answers is
essential for trustworthy decision-making. Previous confidence elicitation methods, which …

Language models (mostly) know what they know

S Kadavath, T Conerly, A Askell, T Henighan… - arxiv preprint arxiv …, 2022 - arxiv.org
We study whether language models can evaluate the validity of their own claims and predict
which questions they will be able to answer correctly. We first show that larger models are …

Teaching models to express their uncertainty in words

S Lin, J Hilton, O Evans - arxiv preprint arxiv:2205.14334, 2022 - arxiv.org
We show that a GPT-3 model can learn to express uncertainty about its own answers in
natural language--without use of model logits. When given a question, the model generates …

A primer on Bayesian neural networks: review and debates

J Arbel, K Pitas, M Vladimirova, V Fortuin - arxiv preprint arxiv:2309.16314, 2023 - arxiv.org
Neural networks have achieved remarkable performance across various problem domains,
but their widespread applicability is hindered by inherent limitations such as overconfidence …

Benchmarking llms via uncertainty quantification

F Ye, M Yang, J Pang, L Wang… - Advances in …, 2025 - proceedings.neurips.cc
The proliferation of open-source Large Language Models (LLMs) from various institutions
has highlighted the urgent need for comprehensive evaluation methods. However, current …

Navigating the grey area: How expressions of uncertainty and overconfidence affect language models

K Zhou, D Jurafsky, T Hashimoto - arxiv preprint arxiv:2302.13439, 2023 - arxiv.org
The increased deployment of LMs for real-world tasks involving knowledge and facts makes
it important to understand model epistemology: what LMs think they know, and how their …

Localizing objects with self-supervised transformers and no labels

O Siméoni, G Puy, HV Vo, S Roburin, S Gidaris… - arxiv preprint arxiv …, 2021 - arxiv.org
Localizing objects in image collections without supervision can help to avoid expensive
annotation campaigns. We propose a simple approach to this problem, that leverages the …