[HTML][HTML] A review of uncertainty quantification in deep learning: Techniques, applications and challenges
Uncertainty quantification (UQ) methods play a pivotal role in reducing the impact of
uncertainties during both optimization and decision making processes. They have been …
uncertainties during both optimization and decision making processes. They have been …
Understanding metric-related pitfalls in image analysis validation
Validation metrics are key for tracking scientific progress and bridging the current chasm
between artificial intelligence research and its translation into practice. However, increasing …
between artificial intelligence research and its translation into practice. However, increasing …
Holistic evaluation of language models
Language models (LMs) are becoming the foundation for almost all major language
technologies, but their capabilities, limitations, and risks are not well understood. We present …
technologies, but their capabilities, limitations, and risks are not well understood. We present …
Large legal fictions: Profiling legal hallucinations in large language models
Do large language models (LLMs) know the law? LLMs are increasingly being used to
augment legal practice, education, and research, yet their revolutionary potential is …
augment legal practice, education, and research, yet their revolutionary potential is …
Unsolved problems in ml safety
Machine learning (ML) systems are rapidly increasing in size, are acquiring new
capabilities, and are increasingly deployed in high-stakes settings. As with other powerful …
capabilities, and are increasingly deployed in high-stakes settings. As with other powerful …
Revisiting the calibration of modern neural networks
Accurate estimation of predictive uncertainty (model calibration) is essential for the safe
application of neural networks. Many instances of miscalibration in modern neural networks …
application of neural networks. Many instances of miscalibration in modern neural networks …
Understanding Dataset Difficulty with -Usable Information
Estimating the difficulty of a dataset typically involves comparing state-of-the-art models to
humans; the bigger the performance gap, the harder the dataset is said to be. However, this …
humans; the bigger the performance gap, the harder the dataset is said to be. However, this …
Simple and principled uncertainty estimation with deterministic deep learning via distance awareness
Bayesian neural networks (BNN) and deep ensembles are principled approaches to
estimate the predictive uncertainty of a deep learning model. However their practicality in …
estimate the predictive uncertainty of a deep learning model. However their practicality in …
Calibrating deep neural networks using focal loss
Miscalibration--a mismatch between a model's confidence and its correctness--of Deep
Neural Networks (DNNs) makes their predictions hard to rely on. Ideally, we want networks …
Neural Networks (DNNs) makes their predictions hard to rely on. Ideally, we want networks …
Natural adversarial examples
We introduce two challenging datasets that reliably cause machine learning model
performance to substantially degrade. The datasets are collected with a simple adversarial …
performance to substantially degrade. The datasets are collected with a simple adversarial …