Priors in bayesian deep learning: A review
V Fortuin - International Statistical Review, 2022 - Wiley Online Library
While the choice of prior is one of the most critical parts of the Bayesian inference workflow,
recent Bayesian deep learning models have often fallen back on vague priors, such as …
recent Bayesian deep learning models have often fallen back on vague priors, such as …
Reviewing autoencoders for missing data imputation: Technical trends, applications and outcomes
Missing data is a problem often found in real-world datasets and it can degrade the
performance of most machine learning models. Several deep learning techniques have …
performance of most machine learning models. Several deep learning techniques have …
Tabnet: Attentive interpretable tabular learning
We propose a novel high-performance and interpretable canonical deep tabular data
learning architecture, TabNet. TabNet uses sequential attention to choose which features to …
learning architecture, TabNet. TabNet uses sequential attention to choose which features to …
Gp-vae: Deep probabilistic time series imputation
Multivariate time series with missing values are common in areas such as healthcare and
finance, and have grown in number and complexity over the years. This raises the question …
finance, and have grown in number and complexity over the years. This raises the question …
MIWAE: Deep generative modelling and imputation of incomplete data sets
We consider the problem of handling missing data with deep latent variable models
(DLVMs). First, we present a simple technique to train DLVMs when the training set contains …
(DLVMs). First, we present a simple technique to train DLVMs when the training set contains …
Deep end-to-end causal inference
Causal inference is essential for data-driven decision making across domains such as
business engagement, medical treatment and policy making. However, research on causal …
business engagement, medical treatment and policy making. However, research on causal …
Hyperimpute: Generalized iterative imputation with automatic model selection
Consider the problem of imputing missing values in a dataset. One the one hand,
conventional approaches using iterative imputation benefit from the simplicity and …
conventional approaches using iterative imputation benefit from the simplicity and …
Interactive concept bottleneck models
Abstract Concept bottleneck models (CBMs) are interpretable neural networks that first
predict labels for human-interpretable concepts relevant to the prediction task, and then …
predict labels for human-interpretable concepts relevant to the prediction task, and then …
Learning to maximize mutual information for dynamic feature selection
Feature selection helps reduce data acquisition costs in ML, but the standard approach is to
train models with static feature subsets. Here, we consider the dynamic feature selection …
train models with static feature subsets. Here, we consider the dynamic feature selection …
Variational Bayesian optimal experimental design
Bayesian optimal experimental design (BOED) is a principled framework for making efficient
use of limited experimental resources. Unfortunately, its applicability is hampered by the …
use of limited experimental resources. Unfortunately, its applicability is hampered by the …