A survey of contextual optimization methods for decision-making under uncertainty

U Sadana, A Chenreddy, E Delage, A Forel… - European Journal of …, 2024 - Elsevier
Recently there has been a surge of interest in operations research (OR) and the machine
learning (ML) community in combining prediction algorithms and optimization techniques to …

The elements of differentiable programming

M Blondel, V Roulet - arxiv preprint arxiv:2403.14606, 2024 - arxiv.org
Artificial intelligence has recently experienced remarkable advances, fueled by large
models, vast datasets, accelerated hardware, and, last but not least, the transformative …

On neural differential equations

P Kidger - arxiv preprint arxiv:2202.02435, 2022 - arxiv.org
The conjoining of dynamical systems and deep learning has become a topic of great
interest. In particular, neural differential equations (NDEs) demonstrate that neural networks …

Craft: Concept recursive activation factorization for explainability

T Fel, A Picard, L Bethune, T Boissin… - Proceedings of the …, 2023 - openaccess.thecvf.com
Attribution methods are a popular class of explainability methods that use heatmaps to
depict the most important areas of an image that drive a model decision. Nevertheless …

Theseus: A library for differentiable nonlinear optimization

L Pineda, T Fan, M Monge… - Advances in …, 2022 - proceedings.neurips.cc
We present Theseus, an efficient application-agnostic open source library for differentiable
nonlinear least squares (DNLS) optimization built on PyTorch, providing a common …

A graph-based methodology for constructing computational models that automates adjoint-based sensitivity analysis

V Gandarillas, AJ Joshy, MZ Sperry, AK Ivanov… - Structural and …, 2024 - Springer
The adjoint method provides an efficient way to compute sensitivities for system models with
a large number of inputs. However, implementing the adjoint method requires significant …

Synergies between disentanglement and sparsity: Generalization and identifiability in multi-task learning

S Lachapelle, T Deleu, D Mahajan… - International …, 2023 - proceedings.mlr.press
Although disentangled representations are often said to be beneficial for downstream tasks,
current empirical and theoretical understanding is limited. In this work, we provide evidence …

Linear adversarial concept erasure

S Ravfogel, M Twiton, Y Goldberg… - … on Machine Learning, 2022 - proceedings.mlr.press
Modern neural models trained on textual data rely on pre-trained representations that
emerge without direct supervision. As these representations are increasingly being used in …

A-nesi: A scalable approximate method for probabilistic neurosymbolic inference

E van Krieken, T Thanapalasingam… - Advances in …, 2023 - proceedings.neurips.cc
We study the problem of combining neural networks with symbolic reasoning. Recently
introduced frameworks for Probabilistic Neurosymbolic Learning (PNL), such as …

Sinkformers: Transformers with doubly stochastic attention

ME Sander, P Ablin, M Blondel… - … Conference on Artificial …, 2022 - proceedings.mlr.press
Attention based models such as Transformers involve pairwise interactions between data
points, modeled with a learnable attention matrix. Importantly, this attention matrix is …