The Bayesian learning rule

ME Khan, H Rue - Journal of Machine Learning Research, 2023 - jmlr.org
We show that many machine-learning algorithms are specific instances of a single algorithm
called the Bayesian learning rule. The rule, derived from Bayesian principles, yields a wide …

PRIOR: Personalized Prior for Reactivating the Information Overlooked in Federated Learning.

M Shi, Y Zhou, K Wang, H Zhang… - Advances in Neural …, 2023 - proceedings.neurips.cc
Classical federated learning (FL) enables training machine learning models without sharing
data for privacy preservation, but heterogeneous data characteristic degrades the …

Sharp global convergence guarantees for iterative nonconvex optimization with random data

KA Chandrasekher, A Pananjady… - The Annals of …, 2023 - projecteuclid.org
Sharp global convergence guarantees for iterative nonconvex optimization with random data
Page 1 The Annals of Statistics 2023, Vol. 51, No. 1, 179–210 https://doi.org/10.1214/22-AOS2246 …

Mirror descent with relative smoothness in measure spaces, with application to sinkhorn and em

PC Aubin-Frankowski, A Korba… - Advances in Neural …, 2022 - proceedings.neurips.cc
Many problems in machine learning can be formulated as optimizing a convex functional
over a vector space of measures. This paper studies the convergence of the mirror descent …

Stochastic approximation beyond gradient for signal processing and machine learning

A Dieuleveut, G Fort, E Moulines… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Stochastic Approximation (SA) is a classical algorithm that has had since the early days a
huge impact on signal processing, and nowadays on machine learning, due to the necessity …

Theoretical guarantees for variational inference with fixed-variance mixture of gaussians

T Huix, A Korba, A Durmus, E Moulines - arxiv preprint arxiv:2406.04012, 2024 - arxiv.org
Variational inference (VI) is a popular approach in Bayesian inference, that looks for the best
approximation of the posterior distribution within a parametric family, minimizing a loss that …

Federated-EM with heterogeneity mitigation and variance reduction

A Dieuleveut, G Fort, E Moulines… - Advances in Neural …, 2021 - proceedings.neurips.cc
Abstract The Expectation Maximization (EM) algorithm is the default algorithm for inference
in latent variable models. As in any other field of machine learning, applications of latent …

A Bregman proximal perspective on classical and quantum Blahut-Arimoto algorithms

K He, J Saunderson, H Fawzi - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
The Blahut-Arimoto algorithm is a well-known method to compute classical channel
capacities and rate-distortion functions. Recent works have extended this algorithm to …

Sharp global convergence guarantees for iterative nonconvex optimization: A gaussian process perspective

KA Chandrasekher, A Pananjady… - arxiv preprint arxiv …, 2021 - arxiv.org
We consider a general class of regression models with normally distributed covariates, and
the associated nonconvex problem of fitting these models from data. We develop a general …

EM++: A parameter learning framework for stochastic switching systems

R Wang, A Bodard, M Schuurmans… - arxiv preprint arxiv …, 2024 - arxiv.org
This paper proposes a general switching dynamical system model, and a custom
majorization-minimization-based algorithm EM++ for identifying its parameters. For certain …