Preconditioned stochastic gradient Langevin dynamics for deep neural networks

C Li, C Chen, D Carlson, L Carin - … of the AAAI conference on artificial …, 2016 - ojs.aaai.org
Effective training of deep neural networks suffers from two main issues. The first is that the
parameter space of these models exhibit pathological curvature. Recent methods address …

Software reliability prediction: A survey

S Oveisi, A Moeini, S Mirzaei… - Quality and Reliability …, 2023 - Wiley Online Library
Softwares play an important role in controlling complex systems. Monitoring the proper
functioning of the components of such systems is the principal role of softwares. Often, a …

Bridging the gap between stochastic gradient MCMC and stochastic optimization

C Chen, D Carlson, Z Gan, C Li… - Artificial Intelligence …, 2016 - proceedings.mlr.press
Abstract Stochastic gradient Markov chain Monte Carlo (SG-MCMC) methods are Bayesian
analogs to popular stochastic optimization methods; however, this connection is not well …

Learning weight uncertainty with stochastic gradient mcmc for shape classification

C Li, A Stevens, C Chen, Y Pu… - Proceedings of the …, 2016 - openaccess.thecvf.com
Learning the representation of shape cues in 2D & 3D objects for recognition is a
fundamental task in computer vision. Deep neural networks (DNNs) have shown promising …

Stochastic gradient MCMC with repulsive forces

V Gallego, DR Insua - arxiv preprint arxiv:1812.00071, 2018 - arxiv.org
We propose a unifying view of two different Bayesian inference algorithms, Stochastic
Gradient Markov Chain Monte Carlo (SG-MCMC) and Stein Variational Gradient Descent …

Robust SGLD algorithm for solving non-convex distributionally robust optimisation problems

A Neufeld, MNC En, Y Zhang - arxiv preprint arxiv:2403.09532, 2024 - arxiv.org
In this paper we develop a Stochastic Gradient Langevin Dynamics (SGLD) algorithm
tailored for solving a certain class of non-convex distributionally robust optimisation …

Scalable bayesian learning of recurrent neural networks for language modeling

Z Gan, C Li, C Chen, Y Pu, Q Su, L Carin - arxiv preprint arxiv:1611.08034, 2016 - arxiv.org
Recurrent neural networks (RNNs) have shown promising performance for language
modeling. However, traditional training of RNNs using back-propagation through time often …

Are you using test log-likelihood correctly?

SK Deshpande, S Ghosh, TD Nguyen… - arxiv preprint arxiv …, 2022 - arxiv.org
Test log-likelihood is commonly used to compare different models of the same data or
different approximate inference algorithms for fitting the same probabilistic model. We …

A unifying and canonical description of measure-preserving diffusions

A Barp, S Takao, M Betancourt, A Arnaudon… - arxiv preprint arxiv …, 2021 - arxiv.org
A complete recipe of measure-preserving diffusions in Euclidean space was recently
derived unifying several MCMC algorithms into a single framework. In this paper, we …

Stochastic bouncy particle sampler

A Pakman, D Gilboa, D Carlson… - … on Machine Learning, 2017 - proceedings.mlr.press
We introduce a stochastic version of the non-reversible, rejection-free Bouncy Particle
Sampler (BPS), a Markov process whose sample trajectories are piecewise linear, to …