Literature DB >> 34556969

A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions.

Wei Deng1, Guang Lin2, Faming Liang3.   

Abstract

We propose an adaptively weighted stochastic gradient Langevin dynamics algorithm (SGLD), so-called contour stochastic gradient Langevin dynamics (CSGLD), for Bayesian learning in big data statistics. The proposed algorithm is essentially a scalable dynamic importance sampler, which automatically flattens the target distribution such that the simulation for a multi-modal distribution can be greatly facilitated. Theoretically, we prove a stability condition and establish the asymptotic convergence of the self-adapting parameter to a unique fixed-point, regardless of the non-convexity of the original energy function; we also present an error analysis for the weighted averaging estimators. Empirically, the CSGLD algorithm is tested on multiple benchmark datasets including CIFAR10 and CIFAR100. The numerical results indicate its superiority over the existing state-of-the-art algorithms in training deep neural networks.

Entities:  

Year:  2020        PMID: 34556969      PMCID: PMC8457681     

Source DB:  PubMed          Journal:  Adv Neural Inf Process Syst        ISSN: 1049-5258


  4 in total

1.  Efficient, multiple-range random walk algorithm to calculate the density of states.

Authors:  F Wang; D P Landau
Journal:  Phys Rev Lett       Date:  2001-03-05       Impact factor: 9.161

2.  Replica Monte Carlo simulation of spin glasses.

Authors: 
Journal:  Phys Rev Lett       Date:  1986-11-24       Impact factor: 9.161

3.  Optimization by simulated annealing.

Authors:  S Kirkpatrick; C D Gelatt; M P Vecchi
Journal:  Science       Date:  1983-05-13       Impact factor: 47.728

4.  Non-convex Learning via Replica Exchange Stochastic Gradient MCMC.

Authors:  Wei Deng; Qi Feng; Liyao Gao; Faming Liang; Guang Lin
Journal:  Proc Mach Learn Res       Date:  2020-07
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.