Literature DB >> 19018704

Justifying and generalizing contrastive divergence.

Yoshua Bengio1, Olivier Delalleau.   

Abstract

We study an expansion of the log likelihood in undirected graphical models such as the restricted Boltzmann machine (RBM), where each term in the expansion is associated with a sample in a Gibbs chain alternating between two random variables (the visible vector and the hidden vector in RBMs). We are particularly interested in estimators of the gradient of the log likelihood obtained through this expansion. We show that its residual term converges to zero, justifying the use of a truncation--running only a short Gibbs chain, which is the main idea behind the contrastive divergence (CD) estimator of the log-likelihood gradient. By truncating even more, we obtain a stochastic reconstruction error, related through a mean-field approximation to the reconstruction error often used to train autoassociators and stacked autoassociators. The derivation is not specific to the particular parametric forms used in RBMs and requires only convergence of the Gibbs chain. We present theoretical and empirical evidence linking the number of Gibbs steps k and the magnitude of the RBM parameters to the bias in the CD estimator. These experiments also suggest that the sign of the CD estimator is correct most of the time, even when the bias is large, so that CD-k is a good descent direction even for small k.

Entities:  

Mesh:

Year:  2009        PMID: 19018704     DOI: 10.1162/neco.2008.11-07-647

Source DB:  PubMed          Journal:  Neural Comput        ISSN: 0899-7667            Impact factor:   2.026


  7 in total

1.  Scalable estimation strategies based on stochastic approximations: Classical results and new insights.

Authors:  Edoardo M Airoldi; Panos Toulis
Journal:  Stat Comput       Date:  2015-07-01       Impact factor: 2.559

2.  Deep Learning for Classification of Normal Swallows in Adults.

Authors:  Joshua M Dudik; James L Coyle; Amro El-Jaroudi; Zhi-Hong Mao; Mingui Sun; Ervin Sejdić
Journal:  Neurocomputing       Date:  2018-01-31       Impact factor: 5.719

3.  Variable but not random: temporal pattern coding in a songbird brain area necessary for song modification.

Authors:  S E Palmer; B D Wright; A J Doupe; M H Kao
Journal:  J Neurophysiol       Date:  2020-12-09       Impact factor: 2.714

4.  Temperature based Restricted Boltzmann Machines.

Authors:  Guoqi Li; Lei Deng; Yi Xu; Changyun Wen; Wei Wang; Jing Pei; Luping Shi
Journal:  Sci Rep       Date:  2016-01-13       Impact factor: 4.379

5.  Three learning stages and accuracy-efficiency tradeoff of restricted Boltzmann machines.

Authors:  Lennart Dabelow; Masahito Ueda
Journal:  Nat Commun       Date:  2022-09-17       Impact factor: 17.694

6.  Mirrored STDP Implements Autoencoder Learning in a Network of Spiking Neurons.

Authors:  Kendra S Burbank
Journal:  PLoS Comput Biol       Date:  2015-12-03       Impact factor: 4.475

7.  msiDBN: a method of identifying critical proteins in dynamic PPI networks.

Authors:  Yuan Zhang; Nan Du; Kang Li; Jinchao Feng; Kebin Jia; Aidong Zhang
Journal:  Biomed Res Int       Date:  2014-04-02       Impact factor: 3.411

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.