(23th-December-2020)
• The walk-back training procedure was proposed by ( ) as a way Bengio et al. 2013c to accelerate the convergence of generative training of denoisingautoencoders. Instead of performing a one-step encode-decode reconstruction, this procedure consists in alternative multiple stochastic encode-decode steps (as in the generative Markov chain) initialized at a training example (just like with the contrastive divergence algorithm, described in section ) and penalizing the last probabilistic 18.2 reconstructions (or all of the reconstructions along the way). Training with k steps is equivalent (in the sense of achieving the same stationary distribution) as training with one step, but practically has the advantage that spurious modes further from the data can be removed more efficiently.
• Generative stochastic networks or GSNs ( , ) are generalizaBengio et al. 2014 tions of denoisingautoencoders that include latent variables h in the generative Markov chain, in addition to the visible variables (usually denoted ). x A GSN isparametrized by two conditional probability distributions which specify one step of the Markov chain:
• 1. p(x( ) k | h( ) k ) tells how to generate the next visible variable given the current latent state. Such a “reconstruction distribution” is also found in denoisingautoencoders, RBMs, DBNs and DBMs.
• 2. p(h( ) k | h( 1) k− ,x( 1) k− ) tells how to update the latent state variable, given the previous latent state and visible variable.
• Discriminant GSNs
Comments