Header menu link for other important links
X
Similarity-based contrastive divergence methods for energy-based deep learning models
Published in Asian Conference on Machine Learning
2015
Pages: 391 - 406
Abstract
Energy-based deep learning models like Restricted Boltzmann Machines are increasinglyused for real-world applications. However, all these models inherently depend on the Contrastive Divergence (CD) method for training and maximization of log likelihood of generating the given data distribution. CD, which internally uses Gibbs sampling, often does not perform well due to issues such as biased samples, poor mixing of Markov chains and highmass probability modes. Variants of CD such as PCD, Fast PCD and Tempered MCMC have been proposed to address this issue. In this work, we propose a new approach to CD-based methods, called Diss-CD, which uses dissimilar data to allow the Markov chain to explore new modes in the probability space. This method can be used with all variants of CD (or PCD), and across all energy-based deep learning models. Our experiments on using this approach on standard datasets including MNIST, Caltech-101 Silhouette and Synthetic Transformations, demonstrate the promise of this approach, showing fast convergence of error in learning and also a better approximation of log likelihood of the data. © 2015 A.R. Sankar & V.N. Balasubramanian.
About the journal
JournalACML 2015 - 7th Asian Conference on Machine Learning
PublisherAsian Conference on Machine Learning