Training Products of Experts by Minimizing Contrastive Divergence
From MaRDI portal
Publication:3149516
DOI10.1162/089976602760128018zbMath1010.68111OpenAlexW2116064496WikidataQ34144628 ScholiaQ34144628MaRDI QIDQ3149516
Publication date: 25 September 2002
Published in: Neural Computation (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1162/089976602760128018
Learning and adaptive systems in artificial intelligence (68T05) Theory of languages and software systems (knowledge-based systems, expert systems, etc.) for artificial intelligence (68T35)
Related Items
A transport equation approach for deep neural networks with quenched random weights, Equilibrium and non-equilibrium regimes in the learning of restricted Boltzmann machines*, Mean-field inference methods for neural networks, Legendre equivalences of spherical Boltzmann machines, Features Reweighting and Selection in ligand-based Virtual Screening for Molecular Similarity Searching Based on Deep Belief Networks, Justifying and Generalizing Contrastive Divergence, Enhanced Gradient for Training Restricted Boltzmann Machines, Generative modeling via tree tensor network states, A quantitative analysis of a generalized Hopfield model that stores and retrieves mismatched memory patterns, Realizing number recognition with simulated quantum semi-restricted Boltzmann machine, Thermodynamics of bidirectional associative memories, Bandgap optimization in combinatorial graphs with tailored ground states: application in quantum annealing, Identifiability of Hierarchical Latent Attribute Models, The emergence of a concept in shallow neural networks, On the effective initialisation for restricted Boltzmann machines via duality with Hopfield model, Image Denoising: The Deep Learning Revolution and Beyond—A Survey Paper, Bayesian mixture variational autoencoders for multi-modal learning, Minimax formula for the replica symmetric free energy of deep restricted Boltzmann machines, Bimodal variational autoencoder for audiovisual speech recognition, Finite mixtures of ERGMs for modeling ensembles of networks, Distributed computation for marginal likelihood based model choice, Unnamed Item, Energy-Based Models with Applications to Speech and Language Processing, The role of mutual information in variational classifiers, An introduction to machine learning: a perspective from statistical physics, Correlated product of experts for sparse Gaussian process regression, High-dimensional density estimation with tensorizing flow, Scalable Bayesian optimization with generalized product of experts, Infinite Swapping Algorithm for Training Restricted Boltzmann Machines, Parameter Learning for Alpha Integration, Dissociable Forms of Repetition Priming: A Computational Model, Ensemble Kalman Methods for High-Dimensional Hierarchical Dynamic Space-Time Models, Efficient Training of Convolutional Deep Belief Networks in the Frequency Domain for Application to High-Resolution 2D and 3D Images, A Novel Parameter Estimation Method for Boltzmann Machines, Rhythmic Inhibition Allows Neural Networks to Search for Maximally Consistent States, An Infinite Restricted Boltzmann Machine, Optimal Curiosity-Driven Modular Incremental Slow Feature Analysis, Neural Quadratic Discriminant Analysis: Nonlinear Decoding with V1-Like Computation, Simultaneous Estimation of Nongaussian Components and Their Correlation Structure, Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review, Implementing a Bayes Filter in a Neural Circuit: The Case of Unknown Stimulus Dynamics, Convergence of Markovian stochastic approximation for Markov random fields with hidden variables, Direct Learning of Sparse Changes in Markov Networks by Density Ratio Estimation, Universal Approximation Depth and Errors of Narrow Belief Networks with Discrete Units, Semi-parametric estimation of the variogram scale parameter of a Gaussian process with stationary increments, Decreasing the Size of the Restricted Boltzmann Machine, Semisupervised Deep Stacking Network with Adaptive Learning Rate Strategy for Motor Imagery EEG Recognition, A Learning Framework for Winner-Take-All Networks with Stochastic Synapses, Unnamed Item, A complete restricted Boltzmann machine on an adiabatic quantum computer, Integration of Stochastic Models by Minimizing α-Divergence, A Connection Between Score Matching and Denoising Autoencoders, Statistical significance of the Netflix challenge, The Poisson transform for unnormalised statistical models, Scalable estimation strategies based on stochastic approximations: classical results and new insights, A Principle for Learning Egocentric-Allocentric Transformation, Learning Compositional Representations of Interacting Systems with Restricted Boltzmann Machines: Comparative Study of Lattice Proteins, Topographic Product Models Applied to Natural Scene Statistics, A Fast Learning Algorithm for Deep Belief Nets, Unnamed Item, Dynamic Programming Algorithms as Products of Weighted Logic Programs, A Generalization of Spatial Monte Carlo Integration, When Does a Mixture of Products Contain a Product of Mixtures?, Legendre decomposition for tensors*, Learning performance in inverse Ising problems with sparse teacher couplings, A dynamical mean-field theory for learning in restricted Boltzmann machines, Unnamed Item, Unnamed Item, Reconstruction of pairwise interactions using energy-based models*, Gaussian-spherical restricted Boltzmann machines, Replica analysis of the lattice-gas restricted Boltzmann machine partition function, Restricted Boltzmann machine based on a Fermi sea, Quantum state tomography of multi-qubit systems — a comparative study, On better training the infinite restricted Boltzmann machines, Tractable Multivariate Binary Density Estimation and the Restricted Boltzmann Forest, A Two-Layer Model of Natural Stimuli Estimated with Score Matching, Learning undirected graphical models using persistent sequential Monte Carlo, Characterizations of non-normalized discrete probability distributions and their application in statistics, Deterministic replica-exchange method without pseudo random numbers for simulations of complex systems, Interoperability of statistical models in pandemic preparedness: principles and reality, Fields of experts, Parameter inference in a probabilistic model from data: regulation of transition rate in the Monte Carlo method, A topological insight into restricted Boltzmann machines, Cost-sensitive boosting algorithms: do we really need them?, Rényi divergence minimization based co-regularized multiview clustering, Spatiotemporal-textual point processes for crime linkage detection, Quantum machine learning: a classical perspective, An Efficient Learning Procedure for Deep Boltzmann Machines, Learning Where to Attend with Deep Architectures for Image Tracking, Using contrastive divergence to seed Monte Carlo MLE for exponential-family random graph models, Nested aggregation of experts using inducing points for approximated Gaussian process regression, Preference relation-based Markov random fields for recommender systems, Topologically ordered feature extraction based on sparse group restricted Boltzmann machines, Convergence analysis of contrastive divergence algorithm based on gradient method with errors, Hierarchical recognition system for target recognition from sparse representations, Exploiting symmetries for scaling loopy belief propagation and relational training, Continual curiosity-driven skill acquisition from high-dimensional video inputs for humanoid robots, Probabilistic Segmentation of Musical Sequences Using Restricted Boltzmann Machines, Graph-based composite local Bregman divergences on discrete sample spaces, Necessary and sufficient conditions of proper estimators based on self density ratio for unnormalized statistical models, Multilayer bootstrap networks, On the importance of hidden bias and hidden entropy in representational efficiency of the Gaussian-bipolar restricted Boltzmann machines, Computational theories on the function of theta oscillations, An adaptive deep Q-learning strategy for handwritten digit recognition, Accelerating deep learning with memcomputing, A Multiclass Classification Method Based on Decoding of Binary Classifiers, Dynamical analysis of contrastive divergence learning: restricted Boltzmann machines with Gaussian visible units, Expected energy-based restricted Boltzmann machine for classification, The flip-the-state transition operator for restricted Boltzmann machines, Properties and comparison of some kriging sub-model aggregation methods, Fixed-domain asymptotic properties of maximum composite likelihood estimators for Gaussian processes, Independent component analysis: recent advances, Deep multimodal fusion: a hybrid approach, Entropic herding, Training restricted Boltzmann machines: an introduction, Stein's method meets computational statistics: a review of some recent developments, Preserving differential privacy in convolutional deep belief networks, Probabilistic joint image segmentation and labeling by figure-ground composition, Gaussian processes for computer experiments, Nested kriging predictions for datasets with a large number of observations, Contrastive Divergence in Gaussian Diffusions, Distributed probabilistic inferencing in sensor networks using variational approximation, Deep, Narrow Sigmoid Belief Networks Are Universal Approximators, Evolutionary algorithms for solving multi-objective travelling salesman problem, Analyzing market baskets by restricted Boltzmann machines, Search for the global extremum using the correlation indicator for neural networks supervised learning, The Architectures of Geoffrey Hinton, Bilevel Optimization with Nonsmooth Lower Level Problems, Conditional random fields for pattern recognition applied to structured data, Modeling documents with Event Model, Machine learning algorithms based on generalized Gibbs ensembles, A bound for the convergence rate of parallel tempering for sampling restricted Boltzmann machines, Unifying neural-network quantum states and correlator product states via tensor networks, Thermodynamics of restricted Boltzmann machines and related learning dynamics, Convergence of contrastive divergence algorithm in exponential family, Graph-induced restricted Boltzmann machines for document modeling, Storing, learning and retrieving biased patterns, Learning to Represent Spatial Transformations with Factored Higher-Order Boltzmann Machines, Scalability of using restricted Boltzmann machines for combinatorial optimization, Joining and splitting models with Markov melding, Composite likelihood estimation for a Gaussian process under fixed domain asymptotics, Deep Belief Networks Are Compact Universal Approximators, A new dual wing harmonium model for document retrieval, Astronomical image restoration using variational methods and model combination, Unnamed Item, Generalized darting Monte Carlo, Comparing Classification Methods for Longitudinal fMRI Studies, High-dimensional variable selection with the plaid mixture model for clustering, Multifractal Analysis of Multivariate Images Using Gamma Markov Random Field Priors, Neural Decoding with Hierarchical Generative Models, Algorithms for estimating the partition function of restricted Boltzmann machines, Class sparsity signature based restricted Boltzmann machine, Discovering the impact of hidden layer parameters on non-iterative training of feed-forward neural networks, Spatial disease mapping using directed acyclic graph auto-regressive (DAGAR) models, Distances and means of direct similarities, Learning sparse FRAME models for natural image patterns, Deep neural mapping support vector machines, The shape Boltzmann machine: a strong model of object shape, Modeling the Correlated Activity of Neural Populations: A Review, On dissipative symplectic integration with applications to gradient-based optimization, Learning a Generative Model of Images by Factoring Appearance and Shape, Bounding the Bias of Contrastive Divergence Learning, Some extensions of score matching, Non-parametric learning of lifted restricted Boltzmann machines, Optimization of neural network training for image recognition based on trigonometric polynomial approximation, Incremental Slow Feature Analysis: Adaptive Low-Complexity Slow Feature Updating from High-Dimensional Input Streams, Consistency of Pseudolikelihood Estimation of Fully Visible Boltzmann Machines, Learning algorithm of Boltzmann machine based on spatial Monte Carlo integration method, Building a telescope to look into high-dimensional image spaces, A tale of three probabilistic families: Discriminative, descriptive, and generative models, Learning large \(Q\)-matrix by restricted Boltzmann machines, Unnamed Item, A Maximum-Likelihood Interpretation for Slow Feature Analysis
Cites Work