Helmholtz machine

Revision as of 08:04, 23 February 2025 by imported>Cosmia Nebula (MDL paper)
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

Template:Short description The Helmholtz machine (named after Hermann von Helmholtz and his concept of Helmholtz free energy) is a type of artificial neural network that can account for the hidden structure of a set of data by being trained to create a generative model of the original set of data.<ref name=“nc95“>Template:Cite journal Template:Closed access</ref><ref>Template:Cite journal</ref> The hope is that by learning economical representations of the data, the underlying structure of the generative model should reasonably approximate the hidden structure of the data set. A Helmholtz machine contains two networks, a bottom-up recognition network that takes the data as input and produces a distribution over hidden variables, and a top-down "generative" network that generates values of the hidden variables and the data itself. At the time, Helmholtz machines were one of a handful of learning architectures that used feedback as well as feedforward to ensure quality of learned models.<ref>Luttrell S.P. (1994). A Bayesian analysis of self-organizing maps. Neural Computation. 1994 Sep 1;6(5):767-94.[1]</ref>

Helmholtz machines are usually trained using an unsupervised learning algorithm, such as the wake-sleep algorithm.<ref>Template:Cite journal Template:Closed access</ref> They are a precursor to variational autoencoders, which are instead trained using backpropagation. Helmholtz machines may also be used in applications requiring a supervised learning algorithm (e.g. character recognition, or position-invariant recognition of an object within a field).

See alsoEdit

ReferencesEdit

Template:Reflist

External linksEdit


Template:Compu-AI-stub