<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://en.formulasearchengine.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=129.129.157.152</id>
	<title>formulasearchengine - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://en.formulasearchengine.com/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=129.129.157.152"/>
	<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/wiki/Special:Contributions/129.129.157.152"/>
	<updated>2026-05-02T00:58:03Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.0-wmf.28</generator>
	<entry>
		<id>https://en.formulasearchengine.com/index.php?title=Muon_spin_spectroscopy&amp;diff=10557</id>
		<title>Muon spin spectroscopy</title>
		<link rel="alternate" type="text/html" href="https://en.formulasearchengine.com/index.php?title=Muon_spin_spectroscopy&amp;diff=10557"/>
		<updated>2014-01-07T09:06:00Z</updated>

		<summary type="html">&lt;p&gt;129.129.157.152: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Variational message passing&#039;&#039;&#039; (VMP) is an [[approximate inference]] technique for continuous- or discrete-valued [[Bayesian networks]], with [[conjugate exponents|conjugate-exponential]] parents, developed by John Winn. VMP was developed as a means of generalizing the approximate [[Variational Bayesian methods|variational methods]] used by such techniques as [[Latent Dirichlet allocation]] and works by updating an approximate distribution at each node through messages in the node&#039;s [[Markov blanket]].&lt;br /&gt;
&lt;br /&gt;
==Likelihood Lower Bound==&lt;br /&gt;
&lt;br /&gt;
Given some set of hidden variables &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; and observed variables &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;, the goal of approximate inference is to lower-bound the probability that a graphical model is in the configuration &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;.  Over some probability distribution &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; (to be defined later), &lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; \ln P(V) = \sum_H Q(H) \ln \frac{P(H,V)}{P(H|V)} = \sum_{H} Q(H) \Bigg[ \ln \frac{P(H,V)}{Q(H)} - \ln \frac{P(H|V)}{Q(H)} \Bigg]  &amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
So, if we define our lower bound to be &lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; L(Q) = \sum_{H} Q(H) \ln \frac{P(H,V)}{Q(H)} &amp;lt;/math&amp;gt;,&lt;br /&gt;
&lt;br /&gt;
then the likelihood is simply this bound plus the [[relative entropy]] between &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt;.  Because the relative entropy is non-negative, the function &amp;lt;math&amp;gt;L&amp;lt;/math&amp;gt; defined above is indeed a lower bound of the log likelihood of our observation &amp;lt;math&amp;gt;V&amp;lt;/math&amp;gt;.  The distribution &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; will have a simpler character than that of &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; because marginalizing over &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; is intractable for all but the simplest of [[graphical models]].  In particular, VMP uses a factorized distribution &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt; Q(H) = \prod_i Q_i(H_i), &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;H_i&amp;lt;/math&amp;gt; is a disjoint part of the graphical model.&lt;br /&gt;
&lt;br /&gt;
==Determining the Update Rule==&lt;br /&gt;
&lt;br /&gt;
The likelihood estimate needs to be as large as possible; because it&#039;s a lower bound, getting closer &amp;lt;math&amp;gt;\log P&amp;lt;/math&amp;gt; improves the approximation of the log likelihood.  By substituting in the factorized version of &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;L(Q)&amp;lt;/math&amp;gt;, parameterized over the hidden nodes &amp;lt;math&amp;gt;H_i&amp;lt;/math&amp;gt; as above, is simply the negative [[relative entropy]] between &amp;lt;math&amp;gt;Q_j&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q_j^*&amp;lt;/math&amp;gt; plus other terms independent of &amp;lt;math&amp;gt;Q_j&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;Q_j^*&amp;lt;/math&amp;gt; is defined as&lt;br /&gt;
&lt;br /&gt;
:&amp;lt;math&amp;gt;Q_j^*(H_j) = \frac{1}{Z} e^{\mathbb{E}_{-j}\{\ln P(H,V)\}} &amp;lt;/math&amp;gt;,&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;\mathbb{E}_{-j}\{\ln P(H,V)\}&amp;lt;/math&amp;gt; is the expectation over all distributions &amp;lt;math&amp;gt;Q_i&amp;lt;/math&amp;gt; except &amp;lt;math&amp;gt;Q_j&amp;lt;/math&amp;gt;.  Thus, if we set &amp;lt;math&amp;gt;Q_j&amp;lt;/math&amp;gt; to be &amp;lt;math&amp;gt;Q_j^*&amp;lt;/math&amp;gt;, the bound &amp;lt;math&amp;gt;L&amp;lt;/math&amp;gt; is maximized.&lt;br /&gt;
&lt;br /&gt;
==Messages in Variational Message Passing==&lt;br /&gt;
&lt;br /&gt;
Parents send their children the expectation of their [[sufficient statistic]] while children send their parents their [[natural parameter]], which also requires messages to be sent from the co-parents of the node.&lt;br /&gt;
&lt;br /&gt;
==Relationship to Exponential Families==&lt;br /&gt;
&lt;br /&gt;
Because all nodes in VMP come from [[Exponential family|exponential families]] and all parents of nodes are [[Conjugate prior|conjugate]] to their children nodes, the expectation of the [[sufficient statistic]] can be computed from the [[normalization factor]].&lt;br /&gt;
&lt;br /&gt;
==VMP Algorithm==&lt;br /&gt;
&lt;br /&gt;
The algorithm begins by computing the expected value of the sufficient statistics for that vector.  Then, until the likelihood converges to a stable value (this is usually accomplished by setting a small threshold value and running the algorithm until it increases by less than that threshold value), do the following at each node:&lt;br /&gt;
# Get all messages from parents&lt;br /&gt;
# Get all messages from children (this might require the children to get messages from the co-parents)&lt;br /&gt;
# Compute the expected value of the nodes sufficient statistics&lt;br /&gt;
&lt;br /&gt;
==Constraints==&lt;br /&gt;
&lt;br /&gt;
Because every child must be conjugate to its parent, this limits the types of distributions that can be used in the model.  For example, the parents of a [[Gaussian distribution]] must be a [[Gaussian distribution]] (corresponding to the [[Mean]]) and a [[gamma distribution]] (corresponding to the precision, or one over &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt; in more common parameterizations).  Discrete variables can have [[Dirichlet distribution|Dirichlet]] parents, and [[Poisson distribution|Poisson]] and [[Exponential distribution|exponential]] nodes must have [[gamma distribution|gamma]] parents.  However, if the data can be modeled in this manner, VMP offers a generalized framework for providing inference.&lt;br /&gt;
&lt;br /&gt;
==External links==&lt;br /&gt;
* [http://research.microsoft.com/infernet Infer.NET]: an inference framework which includes an implementation of VMP with examples.&lt;br /&gt;
* [http://dimple.probprog.org/ dimple]: an open-source inference system supporting VMP.&lt;br /&gt;
* An [http://vibes.sourceforge.net/ older implementation] of VMP with usage examples.&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*{{cite journal |first1=J.M. |last1=Winn |first2=C. |last2=Bishop |title=Variational Message Passing |journal=Journal of Machine Learning Research |volume=6 |pages=661–694 |year=2005 |url=http://johnwinn.org/Publications/papers/VMP2005.pdf |format=PDF}}&lt;br /&gt;
*{{Cite thesis |type=PhD  |title=Variational Algorithms for Approximate Bayesian Inference |url=http://www.cs.toronto.edu/~beal/thesis/beal03.pdf |last=Beal |first=M.J. |year=2003 |publisher=[http://www.gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit], University College London }}&lt;br /&gt;
&lt;br /&gt;
[[Category:Bayesian networks]]&lt;/div&gt;</summary>
		<author><name>129.129.157.152</name></author>
	</entry>
</feed>