Content deleted Content added
Nick Number (talk | contribs) m repaired link(s) to disambiguation page - (you can help) |
→Guassian graphical models of protein structures: select disambig for partition function, recip of Z is already in formula above |
||
Line 53:
:<math>f(\Theta=D) = \frac{1}{Z} \exp\left\{-\frac{1}{2}(D-\mu)^T\Sigma^{-1}(D-\mu)\right\}</math>
Where <math>Z =
To learn the graph structure as a multivariate Gaussian graphical model, we can use either [[L-1 regularization]], or [[neghborhood selection]] algorithms. These algorithms simultaneously learn a graph structure and the edge strength of the connected nodes. An edge strength corresponds to the potential function defined on the corresponding two-node [[clique]]. We use a training set of a number of PDB structures to learn the <math>\mu</math> and <math>\Sigma^{-1}</math>.
Once the model is learned, we can repeat the same step as in the discrete case, to get the density functions at each node, and use analytical form to calculate the free energy. Here, the [[Partition function (mathematics)|partition function]] already has a [[closed form]], so the [[inference]], at least for the Gaussian graphical models is trivial. If the analytical form of the partition function is not available, [[particle filtering]] or [[expectation propagation]] can be used to approximate ''Z'', and then perform the inference and calculate free energy.
{{No footnotes|date=August 2010}}
|