[PGM] Stanford probability graph model (Probabilistic graphical model)-Lecture 2: template models and structured CPDs

Source: Internet
Author: User

The probabilistic graphical model series is explained by Daphne Koller In the probabilistic graphical model of the Stanford open course. Https://class.coursera.org/pgm-2012-002/class/index)

Main contents include (reprinted please indicate the original source http://blog.csdn.net/yangliuy)

1. probabilistic Graph Model Representation and deformation of Bayesian Networks and Markov networks.

2. Reasoning and inference methods, including Exact Inference (variable elimination, clique trees) and approximate inference (belief propagation message passing, Markov Chain Monte Carlo methods ).

3. Learning Methods for parameters and structures in the probability graph model.

4. Use the probability graph model for Statistical Decision modeling.

Lecture 2. template models and structured CPPS.

1 template models

The graphic layout model is a more compact description of the graphic model. Template variables are repeated variables in the graph model multiple times, such as the IQ of multiple students and the difficulty of multiple courses. The template graph model describes how the template variables inherit Dependencies from the template. Typical templatemodels include the dynamic Bayesian Model DBN, Hidden Markov Model hmm, and platemodels. The Dynamic Bayesian model mainly introduces Markov hypothesis and time immutability in Bayesian Networks. These models will be further introduced in the next few lectures. Let's take a look at the exercises of templatemodels.

2 CPD

CPD is the conditional probability distribution. The table method is not suitable for CPD. If a node has K parents, the table has O (2exp (k) rows. The general CPD definition is given below

It mainly includes the following models:

The deterministic CPD is also called context specific independence, as shown in the following formula:

Given the value of variable C, the random variable X is independent from the random variable Y condition, given Z.

3 treestructured CPD

CPD in the tree structure, different path selection in the tree structure will lead to different conditional independence. In the following example, given J and C1, that is, if someone is known to have selected recommendation letter 1, whether or not the work can be obtained depends only on the Validity probability distribution of recommendation letter L1, then the L2 → J Path becomes spurious, therefore, there is no active trails between L1 and L2. L1 and L2 are independent of each other. We call L1 and L2 belong to D-separated, and the conditional independence relationship given J and context c = C1. is also called CSI-separation, that is, context-related conditions are independent.

 

4 noisy or CPD

As shown in, the y node has a parent node Z1 to ZK, or relationship, that is, only Z1 to ZK do not occur, then y = 0, whether Zi is affected by the random variable XI. If we regard y as the Professor's recommendation letter, then X1 to XK can be seen as a student's performance, the probability of being admired by the professor when the student's item I is excellent is λ I. This figure means that if the professor appreciates at least one of the students' excellent performances, there will be a letter of recommendation. So what is z0? This probability is called leak probability. It can be considered that the professor is in a good mood one day. Even if every performance of a student is messy, he also gives a letter of recommendation, that is, this variable is not affected by the student's performance Xi, it depends entirely on the professor. So the probability of no recommendation letter is that the probability that each item of the student performs poorly multiplied by the probability that the professor is in a bad mood. Therefore, the conditional probability formula shown below is shown.

5 sigmoid CPD

Clearly understand the sigmoid function. This conditional probability is well understood, for example:

 

Exercise questions

6. Linear Gaussian Model

Y follows the Gaussian distribution, and its mean value is a linear combination of the father node XI. The variance is irrelevant to the father node.

In a linear Gaussian Model, if the coefficients in the linear combination of mean values depend on the random variable A, the variance and the random variable A are called conditional linear Gaussian models.

 

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.