July algorithm-December machine learning Online Class-14th lesson Note-em algorithm

Source: Internet
Author: User

July Algorithm-December machine Learning online Class -14th lesson Note-em Algorithm

July algorithm (julyedu.com) December machine Learning Online class study note http://www.julyedu.com

?

EM expection Maxium Desired Maximum

1 cited examples

1000 people, Statistics height, 1.75,1.62,1.94, how many men and women, each height corresponds to the male and female

?

1.1 How to calculate? Estimating mean and variance using maximum likelihood estimation

The results of the above conclusions and moment estimates are consistent,

Namely: The mean value of the sample is the mean value of the Gaussian distribution, the pseudo-variance of the sample is the variance of the Gaussian distribution.

If it's a Gaussian distribution, you can use this calculation, the mean value and the variance

?

setting, the height of men and women obeys two Gaussian distributions

The random variable x is a mixture of K Gaussian distributions, if a series of samples of random variable x are observed x1,x2,..., xn,

The target function is: logarithmic likelihood function

Because of the addition of the logarithm function, it is impossible to directly derive the maximal value directly by the derivation solution equation. Divided into two steps

?

1.2 Intuitive understanding of EM

STEP1: estimated data from which group

STEP2: estimate parameters for each component

1.3 Gauss-Proof

Some of the pictures can be learned: To do the great expectations, and constantly do the great Expectations

?

Using Jensen inequalities (convex optimization can be used directly)

Qi is a certain distribution of Z, Qi≥0, with:

Further analysis: There is the above equation can be learned, p,q in direct proportion;

You can get the overall frame diagram of EM.

?

1.4 Derivation of GMM from theoretical formulae

Random variable x is a K Gaussian distribution mixed with the probability of each Gaussian distribution is φ1φ2 ... Φk, the mean value of the first Gaussian distribution is μi and the variance is ōi. If a series of samples of random variable x are observed x1,x2,..., xn, the parameter π,μ,σ is estimated.

E-step, and M-step, respectively.

E-step:

M-step: (First write the expectation, the maximum value)

The mean value is biased, so that the equation equals 0, the mean value of the solution:

Similarly, the value of the variance can be obtained.

After the variance and the mean are obtained, the φ1+φ2+ of φ is biased, and the constraint condition is ... Φk=1, which is the constrained condition with the equation to find the extremum, the Lagrange multiplier method

The final formula is the same as that of the original Euler-like interpretation.

For all data points, it can be considered that the component K generates these points. Component k is a standard Gaussian distribution.

Methods with implicit variables: em+ variational

?

July algorithm-December machine learning Online Class-14th lesson Note-em algorithm

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.