For the naive thought to practise my academic 中文版 skill, the rest of my notes would be wrriten in my terrrible 中文版. Xd
If you have any kind of uncomfortable feel, please close this window and refer to the original edition from Mr Lin.
I'll be really appriciate for your understanding.
OK, so much for this.
We discussed the powerful tool Dual SVM with a hard bound in the last class, which helps us to better understand the Meani Ng of SVM.
But we didn't solve the problem brought from a big ~d~
One large ~d~ my cause disaster when caculating Qd, which are the bottlenect of our model.
Here we introduce one tool called kernel function to better our situation.
If we can get the result of a specific kind of function with the parameters x and X ', we can cut down the process in Compu Ting and give the output diectly for the input.
That's what we call kernel function.
We use a 2nd order polynomial transform to illustrate.
The idea to make the computing sampler are to deal with xTx ' and the polynomial rrelationship at the same time.
So here we get the kernel function whose input is x and X ':
For the above question, we can apply the kernel function:
Quadratic coefficient q n,m = y n y m z n T z m = y n y m K (x N, x m) to get the Matrix Qd.
So, we need not to de the caculation in space of Z, but we could use KERNEL FUNCTION to get znt*zm used xn and XM.
Kernel Trick:plug in efficient Kernel function to avoid dependence on d?
So if we give the This method a name called Kernel SVM:
Let us come back to the 2nd polynomial, if we add some factor into expansion equation, we may get some new kernel functio N:
From the aspect of geometry, different kernel means different geometry distance, which affects the appearance of mapping, The defination of margin.
Now, we already has the common represention of polynomial kernel:
Particularly, Q = 1 means the linear condiction.
Then, what about infinite dimensionalφ (x)? Yes, it is also avalible.
Linear combination of Gaussians centered at SVs Xn, also called Radial Basis Function (RBF) kernel
The value of gamma determine the the extent of the tip for the Gaussian function, which means the shadow of overfiting are Still alive.
In the end, Mr Lin compares these three kernel functions and shows their pros and cons, which is in line with the Intuit Ion.
An important point was that one potential manual kernel has to be"ZZ must always be positive semi-definite"
Machine learning techniques-3-dual Support Vector Machine