Wunda Deep Learning Special Course 3 study notes/week1/setting up ML application

Source: Internet
Author: User

Applying ML is a highly iterative process

Idea->code->experment->, .....

To constantly adjust the hyper-parameters.

Train/dev/test Sets

Datasets are typically divided into train/dev/test sets.

    • Training set: Training for Models

    • Hold-out Cross Validation Set/developmet set: For testing, tuning model hyper-parameters

    • Test set: for final evaluation

Previous ML problem: Data scale at w level, usually 70/30 divided by train/test set or 60/20/20 scale.

Now the Big Data Age: Data scale in the hundred W, trend is the proportion of dev/test set is reduced, because the dev set only needs to be large enough to judge the merits and demerits of different models, the test set is similar. If there is 100w data set, the ratio of 98/1/1 is a more reasonable division.

Experience:

    1. Make sure that the dev/test set comes from the same distribution, which helps to improve the efficiency of the algorithm.
    2. The training set can take a page crawl and so on to obtain a larger amount of data.
    3. There is no test set available, only with the dev set to adjust.
Bias/variance

You can determine whether the model has a high bias/variance problem by train set error and dev set error.

Suppose, for example, that Bayes error is about 0%, which means that humans can be almost completely accurately identified, and the train and dev sets come from the same distribution.

Train Set Error Dev Set Error Results
1% 11% High Variance
15% 16 High bias
15% 30% High Variance & High bias
0.5% 1% Low Variance & Low bias
Some Basic Recipe (Basic guidelines)

STEP1 High bias? -Greater network, longer training time; (Other NN architectures)

STEP2 High variance? -more data, regularization; (Other NN architectures)

In the age of deep learning, there is no need to think too much about the trade-offs between bias and variance, and if reasonable regularization is adopted, a larger network will reduce bias without increasing variance, if there is a large enough network, Getting more data will hardly increase bias in reducing variance, which is one of the most effective reasons for deep learning in the field of supervised learning.

Wunda Deep Learning Special Course 3 study notes/week1/setting up ML application

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.