Intermediary transaction http://www.aliyun.com/zixun/aggregation/6858.html ">seo diagnose Taobao guest cloud host technology Hall
In a website, robots.txt is an important document, each site should have a set of correct robots.txt.
Before search engines can crawl your site's files, they will first take a look at the settings of your site's robots.txt file to see which files and directories you allow it to crawl. So how do you configure your robots.txt file?
Let's look at an example:
# robots.txt file Start
# Exclude Files from all Robots:
User: *
Disallow:/admin_login/
Disallow:/admin/
Disallow:/admin.htm
Disallow:/admin.aspx
# End Robots.txt File
There is a # number are annotations, easy to read.
User is the search engine spider, followed by the * number, indicating that all spiders are effective.
Disallow means no crawl, a directory or file that is not allowed to crawl.
Edit this file, then save it in the root of your site (must be the root directory, of course, you can adjust), then the search engine will see.
If you do not have robots.txt files, in your website access log files, you will see Spiders access to robots.txt files unsuccessful records.
OK, now go and set up your robots.txt file.