Intermediary transaction http://www.aliyun.com/zixun/aggregation/6858.html ">seo diagnose Taobao guest cloud host technology Hall
What is a Web site? It is a protocol, not a command.
As a learning or to engage in SEO optimization friends must know to search spiders, but for this search spiders follow the agreement, but not often heard. Robots are the first files to view when crawling a Web site. It is used to tell spiders what files in the server space can be crawled, and what files do not need to be crawled. Also because of this, using a good robots is very easy for your site to claim power, that is how to write it?
1. Avoid site duplication page
Many web sites offer a wide range of browsing versions, although in a large extent user-friendly, improve the user experience, but the spider has caused a little difficulty, because spiders are difficult to identify primary and secondary, once let it think you in malicious repetition, so light down right, heavy K station pull hair ah.
User: *
Disallow:/sitemap/"No Spiders crawl Text page"
Extended reading: "Huizhou SEO blog K station Mystery" (Ye Jianhui also at the beginning of the blog on the start of a similar problem and was K, specific to see the link)
2. Secure website
Maybe some people have big question marks. Robots and Web site security how to pull up the relationship? This is to say that hackers, many low-level hackers is through the search default background landing to achieve the goal of the invasion site, so security has to prevent AH.
User: *
Disallow:/admin/"prohibit spiders from crawling all files under Admin directory"
Extended reading: "WordPress program Security Settings Skills" (I have been attacked, so even the details of security issues can not be ignored, specific view Ye Jianhui blog)
3. Prevent links from being "stolen"
General hotlinking is not much, but once by the search engine "hotlinking", then the hundred mega Broadband I am afraid that too, so if not do the picture site, but also want to avoid being hotlinking, then can be shielded.
User: *
disallow:.jpg$
4. Submit Sitemap
SEO for the site now to optimize the friend, all know to do site map, but there will always be some friends do not know to submit, just think on the page add a link, then finished. In fact, the robots have to submit this function, a bit to Huizhou SEO Blog for example.
Sitemap:http://www.huizhouseo.cn/sitemap_baidu.xml
Sitemap:http://www.huizhouseo.cn/sitemap.html (Note that the s of the Sitemap needs to be uppercase)
5. Prohibit a certain level two domain name crawl
Some websites will provide some VIP privilege service to the member, but do not want this service to be retrieved by the search engine (same, CSS and JS and so on level two directory files do not need to be crawled also so)
User: *
Disallow:/
The above 5 point robots, I believe it is enough to let SEO beginners to search engine spider control ability to improve a lot, as if Baidu said, we and search engine should be friends, so indispensable to increase some communication, so as to eliminate some unnecessary diaphragm, right! This article by Huizhou SEO blog http:// www.huizhouseo.cn/seojichu/766.html Ye Jianhui feeds, reprint please specify!