Intermediary transaction http://www.aliyun.com/zixun/aggregation/6858.html ">seo diagnose Taobao guest cloud host technology Hall
Spiders crawling often determines the site included, so in peacetime optimization work, we will be more concerned about IIS log, through the observation log to get spider crawling site dynamics, and through the analysis of spider dynamic to monitor the progress of the site optimization. But most of the time webmaster will see their own IIS log shows spiders crawling very not smooth, although the site is considered to optimize the perfect. What are the reasons for this?
There is a reason, the spider is just a robot, it is composed of program code, must follow its provisions to crawl the site, if the site has some site traps, it will make spiders crawl become not smooth, resulting in a series of problems. In the end which Web site traps become spiders crawling stumbling block? The following author to talk about it.
"Trap One": Site map is incorrect
Site Map is a very useful tool for Web sites, for users and spiders, a complete, correct site map can well identify the entire site's architecture, so as to better browse and crawl the site. Because some webmaster for the code is not familiar with, and the structure of the site is not familiar with, casually selected a bad authority of the tool to produce a incomplete or incorrect site map, the final result is to let the spider crawl into which, eventually "lost".
# #解决方法: Choose an authoritative good tool to make site map, such as webmaster tools, Baidu Webmaster Platform tools, etc., if they are more familiar with the site, it is best to hand-made, and personal testing, to ensure good site map of the correctness and integrity.
"Trap Two": The number of dead links in the site is huge
The so-called dead link is the return code of 404 error page links, such links are usually generated after the site revision, or the site after the replacement of the domain name. The existence of dead links is very bad for the user experience and spider crawling. Dead link is no doubt is to block the spider crawling, when the spiders encounter these should not appear dead link, will generate distrust of the site, will eventually give up crawling site.
# #解决办法:
(1) Submit dead link. The use of Baidu Webmaster platform "dead Chain submission" tool to submit the site dead chain, specific instructions please refer to Baidu Webmaster tools.
(2) Redirect or delete dead links, if the location of the dead link is not very important, you can redirect it to the home page, if the location is more important, the number is also more, you can choose to delete the dead link, you can use the tool to delete.
"Trap three": URL contains too many parameters
Although Baidu optimization guide official shows that the current Baidu search robot can also be like Google robot included dynamic parameter site, but static URL of the site will always be more than dynamic URL included more advantages. Therefore, if your site is the same as the following URL, it may cause spiders crawling not smooth:
# #解决办法: Before choosing to do the station program, we must consider whether the program supports static web site URLs, and in the future maintenance should also pay attention to the site URL is really static, as far as possible to discard the dynamic Web site containing parameters of the practice.
"Trap Four": Too much anchor text on the website
Site anchor text too much, resulting in chain sprocket. External sprocket I believe everyone will know, but in fact, internal links can also produce sprocket. Many webmaster in order to enhance the keyword ranking, regardless of the use of too much anchor text, and then caused the link between the page reincarnation effect, and finally let the spider into the bottomless pit, can not come out.
# #解决办法: Clear the past, with links to the intersection of the anchor text, and try to do as much as necessary to add anchor text, abuse of anchor text will only make Baidu spider more step does not recognize your site.
Baidu search engine Spider is always just a program robot, its intelligence is never inferior to us, but we should respect it, provide it with a crawling smooth web site environment, so as to ensure that our website can continue to develop healthily. This article by the Beijing Hospital Registration network http://www.bjghjgw.com feeds, hoped that the reprint friend adds a link, thanks everybody support!