Spiders crawling not smooth clean web site traps is the key

Source: Internet
Author: User
Keywords Traps often

Intermediary transaction http://www.aliyun.com/zixun/aggregation/6858.html ">seo diagnose Taobao guest cloud host technology Hall

Spiders crawling often determines the site included, so in peacetime optimization work, we will be more concerned about IIS log, through the observation log to get spider crawling site dynamics, and through the analysis of spider dynamic to monitor the progress of the site optimization. But most of the time webmaster will see their own IIS log shows spiders crawling very not smooth, although the site is considered to optimize the perfect. What are the reasons for this?

There is a reason, the spider is just a robot, it is composed of program code, must follow its provisions to crawl the site, if the site has some site traps, it will make spiders crawl become not smooth, resulting in a series of problems. In the end which Web site traps become spiders crawling stumbling block? The following author to talk about it.

"Trap One": Site map is incorrect

Site Map is a very useful tool for Web sites, for users and spiders, a complete, correct site map can well identify the entire site's architecture, so as to better browse and crawl the site. Because some webmaster for the code is not familiar with, and the structure of the site is not familiar with, casually selected a bad authority of the tool to produce a incomplete or incorrect site map, the final result is to let the spider crawl into which, eventually "lost".

# #解决方法: Choose an authoritative good tool to make site map, such as webmaster tools, Baidu Webmaster Platform tools, etc., if they are more familiar with the site, it is best to hand-made, and personal testing, to ensure good site map of the correctness and integrity.

"Trap Two": The number of dead links in the site is huge

The so-called dead link is the return code of 404 error page links, such links are usually generated after the site revision, or the site after the replacement of the domain name. The existence of dead links is very bad for the user experience and spider crawling. Dead link is no doubt is to block the spider crawling, when the spiders encounter these should not appear dead link, will generate distrust of the site, will eventually give up crawling site.

# #解决办法:

(1) Submit dead link. The use of Baidu Webmaster platform "dead Chain submission" tool to submit the site dead chain, specific instructions please refer to Baidu Webmaster tools.

(2) Redirect or delete dead links, if the location of the dead link is not very important, you can redirect it to the home page, if the location is more important, the number is also more, you can choose to delete the dead link, you can use the tool to delete.

"Trap three": URL contains too many parameters

Although Baidu optimization guide official shows that the current Baidu search robot can also be like Google robot included dynamic parameter site, but static URL of the site will always be more than dynamic URL included more advantages. Therefore, if your site is the same as the following URL, it may cause spiders crawling not smooth:

# #解决办法: Before choosing to do the station program, we must consider whether the program supports static web site URLs, and in the future maintenance should also pay attention to the site URL is really static, as far as possible to discard the dynamic Web site containing parameters of the practice.

"Trap Four": Too much anchor text on the website

Site anchor text too much, resulting in chain sprocket. External sprocket I believe everyone will know, but in fact, internal links can also produce sprocket. Many webmaster in order to enhance the keyword ranking, regardless of the use of too much anchor text, and then caused the link between the page reincarnation effect, and finally let the spider into the bottomless pit, can not come out.

# #解决办法: Clear the past, with links to the intersection of the anchor text, and try to do as much as necessary to add anchor text, abuse of anchor text will only make Baidu spider more step does not recognize your site.

Baidu search engine Spider is always just a program robot, its intelligence is never inferior to us, but we should respect it, provide it with a crawling smooth web site environment, so as to ensure that our website can continue to develop healthily. This article by the Beijing Hospital Registration network http://www.bjghjgw.com feeds, hoped that the reprint friend adds a link, thanks everybody support!

Related Article

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.