Website duplicate content Baidu included duplicate page how to do

Source: Internet
Author: User

In search engine inclusion, we may find that the same content search engine included more than once, including "static pages, there are dynamic or pseudo-static pages." Seemingly increased the amount of the site is a good thing, but in fact, "repeating content" of the SEO optimization and not good, the result of the final, the light repeat snapshot is recycled deleted, the weight of the site directly affect the weight.


Today for your detailed analysis:
1. Why will the same page of the website be reproduced?
2. What is the impact of re-inclusion on SEO optimization?
3. How do we deal with the problem of repeated collection of websites?
  
From the above example, we can see that the homepage of the site in addition to static page collection, there are many suffixes with "? qz_gdt=" Dynamic page of the collection, but open the link we can find, are pointing to the homepage of the site. So in the face of the site repeated inclusion of the problem, what should we do?


We have to figure out the following three points:


A: What is the reason for the repeated inclusion of Web pages?
In the site optimization process, the search engine does not like to repeat the content of the collection, has been emphasizing the original, that is why the same page of the site will be repeated included?
The general suffix appears a question mark or other string, is because Baidu included the article at the same time, will give the article add a label, but visit the site, will automatically jump back to the original site, home page or inside pages, as for the link in the page is included in the homepage there may be because the article set on the first page display, So the final link address points to the home page.
1. Search engine identification problem-The search engine is just a set of rules for people to write, do not have sufficient recognition, so that some of the content of the URL address is reproduced.
2. Programming problems-some programs in the design for "preview" to facilitate the presence of dynamic pages in the background. However, due to improper design, the search engine still through the index of the directory can find this existence, and then indexed.
3. Template problem-in the template making often easy to ignore this problem, the final template after the completion of both dynamic pages, and static pages, so how can not be repeated by the search engine collection it?
4. Website revision-The website revision may lead to some URL duplicate collection, this can look at the previous "How to make the website revision" this article.


II: What will be the impact of repeated site pages?
In many cases, webmasters want their site to include more, but the same page repeated several times, it seems to increase the amount of included in the short term, but not too long to repeat the page will be deleted, after all, the same content exists in the search database will only occupy space. So, especially some new stations will find today included a lot, tomorrow to see a lot less.
Generally in the search engine after the collection will be redundant dynamic page culling, but if frequent such repeated inclusion culling, culling, may directly affect the spider crawl frequency, reduce the Site page collection, thus affecting the ranking of keywords. SEO optimization does not suggest that sites have the same content of multiple URL address links.


Third: How to deal with the repeated collection of Web pages?
We can proceed from the following four aspects:


1.URL standardization
In the site construction process as far as possible to unify the URL address, do not use dynamic page links, one is the search engine favor static page, on the other hand can also prevent the same content by the search engine repeated inclusion.


2. Using robots.txt to mask dynamic files
Use robots.txt for dynamic file masking, such as: "Disallow:/?" The root directory is blocked with "? "The page, this can be self-flexible settings.


3. Clear the cache in a timely manner
Each site will have two URLs, the general site after the update of the article will need to generate static URLs, the existence of a dynamic, the natural generation of static will be more than a URL, so if not clear the cache in time, it may make the same page static, dynamic URL URLs are included, Because the search cited the two URLs of the same article as independent, naturally two are included. Therefore, the timely elimination of the cache can greatly reduce the page is repeated.


4. Use Sitemap to guide spiders
The role of the site map is not only the whole site of the content of the induction, but also allows spiders to crawl more smoothly and unimpeded, orderly crawl the various pages within the site, which can also reduce the probability of the Web page repeated collection of spiders.

Website duplicate content Baidu included duplicate page how to do

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.