Information Graph for Content replication

Source: Internet
Author: User
Keywords Search engines well known
Tags .url blog content copy get helpful high import

Absrtact: As we all know, content duplication is a big taboo in search engine optimization, often reproduced articles is difficult to get high weight, so AdSense often use false original, the following is part of the information on the copy of the chart, I believe that you will be helpful

As we all know, content duplication is a big taboo in search engine optimization, often reproduced articles is difficult to get high weight, so the webmaster often use false original, the following is part of the information on the copy of the chart, I believe that you will be helpful.

#1 copy content from a blog:

The image above is mainly for the blog this type of Web site media, we use WordPress often the article without reservation to display on the home page, rather than using the output summary (as Semwatch), according to Randfish observation, in fact, this is the search engine mistakenly think that the content duplication.

Duplicate content caused by #2 URL parameter

How does the URL parameter like session ids,tracking IDs cause content duplication? Googlewebmastercentral (need to flip the wall) revealed that the same product page, if the search engine crawler crawl can be through a variety of links to crawl the same product page, there will be the following several negative effects:

1. Multiple URLs will dilute the breadth of the links. For example, the product page of the previous image, if there are 50 import links, it is possible to form a 3 kinds of import URL path, rather than a unique URL, which is equal to the transfer of the import link weight spread to 3 different links.

2. Search results may present unfriendly URLs (such as a large string of long session id,tracking IDs). Thus in Serp, reduce the user to this page clear Understanding degree (English URL such as Semwatch.org/sem, not only has the search engine friendly, more important is the user experience friendliness), does not favor the brand modelling.

#3 search engine's attitude towards content duplication

From the search Engine land to the above figure, the general search engine through 4 steps to identify the content is original or copy:

1. Discovery. When a search engine crawler discovers new content, he will immediately compare it with the content previously included to ensure that the content is original.

2. Discard. First, the search engine will discard pages that include those from the link factory, the MFA site (Made for Adense) and blacklisted IP

3. Anatomy. The next step is to analyze the inbound links of each page to determine the quality and source of the links.

4. Decision. The last is to review the previous pages and related links, decide which page is absolutely original.

#4 keyword Dismantling

Search engines will spiders crawl through one of your specific pages 4 or 40 different pages on a website, this behavior is generally through the content of the page linked to crawl crawl (such as the "snowboard" above), many friends hope that by associating many pages, even if there is no correlation between each other, Many pages of a site use the same keyword over and over again to improve rankings. But in fact, this kind of behavior is not helpful for ranking.

#5 how to deal with content duplication

The difference here is not just through the word "snowboard" but through a variety of, valuable and unique keywords (such as discount skis, child skis, etc.) linked to the original content. This way the search engine can easily determine the relevance of the page to other pages, not only based on search engine friendliness, but also on the user experience and the future information architecture of the site.

Canonical label

Source:Dynamical.biz

Source:SEOmoz.org

In fact, the above two pictures are related to a problem, that is, Web site normalization, in response to this problem, Zac predecessors on the previous analysis-Web site standardization of the latest solution, we can go to study.

301 redirect

301 redirects are also an important way to resolve content duplication, and for permanent redirection, search engines prefer 301 redirects. As 301 redirects are implemented, the weight of the old site is automatically delivered to the new site.

When you understand the search engine for the content of repeated judgments after the mechanism, I believe that you will be able to help optimize the work. We often say "the details win", so whether it is network marketing or simple SEM must follow such a principle and details, before winning.

Chart collection: Ann Smarty, original link This article first SEM Watch, reprint please indicate the source

Related Article

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.