Original content Judgment search engine blind area

Source: Internet
Author: User

August 10, Baidu Webmaster Outpatient Open Day, a lot of webmaster mentioned the site original content by Baidu was judged as reproduced, and the reprint site to determine the original issue, and Lee also said that the site will be the original content of the identification of the improved algorithm. Details see: August 10 Baidu Webmaster Outpatient Open Day Issue collection.

Then on August 11, Google announced that the search results will be a major adjustment to reduce the ranking of infringing sites, Google after receiving copyright owners on the site a large number of infringement allegations, will reduce the site in Google's search results rankings. Search engine as the Internet's most important traffic portal, shoulder the link between the site and users, users, search engines, websites become a symbiotic system of mutual benefit. However, as a search engine difficult to determine what the original content of the user and the site will be affected?

  User article:

Users have long been accustomed to searching for their own information through search engines, the search engine is the user's tool, the general user is also very little concerned about the source of information, only concerned about whether the information itself to meet the needs. That is, the vast majority of users do not care about the creators of information, and do not care which site. If the user through the search engine to reach Sina or Sohu and other large websites, virtually strengthen the trust of information. Even if the information is not original for a large web site.

  Search Engine Chapter:

The object of Search engine service is user, not website; The first task of search engine is to provide valuable information to users. The general user does not care about the original information, however, search engine in order to maintain the fairness of the search results and the entire network environment order, must be in the complex network to determine the original and reprint, to a large number of high-quality original content of the site higher weight and better rankings, bring more traffic, to a large number of copied site down right. But all this premise is the search engine can accurately identify original and reprint, not only to Baidu, even Google, also failed to make people satisfied with the point.

Aside from the technical background, pure from the perspective of common sense, the first thought is the time, original certainly than the content of the reprint appeared earlier. If the article 1 by the website A original, was reproduced by Sina, other major sites are reproduced from Sina, a time on the network appeared a large number of reprint. According to statistics, the search engine only found a collection of the entire internet about 30% of the data, and this is already an astronomical, in the case of limited resources, search engines can not do real-time monitoring of the entire Internet information, so, Sina became article 1 of the original.

Secondly, according to the copyright of the article, more and more articles will have a copyright, although a lot of the reprint will remove this information, but, there is always someone left this paragraph of information, spiders according to the copyright information to identify the original and reproduced. Theoretically feasible, in fact, many unscrupulous reprint not only removed the original copyright information, and even add their own copyright, which caused the same article a number of original source of the phenomenon.

In addition, search engine work has to crawl, participle, denoising, refining keywords and other processes, and the content of the article is not much related to the copyright information is often stripped in the process of denoising. Even if the search engine adjusts the algorithm, retains the copyright information, similarly cannot avoid the afore-mentioned not to reproduce the predicament.

 Website article:

Site to ensure that they produce a large number of high-quality information to please search engines, but also a good user experience to retain users, and often by the search engine bullying, like a doormat xiaoxifu. Many webmasters often complain about search engines, but ignore a point, the search engine's customers are users, not the site, search engines do not have the obligation to give the site traffic.

But the search engine's unfairness and technical flaws bring terrible Matthew effect, big websites getting stronger and smaller websites getting weaker. Large sites at random plagiarism still strong rankings, traffic continues. Small web site hard production, but finally for others to do wedding clothes, ranked No, empty.

You can see that the search engine users do not care about information producers, but this is related to the search engine upstream customers-the site's dead and alive, related to the entire Internet environment, but also search engines must overcome the blind zone.



Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.