The three sites and the two sites were launched low in September (Station A and Station B respectively), basically all of which were collected for pseudo-original Publishing, yesterday, I also started to get a website that is classified separately. It is also post-collection of pseudo-original (Station C ).
The current situation is: Station a B is not indexed on Baidu, but dozens of articles have been indexed by Soso and Google. Neither Site C is included.
After reading the log of March October 5, I found the website a was crawled by a hundred-degree spider, and the address was robots.txt. Then I went away and never came again. First, Google searches for robots.txt and crawls several other pages. Because my website does not have this file, I just sent a robots.txt file.
1: This spider crawls at AM.
2: There are a lot of Google spider crawlers that crawl duplicate pages. For a large page, it will be crawled by n different spider (I found that a page is crawled by more than 10 IP addresses. These IP addresses are searched by Google China, the page is large ..)
Because these sites are not included by Baidu at present, I initially suspected that Baidu found that I was a pseudo-original, but after today's log analysis, Baidu did not come to crawl my page, I have bought all my domain names for Record Filing. I don't know if my domain names are not indexed by Baidu. Is this affected?
I plan to buy a new domain name. I have no record filing test ..
For today's analysis, I will first tease robots.txt to see the effect of tomorrow .. I will share my recent achievements with you and come to the law to make progress.