Do SEO no one does not know spiders, robot.txt file is the file and spider communication, many people will say how to let spiders crawl, but very few people know to use robot.txt to control spiders, robot.txt use good words spider can be used for you. Let me introduce how to manipulate spiders.
One. Screen the same page
Search engines are a thing of the same, will not be included in duplicate or existing pages, even if included in the later will be deleted, so at first we can open the file, write to prohibit the spider to read the code of the page, to the spider to the next good impression, while the impression of the branch greatly reduced points, Affect future spiders to crawl sites.
Two. Maintain website security
Use a robots file can protect the site security, some people may ask how to protect the site, if you do not need a file to screen the Spider Crawl Admin background page or the following files, spiders will default to include these pages, so that hackers can take advantage of.
Three. hotlinking
General Hotlinking also on a few people, but once by the search engine "hotlinking" That guy you 100M broadband also unbearable, if you do is not a picture site, and do not want to be search engine "hotlinking" your site pictures.
Four. Site Map submission
Robot files also support the site map submission function, submitted a site map to make spiders better crawl site content, but a majority of people do not submit site map, although there is no great impact on the proposed submission.
Six. No two-level domain name crawl
Some websites will provide some special services to VIP members, but do not want the service to be retrieved by search engines.
These are enough for you to control the spider a lot, like people need to communicate between, we and spiders need to communicate. This article by Schindler SEO published in Fuzhou Printing network http://www.fzywzx.com, reproduced please keep the connection!