In our optimization process, we will inevitably encounter some problems that plague us, these issues will be related to our optimization strategy. What do we do with the chain effect? Does our space have instability that the naked eye cannot see? Which of our content is more favored? Search engine spiders crawl our site frequency? These issues will be related to our optimization analysis and strategy formulation. So what channels can better analyze these problems? We can analyze our site's IIS log files.
One: The importance of IIS for our SEO
1: Through the site of the IIS log we can understand the search engine on our site crawling situation, including the search engine crawling trajectory and crawling amount, and then analysis of our construction of the chain on the impact of crawling, we put the chain as a search engine spiders into our site entrance, IIS records also record the search engine spiders crawl in from the portal.
2: The site update frequency and search engine crawl frequency has a certain relationship, in general, the frequency of updates is high the search engine crawling more frequently. We can use the search engine crawl frequency in the IIS log for content update fine-tuning.
3: Through the IIS log can reflect the site's space there are some of the current appearance can not recognize errors, and these errors can be the first time by the IIS log reflected. We can find the problem and solve the problem early through the log.
4: At the same time through the IIS log we can analyze our content, what is the search engine is favored, what is the search engine to look. This data is used to fine-tune the content.
Two: How to get the site IIS logs and settings considerations
1: First of all, our host needs to support the site of the IIS log download, we need to buy with our host provider to communicate clearly whether or not to support. If supported, it is generally in the weblog file that we can get the site's log files directly from this file.
2: At the same time we need to note that the author suggested that the production of the diary we had better set it once per hour. If the site is less content of small and medium sized sites can be set to one day. If we have a lot of content, set to build once a day, then the log file may appear to be too large file.
Three: How to analyze IIS logs
1: Analysis of the disintegration of the log file
We can use Notepad to open the log file, while Ctrl search the Baiduspider and Googlebot these two search engine spiders. As shown in the following figure
Baidu Spider
Google Spider
We can analyze it in sections.
2012-04-5 00:47:10 refers to the search engine spider crawling this page point in time
116.255.109.37 This address refers to the IP address of our site
Get represents the event, located behind the spider crawling page, and/represents the homepage.
220.187.51.144 This IP refers to the search engine spider's server IP address. And this address is sometimes not the real search engine spider's IP address, then how do we distinguish it? I share one of my methods. We can open the CMD window in the computer. Enter the NSLOOKUP+IP address inside. If it is true the spider will have its own spider server. As shown in the following figure
Real spiders
Fake spider
And why will there be added spider IP? That is because someone posing as a search engine spider to your site to collect your content. And if there is excessive acquisition will consume your host resources. We can block them based on one fake IP.
200 0 0 is a status code, different status codes have different meanings, such as 200 means the download succeeded, 500 indicates the server timed out, and so on. To some extent, these reflect the situation of our host.
2: We can through the log analysis of the spider crawling the most frequent pages, record down, and analyze their search engine spiders favored reasons.
3: Sometimes our URL address is not uniform appear with the URL address with a slash and no slash problem, search engine spiders will automatically recognize the 301 jump to the page with a slash, here we found that the search engine is able to determine our catalog, so we want to unify our URLs.
4: Search engine Spiders for our Site page crawl is based on the weight of the descending graded, general order of the Home page, directory page, inside page.
The IIS log is the key to our analysis of the status of the site, is one of the key indicators of our optimization strategy. We need to analyze it often and summarize it. Reasonable use of good site resources. This article is edited by free film http://www.ynt01.com/, reprint please keep the source.
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.