Intermediary transaction SEO diagnosis Taobao guest Cloud host technology Hall
Published by: John Mueller, webmaster trend researcher
Original: Advanced Website diagnostics with Google webmaster Tools
Running a Web site is a complex business, so we provide webmaster tools to help you identify potential problems and prevent them from evolving into real problems. Some of these findings are relatively small (for example, with duplicate titles and descriptions), but there are some big problems (such as the inability to access your site). Although webmaster tools cannot tell you exactly what specific changes you need to make, it can help you uncover potential problems that need to be addressed.
Let's take a look at some of the examples we've encountered in Webmaster support forums:
Does your server treat Googlebot as a regular site visitor?
Although Googlebot tries to act like a regular user, some servers may still be confused and respond in a rather strange way. For example, when a user accesses Googlebot for user, while your server can work flawlessly for most of the time, there are some servers running IIS that return server errors (or some other actions related to server errors). In the Webmaster support forums, we have encountered 302 errors when the IIS server returned a 500 error (server error) and 404 error (File not found) appears in the "Network Crawl" diagnostics section, as well as when submitting a sitemap file. If your server is redirected to the error page, you should make sure that we can crawl the error page and make sure the server returns the correct error code. When you do this, we can display these errors in Webmaster tools. If you would like to know more about this issue and possible solutions, please refer to:
Http://todotnet.com/archive/0001/01/01/7472.aspx and http://www.kowitz.net/archive/2006/12/11/ Asp.net-2.0-mozilla-browser-detection-hole.aspx.
If your site is hosted on a server running Microsoft IIS, keep in mind that URLs are naturally case-sensitive (as we do with them). This includes the URLs that appear in the robots.txt file. If your server works in a case-insensitive manner, you should handle this section with care. For example, "Disallow:/paris" will prevent us from accessing/paris, not/paris.
Are there any systematic corrupted links to your site?
Modern content Management Systems (CMS) often lead to problems that can often affect a large number of web pages. Sometimes these questions are straightforward, and you can find them when you browse them, and sometimes it's hard to find them on your own. If a problem similar to the above results in a large number of corrupted links, they typically appear in the Network Crawl Diagnostics section of your Webmaster Tools account (listing the corrupted URLs that normally return the 404 code). In a recent case, a Web site's RSS feed produced a small coding error, resulting in more than 60,000 bad URLs appearing on the site's Webmaster Tools account. You can understand that we prefer to spend our time crawling content instead of these 404 error pages:).
Does your site redirect users to other places?
Many sites have their own focus on a specific area of the user base. We find that such sites sometimes redirect visitors from other geographies to a different page. However, keep in mind that Googlebot does not necessarily come from the destination area you expect, so it may be redirected to another page. This may mean that Googlebot cannot access your home page. If this happens, webmaster tools may have problems validating your site, causing your site to fail validation. Of course, this is not the only reason the site won't be able to pass validation, but if your site does this often, it's best to investigate why. In this case, be sure to treat googletbot in the same way as the user in that region, or you can be considered "content camouflage."
Is your server inaccessible when we crawl?
Even the best sites can do this--server downtime and firewall over protection. If this happens when we try to access your site, we can't crawl your site, and you may not know we've tried to crawl your site. Fortunately, we are tracking these issues and when we are unable to access your site, you can find the "network inaccessible" and "Robots.txt files inaccessible" errors in the Administrator tools account.
Is your website hacked?
Hackers sometimes add strange, hidden content and links away from the topic on the offending page. If it's hidden, you might not notice it right away. Still, it could be a big problem. Although the message center will warn you about some types of hidden text, you'd better pay more attention to it. "What Googlebot sees" under the Webmaster Tools statistics item displays keywords about your site's content. If you see some completely unrelated keywords there, your site is likely to be hacked and you might want to investigate what's going on as soon as possible. You can also use Google News (Google Alerts) or use [site:example.com spam keywords] for inquiries, where "junk keyword" refers to some like pornography, Viagra, tramadol (tramadol, a sedative), Sex and other words that you don't normally appear on your website. If you find that your site is really being attacked, I recommend that you refer to another post on our blog-my website is hacked Off, what should I do?
Webmaster tools can find a lot of questions about the site; these are just some of the more common problems we've encountered recently. Because there are some problems that are really hard to find, before they really become a problem, you might want to look at your Webmaster tools account to see if you can spot some problems. If you find a problem that you really can't solve, why not post on the Webmaster support forums to ask for expert help?
Did you check your site today?