For an SEOr, the server IIS log is a very important optimization reference log, because we can see the status of search engine spider crawling from here, and we can also understand some conditions of the website itself, and we can also analyze the behavior of some users. It is not necessary to use some third-party code for statistics. Of course, some IIS logs are restricted by IDC space providers and must be opened before they can be viewed. Otherwise, they cannot be viewed. If this is the case, it is recommended that you download them online. Some source code to install. Next, the author will explain in detail how server IIS can help the website!
(1) Check the number of times the spider crawls.
A search engine spider is a robot used to crawl the content of our site and submit the crawled content to the search engine database. Therefore, we know the number of times the spider has crawled and indirectly know whether our site has been searched. Engine favor. In addition, by comparing the origins of spiders, we can fully understand which external links are more useful to our site and which spiders value more. We can also understand which pages of our site content spiders prefer and always imitate them. .
For those harmful search engine spiders, we need to block them, because after all, there are many types of spiders. If one day we find many unknown spiders crawling on our website in large numbers, then we should block them. In order to avoid losing authority, the author also calls on everyone to try not to use software that attracts spiders to crawl their own websites.
A large number of search engine spiders crawl our website, which undoubtedly takes up a certain amount of resources on our server. However, a lot of spider crawling is very helpful to the weight of our site, so I suggest that everyone must choose a better server. , to avoid collapse at critical moments. In addition, we learned from some official announcements that many hackers use spider crawling mode to steal website data resources, so we also need to pay attention to these!
(2) Check the extent to which the spider crawls the page.
Generally speaking, what spiders value most is our website homepage, so homepage snapshots are generally updated most frequently. And if the internal pages are also frequent, it will achieve the effect of what we often call instant collection. If some of our pages are not crawled by spiders, we can check through the ISS log whether we have banned spider crawling. . In addition, the author understands that many webmasters usually use the homepage address in the process of making external links. Here, the author calls on everyone to also make more external links to columns and article pages. This will be of great help to our inclusion. And from these we can also understand some conditions of our site, such as which pages the spider enters from, which pages are crawled relatively frequently, which pages are crawled once but are not crawled again or included, so we summarize After that, you can more accurately understand what content spiders prefer, and you can also see whether spiders are interested in our content or are crawling because of external links.
(3) Analysis of website http code.
When a spider crawls and crawls the content of our website, it usually leaves an http status code. When it returns, it usually displays 200, which means it may not be released directly, which means a review period is required.
The following are two self-questions and answers that the author has made about this area. I hope it will be helpful to all webmasters.
1. When the page does not exist on the website, should it return 404 or 200?
First of all, the answer is definitely to return a 404 value, because everyone knows that the 404 page can tell the search engine that this page is an error page and cannot be accessed, but if it is 200, it is different, because it means that this page can be accessed However, when the spider crawls it, it is found to be inaccessible. If there are too many such pages, it will directly cause our site to be demoted by search engines or even K.
2. When the website is under construction or filing, which status code should be returned?
The answer is 503 status, because 503 can tell search engines that our site is temporarily inaccessible and will be restored within a certain period of time. If it is other status codes, search engines may not access it, especially 404 page, which will cause search engines to directly think that the website no longer exists.
The 404 page is actually very helpful to our site, so I suggest you remember to make a 404 page for your site.
(4) Use some professional tools for log analysis
PV value is undoubtedly an expression of user experience for a website. When the bounce rate of our site is too high, then the website is in a state that cannot be opened or the content is poor, then the website ranking is self-evident. , it is impossible to see which pages have higher traffic, so that we can understand the needs of users through exploration and make improvements. But if our website cannot be opened for a long time, or the access is slow, then we can check from the logs whether it is attacked by some malicious traffic. Faced with such a situation, we must either compromise, report the case, or directly change the website. server.
For a website, if users don’t click on your site, it means that your site is not attractive, and search engines will also think that your site is not a good website, so it is very important to do a good user experience studio. . Finally, I recommend two more professional IIS log analysis tools:
1. .awstats,
2. Webalizer
These two tools can also analyze some status codes (⊙o⊙) of the website!
Summary: The IIS log of a website is very helpful in all aspects of website optimization and user experience. It allows us to learn more about our site and pay attention to some small details that we have overlooked. In this way, our site will naturally be favored by search engines, and we can submit it to get some corresponding understanding before our site is K or demoted. At this time, we can make targeted improvements, which can often be achieved. Avoided. This article was originally shared by Name Network http://www.name2012.com . I hope that friends who reprint it will remember to retain the link and copyright. I would like to express my gratitude to you. Well, that’s all for sharing with you today. I will communicate with you more on this platform in the future, so see you next time.
Editor in charge: Chen Long Author's personal space at name2012.com