The website IIS log is one of the very important things, because there you can check the crawling status of search engine robots, you can also learn about your own website, and you can analyze the origin of certain users. You don’t necessarily need to use traffic statistics code to do it. Calculation, but looking at the website IIS logs, some space requirements are limited, and you need to be notified to activate them. Some do not support it. You can also download some code from the Internet to install it. The website IIS log stubs also need space. If the website capacity is not very large, , it will all be exceeded all of a sudden, so you need to find a better space provider to host the website. Then let me talk about how website log analysis can help optimization?
1. Understand the number of times search engine spiders crawl
Spiders are robots sent by search engines to crawl content. Only by knowing the number of spider crawls can we know whether our website is liked by search engines. Otherwise, there is nothing to continue crawling on this website. We can compare the number of crawls and the previous four weeks. Operation comparison can tell how many spiders come and what things cause external links to the website or updates to the website, so as to adjust the web pages. Spiders like original content. If they are all directly copied and pasted reprinted content, spiders may not be able to do so next time. Will come again, I feel like this website is a mirror site of a certain website.
For harmful spiders, if we block (blocked spider IP) spiders, there are many kinds of spiders. If one day we find that an unknown spider IP website appears many times but is demoted or K, then we must ban it. Spider IP access.
A large number of spider visits can increase the server's resources. Frequent spider visits are still helpful to the website, but the resource consumption is also very large, so you need to find a good space provider to put it, otherwise the server will collapse all of a sudden, and it may not be there anymore. Notify you that your website has been deleted.
Baidu Club: Looking at the identification of fake spiders (in order to steal other data), the most important thing is that Baidu spiders always display IPs from Beijing. If they are from other places, they are not real spiders, and they may also be demoted and K-ed. Spider, check carefully.
2. Website page crawling degree
If you like to crawl the home page, the snapshots will be taken every other day. If the internal pages are frequent, they will be captured in seconds. If a certain part of the page is not crawled, then check whether spider crawling is disabled and external links are made. Sometimes we can’t just do most of the home pages, but also do the inner pages properly, otherwise the inclusion will be very low, and we can also know what problems our website has, which pages the spiders mainly crawl or which pages are crawled more frequently. We need to combine and analyze which pages are not crawled, and compare which directory has more and less content (ask the space provider to provide IIS logs). We also need to check the changes and the crawling situation of search engines in different periods. It's different, whether it's because of the reprint or the external link.
3. Analysis of http status code
Spiders generally leave an http status code after crawling. The return of 200 does not mean that it will be released directly. Some are released after a week of updates, and some are released after a month. As long as this code is returned, the page will generally be released. .
Two questions:
1. Should our error page return 404 or 200?
It should return 404 four correct, because only such search engines know that the webpage cannot be accessed. If it is 200, it may cause the situation of being demoted or K, which means that the webpage can still be crawled. Once a large number of All of them cannot be accessed, and they will be punished, so you must prepare a 404 page.
2. If our website needs to be filed and we are under construction, which status code should we return: 500, 400, 404, 500, or 503?
To return 503 (503 tells the search engine that it is temporarily unavailable and will be restored soon. If other status codes are returned, the search engine may not access it. The 404 page means that this page no longer exists, and the search engine will think that you The website no longer exists, so it will be deleted directly. For the time being, it will continue to be crawled next time.)
The website does a 301 permanent redirect but does not return the code. Then you have to check whether the settings are correct, otherwise the weight will not be transferred to the new domain name. Everything must be done to be foolproof.
4. Professional log analysis tools
Allows us to know the pv access value and offensive access value
PV is a reflection of retaining users. If the bounce rate is too high, then the website cannot be opened or the content is not readable. The website cannot escape the fate of being ranked down. You can also check which pages are the most visited. In this way You can find out the needs of users and improve them. If the website cannot be opened for a long time or is opened very slowly, we have to check the logs to see if there are a large number of visits from unknown IPs. It may be that it has been attacked, so we can only compromise or report the case, otherwise we will change to another server, but Changing the soup but not the medicine will still not solve the problem. It is best to take legal measures to protect your own interests.
(The quality of a website is directly determined by the user’s clicks)
If users don’t click on your website, it means that your website is not convincing or attractive, and it is not a good website. Except for marginal and illegal websites, your product descriptions are not detailed enough, the pictures are not clear enough, and the customer service is not strong enough. , then who will continue to stay on your website? To do a good job in user experience is a homework.
Log analysis tools:
(1).awstats、
(2).Webalizer
It can also analyze the status code of the website
Website IIS logs are still very helpful for optimization. Don’t ignore any details. IIS logs can not only know whether your website is helpful to users, but also let search engines know whether it is suitable for their liking. It can also know some conditions of the website. The signs of being demoted and being raped can be found in the IIS log code of the website. Please indicate the source of www.bole110.com when reprinting. Thank you for your cooperation!
Editor in charge: Chen Long Author wangruolun's personal space