Baidu inclusion is always a key topic discussed by SEOER. In the website optimization war, stable inclusion directly affects the website's traffic and income. Why is the snapshot of the homepage of the website updated very quickly, but the internal pages are not included? I believe many friends are looking for answers to the same questions as me. This article will discuss the reasons why the internal pages of the website are not included in Baidu and the countermeasures.
Reason 1: Baidu has indexed but has not yet released the ranking. Baidu's inclusion of articles requires review time. The spider crawls the successfully indexed content and enters it into the database. It needs to be reviewed twice to see if it is SPAM before it is released to participate in the ranking. The release time is directly related to the weight of the website. Baidu updates it periodically before releasing it for inclusion. This is the most common phenomenon for new sites.
Countermeasures: What we have to do is to ensure regular and quantitative updates and observe website logs. As long as the spider crawls normally, wait patiently and there will be surprises.
Reason 2: Robots setting problem causes internal pages not to be crawled and indexed by engine spiders
Countermeasures: Check whether grammatical errors block spider crawling, including the settings of tags such as meta robots and nofollow.
Reason 3: The originality of the article is a problem. The setting of the header tag causes similar duplication of pages, or there are too many collected pages. Since there is already a large amount of similar information in the Baidu database, it is not included. Long-term operation will cause Baidu to lose authority.
Countermeasures: The originality of the article should be maintained (I believe all friends are familiar with pseudo-original articles). Quality is more important than quantity. It is best to update it regularly every day so that Baidu will have a good habit of being punctual in crawling your website.
Reason 4: There are problems with the internal links of the website. For example, there are a large number of wrong links, bad website links, spiders cannot crawl smoothly, and bad URL links cause the website to lose authority.
Countermeasures: Check for incorrect links on various important pages such as homepage and channels (you can observe the crawling errors of Google Webmaster Tools), check bad links in message boards, comments and other netizen interactive functions to avoid interlinking with bad websites.
Reason 5: The website used black hat methods. Baidu identified the deceptive methods and demoted the entire site.
Countermeasures: Stay away from black hat behaviors such as keyword accumulation, mass posting, and black links. The long-term stability of the website is less complicated. Once Baidu takes action, it will not be difficult to recover!
Reason 6: The server is inaccessible, slow, hung, etc. When Baidu Spider indexed content, it was repeatedly identified that the website had a large amount of garbled code or the server was inaccessible, and the website was included in the watch list.
Countermeasures: Optimize website code to improve page reading speed, choose a responsible safe space service provider, mature website programs and regularly update patches, and regularly track traces of modified files.
In addition to doing the above 6 points in a down-to-earth manner, Luotuo specifically proposed a few points worth doing: 1. Make a site map of the entire site on the homepage to allow spiders to crawl to the inner pages smoothly; 2. The relevant link calls on the inner pages must be regular. , to make the theme of each page more prominent; 3. External links can be appropriately used to guide spiders on internal pages. Grasp every detail so that you can rest assured and believe that Baidu will not stay away from including our internal pages.
Author: Luo Tuo, blog ( www.seo173.com ) insists on the road of SEO, welcome to communicate!
Thanks to Luo Tuo for your contribution