1. robots.txt & meta robots
Most websites now use CMS, and content that does not need to be indexed must exist, so robots.txt is extremely important. In addition, content that cannot be fully controlled by robots.txt is best blocked by meta robots from search engine indexing. Don’t think that the more included, the better. The more irrelevant content a search engine includes from a website, the greater the possibility that the weight of relevant content will be dispersed.
2. 301 redirection and domain name unification
Content that needs to be redirected must use 301 redirection. 302 is relatively sensitive to Google and can easily be judged as cheating. Most page hijacking is carried out through 302. Regarding the issue of domain name unification, in addition to redirecting www.xxx.com to xxx.com, we should also ensure that www.xxx.com/a.html can be successfully redirected to xxx.com/a.html. This problem is mainly for Baidu, and Google can solve it in disguise through webmaster tools.
3. Relative paths and management entries
As mentioned earlier, most websites use cms. It is best to use absolute paths for some links on the CMS template, otherwise there will be many more 404s as the paths change; many CMS have reserved administrator login entrances, which will be crawled by search engines if this entrance is left on the page. , even if it is blocked from indexing, search engines such as Yahoo and Baidu will occasionally include it in the scope of inclusion. Therefore, it is best to ensure the isolation of background related links, that is, to delete the entrance to the background from the page. At the same time, reserving the administrator link is also a security risk and can easily be exploited by people with ulterior motives.
Source: http://yeeseo.com/html/3-neglected-seo-details.html