How To Fix Google Search Console Indexing Issues
Most issues site owners have is that their sites cannot be crawled by google or that it can't be Indexed. This is categorized as index issues.
FIXING INDEXING ISSUES ON SEARCH CONSOLE
Below are the most common indexing errors and some tips on how you can solve them:
- Submitted URL marked ‘noindex’ issue can be solved by removing the noindex tag from the page's HTML code, or deleting a noindex header from the HTTP request.
- Submitted URL not found (404) means that the page does not exist, and the server redirected to 404 status code. Check if the content was relocated, and set up a 301 redirect to a new location.
- Submitted URL seems to be a Soft 404. This error appears when your server labeled the page with OK status, but Google decided the page is 404 (not found). This may occur because there’s little content on the page, or because the page moved to a new location. Check if the page has good comprehensive content and add some if it’s thin. Or set up a 301 redirect if the content was moved.
- Submitted URL blocked by robots.txt error can be solved by running the robots.txt tester tool on the URL, and updating the robots.txt file on your website to change or delete the rule.
- Submitted URL returns unauthorized request (401) means that Google cannot access your page without verification. You can either remove authorization requirements or let Googlebot access the page by verifying identity.
- Submitted URL returned 403 error happens when Google has no credentials to perform authorized access. If you want to get this page indexed, allow anonymous access.
Valid with warning
Indexed, though blocked by robots.txt warning appears when the page is indexed by Google despite being blocked by your robots.txt file. How to fix this issue? Decide on whether you want to block this page or not. If you want to block it, then add the noindex tag to the page, limit access to the page by login request, or remove the page by going to Index > Removals > New request.
Note: many SEOs mistakenly assume that robots.txt is the right mechanism to hide the page from Google. This is not true — robots.txt serves mainly to prevent overloading your website with requests. If you block the page with robots.txt, Google will still display it in search results.
Indexed without content means the page is indexed, but, for some reason, Google cannot read the content. This may happen because the page is cloaked, or the format of the page is not recognized by Google. To fix this issue, check the code of your page and follow Google’s tips on how to make your website accessible for users and search engines.