On-Page SEO Guide for Blogger Blogs: Easy Tips for Higher Ranking

If you're managing a website, encountering the "Blocked by Robots.txt" issue in Google Search Console can be frustrating. This error prevents search engines from crawling specific pages on your site, potentially harming your site's SEO performance. In this detailed guide, we’ll explain what this issue means, why it happens, and how you can fix it step-by-step.
When Google Search Console reports that a page is blocked by robots.txt, it means that the page has been disallowed from being crawled by search engine bots. This happens due to the directives specified in your website’s robots.txt
file.
The robots.txt file is a text file that tells search engines which pages or files they can or cannot crawl. While it’s a powerful tool, improper configuration can block essential pages from being indexed by Google, affecting your site's visibility in search results.
Here are the most common reasons for this issue:
Before fixing the issue, you need to identify which pages are affected. Here’s how:
Log in to Google Search Console
Navigate to the 'Indexing' Tab
Filter for Blocked Pages
Inspect the Affected URL
The robots.txt file is located at:
https://yourwebsite.com/robots.txt
To view it:
yourwebsite.com
with your domain).User-agent: *
Disallow: /admin/
Disallow: /private/
Look for any disallow rules blocking the affected pages. For example:
Disallow: /important-page/
If this page is essential for SEO, you’ll need to remove or modify the rule.
To fix the issue:
Access Your Robots.txt File
Modify or Remove the Blocking Rule
For example, if the blocked page is /about-us/
, change:
Disallow: /about-us/
To:
Allow: /about-us/
Save and Upload the Updated File
Use Google’s Robots.txt Tester tool:
If you're using WordPress, follow these steps:
Install an SEO Plugin
Popular plugins like Yoast SEO or Rank Math allow you to edit robots.txt easily.
Edit Robots.txt via the Plugin
Save Changes
If your robots.txt file isn’t blocking the page, the issue might lie elsewhere:
Check if the page has a noindex
meta tag.
<meta name="robots" content="noindex">
If present, remove or modify it to:
<meta name="robots" content="index, follow">
Ensure your server isn’t blocking bots. Common issues include:
Contact your hosting provider to resolve these.
Allow Important Pages
Always allow pages like:
Allow: /
Block Irrelevant Pages
Block non-SEO-critical pages like admin panels:
Disallow: /wp-admin/
Use Wildcards for Efficiency
To block similar pages:
Disallow: /private/*
Regularly Audit Your Robots.txt File
Periodically review your robots.txt file for outdated or unnecessary rules.
Yes, but it’s not recommended. Without a robots.txt file, bots will crawl everything, including irrelevant pages.
Google usually re-crawls within a few days, but you can speed it up by requesting re-indexing in Search Console.
No. Some bots, like malicious scrapers, ignore robots.txt directives.
Fixing the 'Blocked by Robots.txt' issue in Google Search Console is essential to ensure your site is fully crawlable and indexable by search engines. By understanding and properly configuring your robots.txt file, you can prevent unnecessary blocks and improve your site’s SEO performance.
Regularly audit your robots.txt file and use tools like Google Search Console to monitor for errors. A well-maintained robots.txt file is a cornerstone of any successful SEO strategy.
For more SEO tips and tools, check out Google's Robots.txt Documentation.
Comments
Post a Comment
Thanks