Webflow sync, pageviews & more.
NEW

Has anyone experienced the issue of Googlebot being blocked by robots.txt in Webflow, even though access has been allowed in the file?

TL;DR
  • Check and update the robots.txt in Webflow's SEO settings, then re-publish the site to apply changes.
  • Use Google Search Console to verify what Googlebot sees and ensure you're testing the correct domain.
  • Avoid domain conflicts by properly setting up canonical URLs and redirects between custom and webflow.io domains.
  • Account for caching delays and re-submit for indexing in Google Search Console if necessary.
  • Confirm noindex settings aren't enabled in page-level SEO settings.
  • Ensure you're publishing changes to the correct live environment, not just staging.

Yes, some users have experienced Googlebot being blocked by the robots.txt in Webflow, even when it appears correctly configured. This issue is usually related to publishing mismatches or staged changes not being live.

1. Confirm Your Live robots.txt Content

  • Go to Project Settings → SEO tab to check the robots.txt content.
  • Make sure it includes User-agent: * and Allow: / (or your intended allow/disallow rules).
  • Re-publish your site after making any changes in the SEO tab. Webflow does not automatically republish when this file is edited.

2. Validate With Google’s Tools

  • Use Google Search Console’s robots.txt Tester to check what Googlebot is seeing live. This ensures that changes are actually live and correctly formatted.
  • Make sure to test the correct domain (e.g., staging vs. production).

3. Avoid Multiple Domain Conflicts

  • If you're using both the webflow.io subdomain and a custom domain, ensure you're only indexing one version. Googlebot might try to crawl the webflow.io domain if canonical tags or redirects aren’t properly set.
  • Webflow often includes a disallow directive for the webflow.io subdomain (Disallow: /), which is intentional.

4. Clear Cache and Crawl Delays

  • Googlebot cache might delay the detection of a newly updated robots.txt. Allow a short period before rechecking.
  • If urgent, re-submit in Google Search Console and trigger a URL Inspection → Request indexing.

5. Check for Page-Level Meta Robots Tags

  • A page can be blocked by a meta robots noindex tag or header even if it’s allowed in robots.txt.
  • In Webflow, verify in Page Settings that “Hide this page from search engines” is not enabled.

6. Verify with Webflow Hosting

  • Ensure you're checking the correct environment. Edits in Webflow’s robots.txt field apply only to the domains you publish to.
  • If you've only published to staging or didn’t publish changes to the custom domain, Googlebot will still see the old or incorrect file.

Summary

Even with a correct robots.txt in Webflow, Googlebot can be blocked due to publishing delays, caching, or multiple-domain setup. Always verify with Google’s tools and make sure you’ve published the latest SEO setting changes to the production site.

Rate this answer

Other Webflow Questions