🤖 SEO Tools

Robots.txt Generator

Create a perfect robots.txt file for your website in seconds. Control crawlers, block bad bots, add your sitemap — then download and upload instantly.

Quick presets
Custom rule builder
Bad bot blocker
Live validation
Instant download
⚡ Quick Presets
Default
Allow all bots
🚫
Block All
Block every crawler
🔍
Google Only
Allow only Googlebot
🟦
WordPress
Standard WP setup
🛒
Shopify
E-commerce setup
🤖
Block AI Bots
Block AI scrapers
🔧 Rule Builder
⚙️ Global Settings
Block bad bots
Block AI training bots
Add comments

What Is a robots.txt File?

A robots.txt file is a plain text file placed at the root of your website — accessible at https://yoursite.com/robots.txt — that instructs search engine crawlers and bots which pages or sections of your site they are allowed or not allowed to visit. It follows the Robots Exclusion Protocol (REP), an industry-standard used by Google, Bing, Yahoo, and thousands of other crawlers since 1994. Every time a search engine bot visits your website for the first time, the very first file it requests is your robots.txt. Without one — or with a misconfigured one — you risk wasting Google's crawl budget on unimportant pages, accidentally exposing sensitive admin areas, or even blocking your entire site from being indexed.

Why Does Your Website Need a robots.txt?

A properly configured robots.txt is one of the foundational elements of technical SEO. Here is why every website needs one:

  • Protects crawl budget — Google allocates a limited crawl budget per site. Wasting it on login pages, tag archives, or search result pages means important content gets crawled less often.
  • Hides sensitive areas — Admin panels, staging paths, internal APIs, and private directories should never be indexed.
  • Blocks bad bots — Scrapers, spam bots, and AI training crawlers consume server bandwidth without benefiting your site.
  • Points to your sitemap — Including your Sitemap: directive helps Google discover all your pages faster.
  • Prevents duplicate content — Blocking URL parameters and faceted navigation prevents Google from crawling thousands of duplicate versions of the same page.
📍

Where Does robots.txt Live & How to Upload It?

Your robots.txt file must live in the root directory of your website — the same folder that contains your homepage. It is only valid at the top-level domain. Key rules:

  • Correct location: https://yoursite.com/robots.txt
  • Wrong location: https://yoursite.com/blog/robots.txt — crawlers won't find it here
  • Upload via FTP/SFTP using FileZilla or cPanel File Manager
  • WordPress users: Use Yoast SEO or Rank Math to manage it from the dashboard without FTP
  • Shopify users: robots.txt is auto-generated but can be customized via Liquid templates
  • Verify it works by visiting yoursite.com/robots.txt in a browser after upload
⚠️

Common robots.txt Mistakes That Hurt SEO

A misconfigured robots.txt can cause serious SEO damage. These are the most common mistakes:

  • Blocking CSS and JS files — Google needs to render these to understand your page layout. Never disallow /wp-content/ entirely.
  • Disallowing your entire siteDisallow: / for all bots means Google cannot index anything.
  • Using robots.txt to hide sensitive data — Disallowed URLs are still visible in robots.txt itself. Use server authentication for truly private content.
  • Forgetting the sitemap directive — Always add Sitemap: to help Google discover your pages.
  • Wrong path syntax — Paths are case-sensitive. /Admin/ and /admin/ are treated differently.
🆚

robots.txt vs Noindex — What's the Difference?

These two are often confused but serve very different purposes:

  • robots.txt Disallow — Tells crawlers not to visit the URL. The page is never crawled but Google may still know it exists from links. It cannot read the noindex tag if it can't crawl the page.
  • Noindex meta tag — Allows crawling but tells Google not to index the page in search results. The crawler visits the page, reads the tag, and excludes it from its index.
  • Use robots.txt to block admin areas, internal tools, low-quality pages you don't want wasting crawl budget.
  • Use Noindex for pages you want crawled but not shown in search results, like thank-you pages, filtered category pages, or paginated archives.

📖 robots.txt Syntax Reference

# Comment — ignored by crawlers User-agent: * # Applies to ALL bots Disallow: /admin/ # Block /admin/ from all bots Disallow: /private/ # Block /private/ from all bots Allow: /admin/public/ # Allow a sub-path within a blocked path Crawl-delay: 2 # Wait 2 seconds between requests User-agent: Googlebot # Applies only to Googlebot Allow: / # Allow everything for Googlebot Sitemap: https://yoursite.com/sitemap.xml

How to Use This robots.txt Generator

Step 1

Pick a Preset

Select a preset that matches your website platform — Default, WordPress, Shopify, Google Only, or Block AI Bots — to instantly populate the rule builder.

Step 2

Customize Rules

Add or remove Allow/Disallow rules for each User-agent. Use the rule builder to target specific bots and paths with precision.

Step 3

Add Sitemap URL

Enter your sitemap URL (e.g. https://yoursite.com/sitemap.xml) so Google can discover all your pages directly from robots.txt.

Step 4

Generate & Validate

Click Generate to create your file. The live validator checks for common errors and warnings before you download.

Step 5

Download & Upload

Download the robots.txt file and upload it to the root of your website via FTP, cPanel, or your CMS dashboard.

Step 6

Test in Google Search Console

Use the robots.txt Tester in Google Search Console to verify your rules work correctly before they go live.

What to Block & What to Allow — By Website Type

Website TypeBlock These PathsAlways Allow
WordPress Blog /wp-admin/, /?s=, /tag/, /page/ /wp-admin/admin-ajax.php, /wp-content/
E-commerce (Shopify) /cart, /checkout, /account, /orders All product & category pages
SaaS / Web App /dashboard/, /api/, /settings/, /login Marketing & landing pages
News / Media /print/, /amp/ (if duplicate), /search All article & category pages
Portfolio / Brochure /staging/, /admin/, /cgi-bin/ All public-facing pages

✅ Why Use WebTigers robots.txt Generator?

Instant & Visual

See your robots.txt file built in real-time as you configure rules. No manual editing required.

Built-in Validator

Live validation catches common mistakes before you upload — like accidentally blocking your whole site.

🤖

Bad Bot Blocker

One click adds disallow rules for 30+ known scraper bots, spam crawlers, and AI training bots.

🏗️

Platform Presets

WordPress, Shopify, and other CMS-specific presets get you the right configuration instantly.

🔒

100% Private

Everything runs in your browser. Your site URL and rules are never sent to any server.

🆓

Free Forever

No account, no watermark, no limits. Generate and download as many robots.txt files as you need.

Frequently Asked Questions About robots.txt

What is a robots.txt file and why do I need it? +
A robots.txt file is a plain text file at your website root that tells search engine crawlers which pages to crawl and which to skip. You need it to protect crawl budget, keep sensitive pages out of search results, block bad bots, and point search engines to your sitemap. Without it, Google may waste time crawling admin pages instead of your important content.
Does robots.txt affect Google rankings? +
Yes, indirectly. A correct robots.txt helps Google allocate its crawl budget efficiently — directing it to your important pages. Blocking unnecessary pages like admin areas, duplicate content, and filter pages means Google spends more time crawling your valuable content. However, accidentally blocking important pages with Disallow: / can completely prevent them from appearing in search results.
Where do I upload my robots.txt file? +
Upload your robots.txt to the root directory of your website — the same folder that contains your index.html or homepage file. It must be accessible at exactly https://yourdomain.com/robots.txt. You can upload via FTP using FileZilla, via your hosting cPanel File Manager, or through your CMS dashboard (Yoast SEO for WordPress, etc.). After uploading, visit the URL in a browser to confirm it is accessible.
What is the difference between Disallow and Noindex? +
Disallow in robots.txt tells crawlers not to visit a URL at all — they will not crawl it. Noindex is a meta tag placed on the page itself that tells crawlers not to include the page in search results even if they do crawl it. An important implication: if you Disallow a URL, Google cannot read its noindex tag — so Google may still list the URL in search results based on links pointing to it. For pages you want excluded from search, use noindex without blocking in robots.txt.
Can I block specific bots like GPTBot or CCBot? +
Yes. You can add specific User-agent rules for known AI and scraper bots. Common ones include GPTBot (OpenAI), CCBot (Common Crawl), Google-Extended (Google AI training), anthropic-ai (Anthropic/Claude), and Bytespider (TikTok). Use our Bad Bot Blocker or Block AI Bots preset to add disallow rules for all of them automatically with one click.
Does robots.txt block hackers or hide sensitive data? +
No — and this is a common dangerous misconception. robots.txt is a public file that anyone can read by visiting yoursite.com/robots.txt. Ironically, listing sensitive paths in robots.txt can reveal them to hackers looking for admin panels or private directories. To actually protect sensitive content, use server-level authentication (password protection), firewalls, or access control rules — not robots.txt.
How do I test if my robots.txt is working correctly? +
The most reliable way is to use the robots.txt Tester in Google Search Console (under Settings → robots.txt). It lets you enter a URL and see whether Googlebot can access it under your current rules. You can also use online validators like search.google.com/search-console/robots-testing-tool. After any change, always test a few of your most important URLs to make sure they are not accidentally blocked.