SEO 工具

Robots.txt生成器

为网络爬虫生成带有自定义允许和禁止规则的robots.txt文件

Global Parameters

Rule Group #1

Source
# Generated by freetool24.com

User-agent: *
Disallow: /admin/
Disallow: /private/
Allow: /

Sitemap: https://example.com/sitemap.xml
Crawler Guide

User-agent: *
Applies the rule to all bots and web crawlers globally.

Disallow: /
Prevents the targeted bot from crawling any page on the site.

Free Robots.txt Generator

A robots.txt file is a plain-text file at the root of your website that instructs crawlers — Googlebot, Bingbot, and others — which pages to crawl or skip. This robots.txt generator lets you build valid Allow, Disallow, User-agent, and Sitemap directives visually, then copy the file instantly.

How to create a robots.txt file

  1. 1

    Set your Sitemap URL

    Enter your sitemap address, such as https://yoursite.com/sitemap.xml. Search engines use it to discover your most important URLs.

  2. 2

    Choose a User-agent

    Use * to target all crawlers, or name a specific bot such as Googlebot, Bingbot, Googlebot-Image, or GPTBot.

  3. 3

    Add Disallow rules

    List the paths you want crawlers to skip, such as /admin/, /private/, /cart/, or internal search results.

  4. 4

    Add Allow rules

    Use Allow to reopen a specific sub-path inside a blocked folder, such as allowing /admin/help/ while blocking /admin/.

  5. 5

    Copy and deploy

    Copy the generated robots.txt file and upload it to the root of your domain: https://yourdomain.com/robots.txt.

Common robots.txt templates

Allow all crawlers

User-agent: * Disallow: — gives search engines full crawl access.

Block a private folder

User-agent: * Disallow: /admin/ — keeps admin or staging sections out of crawler paths.

Add sitemap discovery

Sitemap: https://example.com/sitemap.xml — points crawlers to your canonical sitemap.

Target one bot

User-agent: Googlebot Disallow: /tmp/ — applies the rule only to Googlebot.

Common mistakes to avoid

Blocking the whole site

Disallow: / blocks crawling for the selected bot. Use it only for private or staging sites.

Using robots.txt for secrets

Blocked URLs are still visible in the file. Do not put private paths or sensitive endpoints there as a security measure.

Forgetting the root location

The file must live at /robots.txt on the host it controls. A robots file in a subfolder is ignored.

Confusing crawl and index

Robots.txt controls crawling. Use noindex on accessible pages when you need to prevent indexing.

FAQ

Does robots.txt prevent indexing?

Robots.txt controls crawling, not indexing. A URL can still appear in search results if Google discovers it from links. Use a noindex meta tag to prevent indexing.

Where does robots.txt go?

At the root of your domain — https://yourdomain.com/robots.txt. Subdirectory placement does not work.

Should I include my sitemap in robots.txt?

Yes. A Sitemap directive helps search engines discover your canonical sitemap URL quickly.

Is robots.txt case-sensitive?

Yes. Disallow: /Admin/ and Disallow: /admin/ are treated as different paths by most crawlers.

没找到需要的?

我们根据社区反馈开发免费工具。欢迎提出能改善您工作流的工具建议!

Robots.txt生成器 — 免费工具在线 | FreeTool24 | FreeTool24