Sitemap Generator for Search & AI Crawlers

Enter your domain. We crawl up to 500 pages, skip noindex content, assign depth-based priority, check AI crawler access, and build a clean XML sitemap ready for Google Search Console.

Enter your domain

We crawl up to 500 pages, skip noindex content, check AI crawler access, and build a clean XML sitemap with depth-based priority.

Try an example

Crawls up to 500 pages

We discover pages by following internal links from your homepage.

Depth-based priority

Homepage gets 1.0, top pages 0.8, deeper pages 0.6-0.5 automatically.

Smart filtering

Automatically skips noindex pages, non-HTML content, and asset files.

AI crawler check

Shows which AI bots can access your content based on robots.txt.

How it works

Get a sitemap in three steps

Enter your domain

Type your website URL. We start crawling from your homepage, checking robots.txt for AI crawler rules and sitemap declarations.

We crawl and analyze

Our crawler discovers up to 500 pages, follows internal links, skips noindex content, and assigns priority based on URL depth.

Download and submit

Get a standards-compliant sitemap.xml with one click. Upload to your server, declare in robots.txt, and submit to Search Console.

Common questions

Frequently asked questions

What is an XML sitemap?

An XML sitemap is a structured file that lists all the important URLs on your website. It helps search engines like Google and AI crawlers like GPTBot discover your pages more efficiently than relying on link-following alone. Think of it as a directory of your site's content.

How many pages can this tool crawl?

This tool crawls up to 500 pages from your domain, following internal links from the homepage. It automatically skips noindex pages, non-HTML content, and asset files. For sites with more than 500 pages, consider your CMS's built-in sitemap generator or a dedicated crawl tool.

How does depth-based priority work?

The generator assigns priority values based on URL depth: the homepage gets 1.0, top-level pages (/about, /pricing) get 0.8, second-level pages (/blog/post) get 0.6, and deeper pages get 0.5. This helps crawlers understand your site hierarchy and prioritize important pages.

Where do I put the sitemap.xml file?

Place it in the root directory of your website (e.g., yoursite.com/sitemap.xml). Then declare its location in your robots.txt file with: Sitemap: https://yoursite.com/sitemap.xml. Finally, submit it through Google Search Console for fastest indexation.

Does this tool check AI crawler access?

Yes. We analyze your robots.txt to show which AI crawlers (GPTBot, ClaudeBot, PerplexityBot, etc.) can access your site. If AI crawlers are blocked, your content won't appear in AI-generated answers — a growing discovery channel you may be missing.

Will this tool find all my pages?

It discovers pages by following links from your homepage. Pages that aren't linked from anywhere (orphan pages) won't be found. For comprehensive coverage, ensure all important pages are linked from your navigation or content. After generating, use our Sitemap Validator to check coverage.

Does the generated sitemap include images or videos?

This version generates a standard XML sitemap with URL locations, lastmod dates, changefreq, and priority. Image sitemaps (using the image:image namespace) and video sitemaps require additional metadata that this tool doesn't currently extract.

How often should I regenerate my sitemap?

Regenerate whenever you add, remove, or significantly update pages. For most sites, the best approach is to use your CMS's automatic sitemap generation, which updates in real-time. This tool is most useful for sites without built-in sitemap support or as a quick audit of what a crawler can find on your site.

Next step

Already have a sitemap? Audit it.

Check for crawl budget waste, AI crawler blocks, lastmod issues, and URL consistency problems.