Sitemap Tools
Find sitemaps for any domain, or validate one you already have.
Checks robots.txt for declared sitemaps, then tries common locations like /sitemap.xml, /sitemap_index.xml, and 9 others. Also checks for HTML <link rel="sitemap"> tags.
Validates XML structure, lastmod dates, duplicates, cross-domain URLs, and spot-checks the first 25 URLs return 200.
Sitemaps found for
Statistics
Sampled URL status
Child sitemaps (sitemap-index)
Why this matters
Sitemaps tell search engines what to crawl
If your sitemap has 404 URLs in it, broken redirects, or URLs from other domains, Google trusts your sitemap less. AI engines also use sitemaps to discover pages.
The lastmod date is the most ignored signal
Pages with accurate <lastmod> dates get re-crawled faster. If you never update lastmod, search engines stop checking back.
Common bugs we catch
- Invalid XML (one malformed tag breaks the whole file)
- Duplicate URLs (same URL listed twice — wastes crawl budget)
- Cross-domain URLs (sitemap.example.com listing other-site.com URLs)
- URLs returning 404 / 5xx (the page in your sitemap is broken)
- URLs that redirect (sitemaps should list the FINAL URL)
- Invalid lastmod dates (must be ISO 8601)
- Files exceeding the 50,000-URL or 50 MB limits
For AEO
AI engines like Perplexity actively read sitemaps to discover content. A clean sitemap = better AI engine indexing. Pair this with a properly-formatted llms.txt and you're well above average.