Sitemap Tools

Find sitemaps for any domain, or validate one you already have.

Checks robots.txt for declared sitemaps, then tries common locations like /sitemap.xml, /sitemap_index.xml, and 9 others. Also checks for HTML <link rel="sitemap"> tags.

Why this matters

Sitemaps tell search engines what to crawl

If your sitemap has 404 URLs in it, broken redirects, or URLs from other domains, Google trusts your sitemap less. AI engines also use sitemaps to discover pages.

The lastmod date is the most ignored signal

Pages with accurate <lastmod> dates get re-crawled faster. If you never update lastmod, search engines stop checking back.

Common bugs we catch

  • Invalid XML (one malformed tag breaks the whole file)
  • Duplicate URLs (same URL listed twice — wastes crawl budget)
  • Cross-domain URLs (sitemap.example.com listing other-site.com URLs)
  • URLs returning 404 / 5xx (the page in your sitemap is broken)
  • URLs that redirect (sitemaps should list the FINAL URL)
  • Invalid lastmod dates (must be ISO 8601)
  • Files exceeding the 50,000-URL or 50 MB limits

For AEO

AI engines like Perplexity actively read sitemaps to discover content. A clean sitemap = better AI engine indexing. Pair this with a properly-formatted llms.txt and you're well above average.