Sitemap Analyzer
Discover any site's sitemap via robots.txt, extract every URL with lastmod and priority, and pull per-page metadata for research.
What is an XML Sitemap?
An XML sitemap is a structured file that lists every URL on a website along with optional metadata: when each page was last modified, how often it changes, and a relative priority. The format follows the sitemaps.org 0.9 protocol — a single
<urlset> with up to 50,000 URLs and a 50 MB ceiling, optionally split across multiple files via a <sitemapindex>. Sites declare their sitemap location in /robots.txt with a Sitemap:directive (RFC 9309 makes this directive global, not user-agent specific) — Google, Bing, and other crawlers read it to discover URLs faster than they would by following links alone. Sitemaps are also a research goldmine: a site's sitemap is a one-glance map of every page it considers indexable, which makes it the fastest way to explore a competitor's structure or audit your own coverage. If you need to create a sitemap, use our generator instead. For a deep on-page audit of any URL discovered here, run the Website SEO Checker.How to use this tool
- 1Paste any website URLEnter the homepage or any URL on the target domain. We derive the origin and look up its robots.txt and sitemap files server-side.
- 2Review the discovery panelSee which sitemaps were declared in robots.txt, which were fetched, and how each was classified — URL set or sitemap index.
- 3Filter and sort URLsUse the URL filter, changefreq selector, and lastmod date range to narrow down the table. Click any column header to sort.
- 4Extract metadata per rowClick Extract on a URL to fetch its title, meta description, og/twitter cards, canonical, x-robots-tag, and word count. Cached after the first click.
- 5Export to CSVUse the Export menu to download the full URL list, or URL list plus the metadata you've extracted. Both formats open in Excel and Google Sheets.
