Free Sitemap Finder and Checker
Discover and analyze all sitemaps on any website. Check robots.txt declarations, count URLs, and verify sitemap health — completely free.
Why Use Our Sitemap Checker?
Quickly discover and analyze all sitemaps on any website. Perfect for SEO audits, site migrations, and technical SEO analysis.
Auto-Discovery
Automatically finds sitemaps by checking robots.txt and common paths like /sitemap.xml and /wp-sitemap.xml.
Index Support
Handles sitemap index files recursively, discovering all nested sitemaps within your site structure.
Robots.txt Check
Verifies if your sitemaps are properly declared in robots.txt for optimal search engine crawling.
URL Counting
Counts total URLs across all sitemaps to help you understand your site's indexed content scope.
Last Modified Dates
Shows when each sitemap was last updated, helping identify stale or outdated sitemaps.
Export Reports
Download detailed sitemap reports or copy URLs directly to clipboard for further analysis.
The Complete Guide to XML Sitemaps
What is an XML Sitemap?
An XML sitemap is a file that lists all the important URLs on your website that you want search engines to crawl and index. Think of it as a roadmap that helps search engine bots like Googlebot discover and understand the structure of your site more efficiently.
While search engines can discover pages through links, a sitemap provides additional metadata like when a page was last modified, how often it changes, and its relative priority compared to other pages. This helps search engines crawl your site more intelligently and keep their index up to date.
Sitemaps are especially important for large sites, new sites with few external links, sites with rich media content, or sites with pages that aren't well linked internally.
Types of Sitemaps
XML Sitemap (urlset)
The standard sitemap format containing a list of URLs with optional metadata like lastmod, changefreq, and priority. Limited to 50,000 URLs and 50MB uncompressed.
Sitemap Index (sitemapindex)
A master file that references multiple sitemap files. Essential for large sites that exceed the 50,000 URL limit. Can reference up to 50,000 individual sitemaps.
Image Sitemap
Extended format that includes image information within URL entries. Helps Google discover images that might not be found through regular crawling, especially those loaded via JavaScript.
Video Sitemap
Contains video-specific information like title, description, duration, and thumbnail URL. Helps videos appear in Google Video search and rich results.
News Sitemap
Required for Google News publishers. Contains article-specific data like publication date, title, and keywords. Only includes content from the past 48 hours.
Declaring Sitemaps in Robots.txt
The recommended way to help search engines find your sitemaps is to declare them in your robots.txt file. This file is one of the first things search engine bots look for when crawling a site.
User-agent: *
Allow: /
Sitemap: https://example.com/sitemap.xml
Sitemap: https://example.com/sitemap-posts.xmlYou can declare multiple sitemaps in robots.txt. This is especially useful if you have separate sitemaps for different content types like posts, pages, products, or media. Our checker specifically looks for these declarations first.
Common Sitemap Issues
- Not in robots.txt: If sitemaps aren't declared, search engines rely on common paths. Always declare in robots.txt for reliable discovery.
- Stale content: Outdated lastmod dates or including noindexed/redirected URLs wastes crawl budget and sends mixed signals.
- Size limits exceeded: Sitemaps over 50MB or 50,000 URLs won't be processed. Use sitemap index files instead.
- XML errors: Malformed XML prevents parsing. Validate your sitemaps after any changes to your sitemap generation logic.
- Wrong URLs: Including URLs that return 404, are blocked by robots.txt, or are non-canonical dilutes sitemap effectiveness.
Sitemap Best Practices
- Keep it current
Update lastmod dates only when content significantly changes. Automatic daily updates for all pages are counterproductive.
- Include canonical URLs only
Only include URLs you want indexed. Skip paginated pages, filtered views, and URLs with noindex tags.
- Use consistent URLs
Match the URL format used in your canonical tags. Don't mix www/non-www or http/https versions.
- Submit to search consoles
While robots.txt helps discovery, submitting directly to Google Search Console and Bing Webmaster Tools ensures processing.
- Segment large sites
Use separate sitemaps for posts, pages, products, and media. This helps with debugging and monitoring.
Quick Tips
- •Declare sitemaps in robots.txt
- •Keep lastmod dates accurate
- •Only include indexable URLs
- •Use sitemap index for large sites
- •Submit to search consoles
- •Validate XML regularly
Sitemap Limits
50,000 per sitemap
50MB uncompressed
50,000 sitemaps per index
Did You Know?
Google processes billions of URLs from sitemaps daily. Sites with accurate, well-structured sitemaps see up to 50% faster indexing of new and updated content compared to those without.
Check Sitemaps in 3 Simple Steps
From URL to complete sitemap analysis in seconds. Discover all sitemaps and verify they're properly configured.
Enter Website URL
- Paste any website URL or domain
- Works with any website, CMS, or platform
Discover Sitemaps
- Tool checks robots.txt and common paths
- Parses sitemap indexes recursively
Analyze & Export
- View URL counts and last modified dates
- Copy or download sitemap report
Explore Free Tools
View AllKeyword Research
Discover high-value keywords and search volumes for your content strategy...
Page Speed Analyzer
Analyze your website's performance and get optimization suggestions...
XML Sitemap Generator
Generate a standards-compliant sitemap.xml for your website in seconds...
Have Questions?
Clear answers to common questions about our tools,
data, and how the analysis works.
A sitemap checker is a tool that discovers and analyzes XML sitemaps on any website. It checks robots.txt for sitemap declarations, finds sitemaps at common paths, parses the XML content, and reports URL counts and last modified dates.
The tool first checks the website's robots.txt file for Sitemap: directives. If none are found, it tries common paths like /sitemap.xml, /wp-sitemap.xml, and /sitemap_index.xml. When a sitemap is found, it parses the XML to extract URL counts and metadata.
A sitemap index is a master file that references multiple individual sitemaps. It's used when a site has more than 50,000 URLs or needs to organize sitemaps by content type. Our tool detects index files and processes the child sitemaps they reference.
Declaring sitemaps in robots.txt helps search engines discover them reliably. Robots.txt is one of the first files search engine bots check when crawling a site. Without this declaration, search engines may not find sitemaps at non-standard paths.
The URL count shows how many pages are listed in each sitemap. For regular sitemaps, this is the number of <url> entries. For sitemap index files, it shows how many child sitemaps are referenced. The total URLs excludes index counts to avoid double-counting.
The lastmod date shows when a sitemap was last modified. For sitemap index files, it often indicates when child sitemaps were updated. Accurate lastmod dates help search engines prioritize crawling of recently updated content.
Yes, our sitemap finder and checker is completely free to use. You can analyze up to 10 websites per year. Simply sign up for a free account to start checking sitemaps.
Yes! You can copy all sitemap URLs to your clipboard with one click, or download a complete analysis report as a text file. The report includes the website URL, sitemap counts, URL totals, and details for each discovered sitemap.
Scale SEO & AI Insights Across Your Organization
Gain complete visibility into SEO and AI performance across your enterprise, with actionable insights for every team and brand.
Achieved 3% growth in Domain Rating
Achieved 1% growth in Domain Rating
Achieved 5% growth in Domain Rating
Achieved 2% growth in Domain Rating
Achieved 2% growth in Domain Rating
Achieved 4% growth in Domain Rating
Achieved 6% growth in Domain Rating