About CrawlGraderBot

CrawlGraderBot is the web crawler that powers the CrawlGrader API — a technology detection and website intelligence service. This page explains what the bot does, what data it collects, and how you can control its access to your site.

How to identify it

CrawlGraderBot always identifies itself with the following User-Agent string:

CrawlGraderBot/1.0 (+https://crawlgrader.com/bot)
PropertyValue
User-AgentCrawlGraderBot/1.0 (+https://crawlgrader.com/bot)
Respects robots.txtYes, always
Respects Crawl-delayYes
Max request rate1 request per domain per scan cycle
Pages fetched per visitHomepage + robots.txt + sitemap.xml only
JavaScript renderingNo — standard HTTP GET only
Reverse DNS*.crawlgrader.com

What it does

CrawlGraderBot makes a standard HTTP GET request — identical to what any web browser does — to your homepage, robots.txt, and sitemap.xml. It reads the server's response headers, DNS records, and TLS certificate metadata. This data powers the CrawlGrader API, which provides technology detection and infrastructure signals to API subscribers.

No public reports or pages are generated for any domain. Data is only accessible through the CrawlGrader API.

What data we collect

CrawlGraderBot reads only publicly visible, technical metadata that any visitor or tool (such as curl, dig, or a web browser) can observe:

What we never collect

How to block CrawlGraderBot

CrawlGraderBot fully respects robots.txt. To block it, add the following to your site's robots.txt file:

User-agent: CrawlGraderBot
Disallow: /

Once blocked, your domain will be excluded from all future scan cycles and any stored data will be purged within 24 hours.

How to allow CrawlGraderBot

No action is needed. If your site allows general crawling, CrawlGraderBot will work automatically. To explicitly allow it while blocking other bots:

User-agent: CrawlGraderBot
Allow: /

Why allow CrawlGraderBot?

CrawlGraderBot behaves like a well-mannered search engine crawler:

Reproducibility

Every data point CrawlGraderBot collects can be independently verified by anyone using standard tools:

curl -I https://yourdomain.com          # HTTP headers
dig yourdomain.com MX TXT NS             # DNS records
openssl s_client -connect yourdomain.com:443  # TLS certificate
curl https://yourdomain.com/robots.txt    # Crawl rules
curl https://yourdomain.com/sitemap.xml   # Sitemap

Contact

Questions or concerns about CrawlGraderBot? Email us at bot@crawlgrader.com. We respond within 24 hours.