
Crawler visits do not guarantee citations. However, if your key pages are not being crawled, it is much harder for them to be cited consistently.
What the Crawlers dashboard shows
When Crawlers is connected, you’ll typically see:- Total visits from AI crawlers/bots in the selected time range, with a comparison versus the previous period.
- A platform breakdown of crawler visits (for example: OpenAI, Anthropic, Google AI, Microsoft, Perplexity).
- A trend chart so you can spot spikes, drops, and steady growth.
- A Recent Visits table that shows:
- which pages are being visited most,
- total visits per page,
- and when each page was last visited.
How to read the trend chart
The trend chart is most useful for answering:- “Did crawler activity increase after we published or updated pages?”
- “Did crawler activity drop after a site change?”
- “Which platforms are contributing most to discovery right now?”
How teams use Crawlers (practical workflow)
A simple way to use this page week to week:- Confirm discovery is happening at all. If crawler visits are consistently present, your site is being discovered.
- Validate changes after publishing. After you publish new content or update a key page, check whether that URL starts appearing (or increases) in Recent Visits.
- Prioritize discovery fixes before content scale. If crawler visits are absent or extremely low for key pages, focus on discovery first. Publishing more content typically won’t help until those pages can be found.
- Use it as a leading indicator. Improvements in crawl activity often show up before improvements in owned citations and prominence. If you see crawler visits increasing on key pages, it’s often a sign you’re moving in the right direction.
Connecting Crawlers (Vercel)
Right now, Crawlers supports a Vercel integration. If you haven’t connected it yet, you’ll see an empty state prompting you to connect:- Go to Website → Crawlers
- Click Connect Vercel
- Authorize the integration
We currently support Vercel and plan to add additional hosting/log integrations over time.
Common reasons crawler activity is low
If you are seeing little or no crawler activity, these are the most common causes:- Robots.txt rules block important paths. This can prevent crawlers from accessing your category pages, product pages, or key guides.
- Key pages are noindex or canonicalized away. If your important pages aren’t indexable or are canonicalized to other URLs, crawlers may deprioritize them.
- Sitemaps are missing or incomplete. Crawlers often use sitemaps to discover new and updated pages efficiently.
- Important pages are not reachable via internal linking. If your key pages are not linked from navigation or hubs, crawlers may not find them consistently.
- Performance issues make pages slow or unstable. Slow or error-prone pages can be crawled less frequently.
If discovery is blocked, publishing more content usually does not help. Fix crawlability first.
A practical crawler debugging checklist
Use this checklist when crawl activity looks lower than expected:- Confirm robots.txt allows crawling for key paths. Pay special attention to category, product, pricing, and “source page” URLs.
- Confirm your most important pages are linked from navigation and/or category hubs. A page that isn’t linked is often a page that won’t be crawled consistently.
- Confirm sitemaps include the pages you want crawled and cited. Make sure new content is in the sitemap and that the sitemap is accessible.
- Fix template-level technical issues surfaced in Website → Pages. If a template affects hundreds of pages, fixing it once can dramatically improve crawl coverage.
- Re-check crawl activity after the next refresh cycle. Crawl changes are not always immediate. Look for a trend over days, not minutes.
Measuring impact
Crawler improvements often show up in this order:- More crawler activity on key pages (especially new/updated guides and hubs)
- More owned pages appearing as citations (Analytics → Citations)
- Improved prominence on prompts that rely on those pages (Analytics → Prompts)