We built an AI-readiness scorer. Then we ran it on our own homepage. We failed one of our own checks.
Our homepage scored 100/A on bot access, sitemap, and llms.txt — and 67/100 on agent readiness. The H1 starts 8,388 HTML characters into the page. Here's why most landing pages fail this check, and what it means for AI search.
We've spent the last few months building seellm.com — server-side observation for AI bot traffic, plus a free /score endpoint that audits any URL for how AI search engines and assistants will see it.
Yesterday we ran it on ourselves.
The result
Overall score: 100 / A.
- All 20 major AI bots allowed in
robots.txt(GPTBot, ClaudeBot, Bytespider, PerplexityBot, Google-Extended, Applebot-Extended, Meta-ExternalAgent, Amazonbot, DuckAssistBot, Mistral-User, xAI, CCBot, YouBot, and the rest) ✅ - robots.txt served, sitemap declared, sitemap valid ✅
- Canonical host resolves, no restrictive
X-Robots-Tagheader ✅ - Spec-conformant
llms.txtandllms-full.txtpublished ✅
So far so good. Then agentReadiness came back needs_fixes (67 / 100):
| Check | Result |
|---|---|
| Homepage HTML is fetchable | ✅ pass |
| Machine-readable summary available (llms.txt) | ✅ pass |
| Initial page within agent-friendly token budget (2,625 tokens) | ✅ pass |
| Useful answer appears near the top | ❌ fail |
| Content has extractable structure (lists/tables) | 🟡 warn |
| Core content not hidden behind JavaScript | 🟡 warn (41 script tags) |
The failing check, verbatim from our own scanner: "Primary heading starts after 8,388 HTML characters."
When ChatGPT-User or Claude-SearchBot fetches our homepage on behalf of a real human prompt, it has to scroll past 8 KB of nav and decorative HTML before reaching the H1 that explains what we do.
The cobbler's children, etc. We're a startup whose entire pitch is "make sure AI search can actually read your site," and our own homepage fails on the most basic agent-readability check.
Why this happens (and probably to you too)
This isn't a SeeLLM-specific problem. We see it across most modern landing pages we audit. The pattern:
- Hero section optimized for human eyeballs. A big animated background, a logo, a top nav, sometimes a banner, sometimes social proof — all rendered before the H1.
- Marketing-design defaults push the H1 down. A typical SaaS landing page hierarchy is: nav → hero visual → headline. The headline is visually first, but in the HTML source it's often the third or fourth major block.
- JavaScript-heavy frameworks compound the problem. React/Next.js apps ship a lot of script tags before the readable content. Agents that don't execute JS — and many still don't, especially during training — see scaffolding, not content.
The result: when an AI assistant has a 2,000–4,000-token budget per page and needs to find your direct answer fast, it gives up before reaching it. Or worse, it summarizes your nav.
What "agent readiness" actually checks
Our /agent-readiness diagnostic looks at a page from an agent's perspective, not a browser's. The six checks we run on the initial server-rendered HTML:
- HTML is fetchable — does the URL return readable HTML at all?
- Machine-readable summary available — is there a
llms.txtpointing to a structured summary? - Token budget — is the initial response within an agent-friendly token count (~5,000 tokens or less for the part agents read)?
- Answer depth — does the H1 + a short answer paragraph appear early in the HTML, or is it buried after long navigation and decoration?
- Extractable structure — are there lists, tables, and clear headings that make the answer easy to lift?
- JavaScript rendering risk — does the readable text exist server-side, or is it injected after hydration?
A page that passes all six is parseable by an AI assistant in a single fetch with high confidence. A page that fails one or more either gets misread, paraphrased instead of cited, or skipped entirely.
The full classification rules are on our methodology page.
The fix we're shipping
Our fix is mundane. Move the H1 + a short answer paragraph above the hero animation in the server-rendered HTML. The hero stays for human visitors; the H1 just shifts up in the source. We're shipping it this week.
The harder lesson: agent readiness is a distinct design constraint from human readability, and most landing pages today are designed only for humans. If you care about AI search as a distribution channel, the same hero that converts human visitors may be the thing keeping you out of ChatGPT and Claude answers.
See your own score
It's a free single-URL audit, no signup, ~30 seconds:
- Run it on our homepage and see the failing check yourself: /agent-readiness/?d=seellm.com
- Run it on your own page:
/agent-readiness— paste any URL - Site-wide audit (bot access + discoverability + AI-readable content):
/score
Try it on your own homepage. If your H1 is more than ~5,000 HTML characters deep, you're probably losing AI surface area without realizing it. If your score is high and you're seeing real ChatGPT/Claude referrals show up in your analytics, we'd love to hear what's on your page that the rest of the industry is getting wrong.
Related reading:
Continue reading
More from the field notes
May 05, 2026
We watched a site's AI search traffic crater 80× in two months. Here's what we saw.
A 90-day case study of one site's crawl-to-click ratio — the share of AI bot fetches that turn into a real human click — and what its collapse from 14.9% to 0.18% means for anyone betting on AI as a distribution channel.
February 01, 2026
Why Google Analytics Can't See AI Visibility
Google Analytics is useful for human sessions, but it cannot show which pages AI systems fetch, revisit, cite, skip, or leave crawled but not cited.
April 30, 2026
What Is Crawled But Not Cited?
Crawled but not cited is the gap between AI systems fetching a page and actually reusing it in answers, citations, referrals, or recommendations.
From reading to action
See which pages AI systems can actually use.
Start with the free AI Visibility Score. When you need page-level evidence, move from static checks to monitoring the pages that matter.