We Ran a 28-Day AI Skills Manifest Experiment. The Data Says: It Doesn't Work.
TL;DR
We deployed llms.txt, a skills manifest (skills.json), and structured content APIs on getaisearchscore.com, then tracked every AI crawler visit for 28 days (14-day baseline + 14-day treatment). Result: zero bots consumed the skills manifest or content APIs. Only GoogleOther fetched llms.txt (2 hits). ChatGPT-User traffic stayed flat (+10%, same 9 pages). Path diversity showed no consistent change across bots. This directly contradicts LightSite AI's published claims of 3x ChatGPT traffic and 5x Q&A visits. Raw data is public at /experiment.
In early 2026, a new idea emerged in the AI search optimization space: if you publish machine-readable files describing your site's content and capabilities, AI crawlers will shift from scattered broad crawling to targeted, efficient consumption. The concept draws on llms.txt (a proposed standard for describing your site to LLMs), plus a “skills manifest” (a JSON file listing API endpoints that bots can call directly).
LightSite AI, a managed GEO service, published claims that deploying these files produced dramatic results: ChatGPT traffic increased 3x, Q&A-focused visits increased 5x, and path diversity dropped from 51.6% to 30% — suggesting bots shifted from crawling random pages to consuming structured APIs.
Their methodology had gaps: no baseline measurement before deployment, no control group, and results observed only inside their proprietary proxy infrastructure. We decided to replicate the experiment on our own production site, with proper before/after measurement.
Experiment Design: 28 Days, Two Phases
Setup
- Site: getaisearchscore.com (Next.js 14 + Fastify, single Hetzner VPS)
- Duration: 28 days (March 10 – April 7, 2026)
- Phase 1 (Days 1–14): Baseline — bot tracking only, no site changes
- Phase 2 (Days 15–28): AI layer deployed — llms.txt, skills.json, /api/ai/* endpoints
- Tracked bots: GPTBot, ChatGPT-User, OAI-SearchBot, ClaudeBot, Googlebot, Bingbot, GoogleOther, Meta-AI
- Live dashboard: /experiment
What We Deployed in Phase 2
- llms.txt and llms-full.txt — machine-readable site descriptions following the proposed standard
- skills.json — a manifest listing 4 “skills” (search_articles, get_article, search_faq, get_methodology) with API endpoint definitions
- /api/ai/* endpoints — structured JSON APIs for blog search, FAQ aggregation, article metadata, and scoring methodology
- Discovery signals —
<link>tags in HTML head, reference in robots.txt, cross-references between files - robots.txt update — explicit
Allow: /api/ai/before the existingDisallow: /api - Canary tokens — unique strings embedded in llms.txt and skills.json to detect if content appeared in LLM outputs
All Phase 2 components were deployed simultaneously on Day 14. This means we cannot attribute effects to individual components — a trade-off we accepted for practical experiment design.
Raw Data
Phase 1: Baseline (March 10–23)
| Bot | Total Hits | Unique Paths | Path Diversity |
|---|---|---|---|
| ChatGPT-User | 289 | 9 | 3.1% |
| Googlebot | 78 | 11 | 14.1% |
| OAI-SearchBot | 54 | 7 | 13.0% |
| GPTBot | 50 | 36 | 72.0% |
| Bingbot | 13 | 8 | 61.5% |
| GoogleOther | 12 | 9 | 75.0% |
| Meta-AI | 4 | 3 | 75.0% |
| Total | 500 | ||
Phase 2: AI Layer Active (March 24 – April 7)
| Bot | Total Hits | Unique Paths | Path Diversity |
|---|---|---|---|
| ChatGPT-User | 317 | 9 | 2.8% |
| Googlebot | 99 | 18 | 18.2% |
| ClaudeBot | 84 | 53 | 63.1% |
| OAI-SearchBot | 41 | 1 | 2.4% |
| Bingbot | 33 | 22 | 66.7% |
| GoogleOther | 27 | 22 | 81.5% |
| GPTBot | 18 | 12 | 66.7% |
| Total | 619 | ||
AI Layer Endpoint Usage (Phase 2 only)
| Endpoint | Bot | Hits |
|---|---|---|
| /llms.txt | GoogleOther | 1 |
| /llms-full.txt | GoogleOther | 1 |
| /skills.json | — | 0 |
| /api/ai/blog | — | 0 |
| /api/ai/faq | — | 0 |
| /api/ai/methodology | — | 0 |
What the Data Shows
Primary finding: AI layer endpoints were ignored
Zero bots consumed skills.json. Zero bots called any /api/ai/ endpoint. Only GoogleOther fetched llms.txt and llms-full.txt (1 hit each). The entire premise of the skills manifest approach — that bots will discover and consume structured APIs — is not supported by our data. Canary tokens embedded in these files were never detected in any LLM output.
ChatGPT-User: The Dominant Bot Was Unaffected
ChatGPT-User accounted for 54% of all bot traffic across both phases (606 of 1,119 total hits). It visited exactly 9 unique paths in both phases — the same pages. Path diversity held steady at 3.1% → 2.8%.
This bot serves live ChatGPT user queries: it fetches specific pages when users ask about topics we cover. It does this regardless of whether we provide a skills manifest. LightSite claimed a 3x ChatGPT traffic increase. We observed +10%, within normal variation.
Path Diversity: No Consistent Pattern
LightSite's central claim was that path diversity decreased after deployment (from scattered crawling to targeted API consumption). Our data shows no consistent direction:
| Bot | Phase 1 | Phase 2 | Direction |
|---|---|---|---|
| GoogleOther | 75.0% | 81.5% | Increased |
| Bingbot | 61.5% | 66.7% | Increased |
| Googlebot | 14.1% | 18.2% | Increased |
| GPTBot | 72.0% | 66.7% | Decreased |
| ChatGPT-User | 3.1% | 2.8% | Flat |
| OAI-SearchBot | 13.0% | 2.4% | Decreased (1 path) |
Three bots increased diversity, two decreased, one stayed flat. OAI-SearchBot's drop is an artifact — it collapsed to visiting a single path, not a sign of targeted API usage.
Other Bot Changes
ClaudeBot appeared only in Phase 2 (84 hits, 53 unique paths). Tempting to attribute to the AI layer, but ClaudeBot did not fetch llms.txt, skills.json, or any /api/ai/ endpoint. Most likely, Anthropic's crawler added our domain to its index during this period.
GPTBot decreased 64% (50 → 18 hits) — the opposite of what the skills manifest hypothesis predicts. Meta-AI disappeared entirely. Both are likely natural crawl scheduling variation.
PerplexityBot was absent throughout both phases — despite Perplexity being an active AI search engine. This may indicate our site is not yet in their crawl queue, or they use a different user-agent string.
Head-to-Head: Our Data vs. LightSite AI Claims
| Metric | LightSite Claimed | Our Result |
|---|---|---|
| ChatGPT traffic change | +200% (3x) | +10% |
| Q&A-focused visits | +400% (5x) | 0 (no API hits) |
| Path diversity (ChatGPT) | 51.6% → 30.0% | 3.1% → 2.8% |
| skills.json consumed | Yes (implied) | No (0 hits) |
Important caveat
Our sites have very different profiles. LightSite AI is a managed GEO platform processing client traffic through a proxy layer; we are a small SaaS blog with ~500 bot hits per two-week period. Higher-traffic sites might see different patterns. However, the core question remains: did any bot consume the manifest and APIs? On our site, the answer is no.
Why the Results Differ
Three structural differences likely explain the gap:
- No baseline in LightSite's experiment. They deployed the skills manifest and observed traffic, but had no pre-deployment comparison period. Any pre-existing bot traffic gets attributed to the manifest.
- Proxy infrastructure confounds. LightSite's system works through their proxy layer, which intercepts and routes bot requests. A proxy can redirect bots to API endpoints by design — that is not bots choosing to use the manifest, it is the proxy routing them there.
- No standard for skills.json. There is no W3C or IETF standard for skills manifests. No browser vendor or AI lab has committed to reading them. Without adoption by the actual AI systems, publishing a manifest is writing a letter that nobody checks their mailbox for.
Limitations
- Single site (N=1). Results may not generalize to sites with different traffic volumes, niches, or domain authority.
- No control group. We compare the same site before/after. External factors cannot be isolated.
- Simultaneous deployment. All Phase 2 components launched at once. We cannot attribute effects to individual components.
- 14 days per phase. Bot crawl cycles may operate on longer timescales. Some effects might take months to materialize.
- Observational, not causal. Even with a change, we could not prove causation without a multi-site controlled experiment.
What This Means for Your AI Search Strategy
- Our data does not support investing in skills manifests. No bot consumed skills.json or used our content APIs. Until a major AI provider confirms support for this format, the engineering effort is not justified.
- llms.txt is low-cost, low-evidence. It takes 15 minutes to deploy and Google's crawler did fetch it. But there is no evidence it changes citation behavior. File under “cannot hurt, might help eventually.”
- Content relevance remains the only proven signal. Our previous research showed content relevance is 62x more predictive of AI citations than structural optimization. This experiment reinforces that finding: bots come for your content, not your metadata files.
- Focus on what bots actually do. ChatGPT-User visits specific pages when real users ask relevant questions. The way to get more AI search traffic is to have more pages that answer more queries well — not to publish manifests describing your existing pages.
Full data is public. View the live experiment dashboard at /experiment. We commit to publishing null results, consistent with our pre-registration approach from the citation study.
Frequently Asked Questions
Does llms.txt actually work for AI search optimization?+
Based on our 28-day experiment, there is no evidence that llms.txt changes AI crawler behavior. GoogleOther fetched it once, but no other bot accessed it. No change in crawl patterns or citation behavior was observed. It's low-cost to deploy (15 minutes) but currently has no demonstrated benefit.
What is a skills manifest (skills.json) and should I implement one?+
A skills manifest is a JSON file that describes your site's capabilities and API endpoints for AI systems. LightSite AI promoted this concept, claiming it changes bot behavior. Our experiment found zero bots accessed skills.json over 14 days. There is no W3C or IETF standard backing this format and no AI lab has committed to supporting it.
Did any AI bot use the structured content APIs?+
No. We deployed four API endpoints (/api/ai/blog, /api/ai/faq, /api/ai/methodology, /api/ai/blog/:slug) and tracked all bot requests for 14 days. Zero hits from any AI crawler. AI bots continue to crawl standard HTML pages, not structured APIs.
How does this compare to LightSite AI's claims?+
LightSite AI claimed 3x ChatGPT traffic increase and 5x Q&A visits after deploying a skills manifest. We saw +10% ChatGPT traffic (within natural variation) and zero API consumption. Key differences: LightSite had no baseline period, their system uses a proxy that routes bot traffic, and their results are not independently reproducible.
Is this experiment statistically valid?+
This is an observational case study on a single site, not a randomized controlled trial. We cannot prove causation in either direction. What we can say: on our site, deploying the full recommended AI discoverability stack produced no measurable change in bot behavior over 14 days. We are transparent about limitations including no control group, single site, and simultaneous deployment.
Alexey Tolmachev
Senior Systems Analyst · AI Search Readiness Researcher
Senior Systems Analyst with 14 years of experience in data architecture, system integration, and technical specification design. Researches how AI search engines process structured data and select citation sources. Creator of the AI Search Readiness Score methodology.
Check Your AI Search Readiness
Get your free AI Search Readiness Score in under 2 minutes. See exactly what to fix so ChatGPT, Perplexity, and Google AI Overviews can find and cite your content.
Scan My Site — FreeNo credit card required.
Related Articles
How to Make Your Documentation AI-Agent Ready in 2026
A practical guide to the dual-audience documentation paradigm: llms.txt, MCP servers, Agent Skills, and the tools that generate them. Includes our own experiment results from implementing these standards on a production SaaS.
12 min read
We Tested Whether AI Search Readiness Score Predicts LLM Citations. It Doesn't.
Pre-registered empirical study: 485 domains, 30 queries, 90 Perplexity runs. AI Search Readiness Score shows zero correlation with citation frequency (r=0.009, p=0.849). Domain Authority is the only significant predictor.
14 min read
Content Relevance Predicts AI Citations — Not SEO Score
Empirical study: content relevance (BM25 + embeddings) predicts AI citations with AUC 0.915. Our 26-check AI Readiness Score adds nothing (p=0.14). 438 domains, 30 queries, 13,140 pairs.
10 min read
Why 90% of AI Search Optimization Advice Is Cargo Cult Science
We tested whether generic SEO scores predict AI citations (they don't: r=0.009). Then we studied Google patents and academic papers to find what actually matters. Here's the science behind AI search optimization.
10 min read
Why AI Crawlers Hate Your JavaScript (And How to Fix It)
How JavaScript-heavy sites block AI search engines from seeing your content, and how to fix the blank page problem in React and Next.js.
7 min read
