We Ran a 28-Day AI Skills Manifest Experiment. The Data Says: It Doesn't Work.

12 min read

TL;DR

We deployed llms.txt, a skills manifest (skills.json), and structured content APIs on getaisearchscore.com, then tracked every AI crawler visit for 28 days (14-day baseline + 14-day treatment). Result: zero bots consumed the skills manifest or content APIs. Only GoogleOther fetched llms.txt (2 hits). ChatGPT-User traffic stayed flat (+10%, same 9 pages). Path diversity showed no consistent change across bots. This directly contradicts LightSite AI's published claims of 3x ChatGPT traffic and 5x Q&A visits. Raw data is public at /experiment.

In early 2026, a new idea emerged in the AI search optimization space: if you publish machine-readable files describing your site's content and capabilities, AI crawlers will shift from scattered broad crawling to targeted, efficient consumption. The concept draws on llms.txt (a proposed standard for describing your site to LLMs), plus a “skills manifest” (a JSON file listing API endpoints that bots can call directly).

LightSite AI, a managed GEO service, published claims that deploying these files produced dramatic results: ChatGPT traffic increased 3x, Q&A-focused visits increased 5x, and path diversity dropped from 51.6% to 30% — suggesting bots shifted from crawling random pages to consuming structured APIs.

Their methodology had gaps: no baseline measurement before deployment, no control group, and results observed only inside their proprietary proxy infrastructure. We decided to replicate the experiment on our own production site, with proper before/after measurement.

Experiment Design: 28 Days, Two Phases

Setup

  • Site: getaisearchscore.com (Next.js 14 + Fastify, single Hetzner VPS)
  • Duration: 28 days (March 10 – April 7, 2026)
  • Phase 1 (Days 1–14): Baseline — bot tracking only, no site changes
  • Phase 2 (Days 15–28): AI layer deployed — llms.txt, skills.json, /api/ai/* endpoints
  • Tracked bots: GPTBot, ChatGPT-User, OAI-SearchBot, ClaudeBot, Googlebot, Bingbot, GoogleOther, Meta-AI
  • Live dashboard: /experiment

What We Deployed in Phase 2

  • llms.txt and llms-full.txt — machine-readable site descriptions following the proposed standard
  • skills.json — a manifest listing 4 “skills” (search_articles, get_article, search_faq, get_methodology) with API endpoint definitions
  • /api/ai/* endpoints — structured JSON APIs for blog search, FAQ aggregation, article metadata, and scoring methodology
  • Discovery signals<link> tags in HTML head, reference in robots.txt, cross-references between files
  • robots.txt update — explicit Allow: /api/ai/ before the existing Disallow: /api
  • Canary tokens — unique strings embedded in llms.txt and skills.json to detect if content appeared in LLM outputs

All Phase 2 components were deployed simultaneously on Day 14. This means we cannot attribute effects to individual components — a trade-off we accepted for practical experiment design.

Raw Data

Phase 1: Baseline (March 10–23)

BotTotal HitsUnique PathsPath Diversity
ChatGPT-User28993.1%
Googlebot781114.1%
OAI-SearchBot54713.0%
GPTBot503672.0%
Bingbot13861.5%
GoogleOther12975.0%
Meta-AI4375.0%
Total500

Phase 2: AI Layer Active (March 24 – April 7)

BotTotal HitsUnique PathsPath Diversity
ChatGPT-User31792.8%
Googlebot991818.2%
ClaudeBot845363.1%
OAI-SearchBot4112.4%
Bingbot332266.7%
GoogleOther272281.5%
GPTBot181266.7%
Total619

AI Layer Endpoint Usage (Phase 2 only)

EndpointBotHits
/llms.txtGoogleOther1
/llms-full.txtGoogleOther1
/skills.json0
/api/ai/blog0
/api/ai/faq0
/api/ai/methodology0

What the Data Shows

Primary finding: AI layer endpoints were ignored

Zero bots consumed skills.json. Zero bots called any /api/ai/ endpoint. Only GoogleOther fetched llms.txt and llms-full.txt (1 hit each). The entire premise of the skills manifest approach — that bots will discover and consume structured APIs — is not supported by our data. Canary tokens embedded in these files were never detected in any LLM output.

ChatGPT-User: The Dominant Bot Was Unaffected

ChatGPT-User accounted for 54% of all bot traffic across both phases (606 of 1,119 total hits). It visited exactly 9 unique paths in both phases — the same pages. Path diversity held steady at 3.1% → 2.8%.

This bot serves live ChatGPT user queries: it fetches specific pages when users ask about topics we cover. It does this regardless of whether we provide a skills manifest. LightSite claimed a 3x ChatGPT traffic increase. We observed +10%, within normal variation.

Path Diversity: No Consistent Pattern

LightSite's central claim was that path diversity decreased after deployment (from scattered crawling to targeted API consumption). Our data shows no consistent direction:

BotPhase 1Phase 2Direction
GoogleOther75.0%81.5%Increased
Bingbot61.5%66.7%Increased
Googlebot14.1%18.2%Increased
GPTBot72.0%66.7%Decreased
ChatGPT-User3.1%2.8%Flat
OAI-SearchBot13.0%2.4%Decreased (1 path)

Three bots increased diversity, two decreased, one stayed flat. OAI-SearchBot's drop is an artifact — it collapsed to visiting a single path, not a sign of targeted API usage.

Other Bot Changes

ClaudeBot appeared only in Phase 2 (84 hits, 53 unique paths). Tempting to attribute to the AI layer, but ClaudeBot did not fetch llms.txt, skills.json, or any /api/ai/ endpoint. Most likely, Anthropic's crawler added our domain to its index during this period.

GPTBot decreased 64% (50 → 18 hits) — the opposite of what the skills manifest hypothesis predicts. Meta-AI disappeared entirely. Both are likely natural crawl scheduling variation.

PerplexityBot was absent throughout both phases — despite Perplexity being an active AI search engine. This may indicate our site is not yet in their crawl queue, or they use a different user-agent string.

Head-to-Head: Our Data vs. LightSite AI Claims

MetricLightSite ClaimedOur Result
ChatGPT traffic change+200% (3x)+10%
Q&A-focused visits+400% (5x)0 (no API hits)
Path diversity (ChatGPT)51.6% → 30.0%3.1% → 2.8%
skills.json consumedYes (implied)No (0 hits)

Important caveat

Our sites have very different profiles. LightSite AI is a managed GEO platform processing client traffic through a proxy layer; we are a small SaaS blog with ~500 bot hits per two-week period. Higher-traffic sites might see different patterns. However, the core question remains: did any bot consume the manifest and APIs? On our site, the answer is no.

Why the Results Differ

Three structural differences likely explain the gap:

  1. No baseline in LightSite's experiment. They deployed the skills manifest and observed traffic, but had no pre-deployment comparison period. Any pre-existing bot traffic gets attributed to the manifest.
  2. Proxy infrastructure confounds. LightSite's system works through their proxy layer, which intercepts and routes bot requests. A proxy can redirect bots to API endpoints by design — that is not bots choosing to use the manifest, it is the proxy routing them there.
  3. No standard for skills.json. There is no W3C or IETF standard for skills manifests. No browser vendor or AI lab has committed to reading them. Without adoption by the actual AI systems, publishing a manifest is writing a letter that nobody checks their mailbox for.

Limitations

  • Single site (N=1). Results may not generalize to sites with different traffic volumes, niches, or domain authority.
  • No control group. We compare the same site before/after. External factors cannot be isolated.
  • Simultaneous deployment. All Phase 2 components launched at once. We cannot attribute effects to individual components.
  • 14 days per phase. Bot crawl cycles may operate on longer timescales. Some effects might take months to materialize.
  • Observational, not causal. Even with a change, we could not prove causation without a multi-site controlled experiment.

What This Means for Your AI Search Strategy

  1. Our data does not support investing in skills manifests. No bot consumed skills.json or used our content APIs. Until a major AI provider confirms support for this format, the engineering effort is not justified.
  2. llms.txt is low-cost, low-evidence. It takes 15 minutes to deploy and Google's crawler did fetch it. But there is no evidence it changes citation behavior. File under “cannot hurt, might help eventually.”
  3. Content relevance remains the only proven signal. Our previous research showed content relevance is 62x more predictive of AI citations than structural optimization. This experiment reinforces that finding: bots come for your content, not your metadata files.
  4. Focus on what bots actually do. ChatGPT-User visits specific pages when real users ask relevant questions. The way to get more AI search traffic is to have more pages that answer more queries well — not to publish manifests describing your existing pages.

Full data is public. View the live experiment dashboard at /experiment. We commit to publishing null results, consistent with our pre-registration approach from the citation study.

Frequently Asked Questions

Does llms.txt actually work for AI search optimization?+

Based on our 28-day experiment, there is no evidence that llms.txt changes AI crawler behavior. GoogleOther fetched it once, but no other bot accessed it. No change in crawl patterns or citation behavior was observed. It's low-cost to deploy (15 minutes) but currently has no demonstrated benefit.

What is a skills manifest (skills.json) and should I implement one?+

A skills manifest is a JSON file that describes your site's capabilities and API endpoints for AI systems. LightSite AI promoted this concept, claiming it changes bot behavior. Our experiment found zero bots accessed skills.json over 14 days. There is no W3C or IETF standard backing this format and no AI lab has committed to supporting it.

Did any AI bot use the structured content APIs?+

No. We deployed four API endpoints (/api/ai/blog, /api/ai/faq, /api/ai/methodology, /api/ai/blog/:slug) and tracked all bot requests for 14 days. Zero hits from any AI crawler. AI bots continue to crawl standard HTML pages, not structured APIs.

How does this compare to LightSite AI's claims?+

LightSite AI claimed 3x ChatGPT traffic increase and 5x Q&A visits after deploying a skills manifest. We saw +10% ChatGPT traffic (within natural variation) and zero API consumption. Key differences: LightSite had no baseline period, their system uses a proxy that routes bot traffic, and their results are not independently reproducible.

Is this experiment statistically valid?+

This is an observational case study on a single site, not a randomized controlled trial. We cannot prove causation in either direction. What we can say: on our site, deploying the full recommended AI discoverability stack produced no measurable change in bot behavior over 14 days. We are transparent about limitations including no control group, single site, and simultaneous deployment.

AT

Alexey Tolmachev

Senior Systems Analyst · AI Search Readiness Researcher

Senior Systems Analyst with 14 years of experience in data architecture, system integration, and technical specification design. Researches how AI search engines process structured data and select citation sources. Creator of the methodology.

Check Your AI Search Readiness

Get your free AI Search Readiness Score in under 2 minutes. See exactly what to fix so ChatGPT, Perplexity, and Google AI Overviews can find and cite your content.

Scan My Site — Free

No credit card required.

Related Articles