Generate a standards-compliant llms.txt file for any website. Makes your site fully readable and citable by AI agents the GEO (Generative Engine Optimization) equivalent of having a great sitemap.
The skill crawls your website using Chrome DevTools, reads your actual pages, and produces a clean llms.txt file in the format specified by Jeremy Howard's llms.txt standard. When AI agents (Claude, ChatGPT, Gemini) visit your site, they read llms.txt first to understand what you are and where to find authoritative content.
Without llms.txt: AI agents guess, hallucinate, or cite competitors instead.
With llms.txt: AI agents cite your product correctly and know exactly where your docs, blog, and key pages live.
If you're inside a website's repo, the skill reads your source files directly pages, routes, blog posts, frontmatter, site config. It writes llms.txt straight to public/ when you approve. No browser required.
Supported frameworks: Next.js (pages + app router), Astro, Nuxt, Gatsby, SvelteKit, Hugo, Jekyll
If you only have the URL, the skill crawls the live site using Chrome DevTools MCP. Falls back to standard web fetch if Chrome isn't available.
Codebase Mode: No extra setup. Just be inside the repo directory.
Live Site Mode:
- Chrome with remote debugging enabled (or any live URL skill will fall back to web fetch)
- Chrome DevTools MCP server configured in your agent (optional, improves JS-rendered sites)
Mac:
open -a "Google Chrome" --args --remote-debugging-port=9222Linux:
google-chrome --remote-debugging-port=9222Windows:
chrome.exe --remote-debugging-port=9222Follow the setup at: https://github.com/ChromeDevTools/chrome-devtools-mcp
Add to your agent's MCP configuration:
{
"mcpServers": {
"chrome-devtools": {
"command": "npx",
"args": ["-y", "@chrome-devtools/mcp-server"],
"env": {
"CHROME_DEBUGGING_PORT": "9222"
}
}
}
}Copy .env.example to .env and fill in:
cp .env.example .envGITHUB_TOKEN and GITHUB_REPO are only needed if you want the agent to automatically open a GitHub PR with the generated file.
Codebase Mode just be inside your project and ask:
"Generate an llms.txt for this site"
"Add llms.txt to this project"
"Make this site readable by AI agents"
The agent will detect your framework, read your pages and blog posts from source, generate llms.txt, and write it to the right directory (e.g. public/llms.txt) after you confirm.
Live Site Mode provide a URL:
"Generate an llms.txt for https://yoursite.com"
"Does https://yoursite.com have an llms.txt? If not, create one."
The agent will:
- Check if
llms.txtalready exists at the domain - Crawl homepage, docs, blog, about, pricing, and API pages
- Generate
llms.txtfollowing the official spec - Optionally generate
llms-full.txtwith full page content - Save the file locally and give you deployment instructions
- Optionally open a GitHub PR if configured
Place llms.txt at your web root so it's accessible at https://yourdomain.com/llms.txt:
| Platform | File Location |
|---|---|
| Next.js / Vercel | /public/llms.txt |
| Astro | /public/llms.txt |
| Nuxt | /public/llms.txt |
| GitHub Pages | Repository root |
| Hugo | /static/llms.txt |
| WordPress | Upload via FTP to web root |
| File | Description |
|---|---|
llms.txt |
Structured link map. LLMs follow links to find content |
llms-full.txt |
Full prose content of key pages. LLMs ingest everything at once |
llms-txt-generator/
├── SKILL.md # Agent instructions
├── README.md # This file
├── .env.example # Environment variables template
├── evals/
│ └── evals.json # Test prompts for skill evaluation
└── references/
├── llms-txt-spec.md # The llms.txt format specification
└── output-template.md # Exact output template with example
MIT