March 7, 2026
llms.txt Explained: The Robots.txt for AI
There is a new file for your website root — and it might be the single most impactful thing you can do for AI visibility in 2026.
What Is llms.txt?
You know robots.txt. It has been around since 1994, telling search engine crawlers which parts of your site they can and cannot access. It is one of the oldest standards on the web, and virtually every website has one.
Now imagine a similar file — but instead of telling crawlers where to go, it tells AI models what your site is about. That is llms.txt.
llms.txt is a proposed standard for providing AI-friendly context about your website. It sits at your domain root (yourdomain.com/llms.txt) and gives language models a structured overview of your site's purpose, content, and organization. Think of it as the executive briefing that a model reads before it tries to understand your individual pages.
The idea is simple: AI models are going to read your website regardless. Without guidance, they have to figure out your site structure, your key offerings, and your content hierarchy on their own. llms.txt gives them a map instead of making them wander.
The Problem It Solves
When a language model encounters your website, it faces a real challenge. Your site might have hundreds or thousands of pages. The model has limited context windows and processing budgets. It cannot read everything, so it has to make choices about what to focus on.
Without llms.txt, those choices are essentially random. The model might index your blog archive but miss your core product pages. It might read an outdated press release and treat that as your current positioning. It might understand your navigation but completely miss the content that actually matters.
The result: AI models form incomplete or inaccurate perceptions of your business. They know some things about you but not the right things. They might answer "What does this company do?" based on a blog post from three years ago instead of your current homepage.
llms.txt solves this by giving models explicit guidance: here is what our site is about, here are the most important pages, and here is how they relate to each other. It is the difference between handing someone a book with no table of contents and handing them one with a clear index.
The Format
The llms.txt format is deliberately simple. It uses Markdown-style conventions that both humans and machines can read easily:
- # Headings define your site name and major sections
- > Blockquotes provide brief descriptions
- - Link text entries list important pages with context
- Sections are separated by blank lines
Here is what a basic llms.txt looks like for a fictional bakery:
- # Sweet & Simple Bakery
- > Artisan bakery in Portland, Oregon. We bake sourdough bread, pastries, and custom cakes using locally sourced organic ingredients.
- ## Products
- - Sourdough Bread: Our signature naturally leavened bread, available in five varieties
- - Custom Cakes: Wedding, birthday, and celebration cakes made to order
- - Daily Pastries: Fresh croissants, danishes, and seasonal specials
- ## About
- - Our Story: Founded in 2018 by two pastry chefs with a mission to make real bread accessible
- - Locations: Two locations in Portland — Pearl District and Division Street
Notice how each entry includes not just the URL but also a brief description. That context is crucial — it tells the AI model what it will find at each page before it visits.
How to Create Your llms.txt
Creating an llms.txt file is straightforward, but doing it well requires some thought. Here is a practical approach:
Start with your identity. The first heading should be your organization name. The blockquote below it should be a one-to-two sentence description of who you are and what you do. Be specific. "Software company" is useless. "AI perception management platform that audits how language models understand your website" tells the model exactly what it needs to know.
Map your key pages. Don't list every page on your site. Focus on the pages that define your business: your core product or service pages, your about page, your pricing page, and any content that differentiates you from competitors. Quality over quantity.
Add descriptive context. For each linked page, include a brief description after a colon. This is the most commonly skipped step and also the most valuable. The description tells the AI what to expect before it visits the page, helping it prioritize and contextualize the content.
Organize by section. Use second-level headings (##) to group related pages. Products, solutions, resources, company — whatever structure makes sense for your site. This mirrors how you think about your content and helps models understand relationships between pages.
Keep it current. An outdated llms.txt is worse than no llms.txt. If you remove a product, update the file. If you launch a new section, add it. Treat it like you would your sitemap — a living document that reflects your current site.
Where to Put It
Place your llms.txt file at the root of your domain: yourdomain.com/llms.txt. This is the conventional location that AI models and tools check first, just like robots.txt.
You should also reference it in your robots.txt file. Add a line like this:
- Sitemap: https://yourdomain.com/sitemap.xml
- # AI model context
- # See https://yourdomain.com/llms.txt
Some sites also add a meta tag or HTTP header pointing to the file, but the root URL convention is the most widely recognized.
There is also an extended variant called llms-full.txt for sites that want to provide more comprehensive content, including full-text descriptions of key pages. The standard llms.txt should be concise (under 50 entries for most sites), while llms-full.txt can be as detailed as you need.
Who Supports llms.txt?
The llms.txt standard is still emerging, but adoption is growing quickly. The specification was proposed by Jeremy Howard and has gained traction across the AI and web development communities.
Several AI companies and tools already look for llms.txt when crawling sites, including multiple major language model providers. The pattern makes intuitive sense — the same way robots.txt became universal because it was simple and useful, llms.txt is gaining adoption because it solves a real problem with minimal effort.
aystos Cockpit checks for llms.txt presence and quality as part of its crawlability audit. The aystos Client can generate and maintain your llms.txt automatically based on your site structure and content. And if you want to learn more about implementation details, our complete llms.txt guide covers the specification in depth.
Why You Should Create One Today
Here is the bottom line: llms.txt takes about 30 minutes to create for most websites, costs nothing to implement, and immediately improves how AI models understand your site. There is no downside.
The sites that implement llms.txt now are establishing their AI perception early — before the standard becomes as universal as robots.txt. They are telling models what matters rather than hoping the models figure it out on their own.
As AI-mediated discovery becomes the primary way people find and evaluate businesses, that early clarity will compound into a real competitive advantage. The models will know who you are, what you do, and why you are different — because you told them directly.
You can create your llms.txt manually, use the aystos Client to generate one automatically, or start with a free Cockpit scan to see how your current AI visibility stacks up. Either way, the file takes minutes to create and could change how millions of AI conversations reference your business.
Generate Your llms.txt with aystos
Scan your site and get an auto-generated llms.txt based on your actual content structure.
Start Free Scan →