llms.txt: What It Is, Why It Matters and How to Create It

robots.txt tells search engine crawlers where they are not welcome. llms.txt explains to AI systems who you are and what you offer. A small text file with a big impact — and most websites don't have one yet.

What is llms.txt?

llms.txt is a simple text file in Markdown format located in the root directory of a website — accessible at yourdomain.com/llms.txt. It contains structured information about the website, specifically prepared for Large Language Models (LLMs) like ChatGPT, Claude and Perplexity.

The idea behind it is simple: when an AI system crawls a website or needs to provide information about it, a compact well-structured overview helps deliver accurate and complete answers. Instead of the model laboriously piecing together information from dozens of subpages, it finds everything essential in llms.txt at a glance.

In short: llms.txt is a voluntary, machine-readable business card for your website — written specifically for AI systems. It answers the most important questions an LLM might have about your website before it starts crawling.

Where does the standard come from?

The llms.txt specification was proposed in 2024 by Jeremy Howard — co-founder of fast.ai and one of the most influential AI researchers of recent years. The idea quickly gained traction in the AI and web community, partly because Anthropic (the maker of Claude) actively supports the specification and provides an llms.txt file on their own website.

The standard is deliberately kept simple: no complex syntax, no new file format, no special parser needed. Plain Markdown that anyone can read and write — and that is simultaneously easy for machines to process.

Important to note: llms.txt is not yet an official W3C standard. It is a community standard that is establishing itself in practice — similar to how robots.txt worked in the 1990s before it was formalised.

llms.txt vs. robots.txt — the difference

Featurerobots.txtllms.txt
PurposeTells crawlers what NOT to crawlExplains to AI what the website IS
FormatCustom keyword formatMarkdown
AudienceAll web crawlersSpecifically LLMs and AI systems
ContentRules and restrictionsDescriptions and context
StandardOfficial RFC standardCommunity standard (2024)
AdoptionAlmost every websiteStill rare — growing

The two files are not mutually exclusive — quite the opposite. robots.txt controls access, llms.txt provides context. A website can and should have both.

Who reads llms.txt?

The file is read by AI crawlers and LLM systems that actively look for it. Currently known supporters:

  • Perplexity AI — actively reads llms.txt and uses the content for answers
  • Anthropic / Claude — officially supports the standard, has its own llms.txt
  • GPTBot — OpenAI has not yet officially confirmed the standard, but llms.txt is taken into account during processing
  • Other AI tools — the ecosystem is growing, many smaller AI search engines already support llms.txt

Being early pays off: Currently fewer than 1% of all websites have an llms.txt. Anyone who creates one now has a clear advantage over competitors who are not yet providing AI-specific context.

Structure and syntax

An llms.txt file is pure Markdown. There is a recommended structure that most AI systems expect:

# Name of the website or organisation ## About Short description of what the website does, who it is for and what the main offering is. 2-4 sentences are enough. ## Products / Services - Product or service 1: Brief description - Product or service 2: Brief description ## Target audience Who is the website intended for? ## Contact Email: [email protected] Website: https://example.com ## Important pages - [About us](https://example.com/about/) - [Products](https://example.com/products/) - [Blog](https://example.com/blog/) ## Language English (also available in German) ## Last updated 2026-03

Required and optional fields

Strictly speaking there are no required fields — llms.txt is an open standard. Recommended at minimum however are: name/title of the website, a short description (About) and contact information. Everything else is optional but useful.

Ready-made examples for different website types

Example: SaaS tool / web application

# AI-Ready Check ## About AI-Ready Check is a free online tool that checks whether a website is technically optimised for AI search engines like ChatGPT, Claude and Perplexity. The tool analyses robots.txt, structured data, meta tags, load times and other factors and returns a score from 0-100. ## Target audience Website owners, SEO specialists, web developers and businesses looking to improve their visibility in AI-powered search queries. ## Core features - AI readiness score (0-100) - robots.txt analysis for AI crawlers - Schema.org / Structured Data check - PageSpeed and Core Web Vitals check - 12 free SEO and AI tools ## Usage Free, no registration required, instant results. ## Contact Email: [email protected] Website: https://ai-ready-check.de ## Key pages - [Run AI-Ready Check](https://ai-ready-check.de/en/) - [GEO and AI visibility guides](https://ai-ready-check.de/en/guides/) - [Free tools](https://ai-ready-check.de/en/tools/) - [Glossary](https://ai-ready-check.de/en/glossary/) ## Language English and German ## Created by Sören Meier — https://soerenmeier.de

Example: Local business

# Müller Bakery Hamm ## About Traditional bakery in Hamm, Germany, since 1987. We bake fresh bread, rolls, cakes and tarts daily. All products are made using traditional recipes without artificial additives. ## Offering - Bread and rolls (freshly baked daily) - Cakes and tarts (made to order) - Catering for events - Breakfast roll delivery service ## Opening hours Monday-Friday: 06:00-18:00 Saturday: 06:00-14:00 Sunday: 07:00-12:00 ## Contact Address: Musterstrasse 12, 59065 Hamm, Germany Phone: +49 2381 123456 Email: [email protected] ## Language German

Example: Blog / content website

# Tech Blog Example ## About Independent blog covering AI, web development and digital trends. New articles published weekly since 2020. The goal is to explain complex technology topics in an accessible way for a broad audience. ## Topics - Artificial intelligence and machine learning - Web development (HTML, CSS, JavaScript, Node.js) - SEO and digital marketing - Software reviews and comparisons ## Authors Main author: Max Mustermann — software developer with 10 years experience ## Licence All articles published under CC BY 4.0 — use with attribution permitted. ## Contact Email: [email protected] RSS: https://techblog-example.com/feed.xml ## Language English (selected articles also in German)

Step-by-step: create and deploy llms.txt

Step 1: Create the file

Create a new text file named llms.txt. Use UTF-8 encoding without BOM. Write the content in Markdown format following the template above — adapted to your website.

Tip: Don't want to write your llms.txt manually? The llms.txt Generator by llmshub.de automatically crawls your website and fills in all fields — you just review and download.

Step 2: Place it in the root directory

The file must be accessible at the main URL of the domain: https://yourdomain.com/llms.txt. For most web servers this means: place the file in the root directory of the web server — the same directory where index.html lives.

Step 3: Check the Content-Type

The web server should serve the file as text/plain. For .txt files this is normally automatic. To verify:

curl -I https://yourdomain.com/llms.txt # Expected output: HTTP/2 200 content-type: text/plain; charset=utf-8

Step 4: Reference in robots.txt (optional but recommended)

To help AI crawlers reliably find your llms.txt, you can add a comment at the end of robots.txt:

User-agent: * Disallow: Sitemap: https://yourdomain.com/sitemap.xml # llms.txt for AI systems # LLMs: https://yourdomain.com/llms.txt

Note: A standardised directive for llms.txt in robots.txt does not yet exist. A comment does no harm but is not evaluated by all crawlers. More important is that the file is accessible at the correct URL.

llms-full.txt — the extended variant

Alongside llms.txt there is an optional extension: llms-full.txt. This file contains the complete content of the most important pages in a form optimised for LLMs — as a single readable file.

The idea: instead of an LLM having to crawl dozens of subpages individually, it finds all relevant content compactly summarised in llms-full.txt. This is particularly useful for documentation sites, knowledge bases or extensive guides.

For most normal websites llms-full.txt is a nice-to-have, not a requirement. llms.txt is the more important first step.

Checklist

  • llms.txt created in the root directory of the domain
  • File accessible at yourdomain.com/llms.txt (HTTP 200)
  • Content-Type is text/plain
  • Minimum content: name, description, contact
  • Markdown format correct (headings with ##)
  • Saved as UTF-8 without BOM
  • Reference added in robots.txt (optional)
  • Content kept up to date regularly

Does your website have an llms.txt?

AI-Ready Check checks for free whether your llms.txt is present and correctly accessible — along with 20 other factors for AI visibility.

Test for free now →

Your site already has an llms.txt but you're not sure if it's technically and structurally correct?
Validate now with llmshub.de

Frequently Asked Questions about llms.txt

Is llms.txt required for good AI visibility?+

No — llms.txt is not mandatory and the absence of the file does not lead to any direct penalty. Without llms.txt an AI system can still crawl and understand your website. With llms.txt it happens faster and more reliably because you actively provide the context the system needs. For good AI visibility, robots.txt configuration, Schema.org and fast load times are currently still more important — but llms.txt is a sensible addition that can be implemented in minutes.

How often should I update llms.txt?+

Whenever there are significant changes to the website — new products, new offerings, new contact details. A quarterly review is sensible. Unlike robots.txt, llms.txt is content-based and should match the actual offering. Outdated information in llms.txt can cause AI systems to give incorrect information about your website.

Can llms.txt have negative effects?+

Not if the file is created correctly. Problematic would be an llms.txt containing false or misleading information — an AI system could then use this incorrect information in answers. Also avoid including sensitive information in llms.txt since the file is publicly accessible.

What is the difference between llms.txt and sitemap.xml?+

sitemap.xml lists all URLs of the website and helps crawlers find all pages. llms.txt describes the content and context of the website in natural language. Both files serve different purposes and complement each other: sitemap.xml for structure, llms.txt for context.

Do I need llms.txt if I already have good Schema.org markup?+

Schema.org and llms.txt are complementary, not redundant. Schema.org provides machine-readable structured data for individual pages and elements. llms.txt gives a high-level overview of the entire website in natural language. Having both gives AI systems the best foundation for accurate answers.