llms.txt is a structured, markdown-formatted index of all Bright Data documentation pages - one entry per page with a short description and a direct link.Format:
# Bright Data Docs## Docs- [Agent Web Access](https://docs.brightdata.com/ai/agents.md): Complete web infrastructure for AI agents- [SERP API Introduction](https://docs.brightdata.com/scraping-automation/serp-api/introduction.md): Real-time search results- [Web Unlocker](https://docs.brightdata.com/scraping-automation/web-unlocker/introduction.md): Bypass bot detection...
Note that every link points to the .md version of the page - clean markdown, no HTML.Use it when:
Loading into an agent’s system prompt for full product awareness
Feeding a retrieval system to decide which doc pages to fetch
Giving a coding agent a map of available products before it starts a task
# Quick previewcurl https://docs.brightdata.com/llms.txt | head -40# Download for offline usecurl -o brightdata-llms.txt https://docs.brightdata.com/llms.txt
llms-full.txt contains the complete text of all Bright Data documentation in a single file - clean markdown, no HTML, no navigation chrome.Use it when:
Building a RAG pipeline over Bright Data docs
Injecting full product knowledge into a long-context model (Gemini 1.5 Pro, Claude, etc.)
Reference the file directly in a prompt - Claude Code will fetch and read it:
Please read https://docs.brightdata.com/llms.txt to understand the availableBright Data products, then help me choose the right API for scraping Amazon product pages.
Then reference it in your CLAUDE.md or system prompt:
# Project contextSee .claude/brightdata-docs.txt for the full Bright Data product reference.
Add as a project rules file so your agent has it in context automatically:
# Save to project rules directorycurl -o .cursor/rules/brightdata.md https://docs.brightdata.com/llms.txt# or for Windsurf:curl -o .windsurf/rules/brightdata.md https://docs.brightdata.com/llms.txt
Now every Cursor Composer or Windsurf Cascade session has Bright Data product awareness built in.
import httpxfrom langchain.text_splitter import MarkdownTextSplitter# Fetch the full docsresponse = httpx.get("https://docs.brightdata.com/llms-full.txt")docs_content = response.text# Split into chunkssplitter = MarkdownTextSplitter(chunk_size=1000, chunk_overlap=100)chunks = splitter.create_documents([docs_content])# Add to your vector storevectorstore.add_documents(chunks)
import httpx# Fetch the indexllms_txt = httpx.get("https://docs.brightdata.com/llms.txt").textmessages = [ { "role": "system", "content": f"""You are a helpful assistant with expertise in Bright Data's web data APIs.Here is the full index of available documentation:{llms_txt}Use this to understand which products and APIs are available, then fetchspecific pages when you need full details on a product.""" }, {"role": "user", "content": "How do I scrape Amazon product pages?"}]
This lets agents fetch specific pages on demand without parsing any HTML.
Recommended pattern for agents: Load llms.txt to understand what’s available → identify the relevant page → fetch that page’s .md URL for full details. This keeps token usage efficient while giving the agent complete information when needed.