skills$openclaw/agent-docs
tylervovan6.4kβ˜…

by tylervovan

agent-docs – OpenClaw Skill

agent-docs is an OpenClaw Skills integration for coding workflows. Create documentation optimized for AI agent consumption. Use when writing SKILL.md files, README files, API docs, or any documentation that will be read by LLMs in context windows. Helps structure content for RAG retrieval, token efficiency, and the Hybrid Context Hierarchy.

6.4k stars3.7k forksSecurity L1
Updated Feb 7, 2026Created Feb 7, 2026coding

Skill Snapshot

nameagent-docs
descriptionCreate documentation optimized for AI agent consumption. Use when writing SKILL.md files, README files, API docs, or any documentation that will be read by LLMs in context windows. Helps structure content for RAG retrieval, token efficiency, and the Hybrid Context Hierarchy. OpenClaw Skills integration.
ownertylervovan
repositorytylervovan/agent-docs
languageMarkdown
licenseMIT
topics
securityL1
installopenclaw add @tylervovan/agent-docs
last updatedFeb 7, 2026

Maintainer

tylervovan

tylervovan

Maintains agent-docs in the OpenClaw Skills directory.

View GitHub profile
File Explorer
4 files
.
references
advanced-patterns.md
5.3 KB
_meta.json
276 B
SKILL.md
4.5 KB
SKILL.md

name: agent-docs description: Create documentation optimized for AI agent consumption. Use when writing SKILL.md files, README files, API docs, or any documentation that will be read by LLMs in context windows. Helps structure content for RAG retrieval, token efficiency, and the Hybrid Context Hierarchy.

Agent Docs

Write documentation that AI agents can efficiently consume. Based on Vercel benchmarks and industry standards (AGENTS.md, llms.txt, CLAUDE.md).

The Hybrid Context Hierarchy

Three-layer architecture for optimal agent performance:

Layer 1: Constitution (Inline)

Always in context. 2,000–4,000 tokens max.

# AGENTS.md
> Context: Next.js 16 | Tailwind | Supabase

## 🚨 CRITICAL
- NO SECRETS in output
- Use `app/` directory ONLY

## πŸ“š DOCS INDEX (use read_file)
- Auth: `docs/auth/llms.txt`
- DB: `docs/db/schema.md`

Include:

  • Security rules, architecture constraints
  • Build/test/lint commands (top for primacy bias)
  • Documentation map (where to find more)

Layer 2: Reference Library (Local Retrieval)

Fetched on demand. 1K–5K token chunks.

  • Framework-specific guides
  • Detailed style guides
  • API schemas

Layer 3: Research Assistant (External)

Gated by allow-lists. Edge cases only.

  • Latest library updates
  • Stack Overflow for obscure errors
  • Third-party llms.txt

Why This Works

Vercel Benchmark (2026):

ApproachPass Rate
Tool-based retrieval53%
Retrieval + prompting79%
Inline AGENTS.md100%

Root cause: Meta-cognitive failure. Agents don't know what they don't knowβ€”they assume training data is sufficient. Inline docs bypass this entirely.

Core Principles

1. Compressed Index > Full Docs

An 8KB compressed index outperforms a 40KB full dump.

Compress to:

  • File paths (where code lives)
  • Function signatures (names + types only)
  • Negative constraints ("Do NOT use X")

RAG systems split at headers. Each section must be self-contained:

## Database Setup          ← Chunk boundary

Prerequisites: PostgreSQL 14+

1. Create database...

Rules:

  • Front-load key info (chunkers truncate)
  • Descriptive headers (agents search by header text)

3. Inline Over Links

Agents can't autonomously browse. Each link = tool call + latency + potential failure.

ApproachToken LoadAgent Success
Full inline~12Kβœ… High
Links only~2K❌ Requires fetching
Hybrid~4K baseβœ… Best of both

4. The "Lost in the Middle" Problem

LLMs have U-shaped attention:

  • Strong: Start of context (primacy)
  • Strong: End of context (recency)
  • Weak: Middle of context

Solution: Put critical rules at TOP of AGENTS.md. Governance first, details later.

5. Signal-to-Noise Ratio

Strip everything that isn't essential:

  • No "Welcome to..." preambles
  • No marketing text
  • No changelogs in core docs

Formats like llms.txt and AGENTS.md mechanically increase SNR.

llms.txt Standard

Machine-readable doc index for agents:

# Project Name

> One-line project description.

## Authentication

- [Setup](docs/auth/setup.md): Environment vars and init
- [Server](docs/auth/server.md): Cookie handling

## Database

- [Schema](docs/db/schema.md): Full Prisma schema

Location: /llms.txt at domain root Companion: /llms-full.txt β€” full concatenated docs, HTML stripped

Security Considerations

Inline = Trusted

AGENTS.md is part of your codebase. Controlled, version-pinned.

External = Attack Surface

  • Indirect prompt injection via hidden text
  • SSRF risks if agents can browse freely
  • Dependency on external uptime

Mitigation: Domain allow-lists, human-in-the-loop for external retrieval.

Anti-Patterns

  1. Pasting 50 pages β€” triggers "Lost in the Middle"
  2. "See external docs" β€” agents can't browse autonomously
  3. Generic advice β€” "Write clean code" (use specific constraints)
  4. TOC-only docs β€” indexes without content
  5. Trusting retrieval alone β€” 53% vs 100% pass rate

Advanced Patterns

For detailed guidance on RAG optimization, multi-framework docs, and API templates, see references/advanced-patterns.md.

Validation Checklist

  • Critical governance at TOP of doc
  • Total inline context under 4K tokens
  • Each H2 section self-contained
  • No external links without inline summary
  • Negative constraints explicit ("Do NOT...")
  • File paths and signatures, not full code
README.md

No README available.

Permissions & Security

Security level L1: Low-risk skills with minimal permissions. Review inputs and outputs before running in production.

### Inline = Trusted AGENTS.md is part of your codebase. Controlled, version-pinned. ### External = Attack Surface - Indirect prompt injection via hidden text - SSRF risks if agents can browse freely - Dependency on external uptime **Mitigation:** Domain allow-lists, human-in-the-loop for external retrieval.

Requirements

  • OpenClaw CLI installed and configured.
  • Language: Markdown
  • License: MIT
  • Topics:

FAQ

How do I install agent-docs?

Run openclaw add @tylervovan/agent-docs in your terminal. This installs agent-docs into your OpenClaw Skills catalog.

Does this skill run locally or in the cloud?

OpenClaw Skills execute locally by default. Review the SKILL.md and permissions before running any skill.

Where can I verify the source code?

The source repository is available at https://github.com/openclaw/skills/tree/main/skills/tylervovan/agent-docs. Review commits and README documentation before installing.