For a more complete example of a SQL agent with query execution, error correction, and validation, see our SQL Agent tutorial. This tutorial focuses on the progressive disclosure pattern which can be applied to any domain.
How it works
Here’s the flow when a user asks for a SQL query: Why progressive disclosure:- Reduces context usage - load only the 2-3 skills needed for a task, not all available skills
- Enables team autonomy - different teams can develop specialized skills independently (similar to other multi-agent architectures)
- Scales efficiently - add dozens or hundreds of skills without overwhelming context
- Simplifies conversation history - single agent with one conversation thread
- Latency: Loading skills on-demand requires additional tool calls, which adds latency to the first request that needs each skill
- Workflow control: Basic implementations rely on prompting to guide skill usage - you cannot enforce hard constraints like “always try skill A before skill B” without custom logic
Setup
Installation
This tutorial requires thelangchain package:
LangSmith
Set up LangSmith to inspect what is happening inside your agent. Then set the following environment variables:Select an LLM
Select a chat model from LangChain’s suite of integrations:- OpenAI
- Anthropic
- Azure
- Google Gemini
- AWS Bedrock
- HuggingFace
1. Define skills
First, define the structure for skills. Each skill has a name, a brief description (shown in the system prompt), and full content (loaded on-demand):View complete skill definitions
View complete skill definitions
2. Create skill loading tool
Create a tool to load full skill content on-demand:load_skill tool returns the full skill content as a string, which becomes part of the conversation as a ToolMessage. For more details on creating and using tools, see the Tools guide.
3. Build skill middleware
Create custom middleware that injects skill descriptions into the system prompt. This middleware makes skills discoverable without loading their full content upfront.This guide demonstrates creating custom middleware. For a comprehensive guide on middleware concepts and patterns, see the custom middleware documentation.
load_skill tool is registered as a class variable, making it available to the agent.
Production consideration: This tutorial loads the skill list in
__init__ for simplicity. In a production system, you may want to load skills in the before_agent hook instead, allowing them to be refreshed periodically to reflect up-to-date changes (e.g., when new skills are added or existing ones are modified). See the before_agent hook documentation for details.4. Create the agent with skill support
Now create the agent with the skill middleware and a checkpointer for state persistence:load_skill to retrieve full skill content when needed. The checkpointer maintains conversation history across turns.
5. Test progressive disclosure
Test the agent with a question that requires skill-specific knowledge:load_skill("sales_analytics") to get the full schema and business logic, and then used that information to write a correct query following the database conventions.
6. Advanced: Add constraints with custom state
Optional: Track loaded skills and enforce tool constraints
Optional: Track loaded skills and enforce tool constraints
You can add constraints to enforce that certain tools are only available after specific skills have been loaded. This requires tracking which skills have been loaded in custom agent state.Create the agent with the middleware that registers the constrained tool:Now if the agent tries to use
Define custom state
First, extend the agent state to track loaded skills:Update load_skill to modify state
Modify theload_skill tool to update state when a skill is loaded:Create constrained tool
Create a tool that’s only usable after a specific skill has been loaded:Update middleware and agent
Update the middleware to use the custom state schema:write_sql_query before loading the required skill, it will receive an error message prompting it to load the appropriate skill (e.g., sales_analytics or inventory_management) first. This ensures the agent has the necessary schema knowledge before attempting to validate queries.Complete example
View complete runnable script
View complete runnable script
Here’s a complete, runnable implementation combining all the pieces from this tutorial:This complete example includes:
- Skill definitions with full database schemas
- The
load_skilltool for on-demand loading SkillMiddlewarethat injects skill descriptions into the system prompt- Agent creation with middleware and checkpointer
- Example usage showing how the agent loads skills and writes SQL queries
- Install required packages:
pip install langchain langchain-openai langgraph - Set your API key (e.g.,
export OPENAI_API_KEY=...) - Replace the model initialization with your preferred LLM provider
Implementation variations
View implementation options and trade-offs
View implementation options and trade-offs
This tutorial implemented skills as in-memory Python dictionaries loaded through tool calls. However, there are several ways to implement progressive disclosure with skills:Storage backends:
- In-memory (this tutorial): Skills defined as Python data structures, fast access, no I/O overhead
- File system (Claude Code approach): Skills as directories with files, discovered via file operations like
read_file - Remote storage: Skills in S3, databases, Notion, or APIs, fetched on-demand
- System prompt listing: Skill descriptions in system prompt (used in this tutorial)
- File-based: Discover skills by scanning directories (Claude Code approach)
- Registry-based: Query a skill registry service or API for available skills
- Dynamic lookup: List available skills via a tool call
- Single load: Load entire skill content in one tool call (used in this tutorial)
- Paginated: Load skill content in multiple pages/chunks for large skills
- Search-based: Search within a specific skill’s content for relevant sections (e.g., using grep/read operations on skill files)
- Hierarchical: Load skill overview first, then drill into specific subsections
- Small skills (< 1K tokens / ~750 words): Can be included directly in system prompt and cached with prompt caching for cost savings and faster responses
- Medium skills (1-10K tokens / ~750-7.5K words): Benefit from on-demand loading to avoid context overhead (this tutorial)
- Large skills (> 10K tokens / ~7.5K words, or > 5-10% of context window): Should use progressive disclosure techniques like pagination, search-based loading, or hierarchical exploration to avoid consuming excessive context
Progressive disclosure and context engineering
Combining with few-shot prompting and other techniques
Combining with few-shot prompting and other techniques
Progressive disclosure is fundamentally a context engineering technique - you’re managing what information is available to the agent and when. This tutorial focused on loading database schemas, but the same principles apply to other types of context.
Combining with few-shot prompting
For the SQL query use case, you could extend progressive disclosure to dynamically load few-shot examples that match the user’s query:Example approach:- User asks: “Find customers who haven’t ordered in 6 months”
- Agent loads
sales_analyticsschema (as shown in this tutorial) - Agent also loads 2-3 relevant example queries (via semantic search or tag-based lookup):
- Query for finding inactive customers
- Query with date-based filtering
- Query joining customers and orders tables
- Agent writes query using both schema knowledge AND example patterns
Next steps
- Learn about middleware for more dynamic agent behaviors
- Explore context engineering techniques for managing agent context
- Explore the handoffs pattern for sequential workflows
- Read the supervisor pattern for parallel task routing
- See multi-agent patterns for other approaches to specialized agents
- Use LangSmith to debug and monitor skill loading