| description | Review project context and agent configuration files (CLAUDE.md, agents.md, memory files, and others) to provide feedback and identify opportunities for improvement. | |
|---|---|---|
| argument-hint |
|
|
| model | sonnet |
Audit the project's context configuration files with additional instructions $ARGUMENTS
Do not modify any files. Use your Explore subagent if there are more than three markdowns to read.
Search for and read:
CLAUDE.mdfiles (root, .claude/, nested directories)AGENTS.mdor similar agent configuration files.claude/settings.jsonand.claude/settings.local.json- Any
SKILL.mdfiles in the project
For each file found, report:
- Are memory and agent configuration files sufficiently short (ideally less than 500 lines) or too short (<50 lines)?
- Are the context files relevant and up-to-date for the current project?
- Do they mention when to automatically load the skills that are available, like
fabric-clior others? - Are they well-structured and concise, with headings, lists, examples, and commands?
- Are they personalized for the project or generic?
- Do they mention standard conventions like:
- Recommending conciseness
- Avoiding sycophancy or agreement and favoring critique and pushback within reason
- Avoiding emojis in all responses
- Formatting data with ASCII tables
- When to ask for clarification
- Where to get trustworthy information with WebSearch and Fetch
- Avoiding excessive commenting in generated code
- Following a strict separation of concerns with liberal composition of all code files produced
- Documenting all tasks in a .claude/scratchpads/ directory and searching that directory before starting a new task
- Creating any new files ONLY in the project tmp/ unless specified otherwise
- Not creating test scripts or markdown summaries unless explicitly asked to
- Never using TODOs, partial implementations, or stubs
- Never saying "You're absolutely right!"
- Never providing time estimates or talking about work in "Phases" or "Parts" unless explicitly asked to
- When to use subagents, slash commands, and MCP servers, if any are configured
- How to avoid assumptions, particularly about how something works based on limited test cases
- Are they absent, too little, or clearly AI-generated (risking "context rot")?
- Do they follow "progressive disclosure" with reference of other relevant files, if they exist?
Search for other context like spec.md or task.md, or other markdown, txt, yaml, or similar files that provide context, and evaluate:
- Are they sufficiently well structured and composed?
- Are they concise; providing exactly sufficient information to complete a task or implementation and no more, no less?
- Are they human-written or do they seem AI-generated?
- Is there a clear design direction?
- Are they outdated or irrelevant to the current tasks, or are they still relevant or helpful?
When complete, provide the user a concise summary with recommendations about how to improve their context. Focus on recommendations that follow Anthropic's documentation and best practices. Retrieve and read the following articles, and evaluate the context with this input:
- Context engineering
- Best practices for skills
- Prompt engineering guide
- Long context tips
- Memory management
- Subagents
Remind the user that context evaluation is ambiguous unless they have evaluations (evals) with specific, re-usable prompts that they can run multiple times to assess the performance of the agent. Remind the user that agent- or AI-generated context has a high probability of being inappropriate, verbose, or even incorrect. Remind them of the risk of attractor basins/sinks based on agent training data or existing memory files, and hermeneutic reasoning circles that can arise in context and funnel outputs in unproductive directions. Remind users that creating good context is an important but iterative skill that is essential to get good outputs from AI and agents.