Nicholai f372ab56de chore: add project configuration and agent files
Add BMAD, Claude, Cursor, and OpenCode configuration directories along with AGENTS.md documentation.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-11-27 04:31:56 -07:00

14 KiB

System Builder Guide

Complete guide for using the /build-context-system command to create custom context-aware AI systems.

Overview

The system builder creates complete .opencode folder architectures tailored to your domain and use cases. It follows research-backed patterns from Stanford and Anthropic for optimal AI agent performance.

Quick Start

/build-context-system

This launches an interactive interview that guides you through:

  1. Domain & Purpose (2-3 questions)
  2. Use Cases & Workflows (3-4 questions)
  3. Complexity & Scale (2-3 questions)
  4. Integration & Tools (2-3 questions)
  5. Review & Confirmation

What You Get

Complete .opencode Structure

.opencode/
├── agent/
│   ├── {domain}-orchestrator.md      # Main coordinator
│   └── subagents/
│       ├── {specialist-1}.md
│       ├── {specialist-2}.md
│       └── {specialist-3}.md
├── context/
│   ├── domain/                       # Core knowledge
│   ├── processes/                    # Workflows
│   ├── standards/                    # Quality rules
│   └── templates/                    # Reusable patterns
├── workflows/
│   ├── {workflow-1}.md
│   └── {workflow-2}.md
├── command/
│   ├── {command-1}.md
│   └── {command-2}.md
├── README.md                         # System overview
├── ARCHITECTURE.md                   # Architecture guide
├── TESTING.md                        # Testing checklist
└── QUICK-START.md                    # Usage examples

Research-Backed Optimizations

All generated agents follow proven patterns:

  • +20% routing accuracy (LLM-based decisions with @ symbol routing)
  • +25% consistency (XML structure with optimal component ordering)
  • 80% context efficiency (3-level context allocation)
  • +17% overall performance (position-sensitive component sequencing)

Interview Process

Phase 1: Domain & Purpose

Question 1: What is your primary domain or industry?

Examples:

  • E-commerce and online retail
  • Data engineering and analytics
  • Customer support and service
  • Content creation and marketing
  • Software development and DevOps
  • Healthcare and medical services
  • Financial services and fintech
  • Education and training

Question 2: What is the primary purpose of your AI system?

Examples:

  • Automate repetitive tasks
  • Coordinate complex workflows
  • Generate content or code
  • Analyze and process data
  • Provide customer support
  • Manage projects and tasks
  • Quality assurance and validation
  • Research and information gathering

Question 3: Who are the primary users?

Examples:

  • Developers and engineers
  • Content creators and marketers
  • Data analysts and scientists
  • Customer support teams
  • Product managers
  • Business executives

Phase 2: Use Cases & Workflows

Question 4: What are your top 3-5 use cases?

Be specific. Good examples:

  • "Process customer orders from multiple channels"
  • "Generate blog posts and social media content"
  • "Analyze sales data and create reports"
  • "Triage and route support tickets"
  • "Review code for security vulnerabilities"

Bad examples:

  • "Do stuff"
  • "Help with work"
  • "Make things better"

Question 5: What is the typical complexity?

For each use case:

  • Simple: Single-step, clear inputs/outputs, no dependencies
  • Moderate: Multi-step process, some decision points, basic coordination
  • Complex: Multi-agent coordination, many decision points, state management

Question 6: Are there dependencies between use cases?

Examples:

  • "Research must happen before content creation"
  • "Validation happens after processing"
  • "All tasks are independent"

Phase 3: Complexity & Scale

Question 7: How many specialized agents do you need?

Guidance:

  • 2-3 agents: Simple domain with focused tasks
  • 4-6 agents: Moderate complexity with distinct specializations
  • 7+ agents: Complex domain with many specialized functions

Question 8: What types of knowledge does your system need?

Categories:

  • Domain knowledge: Core concepts, terminology, business rules, data models
  • Process knowledge: Workflows, procedures, integration patterns, escalation paths
  • Standards knowledge: Quality criteria, validation rules, compliance requirements, error handling
  • Template knowledge: Output formats, common patterns, reusable structures

Question 9: Will your system need to maintain state or history?

Options:

  • Stateless: Each task is independent, no history needed
  • Project-based: Track state within projects or sessions
  • Full history: Maintain complete history and learn from past interactions

Phase 4: Integration & Tools

Question 10: What external tools or platforms will your system integrate with?

Examples:

  • APIs (Stripe, Twilio, SendGrid, etc.)
  • Databases (PostgreSQL, MongoDB, Redis, etc.)
  • Cloud services (AWS, GCP, Azure, etc.)
  • Development tools (GitHub, Jira, Slack, etc.)
  • Analytics platforms (Google Analytics, Mixpanel, etc.)
  • None - standalone system

Question 11: What file operations will your system perform?

Options:

  • Read only: Only read existing files
  • Read/write: Read and create/modify files
  • Full management: Complete file lifecycle management

Question 12: Do you need custom slash commands?

Examples:

  • /process-order {order_id}
  • /generate-report {type} {date_range}
  • /analyze-data {source} {destination}

Phase 5: Review & Confirmation

The system presents a complete architecture summary showing:

  • All components to be created
  • Agent specifications
  • Context file organization
  • Workflow definitions
  • Custom commands
  • Estimated file counts

You can:

  • Proceed - Generate the complete system
  • 🔄 Revise - Adjust specific components
  • Cancel - Start over

System Architecture

Hierarchical Agent Pattern

User Request
     ↓
Main Orchestrator
     ↓
  Analyzes → Allocates Context → Routes
     ↓              ↓                ↓
Subagent A    Subagent B      Subagent C

3-Level Context Allocation

Level 1: Complete Isolation (80% of cases)

  • Context: Task description only
  • Use for: Simple, well-defined operations
  • Performance: 80% reduction in context overhead

Level 2: Filtered Context (20% of cases)

  • Context: Task + relevant domain knowledge
  • Use for: Operations requiring domain expertise
  • Performance: 60% reduction in context overhead

Level 3: Windowed Context (Rare)

  • Context: Task + domain knowledge + historical state
  • Use for: Complex multi-step operations
  • Performance: Optimized for accuracy over speed

Context Organization

Domain Knowledge (context/domain/)

  • Core concepts and definitions
  • Terminology and glossary
  • Business rules and policies
  • Data models and schemas

Process Knowledge (context/processes/)

  • Standard workflows and procedures
  • Integration patterns
  • Edge case handling
  • Escalation paths

Standards Knowledge (context/standards/)

  • Quality criteria and metrics
  • Validation rules
  • Compliance requirements
  • Error handling standards

Template Knowledge (context/templates/)

  • Output format templates
  • Common patterns and structures
  • Reusable components
  • Example artifacts

Generated Components

Main Orchestrator

The orchestrator is your system's "brain" that:

  • Analyzes incoming requests
  • Assesses complexity
  • Allocates appropriate context level
  • Routes to specialized subagents
  • Coordinates workflow execution
  • Validates results
  • Delivers final outputs

Key features:

  • Multi-stage workflow execution
  • Routing intelligence (analyze→allocate→execute)
  • Context engineering (3-level allocation)
  • Validation gates
  • Performance metrics

Specialized Subagents

Each subagent is a specialist that:

  • Handles ONE specific task extremely well
  • Receives complete, explicit instructions
  • Operates statelessly (no conversation history)
  • Returns structured outputs (YAML/JSON)
  • Validates inputs and outputs

Common subagent types:

  • Research Agent: Gathers information from external sources
  • Validation Agent: Validates outputs against standards
  • Processing Agent: Transforms or processes data
  • Generation Agent: Creates content or artifacts
  • Integration Agent: Handles external system integrations

Workflows

Workflows define reusable process patterns:

  • Step-by-step procedures
  • Context dependencies
  • Decision points
  • Success criteria
  • Validation gates

Workflow complexity levels:

  • Simple: 3-5 linear stages
  • Moderate: 5-7 stages with decision trees
  • Complex: 7+ stages with multi-agent coordination

Custom Commands

Slash commands provide user-friendly interfaces:

  • Clear syntax with parameters
  • Agent routing specification
  • Concrete examples
  • Expected output format
  • Usage documentation

Best Practices

Context File Organization

  1. Keep files focused: 50-200 lines per file
  2. Clear naming: Use descriptive names (pricing-rules.md, not rules.md)
  3. No duplication: Each piece of knowledge in exactly one file
  4. Document dependencies: List what other files are needed
  5. Include examples: Every concept should have concrete examples

Agent Design

  1. Single responsibility: Each agent does ONE thing well
  2. Stateless subagents: No conversation history or state
  3. Complete instructions: Every call includes ALL needed information
  4. Explicit outputs: Define exact output format with examples
  5. Validation gates: Validate at critical points

Context Efficiency

  1. Prefer Level 1: Use isolation for 80% of tasks
  2. Selective Level 2: Only when domain knowledge is truly needed
  3. Rare Level 3: Only for complex multi-agent coordination
  4. Load selectively: Only load context files actually needed

Workflow Design

  1. Clear stages: Each stage has clear purpose and output
  2. Prerequisites: Document what must be true before each stage
  3. Checkpoints: Validate at critical points
  4. Success criteria: Define measurable outcomes
  5. Error handling: Plan for failures

Testing Your System

Component Testing

  1. Test orchestrator with simple request
  2. Test each subagent independently
  3. Verify context files load correctly
  4. Test workflows end-to-end
  5. Test custom commands
  6. Validate error handling
  7. Test edge cases

Integration Testing

  1. Multi-agent coordination: Test complex workflows
  2. Context loading: Verify correct files are loaded
  3. Routing logic: Ensure requests route correctly
  4. Validation gates: Check quality checks work
  5. Performance: Measure context efficiency

Quality Validation

Generated systems should score:

  • Agent Quality: 8+/10 (XML optimization)
  • Context Organization: 8+/10 (modularity)
  • Workflow Completeness: 8+/10 (all stages defined)
  • Documentation Clarity: 8+/10 (comprehensive)
  • Overall: 8+/10 (production-ready)

Customization

After generation, you can customize:

  1. Context files: Add your domain-specific knowledge
  2. Workflows: Adjust based on real usage patterns
  3. Validation criteria: Refine quality standards
  4. Agent prompts: Add examples and edge cases
  5. Commands: Create additional slash commands

Performance Expectations

Context Efficiency

  • 80% of tasks use Level 1 context (isolation)
  • 20% of tasks use Level 2 context (filtered)
  • Level 3 context (windowed) is rare

Quality Improvements

  • Routing Accuracy: +20% (LLM-based decisions)
  • Consistency: +25% (XML structure)
  • Context Efficiency: 80% reduction in overhead
  • Overall Performance: +17% improvement

Troubleshooting

Common Issues

Issue: Generated agents don't route correctly Solution: Check @ symbol usage and context level specifications

Issue: Context files are too large Solution: Split into smaller, focused files (50-200 lines)

Issue: Workflows are unclear Solution: Add more detailed steps and examples

Issue: Commands don't work Solution: Verify agent routing in frontmatter

Getting Help

  1. Review generated documentation (README.md, ARCHITECTURE.md)
  2. Check TESTING.md for testing guidance
  3. Review QUICK-START.md for usage examples
  4. Examine template files for patterns
  5. Ask specific questions about components

Examples

E-commerce System

Domain: E-commerce Order Management Agents: order-processor, inventory-checker, payment-handler, shipping-calculator Workflows: simple-order, complex-order, refund-process Commands: /process-order, /check-inventory, /process-refund

Data Pipeline System

Domain: Data Engineering Agents: data-extractor, transformation-engine, quality-validator, data-loader Workflows: standard-etl, complex-transformation, data-quality-check Commands: /run-pipeline, /validate-data, /transform-data

Content Creation System

Domain: Content Marketing Agents: research-assistant, content-generator, quality-validator, publisher Workflows: research-enhanced, multi-platform, quick-post Commands: /create-content, /research-topic, /publish-content

Next Steps

After your system is generated:

  1. Review the documentation: Start with README.md
  2. Test basic functionality: Try simple commands
  3. Customize context: Add your domain knowledge
  4. Run through testing checklist: Ensure quality
  5. Refine based on usage: Iterate and improve

Resources

  • Templates: .opencode/context/system-builder-templates/
  • Examples: See generated systems for reference
  • Documentation: README.md, ARCHITECTURE.md, TESTING.md
  • Patterns: Review template files for best practices

Ready to build your context-aware AI system?

Run: /build-context-system