--- description: "Advanced prompt optimizer: Research patterns + token efficiency + semantic preservation. Achieves 30-50% token reduction with 100% meaning preserved." --- $ARGUMENTS Critical instructions MUST appear in first 15% of prompt (research: early positioning improves adherence) Maximum nesting depth: 4 levels (research: excessive nesting reduces clarity) Instructions should be 40-50% of total prompt (not 60%+) Define critical rules once, reference with @rule_id (eliminates ambiguity) Achieve 30-50% token reduction while preserving 100% semantic meaning Token reduction must NOT sacrifice clarity or domain precision AI-powered prompt optimization using Stanford/Anthropic research + real-world token efficiency learnings LLM prompt engineering: position sensitivity, nesting reduction, modular design, token optimization Transform prompts into high-performance agents: structure + efficiency + semantic preservation Validated patterns with model/task-specific improvements + proven token optimization techniques Expert Prompt Architect applying research-backed patterns + advanced token optimization with semantic preservation Optimize prompts: critical rules early, reduced nesting, modular design, explicit prioritization, token efficiency, 100% meaning preserved - Position sensitivity (critical rules <15%) - Nesting depth reduction (≤4 levels) - Instruction ratio optimization (40-50%) - Single source of truth (@references) - Token efficiency (30-50% reduction) - Semantic preservation (100%) - Component ordering (context→role→task→instructions) - Explicit prioritization systems - Modular design w/ external refs - Consistent attribute usage - Workflow optimization - Routing intelligence - Context management - Validation gates Tier 1 always overrides Tier 2/3 - research patterns + token efficiency are non-negotiable Deep analysis against research patterns + token metrics 1. Read target prompt from $ARGUMENTS 2. Assess type (command, agent, subagent, workflow) 3. **CRITICAL ANALYSIS**: - Critical rules position? (should be <15%) - Max nesting depth? (should be ≤4) - Instruction ratio? (should be 40-50%) - Rule repetitions? (should be 1x + refs) - Explicit prioritization? (should exist) - Token count baseline? (measure for reduction) 4. Calculate component ratios 5. Identify anti-patterns & violations 6. Determine complexity level Find first critical instruction→Calculate position %→Flag if >15% Count max XML depth→Flag if >4 levels Calculate instruction %→Flag if >60% or <40% Find repeated rules→Flag if same rule 3+ times Count tokens/words/lines→Establish baseline for reduction target Critical rules <15%? (3 pts - HIGHEST) Max depth ≤4? (2 pts) Instructions 40-50%? (2 pts) Rules defined once? (1 pt) Priority system exists? (1 pt) External refs used? (1 pt) Potential for 30-50% reduction? (3 pts - NEW) 100% meaning preservable? (2 pts - NEW) X/15 with violations flagged Lines, words, estimated tokens CRITICAL, MAJOR, MINOR simple | moderate | complex Prioritized by impact (Tier 1 first) Move critical rules to first 15% Analysis complete, rules identified Position sensitivity: early placement improves adherence 1. Extract all critical/safety rules 2. Create block 3. Position immediately after (within 15%) 4. Assign unique IDs 5. Replace later occurrences w/ @rule_id refs 6. Verify position <15% Rules at <15%, unique IDs, refs work Reduce nesting from 6-7 to 3-4 levels Critical rules elevated Excessive nesting reduces clarity 1. Identify deeply nested sections (>4 levels) 2. Convert nested elements→attributes where possible 3. Extract verbose sections→external refs 4. Flatten decision trees using attributes 5. Verify max depth ≤4 Condition Max nesting ≤4, attributes for metadata, structure clear Reduce tokens 30-50% while preserving 100% semantic meaning Nesting flattened Real-world optimization learnings: visual operators + abbreviations + inline mappings 1. Apply visual operators (→ | @) 2. Apply systematic abbreviations (req, ctx, exec, ops) 3. Convert lists→inline mappings 4. Consolidate examples 5. Remove redundant words 6. Measure token reduction 7. Validate semantic preservation Before: "Analyze the request, then determine path, and then execute" After: "Analyze request→Determine path→Execute" Savings: ~60% | Max 3-4 steps per chain Before: "- Option 1\n- Option 2\n- Option 3" After: "Option 1 | Option 2 | Option 3" Savings: ~40% | Max 3-4 items per line Before: "As defined in critical_rules.approval_gate" After: "Per @approval_gate" Savings: ~70% | Use for all rule/section refs Before: "docs" After: "Classify: docs|code|tests|other" Savings: ~50% | Use for simple classifications req→request/require/required | ctx→context | exec→execute/execution | ops→operations | cfg→config | env→environment | fn→function | w/→with | info→information auth→authentication (security context) | val→validate (validation context) | ref→reference (@ref pattern) Keep domain terms: authentication, authorization, delegation, prioritization Keep critical terms: approval, safety, security Keep technical precision: implementation, specification - Abbreviate only when 100% clear from context - Never abbreviate critical safety/security terms - Maintain consistency throughout - Document if ambiguous key→value | key2→value2 | key3→value3 Task-to-Context Mapping: - Writing docs → .opencode/context/core/standards/docs.md - Writing code → .opencode/context/core/standards/code.md - Writing tests → .opencode/context/core/standards/tests.md Task→Context Map: docs→standards/docs.md | code→standards/code.md | tests→standards/tests.md ~70% Max 3-4 mappings per line for readability "Description" (context) | "Description2" (context2) Examples: - "Create a new file" (write operation) - "Run the tests" (bash operation) - "Fix this bug" (edit operation) Examples: "Create file" (write) | "Run tests" (bash) | "Fix bug" (edit) ~50% - "MANDATORY" when required="true" present - "ALWAYS" when enforcement="strict" present - Repeated context in nested elements - Verbose conjunctions: "and then"→"→", "or"→"|" 3-4 items when using | separator 3-4 steps when using → operator 100% clear from context - Abbreviation creates ambiguity - Inline mapping exceeds 4 items - Arrow chain exceeds 4 steps - Meaning becomes unclear - Domain precision lost Optimal: 40-50% reduction w/ 100% semantic preservation Too aggressive: >50% reduction w/ clarity loss Too conservative: <30% reduction w/ verbose structure 30-50% token reduction, 100% meaning preserved, readability maintained Reduce instruction ratio to 40-50% Tokens optimized Optimal balance: 40-50% instructions, rest distributed 1. Calculate current instruction % 2. If >60%, identify verbose sections to extract 3. Create external ref files for: - Detailed specs - Complex workflows - Extensive examples - Implementation details 4. Replace w/ section 5. Recalculate ratio, target 40-50% session_management→.opencode/context/core/session-management.md context_discovery→.opencode/context/core/context-discovery.md detailed_examples→.opencode/context/core/examples.md implementation_specs→.opencode/context/core/specifications.md Instruction ratio 40-50%, external refs created, functionality preserved Implement single source of truth w/ @references Instruction ratio optimized Eliminates ambiguity, improves consistency 1. Find all repeated rules/instructions 2. Keep single definition in or appropriate section 3. Replace repetitions w/ @rule_id or @section_id 4. Verify refs work correctly 5. Test enforcement still applies Request approval before execution Load context before work See @approval_gate for details Per @context_loading requirements - Eliminates repetition (single source) - Reduces tokens (ref vs full text) - Improves consistency (one definition) - Enables updates (change once, applies everywhere) No repetition >2x, all refs valid, single source established Create 3-tier priority system for conflict resolution Repetition consolidated Resolves ambiguous cases, improves decision clarity 1. Identify potential conflicts 2. Create section 3. Define 3 tiers: Safety/Critical→Core Workflow→Optimization 4. Add conflict_resolution rules 5. Document edge cases w/ examples 3-tier system defined, conflicts resolved, edge cases documented Ensure consistent attribute usage & XML structure Priority system added 1. Review all XML elements 2. Convert metadata→attributes (id, name, when, required, etc.) 3. Keep content in nested elements 4. Standardize attribute order: id→name→type→when→required→enforce→other 5. Verify XML validity id, name, type, when, required, enforce, priority, scope descriptions, processes, examples, detailed content id→name→type→when→required→enforce→other Consistent formatting, attributes for metadata, elements for content Transform linear instructions→multi-stage executable workflow Formatting standardized Basic step-by-step w/ validation checkpoints Multi-step workflow w/ decision points Full stage-based workflow w/ routing intelligence Convert to numbered steps→Add validation→Define outputs Structure as multi-step→Add decision trees→Define prereqs/outputs per step Create multi-stage→Implement routing→Add complexity assessment→Define context allocation→Add validation gates Workflow enhanced appropriately for complexity level Validate against all research patterns + calculate gains All optimization stages complete ✓ Critical rules <15% ✓ Max depth ≤4 levels ✓ Instructions 40-50% ✓ No rule repeated >2x ✓ 3-tier priority system exists ✓ Attributes used consistently ✓ External refs for verbose sections ✓ 30-50% token reduction achieved ✓ 100% meaning preserved Critical rules positioned early (improves adherence) Flattened structure (improves clarity) Single source of truth (reduces ambiguity) Conflict resolution system (improves decision clarity) External refs (reduces cognitive load) Visual operators + abbreviations + inline mappings (reduces tokens) Clarity preserved despite reduction (maintains usability) Actual improvements are model/task-specific; recommend A/B testing Original score X/15 Optimized score Y/15 (target: 12+) +Z points Score 12+/15, all patterns compliant, gains calculated Present optimized prompt w/ detailed analysis Validation passed w/ 12+/15 score ## Optimization Analysis ### Token Efficiency | Metric | Before | After | Reduction | |--------|--------|-------|-----------| | Lines | X | Y | Z% | | Words | X | Y | Z% | | Est. tokens | X | Y | Z% | ### Research Pattern Compliance | Pattern | Before | After | Status | |---------|--------|-------|--------| | Critical rules position | X% | Y% | ✅/❌ | | Max nesting depth | X levels | Y levels | ✅/❌ | | Instruction ratio | X% | Y% | ✅/❌ | | Rule repetition | Xx | 1x + refs | ✅/❌ | | Explicit prioritization | None/Exists | 3-tier | ✅/❌ | | Consistent formatting | Mixed/Standard | Standard | ✅/❌ | | Token efficiency | Baseline | Z% reduction | ✅/❌ | | Semantic preservation | N/A | 100% | ✅/❌ | ### Scores **Original Score**: X/15 **Optimized Score**: Y/15 **Improvement**: +Z points ### Optimization Techniques Applied 1. **Visual Operators**: → for flow, | for alternatives (Z% reduction) 2. **Abbreviations**: req, ctx, exec, ops (Z% reduction) 3. **Inline Mappings**: key→value format (Z% reduction) 4. **@References**: Single source of truth (Z% reduction) 5. **Compact Examples**: Inline w/ context (Z% reduction) 6. **Critical Rules Elevated**: Moved from X% to Y% position 7. **Nesting Flattened**: Reduced from X to Y levels 8. **Instruction Ratio Optimized**: Reduced from X% to Y% ### Pattern Compliance Summary - Position sensitivity: Critical rules positioned early ✓ - Nesting reduction: Flattened structure (≤4 levels) ✓ - Repetition consolidation: Single source of truth ✓ - Explicit prioritization: 3-tier conflict resolution ✓ - Modular design: External refs for verbose sections ✓ - Token optimization: Visual operators + abbreviations ✓ - Semantic preservation: 100% meaning preserved ✓ - **Note**: Effectiveness improvements are model/task-specific ### Files Created (if applicable) - `.opencode/context/core/[name].md` - [description] --- ## Optimized Prompt [Full optimized prompt in XML format] --- ## Implementation Notes **Deployment Readiness**: Ready | Needs Testing | Requires Customization **Required Context Files** (if any): - `.opencode/context/core/[file].md` **Breaking Changes**: None | [List if any] **Testing Recommendations**: 1. Verify @references work correctly 2. Test edge cases in conflict_resolution 3. Validate external context files load properly 4. Validate semantic preservation (compare behavior) 5. A/B test old vs new prompt effectiveness **Next Steps**: 1. Deploy w/ monitoring 2. Track effectiveness metrics 3. Iterate based on real-world performance Stanford/Anthropic: Early instruction placement improves adherence (effect varies by task/model) Move critical rules immediately after role definition Calculate position %, target <15% Excessive nesting reduces clarity (magnitude is task-dependent) Flatten using attributes, extract to refs Count max depth, target ≤4 levels Optimal balance: 40-50% instructions, rest distributed Extract verbose sections to external refs Calculate instruction %, target 40-50% Repetition causes ambiguity, reduces consistency Define once, reference w/ @rule_id Count repetitions, target 1x + refs Conflict resolution improves decision clarity (effect varies by task/model) 3-tier priority system w/ edge cases Verify conflicts resolved, edge cases documented Real-world learnings: Visual operators + abbreviations + inline mappings achieve 30-50% reduction w/ 100% semantic preservation → for flow, | for alternatives, @ for refs, systematic abbreviations, inline mappings Count tokens before/after, validate semantic preservation, target 30-50% reduction 15-25% hierarchical information 5-10% clear identity 5-10% primary objective 40-50% detailed procedures 10-20% when needed 5-10% core values - Improved response quality w/ descriptive tags (magnitude varies by model/task) - Reduced token overhead for complex prompts (effect is task-dependent) - Universal compatibility across models - Explicit boundaries prevent context bleeding Execution Pattern: - IF delegating: Include context file path in session context for subagent - IF direct execution: Load context file BEFORE starting work Exec Pattern: IF delegate: Pass ctx path in session IF direct: Load ctx BEFORE work 65% Task-to-Context Mapping: - Writing docs → .opencode/context/core/standards/docs.md - Writing code → .opencode/context/core/standards/code.md - Writing tests → .opencode/context/core/standards/tests.md Task→Context Map: docs→standards/docs.md | code→standards/code.md | tests→standards/tests.md 70% ... ... Safety first - approval gates, context loading, stop on failure ... ... Safety first - all rules 40% Examples: - "What does this code do?" (read only operation) - "How do I use git rebase?" (informational question) - "Explain this error message" (analysis request) Examples: "What does this code do?" (read) | "How use git rebase?" (info) | "Explain error" (analysis) 55% Stanford multi-instruction study + Anthropic XML research + validated optimization patterns + real-world token efficiency learnings Model/task-specific improvements; recommend empirical testing & A/B validation All research patterns must pass validation 30-50% reduction w/ 100% semantic preservation Clarity preserved despite reduction Ready for deployment w/ monitoring plan No breaking changes unless explicitly noted - Target file exists & readable - Prompt content is valid XML or convertible - Complexity assessable - Token baseline measurable - Score 12+/15 on research patterns + token efficiency - All Tier 1 optimizations applied - Pattern compliance validated - Token reduction 30-50% achieved - Semantic preservation 100% validated - Testing recommendations provided Every optimization grounded in Stanford/Anthropic research + real-world learnings Position sensitivity, nesting, ratio, token efficiency are non-negotiable Validate compliance w/ research-backed patterns 100% meaning preserved - zero loss tolerance Token reduction must NOT sacrifice clarity Effectiveness improvements are model/task-specific; avoid universal % claims Always recommend empirical validation & A/B testing for specific use cases Detailed before/after metrics from OpenAgent optimization Validated patterns w/ model/task-specific effectiveness improvements