Evolve Playbooks to Match Claude Capabilities
Purpose: Quarterly (or on-demand) review of Claude capability updates and playbook regeneration to maintain alignment and maximize efficiency.
Mindset: Self-healing, self-improving system. Playbooks exist to serve users. As Claude improves, playbooks should improve automatically. Apply /pb-preamble thinking (challenge assumptions about what’s still true) and /pb-design-rules thinking (does every playbook still embody Clarity, Simplicity, Resilience?).
Core Principle: We don’t freeze playbooks at a point-in-time. We evolve them continuously as Claude capabilities improve. This is how we stay efficient.
Resource Hint: opus - Strategic evolution; capability assessment and design decisions.
When to Use
- Quarterly schedule - Feb, May, Aug, Nov (fixed calendar)
- Major Claude version release - When Claude 4.6 → 4.7 drops
- Context limit stress - If hitting session limits regularly
- Latency complaints - If playbooks feel slow
- User feedback - When patterns don’t work in practice
Quarterly Schedule & Operational Framework
Fixed Quarterly Calendar
Evolution cycles run on a fixed quarterly schedule, not ad-hoc:
| Quarter | Cycle Window | Development Period | Evolution Period | Release Date |
|---|---|---|---|---|
| Q1 | Jan 20 - Feb 15 | Jan 1 - Feb 9 | Feb 10 - Feb 15 | Feb 16 (tag vX.Y.0) |
| Q2 | Apr 20 - May 15 | Apr 1 - May 9 | May 10 - May 15 | May 16 (tag vX.Y.0) |
| Q3 | Jul 20 - Aug 15 | Jul 1 - Aug 9 | Aug 10 - Aug 15 | Aug 16 (tag vX.Y.0) |
| Q4 | Oct 20 - Nov 15 | Oct 1 - Nov 9 | Nov 10 - Nov 15 | Nov 16 (tag vX.Y.0) |
Fixed dates enable:
- Team predictability (everyone knows when evolution happens)
- Planning visibility (teams budget for evolution work)
- Consistent rhythm (quarterly on schedule, not whenever convenient)
Evolution Manager Role
Responsibility: One person per quarter manages the evolution cycle end-to-end.
Qualifications:
- Familiar with playbooks and architecture
- Can make judgment calls on evolution priorities
- Access to git tags, GitHub releases, merge permissions
- 4-6 hours of focused time
Responsibilities:
-
Week Before Evolution (Preparation)
- Review capability changes since last quarter
- Run git signals (if not already done)
- Prepare evolution input document
- Schedule team review session (30-45 min)
-
Evolution Period (Monday-Friday)
- Facilitate playbook review and change proposals
- Lead capability analysis with team
- Consolidate findings into prioritized change list
- Manage PR review and approval process
- Ensure testing validates all changes
- Prepare release notes
-
Release Day (Friday)
- Merge evolution PR to main
- Create git tag and GitHub release
- Update project CLAUDE.md
- Post evolution summary to team
-
Post-Release (Following Monday)
- Verify documentation builds correctly
- Run verification checks
- Document any post-release fixes needed
- Plan next quarter’s evolution inputs
Team Coordination
Evolution Review Meeting (Tuesday of evolution week)
- Duration: 45 minutes
- Attendees: Evolution Manager, 1-2 senior engineers, playbook steward
- Agenda:
- Capability changes since last quarter (10 min)
- Git signals analysis (if applicable) (10 min)
- Proposed changes discussion (20 min)
- Approval and prioritization (5 min)
Decision Criteria:
- ✅ Changes based on new Claude capabilities or user feedback
- ✅ Changes improve clarity, simplicity, or efficiency
- ❌ Changes that break established patterns without strong justification
- ❌ Changes that contradict preamble or design rules
Pre-Evolution Checklist
Before starting evolution work, Evolution Manager verifies:
- Current date is within evolution period (e.g., Feb 10-15)
- All capability changes documented (Claude model versions, new features, etc.)
- Git signals run (if applicable, use
python scripts/git-signals.py) - Team knows evolution is happening (Slack/standup announcement)
- Main branch is clean and up to date
- Previous quarter’s changes are stable in production
- Snapshot created (
git tag v-pre-evolution-YYYY-Q[N]) - Evolution input document prepared for review meeting
Quick Start: Run Evolution Cycle
Step 1: Prepare Environment
# Ensure clean state
git status # Must be clean
git checkout main && git pull origin main # On latest main
# Create evolution branch
git checkout -b evolve/$(date +%Y-%m-%d) main
# Load metadata schema
cat .playbook-metadata-schema.yaml # Review schema
# Examples are archived in git history; current commands are your reference
Step 1.5: Snapshot Before Evolution
Critical: Create a snapshot before making changes. This enables safe rollback if anything breaks.
# Create snapshot of current state
python3 scripts/evolution-snapshot.py \
--create "Before Q1 2026 evolution"
# Record the evolution cycle in structured log
python3 scripts/evolution-log.py \
--record-cycle "2026-Q1" \
--trigger quarterly \
--capability-changes "Sonnet 4.6: 30% faster, no cost change"
This creates:
- Git tag as backup (can revert to this if needed)
- Snapshot metadata for audit trail
- Evolution log entry to track this cycle
Step 2: Run Analysis
# Analyze current state
python3 scripts/evolve.py --analyze
# View detailed report
cat todos/evolution-analysis.json | jq '.'
# Check validation
python3 scripts/evolve.py --validate
Step 3: Review Capability Changes
Since last evolution, what has changed?
- Claude model versions: Run
pbai --versionor check recent announcements - Speed improvements: Sonnet faster? Opus cost-effective for more tasks?
- Context windows: Larger windows change what you can keep in main context
- Latency profile: Different models, different speeds
- Reasoning depth: Better reasoning changes what model to use for what task
Document findings in todos/evolution-log.md:
## Evolution Cycle: 2026-Q2
### Capability Changes Since Last Cycle
- Claude Sonnet 4.5 → 4.6: 30% faster, same cost
- Context window: 200K → 200K (no change)
- Reasoning: Better at multi-step planning
### Implications
- Parallelization now viable (Sonnet fast enough)
- Model routing: Haiku can take more routine tasks
- Context efficiency: Still critical (not changed)
Step 4: Audit Playbooks Against New Capabilities
For each major playbook category, ask:
Development playbooks (pb-start, pb-cycle, pb-commit, pb-pr)
- Can Sonnet 4.6 now handle complex design reviews that needed Opus before?
- Are our model hints still accurate?
- Should parallelization be standard pattern?
Review playbooks (pb-review-code, pb-security, pb-voice)
- Should code review default to Sonnet (vs Opus)?
- Is parallel review (multiple agents) now viable?
- Are detection patterns still current?
Planning playbooks (pb-plan, pb-adr, pb-think)
- Does Sonnet 4.6 handle ideation/synthesis better?
- Should we escalate fewer things to Opus?
- Can we simplify playbooks for routine decisions?
Utilities (pb-patterns, pb-guidance, pb-learn)
- Are best practices still current?
- Do patterns still make sense?
- Are examples still best-practice?
Step 5: Propose Changes
Document each opportunity:
### Opportunity 1: Model Routing Update
**Current:** pb-start says "use Sonnet"
**Capability change:** Sonnet 4.6 is 30% faster
**Proposal:** Update model routing to:
- Haiku for file search, status checks (unchanged)
- Sonnet for development (unchanged)
- Opus for security/architecture (unchanged)
**Rationale:** No change needed; Sonnet still correct model
---
### Opportunity 2: Parallel Research Pattern
**Current:** Sequential agent execution in /pb-claude-orchestration
**Capability change:** Sonnet 4.6 fast enough for parallel fan-out
**Proposal:** Add "Parallel Research Pattern" section:
1. Main launches 3 agents simultaneously
2. Each agent explores independently
3. Results merged in synthesis stage
**Impact:** Session runtime -30% for exploration tasks
**Confidence:** High (pattern validated in playbook development)
Step 6: Test Proposed Changes
For each significant change, validate on 2-3 real tasks:
# Example: Test parallel research pattern
# 1. Identify a task that would benefit
# 2. Run with old (sequential) approach
# 3. Time: 15 minutes
# 4. Run with new (parallel) approach
# 5. Time: 10 minutes
# 6. Document: "Parallel X saved Y minutes"
Record results:
### Validation: Parallel Research Pattern
**Task:** Investigate codebase for X feature
**Old pattern (sequential):** 20 min (Agent A) + 15 min (Agent B) = 35 min total
**New pattern (parallel):** max(20 min, 15 min) = 20 min total
**Result:** 43% faster. Impact = HIGH. Implement.
Step 7: Generate Diff and Request Approval
Before applying changes, generate a diff to see exactly what will change:
# Generate detailed diff comparing current to proposed
python3 scripts/evolution-diff.py \
--detailed main HEAD
# Generate markdown report for PR
python3 scripts/evolution-diff.py \
--report main HEAD
This creates todos/evolution-diff-report.md showing:
- Which playbooks are affected
- What fields change (old → new values)
- Why changes are being proposed
GOVERNANCE GATE: Create a PR and request peer review BEFORE applying changes.
# Create feature branch for changes
git checkout -b evolution/$(date +%Y-Q$((($(date +%m)-1)/3+1)))
# Commit proposed changes
git add commands/
git commit -m "evolution: proposed changes for review"
# Push and create PR
git push origin evolution/...
gh pr create --title "evolution(quarterly): Q1 2026" \
--body "See todos/evolution-diff-report.md for details"
Peer review checklist:
- ✅ Capability changes documented accurately
- ✅ Proposed changes make sense given new capabilities
- ✅ No unintended side effects
- ✅ Metadata is consistent (run test suite)
- ✅ Related commands still exist and are reachable
Only proceed after peer approval and merge to main.
Step 7.5: Apply Approved Changes
Once PR is approved and merged to main, apply the changes:
# Example: Update pb-claude-orchestration
# 1. Add "Parallel Research Pattern" section
# 2. Update examples to use parallel where applicable
# 3. Regenerate CLAUDE.md
# 4. Update MEMORY.md with new strategy
# Regenerate metadata-driven files
python3 scripts/evolve.py --generate
# Validate all metadata
python3 scripts/evolve.py --validate
Step 8: Update Metadata
For each playbook that changed:
# Example: Update pb-start metadata
# - Update last_reviewed date
# - Update execution_time_estimate if timing changed
# - Add last_evolved date
# - Update summary if scope changed
# - Update related_commands if topology changed
Run validation:
python3 scripts/evolve.py --validate
Step 9: Regenerate Auto-Generated Files
# Regenerate all auto-generated indices
python3 scripts/evolve.py --generate
# Regenerate project CLAUDE.md
/pb-claude-project
# Regenerate global CLAUDE.md
/pb-claude-global
# Run docs build
mkdocs build --strict
Step 10: Complete Evolution Cycle
# Stage changes
git add commands/ docs/ scripts/ .claude/ CHANGELOG.md
# Commit with evolution note
git commit -m "evolve(quarterly): $(date +%Y-Q$((($(date +%m)-1)/3+1)))"
# Tag release (if this is a versioned release)
git tag -a v2.X.0 -m "v2.X.0: Q1 2026 evolution"
# Record cycle completion
python3 scripts/evolution-log.py \
--complete "2026-Q1" \
--pr <pr-number>
# Push
git push origin main --tags
Step 11: If Evolution Breaks Something (Rollback)
If you discover issues after applying evolution changes:
# List available snapshots
python3 scripts/evolution-snapshot.py --list
# Show details of specific snapshot
python3 scripts/evolution-snapshot.py --show evolution-20260209-HHMMSS
# Rollback to snapshot (interactive confirmation)
python3 scripts/evolution-snapshot.py --rollback evolution-20260209-HHMMSS
# Or force rollback without confirmation
python3 scripts/evolution-snapshot.py --rollback evolution-20260209-HHMMSS --force
# Record the revert in evolution log
python3 scripts/evolution-log.py \
--revert "2026-Q1" \
--reason "Parallel patterns caused context bloat; needs refinement"
# Push rollback commit
git push origin main
Anatomy of a Good Evolution
What Changed?
- New Claude capabilities (model speed, reasoning, capabilities)
- User feedback (patterns that don’t work, confusing guidance)
- Tech debt (playbooks that have become stale)
- New patterns discovered in practice
How to Spot Evolution Opportunities?
Pattern 1: Capability-Execution Mismatch
- You say “use Sonnet for X” but Sonnet 4.6 can now do Y (more complex) just as well
- Fix: Update model hint, regenerate CLAUDE.md
Pattern 2: Manual Work That Could Automate
- You’re manually updating 5 playbooks when you could update metadata + regenerate
- Fix: Metadata-driven auto-generation, one source of truth
Pattern 3: Complexity That Could Simplify
- Playbook has 10 decision trees but Sonnet 4.6 can handle the full decision at once
- Fix: Consolidate into single decision, simpler playbook
Pattern 4: Serialization That Could Parallelize
- You launch Agent A, wait for result, then launch Agent B
- But now both could launch simultaneously, merge results
- Fix: Document parallel pattern, add to orchestration guide
Pattern 5: Context That Could Compress
- Main context has 50K tokens of file content
- Could move to subagent (returns compression summary)
- Fix: Update context strategy in pb-claude-orchestration
What Doesn’t Change?
- Preamble thinking (challenge assumptions, peer collaboration) - timeless
- Design rules (clarity, simplicity, robustness) - timeless
- Atomic commits, quality gates - foundational, not outdated by capability
- Test-first discipline - still best practice
Evolution Log Structure
The evolution system maintains two logs:
1. Structured Audit Log (todos/evolution-audit.json)
Machine-readable JSON format for pattern analysis and automation:
{
"cycles": [
{
"cycle": "2026-Q1",
"started_at": "2026-02-09T12:00:00",
"trigger": "quarterly",
"capability_changes": "Sonnet 4.6: 30% faster, same cost",
"changes": [
{
"command": "pb-claude-orchestration",
"field": "execution_pattern",
"before": "sequential",
"after": "parallel",
"rationale": "Sonnet 4.6 fast enough for concurrent agents"
}
],
"status": "completed",
"snapshot_id": "evolution-20260209-143022",
"pr_number": 42
}
]
}
Use this log to:
- Detect patterns (what fields change most often?)
- Measure impact (did evolution help or hurt?)
- Enable automation (future cycles can suggest changes)
- Audit decisions (why did we make this change?)
# View evolution history
python3 scripts/evolution-log.py --show
# Analyze patterns
python3 scripts/evolution-log.py --analyze
# Export timeline
python3 scripts/evolution-log.py --export
2. Narrative Release Notes (CHANGELOG.md)
Human-readable summary for each release:
## v2.11.0 (2026-05-15) - Q2 Evolution
### Capability Changes
- Sonnet 4.6 → 4.7: +15% reasoning depth
- No speed or cost changes
- New tool: structured output
### Improvements
- Parallel research patterns now standard in exploration tasks
- Model routing optimized (Haiku handles 10 more utility cases)
- Context efficiency improved 12% via better compression
### Metrics
- Average session time: -8% (from 32 min to 29 min)
- Cost per session: -3% (minor optimization)
- User satisfaction: +5% (feedback survey)
Common Evolution Scenarios
Scenario A: Speed Improvement (e.g., Sonnet 4.5 → 4.6)
Signal: “New Sonnet is 30% faster, same cost”
Analysis:
- What was Sonnet+Opus before might be Sonnet-only now
- Parallelization becomes more viable
- Session times drop
Action:
- Revisit model routing decisions
- Test parallelization patterns
- Update execution time estimates
- Document efficiency gains
Scenario B: Context Window Expansion
Signal: “Claude context now 400K tokens (was 200K)”
Analysis:
- Can now keep more files in main context
- Compression strategy becomes optional
- But context efficiency still matters (cost)
Action:
- Update context loading strategy
- Test keeping full codebase in context
- Measure tokens used; may stay selective
- Update MEMORY.md with new patterns
Scenario C: User Feedback (Patterns Don’t Work)
Signal: “This playbook guidance is confusing, I did it differently”
Analysis:
- Reality doesn’t match documentation
- Users are finding better way
- Playbook is stale or unclear
Action:
- Interview users on what worked
- Update playbook with real pattern
- Validate on 3+ users
- Simplify if new pattern is simpler
Scenario D: New Capability (e.g., Tool Use, Custom Models)
Signal: “Claude now supports X”
Analysis:
- This changes what’s possible
- May enable new playbooks or patterns
- May make old patterns obsolete
Action:
- Research capability thoroughly
- Design playbooks for new capability
- Test extensively before releasing
- Document when this capability became available
Evolution Release Strategy
Regular Releases (Every Quarter)
- Run pb-evolve on fixed schedule
- Document capability changes
- Implement small improvements
- Release as minor version bump (v2.X.0)
Emergency Evolution (New Capability)
- Outside normal schedule
- When major capability lands
- Run full pb-evolve cycle
- Release as patch or minor (v2.X.Y)
Versioning
- v2.X.0: Quarterly evolution
- v2.X.Y: Emergency evolution or small fix
- v1.X.0: Major architectural change
Success Criteria for Evolution
Before publishing an evolution cycle, define and verify success metrics:
For Capability-Driven Evolution (e.g., Claude 4.6 release)
Define:
- “What efficiency improvements do we expect?” (e.g., 15% faster sessions)
- “Which playbooks can be simplified?” (list specific commands)
- “Will model routing change?” (document before/after)
Verify:
- Session timing improved by X% (measured on real tasks)
- User satisfaction feedback positive
- Cost per session unchanged or lower
- No regressions in code quality
For User Feedback Evolution (e.g., Patterns don’t work)
Define:
- “What feedback were we acting on?” (reference issue/comment)
- “What’s the new pattern?” (specific changes to command)
- “Who validates the fix?” (team member, user, or self-test)
Verify:
- User can achieve the goal using updated docs/command
- New pattern validated with 2+ real use cases
- Existing related commands still work with new approach
For Technical Debt Evolution (e.g., Stale patterns)
Define:
- “What pattern is now outdated?” (specific reason)
- “What replaces it?” (new approach, with rationale)
- “Is this a breaking change?” (affects users? need migration guide?)
Verify:
- Migration guide written (if breaking)
- Existing projects tested with new approach
- Related commands still integrate properly
Checklist: Before Publishing Evolution
- Success criteria defined (see section above)
- Success criteria verified
- All playbooks validated (python3 scripts/evolve.py –validate)
- No circular cross-references
- Metadata coverage > 95%
- mkdocs build –strict passes
- markdownlint passes
- CHANGELOG updated
- MEMORY.md updated with lessons
- Evolution log entry written
- Tests pass
- Tested on 2-3 real tasks
Rollback Procedures
If evolution introduces issues after merging, follow these steps:
Immediate Response (Within 1 hour of issue discovery)
# 1. Identify the problem
# - Review recent changes
# - Check which playbooks caused the issue
# 2. Assess severity
# - Does this break user workflows? (CRITICAL)
# - Does this cause confusion? (HIGH)
# - Is this a minor clarity issue? (MEDIUM)
# 3. Decide: Fix Forward vs Rollback
# CRITICAL: Rollback immediately
# HIGH: Rollback if fix takes >30 min, fix forward if quick fix available
# MEDIUM: Fix forward (don't rollback for minor issues)
Rolling Back Evolution (If Needed)
# Step 1: Retrieve pre-evolution snapshot
python3 scripts/evolution-snapshot.py --list
# Shows: evolution-20260210-143022, evolution-20260211-091845, etc.
# Step 2: Review what will be restored
python3 scripts/evolution-snapshot.py --show evolution-20260210-143022
# Step 3: Restore (interactive confirmation)
python3 scripts/evolution-snapshot.py --rollback evolution-20260210-143022
# Step 4: Verify restoration
git log --oneline -3
mkdocs build --strict
# Step 5: Record the revert
python3 scripts/evolution-log.py \
--revert "2026-Q1" \
--reason "Caused confusion in pb-guide, needs refinement"
# Step 6: Push rollback commit
git push origin main
# Step 7: Communicate
# Announce rollback in team Slack/standup with reason
Post-Rollback Analysis
After rollback, document:
# Evolution Rollback Report: 2026-Q1
**Date:** Feb 15, 2026
**Reason:** Proposed changes to pb-guide clarity caused more confusion than before
## What Went Wrong
- Change assumed users familiar with concept X (they weren't)
- New section headings created ambiguity about scope
- Examples didn't match current usage patterns
## Learning for Next Cycle
- Earlier user validation before committing large doc changes
- Test changes with actual users (2-3 people) before merging
- Include examples that match documented patterns exactly
## Re-Planning
- Keep current pb-guide as-is for Q1
- Plan more targeted clarity improvements for Q2
- Assign to different reviewer with user feedback focus
Evolution Metrics & Reporting
Measuring Evolution Success
Track these metrics for each evolution cycle:
Quality Metrics:
- ✅ No bugs introduced (zero rollbacks needed)
- ✅ No regressions (existing functionality preserved)
- ✅ Documentation builds successfully
- ✅ All tests pass
Adoption Metrics:
- When was the evolution PR merged? (commits per day unchanged)
- Any user feedback about changes? (watch for GitHub issues)
- Are new patterns being adopted? (track in next cycle’s signals)
Efficiency Metrics:
- Time taken for evolution cycle (hours)
- Lines of code/documentation changed
- Number of playbooks touched
Quarterly Evolution Report Template
Create todos/evolution-report-YYYY-Q[N].md after each cycle:
# Evolution Report: Q1 2026
**Evolution Manager:** [Name]
**Cycle Period:** Feb 10-15, 2026
**Release Date:** Feb 16, 2026
## Capability Changes Assessed
- Claude Sonnet: [version change, if any]
- Claude Opus: [version change, if any]
- New capabilities: [e.g., tool use, structured output]
## Changes Made
### Playbooks Updated
- pb-guide (3 sections clarified)
- pb-cycle (added parallel review pattern)
- pb-git-signals (integrated with evolution planning)
### Impact Assessment
- Breaking changes: 0
- Potentially confusing changes: 0 (no rollbacks needed)
- User-facing improvements: 3
### Metrics
- Evolution time: 4 hours
- Lines changed: 280
- Tests run: 40 (all passed)
## User Feedback (If Any)
- [Positive feedback on changes]
- [Questions or confusion]
- [Suggestions for next cycle]
## Learnings & Improvements for Q2
1. [What went well]
2. [What to improve]
3. [Process improvements]
## Next Quarter Priorities
- [Based on feedback and evolution planning]
Post-Evolution Review
One week after evolution release (e.g., Feb 23), evaluate:
Stability Check
# 1. Verify no regressions
# - No user bug reports related to evolution changes
# - CI/CD still green
# - Deployment still smooth
# 2. Document any minor issues
# - Typos or clarity gaps found by users
# - Add to next quarter's evolution input
# 3. Measure actual impact
# - Did playbook improvements help? (user feedback)
# - Are new patterns being used? (git commits)
# - Did efficiency improve? (session times)
Updating Evolution Log
python3 scripts/evolution-log.py --complete-review "2026-Q1" \
--stability "green" \
--feedback "[user feedback summary]"
Planning Next Cycle
By end of week after evolution:
- Document learnings for next Evolution Manager
- Capture early user feedback for next evolution input
- Update MEMORY.md with patterns discovered
- Plan Q2 evolution inputs
Evolution Tracking System
Central Evolution Dashboard
Maintain todos/evolution-dashboard.md for quarter-at-a-glance status:
# Evolution Dashboard: 2026
## Q1 (Feb 10-15) - COMPLETE
- Evolution Manager: [Name]
- Status: ✅ Released Feb 16
- Capability focus: Sonnet 4.6 performance improvements
- Changes: 3 playbooks, 280 lines
- Impact: No regressions, positive feedback
- Post-review: Stable, metrics good
## Q2 (May 10-15) - UPCOMING
- Evolution Manager: [TBD - assign by April 20]
- Preliminary capability focus: Context window, reasoning improvements
- Estimated changes: TBD
- Key questions: [To be researched in May]
## Q3 (Aug 10-15) - PLANNING
- Evolution Manager: [Rotate from Q1]
- Preliminary focus: TBD
## Q4 (Nov 10-15) - PLANNING
- Evolution Manager: [Rotate from Q2]
- Preliminary focus: TBD
Pre-Evolution Preparation Tracking
30 days before evolution cycle:
# Q2 2026 Evolution Prep (30 days before May 10)
**Timeline:**
- April 10: Evolution Manager assigned, research phase begins
- April 15: Capability analysis draft completed
- April 20: Review meeting scheduled
- May 1: Evolution input document finalized
- May 9: Team review meeting
- May 10: Evolution work begins
**Checklist:**
- [ ] Evolution Manager assigned (person + backup)
- [ ] Capability changes researched
- [ ] Git signals run (if applicable)
- [ ] Evolution meeting scheduled
- [ ] Input document drafted
- [ ] Team notified
Related Commands
/pb-claude-global- Regenerate global CLAUDE.md/pb-claude-project- Regenerate project CLAUDE.md/pb-standards- Quality standards (validated by evolution)/pb-preamble- Thinking philosophy (doesn’t change)/pb-design-rules- Design principles (doesn’t change)
Tips for Sustainable Evolution
- Make metadata source of truth - Everything derives from metadata
- Automate what’s repetitive - scripts/evolve.py handles index generation
- Document rationale - Every change explains why (for future evolution)
- Test before releasing - Validate on real tasks
- Measure impact - Track efficiency gains
- Collect feedback - Users will find patterns that don’t work
- Iterate publicly - Share evolution log so users understand changes
How This Works in Practice
Imagine Sonnet 4.6 is released and it’s 30% faster.
- pb-evolve runs → analyzes capability changes
- Opportunity identified → “Can now parallelize more tasks”
- Pattern validated → tests on real task, confirms 30% speedup
- Playbook updated → adds parallel pattern to pb-claude-orchestration
- Metadata updated → updates execution_time_estimate, last_evolved
- Files regenerated → mkdocs build, scripts/evolve.py –generate
- Committed → git commit, tagged v2.10.0
- Users benefit → faster sessions, happier users, sustainable excellence
This is self-healing DNA in action.
What Gets Evolved?
- Command metadata (last_reviewed, execution_time_estimate, difficulty)
- Model routing decisions (when to use Haiku vs Sonnet vs Opus)
- Execution patterns (when to parallelize, when to serialize)
- Context loading strategy (what to load in main, what to defer)
- Best practices (patterns that work in practice)
- Examples (keep them current)
What Doesn’t Get Evolved?
- Preamble thinking (timeless)
- Design rules (timeless)
- Command structure (breaking change, very rare)
- Commit discipline (timeless)
- Testing standards (timeless)
Last Updated: 2026-02-09 Version: 1.0 (Foundation Release)
Self-improvement is how we stay relevant. When Claude evolves, we evolve. When users teach us better patterns, we implement them. This playbook is never “done”-it’s always improving.