A step-by-step walkthrough of the commands,
workflows, and daily patterns that make it work.
Before You Start
Most of the planning happens before you write a line of code. This is the suggested approach — research first, validate second, build last.
Use Claude Deep Research, Gemini, or manual reading to explore your domain. Save each session as a markdown file. Note key findings, competing approaches, and open questions.
Happens outside the framework. Pure ideation.
Paste your research findings into /deliberate. The agent panel evaluates from architecture, security, UX, and anti-groupthink perspectives. One deliberation per major theme.
Bridges external research into framework-captured reasoning.
Use /discover-projects to find reference implementations on GitHub, then /analyze-project to score their patterns. If 3+ projects implement a pattern well, that’s strong evidence.
Grounds research against actual working code.
Run /deliberate again with all evidence — research, specialist findings, and project analysis scores. Produce ADRs documenting your architectural decisions with full alternatives considered.
Synthesizes everything into traceable decisions.
Run /plan to produce a structured spec. Specialists review it. Once approved, /build_module takes over and builds it with integrated quality gates.
Getting Started
The full framework infrastructure: 11 agents, 16 commands, 7 hooks, and the four-layer capture stack. Add your application code to src/ and start building.
Already have a codebase? Use /onboard instead of cloning. It runs the “takeover” protocol:
Daily Pattern
Build features, fix bugs, refactor. The framework doesn't change how you write code — it changes what happens after.
/reviewThe facilitator assesses risk, assembles specialists, and produces a structured review with a verdict. Takes 30-90 seconds.
Blocking findings must be fixed. Advisory findings are recommendations. The review report tells you which is which.
Run python scripts/quality_gate.py or just commit — the pre-commit hook runs it automatically.
If the quality gate passes and the review is clean, commit your code. The entire review is captured in the discussion archive.
For complex changes, run /walkthrough and /quiz to verify you understand the AI-generated code.
Core Command
The facilitator creates a discussion, assesses risk level, selects 2-5 specialists, dispatches them in parallel, collects findings, synthesizes a verdict, writes a report to docs/reviews/, and seals the discussion.
2-3 agents. QA + 1 domain specialist. Ensemble mode.
3-4 agents. QA + Architecture + domain. Structured Dialogue mode.
4-5 agents. Full panel including Security + Independent Perspective.
Approve · Approve with changes · Request changes · Reject
Building Features
The spec is reviewed by specialists before any code is written. Catching architecture mistakes at the spec stage is 10x cheaper than fixing them in code.
When a build task creates a new module, touches security code, or changes database schema, 2 specialists automatically review the code mid-build. Max 2 rounds per checkpoint.
After building, you get a summary with: tasks completed, checkpoint results, unresolved concerns (if any), and a recommendation for the final /review.
Decision Making
Use /deliberate when you face a decision with multiple valid options and non-obvious trade-offs. The specialists bring different professional perspectives that surface considerations you might miss alone.
Each specialist analyzes the question from their domain (security, performance, architecture, etc.) without seeing others' answers.
Specialists see each other's findings and can refine their positions. Disagreements are surfaced, not hidden.
The facilitator synthesizes all perspectives into a recommendation with explicit trade-offs. You make the final call.
Every deliberation is captured in discussions/ with full event streams. Six months later, you can read exactly why a decision was made.
Quick Reference
Research first (Phase 1), then /deliberate to pressure-test, then /plan to create a spec.
Run /deliberate with both options. The specialist panel will surface trade-offs you might miss.
/discover-projects to find reference repos, then /analyze-project to score their patterns for your project.
/review src/ — the specialist panel reviews from security, architecture, performance, and quality perspectives.
Write a regression test tagged @pytest.mark.regression, add it to the regression ledger, then /review the fix.
/walkthrough for a guided explanation, then /quiz to verify you actually get it before shipping.
/onboard maps your codebase, reverse-engineers decisions, and creates a stabilization plan with a debt ledger.
/retro analyzes all discussions, surfaces recurring patterns, and proposes process improvements. Run /batch-evaluate to audit pending adoptions.
Learning & Growth
Generates a guided reading path through code you (or AI) wrote. Explains decisions, trade-offs, and how components interact. Progressive disclosure from mental model to implementation details.
Bloom's taxonomy quiz on code you're about to ship. Tests understanding at 4 levels: recall, application, analysis, and evaluation. Includes debug scenarios and change-impact questions.
Points the specialist team at an external project to evaluate patterns worth adopting. Produces a scored recommendation report. Patterns scoring 20+/25 are recommended.
Searches GitHub for interesting projects to analyze. Filters by topic, language, or keywords. Checks for AI integration artifacts and ranks candidates for /analyze-project.
Continuous Improvement
Run at the end of each sprint. Queries all discussions from the period, identifies recurring patterns, evaluates adopted patterns (PENDING → CONFIRMED or REVERTED), and proposes process adjustments.
The big one. Quarterly assessment of framework effectiveness: agent scoring, architectural drift, rule updates, decision churn. Drives framework-level evolution.
Quick health check on all 5 pipeline layers. Reports on discussion volume, SQLite index completeness, findings extraction, pattern clustering, and curated memory currency.
/review captures findings. Agent reflections after each discussion note what worked and what didn't.
/retro aggregates findings, evaluates adoptions, surfaces stale advisories, and proposes adjustments.
/meta-review assesses the framework itself. Which agents are valuable? Which rules need updating? Is there architectural drift?
When a pattern proves valuable across multiple reviews, promote it to curated memory with /promote. Requires your explicit approval (Principle #7).
Reviews all PENDING pattern adoptions from /analyze-project runs. Checks whether each pattern was actually implemented, verifies evidence, and presents verdicts for your approval.
Safety Net
Formatting (ruff format) → Linting (ruff check) → Tests (pytest) → Coverage (≥ 80%) → ADR completeness → Review existence → Regression ledger
Scans for 12 patterns (API keys, JWT, AWS, PATs) before any file write. Blocks the write if secrets are detected.
Prevents concurrent agent edits to the same file. Locks auto-expire after 120 seconds.
Every Python file is auto-formatted with ruff after every edit. You never commit unformatted code.
Blocks direct pushes to main. Create a feature branch, open a PR, then merge. Keeps your history clean.
BUILD_STATUS.md is auto-saved before context compaction and auto-loaded on resume. Work-in-progress survives sessions.
Release & Lineage
Full release workflow: quality gate verification, testing checklist, version bump, changelog generation, and rollback strategy. Everything you need to ship with confidence.
Shows how your project relates to the upstream template. Detects drift, validates the manifest, and reports divergence distance. Intentional divergences can be pinned as traits.
When you fork this template for a real project, the framework-lineage.yaml manifest tracks:
The Steward is the framework's institutional memory for genealogy. It knows where your project came from, how far it has diverged, and which divergences are intentional vs. accidental drift.
Reference
The framework adds structure to AI-assisted development
without slowing you down.
AI-Native Agentic Development Framework v2.1 · Diviner Dojo
diviner-dojo@gmail.com