Practical Guide

How to Use the
Agentic Development Framework

A step-by-step walkthrough of the commands,
workflows, and daily patterns that make it work.

Navigate: arrow keys

From Idea to Code: Five Phases

Most of the planning happens before you write a line of code. This is the suggested approach — research first, validate second, build last.

1. Seed 2. Anchor 3. Scout 4. Reconcile 5. Specify

1. Seed — Research Your Idea

Use Claude Deep Research, Gemini, or manual reading to explore your domain. Save each session as a markdown file. Note key findings, competing approaches, and open questions.

Happens outside the framework. Pure ideation.

2. Anchor — Pressure-Test with Specialists

Paste your research findings into /deliberate. The agent panel evaluates from architecture, security, UX, and anti-groupthink perspectives. One deliberation per major theme.

Bridges external research into framework-captured reasoning.

3. Scout — Find Real-World Validation

Use /discover-projects to find reference implementations on GitHub, then /analyze-project to score their patterns. If 3+ projects implement a pattern well, that’s strong evidence.

Grounds research against actual working code.

4. Reconcile — Make Decisions

Run /deliberate again with all evidence — research, specialist findings, and project analysis scores. Produce ADRs documenting your architectural decisions with full alternatives considered.

Synthesizes everything into traceable decisions.

5. Specify — Create the Executable Plan

Run /plan to produce a structured spec. Specialists review it. Once approved, /build_module takes over and builds it with integrated quality gates.

This is where
coding begins.

Setup in 3 Minutes

# 1. Clone and install
cd agent_framework_template
pip install -r requirements.txt
# 2. Initialize the metrics database
python scripts/init_db.py
# 3. Run tests to verify
pytest tests/ -v

Prerequisites

Required:
Python 3.11+ — runtime
Git — version control, hooks, commit gates
Claude Code — the CLI that runs all agents and commands
Recommended:
VS Code — editor with Claude Code integration
GitHub CLI (gh) — for /discover-projects, /ship, PR workflows

What You Get

The full framework infrastructure: 11 agents, 16 commands, 7 hooks, and the four-layer capture stack. Add your application code to src/ and start building.

For Existing Projects

Already have a codebase? Use /onboard instead of cloning. It runs the “takeover” protocol:

1. Maps your existing code structure
2. Reverse-engineers ADRs from your architecture
3. Proposes coding standards based on your patterns
4. Creates a stabilization plan and debt ledger

The Typical Development Flow

Write Code /review Fix Findings Quality Gate Commit

1. Write Your Code

Build features, fix bugs, refactor. The framework doesn't change how you write code — it changes what happens after.

2. Run /review

The facilitator assesses risk, assembles specialists, and produces a structured review with a verdict. Takes 30-90 seconds.

3. Address Findings

Blocking findings must be fixed. Advisory findings are recommendations. The review report tells you which is which.

4. Quality Gate

Run python scripts/quality_gate.py or just commit — the pre-commit hook runs it automatically.

5. Commit

If the quality gate passes and the review is clean, commit your code. The entire review is captured in the discussion archive.

6. Education Gate

For complex changes, run /walkthrough and /quiz to verify you understand the AI-generated code.

/review — Multi-Agent Code Review

# Review specific files
/review src/routes.py src/models.py
# Review an entire directory
/review src/
# Review with context
/review src/auth.py # "focus on the new JWT validation"

What Happens Behind the Scenes

The facilitator creates a discussion, assesses risk level, selects 2-5 specialists, dispatches them in parallel, collects findings, synthesizes a verdict, writes a report to docs/reviews/, and seals the discussion.

Risk-Based Specialist Assembly

LOW

Config, docs, simple fixes

2-3 agents. QA + 1 domain specialist. Ensemble mode.

MED

New features, refactoring

3-4 agents. QA + Architecture + domain. Structured Dialogue mode.

HIGH

Security, architecture changes

4-5 agents. Full panel including Security + Independent Perspective.

Verdicts

Approve · Approve with changes · Request changes · Reject

/plan + /build_module — Spec-Driven Development

Step 1: Plan the feature

/plan "Add user authentication with JWT tokens"
# Produces a structured spec with:
# - Task breakdown
# - File list (new + modified)
# - Architecture decisions
# - Risk assessment
# - Specialist review of the plan
# Waits for your approval before proceeding

Step 2: Build from the spec

/build_module docs/sprints/auth-spec.md
# Executes the spec task-by-task
# Mid-build checkpoint reviews fire automatically
# Tests run after each task
# Education gate activates at the end

Why Spec-First?

The spec is reviewed by specialists before any code is written. Catching architecture mistakes at the spec stage is 10x cheaper than fixing them in code.

Mid-Build Checkpoints

When a build task creates a new module, touches security code, or changes database schema, 2 specialists automatically review the code mid-build. Max 2 rounds per checkpoint.

The Build Summary

After building, you get a summary with: tasks completed, checkpoint results, unresolved concerns (if any), and a recommendation for the final /review.

/deliberate — Structured Multi-Agent Discussion

/deliberate "Should we use SQLAlchemy ORM or raw SQL
  for the new reporting module?"
/deliberate "What's the best approach for handling
  file uploads larger than 100MB?"
/deliberate "Should we split the monolith into
  microservices or keep it modular?"

When to Deliberate

Use /deliberate when you face a decision with multiple valid options and non-obvious trade-offs. The specialists bring different professional perspectives that surface considerations you might miss alone.

What You Get

1

Independent Analysis

Each specialist analyzes the question from their domain (security, performance, architecture, etc.) without seeing others' answers.

2

Cross-Pollination

Specialists see each other's findings and can refine their positions. Disagreements are surfaced, not hidden.

3

Synthesis & Recommendation

The facilitator synthesizes all perspectives into a recommendation with explicit trade-offs. You make the final call.

Every deliberation is captured in discussions/ with full event streams. Six months later, you can read exactly why a decision was made.

What Do I Do When…

“I have an idea but don’t know where to start”

Research first (Phase 1), then /deliberate to pressure-test, then /plan to create a spec.

“I’m choosing between two approaches”

Run /deliberate with both options. The specialist panel will surface trade-offs you might miss.

“I want to see how others solved this”

/discover-projects to find reference repos, then /analyze-project to score their patterns for your project.

“I wrote code and need a second opinion”

/review src/ — the specialist panel reviews from security, architecture, performance, and quality perspectives.

“I found a bug and fixed it”

Write a regression test tagged @pytest.mark.regression, add it to the regression ledger, then /review the fix.

“AI wrote code I don’t fully understand”

/walkthrough for a guided explanation, then /quiz to verify you actually get it before shipping.

“I have an existing project I want to improve”

/onboard maps your codebase, reverse-engineers decisions, and creates a stabilization plan with a debt ledger.

“The sprint is over — what now?”

/retro analyzes all discussions, surfaces recurring patterns, and proposes process improvements. Run /batch-evaluate to audit pending adoptions.

Commands That Make You Better

/walkthrough education

Generates a guided reading path through code you (or AI) wrote. Explains decisions, trade-offs, and how components interact. Progressive disclosure from mental model to implementation details.

/walkthrough src/routes.py

/quiz assessment

Bloom's taxonomy quiz on code you're about to ship. Tests understanding at 4 levels: recall, application, analysis, and evaluation. Includes debug scenarios and change-impact questions.

/quiz src/routes.py

/analyze-project external

Points the specialist team at an external project to evaluate patterns worth adopting. Produces a scored recommendation report. Patterns scoring 20+/25 are recommended.

/analyze-project tiangolo/fastapi

/discover-projects search

Searches GitHub for interesting projects to analyze. Filters by topic, language, or keywords. Checks for AI integration artifacts and ranks candidates for /analyze-project.

/discover-projects "fastapi multi-agent"

The Framework Improves Itself

/retro sprint

Run at the end of each sprint. Queries all discussions from the period, identifies recurring patterns, evaluates adopted patterns (PENDING → CONFIRMED or REVERTED), and proposes process adjustments.

/retro

/meta-review quarterly

The big one. Quarterly assessment of framework effectiveness: agent scoring, architectural drift, rule updates, decision churn. Drives framework-level evolution.

/meta-review

/knowledge-health diagnostic

Quick health check on all 5 pipeline layers. Reports on discussion volume, SQLite index completeness, findings extraction, pattern clustering, and curated memory currency.

/knowledge-health

The Improvement Cadence

D

Daily

/review captures findings. Agent reflections after each discussion note what worked and what didn't.

S

Sprint

/retro aggregates findings, evaluates adoptions, surfaces stale advisories, and proposes adjustments.

Q

Quarterly

/meta-review assesses the framework itself. Which agents are valuable? Which rules need updating? Is there architectural drift?

/promote — Layer 3

When a pattern proves valuable across multiple reviews, promote it to curated memory with /promote. Requires your explicit approval (Principle #7).

/batch-evaluate — Audit

Reviews all PENDING pattern adoptions from /analyze-project runs. Checks whether each pattern was actually implemented, verifies evidence, and presents verdicts for your approval.

The Quality Gate — Your Last Line of Defense

# Run manually
python scripts/quality_gate.py
# Auto-fix formatting and lint
python scripts/quality_gate.py --fix
# Skip specific checks
python scripts/quality_gate.py --skip-reviews
# It also runs automatically on every git commit
Quality Gate: 7/7 passed

What It Checks

Formatting (ruff format) → Linting (ruff check) → Tests (pytest) → Coverage (≥ 80%) → ADR completeness → Review existence → Regression ledger

Hooks That Protect You

!

Secret Detection

Scans for 12 patterns (API keys, JWT, AWS, PATs) before any file write. Blocks the write if secrets are detected.

🔒

File Locking

Prevents concurrent agent edits to the same file. Locks auto-expire after 120 seconds.

Auto-Format

Every Python file is auto-formatted with ruff after every edit. You never commit unformatted code.

Branch Protection

Blocks direct pushes to main. Create a feature branch, open a PR, then merge. Keeps your history clean.

Session Continuity

BUILD_STATUS.md is auto-saved before context compaction and auto-loaded on resume. Work-in-progress survives sessions.

Shipping and Tracking Your Fork

/ship release

Full release workflow: quality gate verification, testing checklist, version bump, changelog generation, and rollback strategy. Everything you need to ship with confidence.

/ship

/lineage drift

Shows how your project relates to the upstream template. Detects drift, validates the manifest, and reports divergence distance. Intentional divergences can be pinned as traits.

/lineage
Drift status: current (distance: 0)
Pinned traits: 0

Template → Project Relationship

When you fork this template for a real project, the framework-lineage.yaml manifest tracks:

Fork point — When you diverged from the template
Drift status — How far you've diverged
Pinned traits — Intentional divergences you want to keep
Custodian — Who approves framework changes

The Steward Agent

The Steward is the framework's institutional memory for genealogy. It knows where your project came from, how far it has diverged, and which divergences are intentional vs. accidental drift.

Command Cheat Sheet

Every Day

/review src/            # Code review
/walkthrough src/file.py  # Explain code
/quiz src/file.py      # Test understanding

Building Features

/plan "description"      # Create spec
/build_module spec.md   # Build from spec
/deliberate "question"  # Discuss trade-offs

Quality

python scripts/quality_gate.py # Run checks
python scripts/quality_gate.py --fix # Auto-fix

Process Improvement

/retro                  # Sprint retro
/meta-review          # Quarterly review
/knowledge-health     # Pipeline check
/promote               # Promote to Layer 3

External Learning

/discover-projects "topic" # Find repos
/analyze-project owner/repo  # Analyze
/batch-evaluate          # Audit adoptions

Release & Lineage

/ship                    # Release workflow
/lineage                # Drift status
/onboard                # Adopt framework

Start Building

The framework adds structure to AI-assisted development
without slowing you down.

# Clone, install, go
git clone https://github.com/Diviner-Dojo/agent_framework_template
cd agent_framework_template
pip install -r requirements.txt
python scripts/init_db.py
# Your first command
/review src/

AI-Native Agentic Development Framework v2.1 · Diviner Dojo
diviner-dojo@gmail.com