AI agents write code fast. Without structure, they ship chaos.
ADD is coordinated agent swarms — test-writers, implementers, reviewers, deployers — that ship verified software as a team. Spec-driven. Test-first. Independently verified. Human-validated.
Website · Install · Quick Start · Agent Teams · Human-in-the-Loop · For PMs · Learning
AI code generation has changed how software gets built, but development practices haven't kept up. Developers and agents operate without structure, leading to specification drift, unpredictable quality, lost knowledge, and unclear handoffs.
ADD brings discipline to AI-native development the same way TDD brought discipline to testing.
TDD gave us tests before code. BDD gave us behavior before tests. ADD gives us coordinated agent teams before everything.
ADD is a structured SDLC methodology where AI agents do the development work — writing tests, implementing features, reviewing code, deploying — while humans architect, decide, and verify the user experience. It's not a tool. It's a way of working.
The six principles:
- Specs before code — Every feature starts as a specification. No spec, no code.
- Tests before implementation — Strict TDD: RED → GREEN → REFACTOR → VERIFY.
- Trust but verify — Sub-agents work autonomously. Orchestrators independently verify. Humans validate UX via screenshots.
- Structured collaboration — Interviews, away mode, decision points — humans and agents have clear protocols.
- Environment awareness — Skills adapt to where you're deploying: local, staging, or production.
- Continuous learning — Agents accumulate knowledge. Retrospectives propagate lessons across projects.
Marketplace (recommended):
claude plugin marketplace add MountainUnicorn/add
claude plugin install add@add-marketplaceSource install:
claude plugin install --source https://github.com/MountainUnicorn/addUpdate an existing install:
claude plugin update add@add-marketplaceNo runtime dependencies. No build step. ADD is pure markdown and JSON — it runs entirely within Claude Code's plugin system.
/add:initADD interviews you about your project (product vision, tech stack, team size, deployment model) and scaffolds the project structure. Takes about 5 minutes.
What gets created:
your-project/
├── .add/
│ ├── config.json # Project configuration
│ ├── learnings.md # Agent knowledge base
│ └── cycles/ # Work cycle tracking
├── docs/
│ ├── prd.md # Product Requirements Document
│ ├── plans/ # Implementation plans
│ └── milestones/ # Milestone tracking
├── specs/ # Feature specifications
└── CLAUDE.md # Project context for Claude
/add:spec "user authentication"ADD runs a structured interview (6-10 questions, ~5 min) and generates a complete feature specification with acceptance criteria, test cases, data models, and edge cases.
/add:plan specs/user-authentication.mdTransforms the spec into an actionable implementation plan with task breakdown, effort estimation, dependency mapping, and risk assessment.
/add:tdd-cycle specs/user-authentication.mdExecutes the full TDD cycle: writes failing tests from the spec (RED), implements minimal code to pass them (GREEN), refactors for quality, then independently verifies everything.
/add:verifyRuns up to 5 levels of quality checks: lint, type checking, unit tests, coverage, and spec compliance. For UI projects, humans validate the user experience via screenshot review.
/add:deployEnvironment-aware deployment with pre-deploy verification and post-deploy smoke tests.
ADD doesn't use a single agent. It dispatches specialized sub-agents — each with scoped tool permissions — then independently verifies their work. Trust but verify.
Each sub-agent is isolated — test-writers can't deploy, reviewers can't edit code. At Beta+ maturity, agents work in parallel via git worktrees: 2-5 concurrent agents with WIP limits that scale with maturity (POC=1, Alpha=2, Beta=4, GA=5).
ADD defines three engagement modes. You choose how much autonomy agents get.
- Guided — Human approves each step. Every file change reviewed before commit. Best for POC maturity and unfamiliar codebases.
- Balanced (default) — Agents execute TDD cycles freely within spec boundaries. Pause at ambiguity or architecture forks. Best for Alpha/Beta.
- Autonomous — Human defines scope and boundaries, then walks away. Agents execute full cycles, commit, verify. Best for GA maturity.
Screenshot validation is part of every mode. After agents build, humans review the actual user experience. Claude Code reads images natively — no Playwright or browser automation required.
The /add:away and /add:back commands power autonomous mode:
- Define scope and boundaries
/add:away— agent builds a work plan- Agent executes autonomously (TDD cycles, commits, verify — all decisions logged)
/add:back— full briefing: what shipped, what's blocked, what needs your decision
AI agents are writing your team's code. ADD gives you the governance, traceability, and progress visibility you need — without slowing anyone down.
- Progress visibility — Hill charts show where every feature stands — still figuring out or actively executing.
- Full traceability — Every line of code traces back through tests, to a spec, to a PRD requirement.
- One dial for rigor — The maturity lifecycle (POC → GA) governs all process. No process debates — just turn the dial.
- Quality gates — 5 automated gates at every stage. Agents can't skip verification.
- Human at the right moments — Agents handle execution. Humans handle judgment — architecture, scope, production approvals.
- Compounding intelligence — Every project makes the next one faster via cross-project knowledge.
ADD organizes work in four nested levels. Each level adds detail.
ROADMAP (Now / Next / Later — no fake dates)
└── MILESTONE (Hill chart: uphill figuring out → downhill executing)
└── CYCLE (Scope-boxed batch — ends when validation passes, not a timer)
└── EXECUTION (TDD: RED → GREEN → REFACTOR → VERIFY)
No fake dates — milestones use Now/Next/Later horizons that reflect actual priority. No sprints — cycles are scope-boxed and end when validation criteria are met. Hill charts show whether features are still being figured out (uphill: SHAPED → SPECCED → PLANNED) or being executed (downhill: IN_PROGRESS → VERIFIED → DONE).
Every ADD project declares a maturity level that governs all process rigor. This is the master control:
POC — A paragraph PRD. Optional specs and TDD. Pre-commit gate only. Serial execution.
Alpha — 1-page PRD. Critical-path specs and TDD. Adds CI gate. Up to 2 parallel agents.
Beta — Full PRD template. All specs required, TDD enforced. Adds pre-deploy gate. 2-4 parallel agents.
GA — Full PRD + architecture docs. Acceptance criteria on every spec. Strict TDD. All 5 quality gates. 3-5 agents via worktrees.
A POC project gets almost no ceremony. A GA project gets exhaustive verification. Promotion happens deliberately — triggered by cycle completion or retrospectives when gap analysis shows readiness.
No code without a spec. ADD enforces a document hierarchy — PRD → Spec → Plan → Tests → Code — so every line of implementation traces back to an approved requirement. The /add:spec command interviews you about the feature and generates acceptance criteria, test cases, data models, API contracts, and edge cases.
ADD enforces the TDD cycle at every level. The /add:tdd-cycle skill orchestrates sub-agents: one writes failing tests (RED), another writes minimal code to pass them (GREEN), a reviewer identifies refactoring opportunities, and a verifier independently confirms everything works. Commits follow the pattern: test: → feat: → refactor:.
For projects with a UI, humans validate the actual user experience by reviewing screenshots. Claude Code reads images natively — no Playwright, Puppeteer, or browser automation required. The human takes screenshots in their browser and shares them for agent review. This is a core part of ADD's trust-but-verify loop.
ADD formalizes five engagement modes between humans and agents:
- Spec interview — Human leads, agent refines. For creating PRDs and specs.
- Quick check — Agent asks, human answers. For clarifying a detail mid-work.
- Decision point — Human decides, agent implements. For architectural choices.
- Review gate — Human reviews, agent awaits. For approving artifacts.
- Status pulse — Agent reports, human adjusts. For progress check-ins.
Interviews follow a structured protocol: questions are estimated upfront ("I have 8 questions, about 5 minutes"), asked one at a time, and prioritized so the most critical questions come first.
Going to lunch? Stepping into a meeting? Tell the agent:
/add:away "back in 2 hours"ADD assesses available work, presents an autonomous work plan for your approval, and gets to work. When you return:
/add:backYou get a briefing: what was completed, what's in progress, what needs your decision. The agent won't deploy to staging/production or start features without specs while you're gone.
Five checkpoints catch issues progressively earlier:
- Pre-commit — Lint + formatting (before write)
- Pre-push — Type checking (before push)
- CI — Unit tests + coverage (on push)
- Pre-deploy — Spec compliance + integration tests (before deploy)
- Post-deploy — Smoke tests + monitoring (after deploy)
At Beta/GA maturity, ADD coordinates multiple agents working in parallel:
- Git worktrees for full isolation between agents
- File reservations to prevent conflicts
- Merge sequence (infrastructure first, then features)
- WIP limits to prevent coordination overhead
- Trust-but-verify — orchestrators independently check sub-agent work
ADD uses cycles (not sprints) — scope-boxed batches of work that end when validation criteria are met, not when a timer expires.
/add:cycle --plan # Plan the next batch of work
/add:cycle --status # Check progress, update hill chart
/add:cycle --complete # Close cycle, capture learningsFeatures in a cycle progress through positions: SHAPED → SPECCED → PLANNED → IN_PROGRESS → VERIFIED → DONE. Hill charts visualize whether work is still being figured out (uphill) or being executed (downhill).
This is what makes ADD compound over time. Agents don't start from zero on each project.
Agents read all three tiers before starting any task. More specific tiers take precedence:
- Tier 1: Plugin-Global (
knowledge/global.md) — Universal ADD best practices. Ships with ADD, read-only. - Tier 2: User-Local (
~/.claude/add/library.md) — Your cross-project wisdom. Promoted during/add:retro. - Tier 3: Project-Specific (
.add/learnings.md) — This project's discoveries. Auto-populated at checkpoints.
Tier 1 ships with ADD — curated best practices about agent coordination, away mode, and methodology. Every ADD user benefits immediately.
Tier 2 follows you across projects (machine-local, not committed):
~/.claude/add/
├── profile.md # Your preferences, conventions, working style
├── library.md # Accumulated wisdom from all your projects
└── projects.json # Index of ADD-managed projects
Tier 3 is committed to git — every team member (human or agent) benefits:
.add/learnings.md
├── Architecture decisions and rationale
├── What worked / what didn't
├── Patterns discovered during development
└── Tool and framework quirks
Project A: discovers "UUID columns must be type uuid, not text"
→ Stored in .add/learnings.md (Tier 3: project-level)
→ Promoted to ~/.claude/add/library.md during /add:retro (Tier 2: cross-project)
Project B: agent reads library before implementing database schema
→ Finds UUID pattern, applies it automatically
→ No one repeats the mistake
ADD maintainers: universal methodology insights
→ Promoted to knowledge/global.md (Tier 1: plugin-global)
→ Every ADD user benefits on next install/update
Knowledge is captured automatically — no human effort required:
- After every
/add:verify— what passed, what failed, why - After every TDD cycle — patterns in test writing and implementation
- After every deployment — environment-specific discoveries
- After every away session — what the agent learned working autonomously
- During
/add:retro— human + agent reflect, promote learnings to cross-project library
/add:retroTwo modes:
- Interactive: Human and agent reflect together (~5 questions). Best insights come from this.
- Agent summary: Quick non-interactive pulse check. Good for solo developers.
Retrospectives update .add/learnings.md, promote patterns to ~/.claude/add/library.md, archive the retro for future reference, and suggest process changes.
ADD is intentionally simple:
- No runtime dependencies — Pure markdown and JSON files
- No build step — Install and use immediately
- No backend — Everything lives in your git repo (
.add/) or locally (~/.claude/add/) - No vendor lock-in — Standard markdown specs and plans work with any tool
- Plugin format — Claude Code
.claude-plugin/plugin.jsonmanifest
The entire plugin is ~60 files of markdown, JSON, and templates. It runs entirely within Claude Code's plugin system using commands, skills, rules, hooks, knowledge, and templates.
ADD's core is zero-dependency, but some features benefit from optional tools:
- Screenshot validation — Used by
/add:verifyand human review. Requires Claude Code's built-in image reading. Fallback: agents describe UI state in text. - Image generation — Used by
/add:infographicand/add:brand-update. Requires an image generation MCP tool (e.g., Vertex AI Imagen). Fallback: SVG generation.
Screenshot validation is a core part of ADD's trust-but-verify loop for projects with a UI. After agents build, humans review screenshots of the actual user experience. Claude Code reads images natively — no Playwright, Puppeteer, or browser automation required. The human takes screenshots in their browser and shares them for review.
Image generation enhances branding and documentation workflows. ADD auto-detects whether an image generation MCP tool is available and adapts accordingly. Without it, /add:infographic produces SVG infographics instead of raster images. No functionality is lost — just a different output format.
Solo developers using Claude Code for daily work who want structure without overhead. ADD scales down — a POC project gets lightweight specs and optional TDD. You get the benefits of methodology without the ceremony.
Small teams (2-10 engineers) looking to standardize how they work with AI agents. ADD gives your team a shared language: specs, plans, quality gates, and a maturity model that grows with your project.
Larger teams and enterprises scaling AI agents across multiple projects. ADD's cross-project learning system means agents get smarter over time. Patterns discovered on Project A automatically inform Project B. The maturity lifecycle gives leadership visibility into project rigor.
Open source maintainers who want contributors (human or AI) to follow consistent practices. ADD's spec-driven approach means every PR traces back to a specification.
ADD works on existing projects, not just new ones. When you run /add:init on an existing codebase:
- Discovery — ADD scans your project structure, detects test frameworks, linters, and conventions
- Non-destructive setup — Adds
.add/alongside your existing structure, never replaces - Maturity assessment — Suggests a starting maturity level based on existing rigor
- Catch-up spike — Identifies gaps and creates a remediation cycle
- Retroactive specs — Seeds learnings from git history
Existing conventions are respected. ADD layers on top of what you already have.
ADD adapts to your deployment reality:
- Tier 1 — Local development only. Personal projects, plugins, CLIs.
- Tier 2 — Local + production. SaaS apps, APIs with a prod server.
- Tier 3 — Full pipeline (dev → staging → prod). Enterprise applications, team projects.
Quality gates, deployment skills, and test matrices all adjust based on your tier.
Commands — 9 slash commands
| Command | Purpose |
|---|---|
/add:init |
Bootstrap ADD in your project via structured interview |
/add:spec |
Create a feature specification through interview |
/add:cycle |
Plan, track, and complete work cycles |
/add:brand |
View project branding — accent color, palette, drift detection |
/add:brand-update |
Update branding materials and audit artifacts |
/add:changelog |
Generate/update CHANGELOG.md from conventional commits |
/add:away |
Declare absence — get autonomous work plan |
/add:back |
Return from absence — get briefing |
/add:retro |
Run a retrospective — capture and promote learnings |
Skills — 9 workflow skills
| Skill | Purpose |
|---|---|
/add:tdd-cycle |
Complete RED → GREEN → REFACTOR → VERIFY cycle |
/add:test-writer |
Write failing tests from spec (RED phase) |
/add:implementer |
Write minimal code to pass tests (GREEN phase) |
/add:reviewer |
Code review for spec compliance (read-only) |
/add:verify |
Run quality gates — lint, types, tests, coverage, spec compliance |
/add:plan |
Create implementation plan from spec |
/add:optimize |
Performance optimization pass |
/add:deploy |
Environment-aware deployment with verification |
/add:infographic |
Generate project infographic SVG from PRD + config |
Rules — 11 auto-loaded behavioral rules
| Rule | What it enforces |
|---|---|
spec-driven |
No code without a spec |
tdd-enforcement |
Strict RED → GREEN → REFACTOR → VERIFY cycle |
human-collaboration |
Interview protocol, engagement modes, away/back |
agent-coordination |
Trust-but-verify, sub-agent isolation, parallel execution |
source-control |
Feature branches, conventional commits, TDD commit pattern |
environment-awareness |
Tier-based behavior (local / local+prod / full pipeline) |
quality-gates |
5-level gate system from pre-commit to post-deploy |
learning |
Automatic checkpoints, knowledge persistence, retro integration |
project-structure |
Standard .add/ layout, cross-project persistence paths |
maturity-lifecycle |
Master dial — governs ALL ADD behavior per maturity level |
design-system |
Silicon Valley Unicorn aesthetic for all generated visuals |
Project Structure Reference
After /add:init, your project gets this structure:
your-project/
├── .add/ # ADD state (git-committed)
│ ├── config.json # Project config
│ ├── learnings.md # Agent knowledge base
│ └── cycles/ # Work cycle tracking
├── docs/
│ ├── prd.md # Product Requirements Document
│ ├── plans/ # Implementation plans
│ └── milestones/ # Milestone tracking
├── specs/ # Feature specifications
├── tests/
│ └── screenshots/ # Visual verification
└── CLAUDE.md # Project context for Claude
Cross-project persistence (machine-local, not committed):
~/.claude/add/
├── profile.md # Your preferences
├── library.md # Cross-project wisdom
└── projects.json # Project index
- v0.1.0 — Complete. Core infrastructure — 6 commands, 8 skills, 10 rules, 10 templates.
- v0.2.0 — Complete. Branding system, image gen detection, auto-changelog, infographic generation.
- v0.4.0 — Next. Adoption & polish —
/add:init --adopt, enhanced interviews, cross-project sync. - v1.0.0 — Planned. Marketplace ready — CI/CD hooks, advanced learnings, team profiles.
MIT
ADD something to your development.
Built with ADD, using ADD. Dog-fooded from day one.
getADD.dev · GitHub