Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
60 changes: 60 additions & 0 deletions definitions/create-expert/perstack.toml
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,66 @@
# β”œβ”€β”€ @create-expert/test-expert β€” single query execution (pure executor, no evaluation)
# └── @create-expert/verify-test β€” artifact inspection + execution + definition review
# =============================================================================
#
# =============================================================================
# Design Principles
#
# 1. Built-in Verification
# - Verifier EXECUTES and TESTS artifacts β€” it is not a code reviewer.
# - One verifier per team, not one per executor.
# - Verifier must be direct child of coordinator, not nested under executor.
# - Verifier needs `exec` in pick list. Without it, verification degrades
# to file reading, which cannot catch runtime failures.
#
# 2. Instruction Quality via Binary Checks
# - Subjective self-checks ("would removing this make output worse?")
# do not work β€” the LLM always judges its own output as necessary.
# - Binary checks work: "code block present?" "library name present?"
# have unambiguous yes/no answers with clear remediation actions.
# - Pattern: structural checks (delegates array, pick list) have always
# been followed. Content checks must follow the same binary pattern.
#
# 3. Domain Agnosticism
# - create-expert must produce experts for ANY domain β€” coding, writing,
# research, design, operations β€” not just software development.
# - Rules and examples in the pipeline must use domain-neutral language.
# - Coding-specific vocabulary in rules biases generated experts toward
# code-only solutions even when the domain is broader.
#
# 4. Plan ↔ write-definition Boundary
# - Plan outputs role assignments and domain knowledge, NOT instruction
# content. Architecture section = tree structure only (name, one-line
# purpose, role). Domain Knowledge section = raw material for
# instructions.
# - write-definition composes instructions by selecting relevant domain
# constraints per expert. It must NOT copy plan details wholesale.
# - Without this boundary, plan bloat leaks directly into instructions.
#
# 5. Failure Conditions
# - Not the inverse of success criteria β€” they are hard reject rules
# derived from deeply understanding the domain.
# - Each must specify: what is wrong, which expert caused it, and
# where to restart.
# - These go into the verifier's instruction so it knows what to reject.
#
# 6. Instruction Content = Domain Constraints Only
# - An instruction should contain ONLY what the LLM cannot derive on
# its own: business rules, quality bars, domain-specific anti-patterns,
# completion criteria, priority rules.
# - Everything else β€” implementation techniques, library choices, file
# structures, well-known algorithms β€” is noise that dilutes the signal.
# - Enforced by a 6-item binary checklist (no code blocks, no library
# names, no file paths, no procedures, no technique explanations,
# ≀15 lines for non-coordinator experts).
#
# 7. Brevity = Focus
# - Verbose instructions dilute LLM attention. The more you write, the
# less each sentence weighs. Short, dense instructions outperform
# long, thorough ones because the LLM concentrates on what matters.
# - This applies at every level: plan.md, instruction fields, and this
# pipeline's own definitions. If it can be said in fewer words, it
# should be.
# =============================================================================

# =============================================================================
# create-expert β€” Coordinator
Expand Down
Loading