Use this template when you want to turn a real-repo pass into a stable, reviewable case study.
The purpose is not marketing copy. The purpose is to preserve enough structured evidence to answer:
- what repository and host were used
- what task family was repeated
- what ExperienceEngine actually did
- whether the intervention created net value
- what should be kept, changed, or retired
Create a case study when at least one of these is true:
- the repository has produced repeated interventions in the same task family
- a Pack has been published from real usage
- compiled host-facing assets have been reviewed or deployed
- benchmark and evidence artifacts show a meaningful trend
Do not create a case study for one-off wiring checks or toy-repo smoke tests.
Before writing a case study, gather:
- repository name or anonymized label
- host used:
OpenClawClaude CodeCodex
- primary task family
- evaluation mode:
liveshadowholdout
- latest benchmark report
- latest evaluation bundle
- latest evidence package
- any published Pack ids used in the pass
- any deployed host artifacts used in the pass
Use the following sections.
Capture:
- repository or anonymized system name
- host used
- date range of the pass
- primary problem area
- why ExperienceEngine was relevant
Example prompts:
- What kind of repeated work was happening?
- Why was this repo a good candidate for ExperienceEngine?
Describe the starting situation before the current pass:
- repeated failures, repeated debugging, or repeated repair patterns
- whether prior experience already existed
- whether the repo already had Packs or compiled artifacts
Keep this concrete. Prefer task-family facts over general impressions.
Describe what was actually used:
inspect --lasthelped/harmed- Pack workflow used or not used
- compile targets used
- deploy targets used
If Packs were involved, record:
- pack id
- version
- whether the pack was newly published or reused
Pull the key numbers from artifacts:
deliveryRatesuppressionRatehelpfulRateharmfulRatenetHelpfulRateverdictsuggestedMode
Also record:
- intervention mode
- injected node count
- any feedback or attribution reason if relevant
Summarize the most useful interventions:
- which task family benefited
- which node patterns helped
- whether the guidance was specific enough to change the agent's execution path
Prefer 2 to 5 concrete observations over a long generic summary.
Record:
- noisy interventions
- drifted deployed artifacts
- weak or confusing guidance
- cases where explicit feedback was necessary
If something should cool or retire, say so directly.
End the case study with a clear next-step decision:
- keep
live - move to
shadow - move to
holdout - publish / extend Pack
- redeploy compiled artifact
- cool / retire specific experience
# Case Study: <repo-or-anonymized-name>
## Context
- Host:
- Date range:
- Primary task family:
- Why ExperienceEngine was used:
## Initial State
- Repeated issue pattern:
- Existing nodes or packs:
- Existing compiled/deployed artifacts:
## ExperienceEngine Usage
- Commands / surfaces used:
- Packs involved:
- Compiled targets involved:
- Deployments involved:
## Evidence Summary
- deliveryRate:
- suppressionRate:
- helpfulRate:
- harmfulRate:
- netHelpfulRate:
- verdict:
- suggestedMode:
## What Helped
- ...
## What Caused Friction
- ...
## Decision
- ...Recommended options:
- keep the generated
case-study.mdfrom evaluation artifacts - or write a curated case study under
docs/development/ - link back to the matching benchmark report, evaluation bundle, and evidence package
If the repository name is sensitive, anonymize it consistently and keep the artifact paths separately.
A good case study should be:
- based on real repository work
- tied to real artifacts
- specific about what helped and what harmed
- short enough to re-read quickly
- actionable enough to inform the next Pack / deploy / mode decision