sdd:plan

安装量: 155
排名: #5557

安装

npx skills add https://github.com/neolabhq/context-engineering-kit --skill sdd:plan

Refine Task Workflow Role You are a task refinement orchestrator. Take a draft task file created by /add-task and refine it through a coordinated multi-agent workflow with quality gates after each phase. Goal This workflow command refines an existing draft task through: Parallel Analysis - Research, codebase analysis, and business analysis in parallel Architecture Synthesis - Combine findings into architectural overview Decomposition - Break into implementation steps with risks Parallelize - Reorganize steps for maximum parallel execution Verify - Add LLM-as-Judge verification sections Promote - Move refined task from draft/ to todo/ All phases include judge validation to prevent error propagation and ensure quality thresholds are met. User Input $ARGUMENTS Command Arguments Parse the following arguments from $ARGUMENTS : Argument Definitions Argument Format Default Description task-file Path to task file Required Path to draft task file (e.g., .specs/tasks/draft/add-validation.feature.md ) --continue --continue [stage] None Continue refining from a specific stage. Stage is optional - resolve from context if not provided. --target-quality --target-quality X.X 3.5 Target threshold value (out of 5.0) for judge pass/fail decisions. --max-iterations --max-iterations N 3 Maximum implementation + judge retry cycles per phase before moving to next stage (regardless of pass/fail). --included-stages --included-stages stage1,stage2,... All stages Comma-separated list of stages to include. --skip --skip stage1,stage2,... None Comma-separated list of stages to exclude. --fast --fast N/A Alias for --target-quality 3.0 --max-iterations 1 --included-stages business analysis,decomposition,verifications --one-shot --one-shot N/A Alias for --included-stages business analysis,decomposition --skip-judges - minimal refinement without quality gates. --human-in-the-loop --human-in-the-loop phase1,phase2,... None Phases after which to pause for human verification. --skip-judges --skip-judges false Skip all judge validation checks - phases proceed without quality gates. --refine --refine false Incremental refinement mode - detect changes against git and re-run only affected stages (top-to-bottom propagation). Stage Names (for --included-stages / --skip ) Stage Name Phase Description research 2a Gather relevant resources, documentation, libraries codebase analysis 2b Identify affected files, interfaces, integration points business analysis 2c Refine description and create acceptance criteria architecture synthesis 3 Synthesize research and analysis into architecture decomposition 4 Break into implementation steps with risks parallelize 5 Reorganize steps for parallel execution verifications 6 Add LLM-as-Judge verification rubrics Configuration Resolution Parse $ARGUMENTS and resolve configuration as follows:

Extract task file path (first positional argument, required)

TASK_FILE = first argument that is a file path (must exist in .specs/tasks/draft/)

Parse alias flags first (they set multiple defaults)

if --fast present: THRESHOLD = 3.0 MAX_ITERATIONS = 1 INCLUDED_STAGES = ["business analysis", "decomposition", "verifications"] if --one-shot present: INCLUDED_STAGES = ["business analysis", "decomposition"] SKIP_JUDGES = true

Initialize defaults

THRESHOLD ?= --target-quality || 3.5 MAX_ITERATIONS ?= --max-iterations || 3 INCLUDED_STAGES ?= --included-stages || ["research", "codebase analysis", "business analysis", "architecture synthesis", "decomposition", "parallelize", "verifications"] SKIP_STAGES = --skip || [] HUMAN_IN_THE_LOOP_PHASES = --human-in-the-loop || [] SKIP_JUDGES = --skip-judges || false REFINE_MODE = --refine || false CONTINUE_STAGE = null if --continue [stage] present: CONTINUE_STAGE = stage or resolve from context

Compute final active stages

ACTIVE_STAGES = INCLUDED_STAGES - SKIP_STAGES Context Resolution for --continue When --continue is used without explicit stage: Stage Resolution: Parse the task file for completion markers (e.g., [x] checkboxes) Identify the last completed phase/judge Resume from the next incomplete phase Refine Mode Behavior ( --refine ) When --refine is used: Change Detection: First check file status: git status --porcelain -- Compare current task file against last git commit: git diff HEAD -- This captures both staged and unstaged changes vs HEAD If file is untracked or has no git history, compare against the original task structure Identify which sections have been modified by the user Look for // comment markers indicating user feedback/corrections Top-to-Bottom Propagation: Determine the earliest modified section (highest in document) Re-run only stages that correspond to or come after the modified section Earlier stages (above the modification) are preserved as-is Section-to-Stage Mapping: Modified Section Re-run From Stage Description / Acceptance Criteria business analysis (Phase 2c) Architecture Overview architecture synthesis (Phase 3) Implementation Process / Steps decomposition (Phase 4) Parallelization / Dependencies parallelize (Phase 5) Verification sections verifications (Phase 6) Refine Execution: Skip research (2a) and codebase analysis (2b) unless explicitly requested Pass user modifications and // comments as additional context to agents Agents should incorporate user feedback while preserving unchanged content Example:

User edited the Architecture Overview section

/plan .specs/tasks/todo/my-task.feature.md --refine

Detects Architecture section changed → re-runs from Phase 3 onwards

Skips: research, codebase analysis, business analysis

Runs: architecture synthesis, decomposition, parallelize, verifications

Human-in-the-Loop Behavior Human verification checkpoints occur: Trigger Conditions: After implementation + judge verification PASS for a phase in HUMAN_IN_THE_LOOP_PHASES After implementation + judge + implementation retry (before the next judge retry) At Checkpoint: Display current phase results summary Display generated artifacts with paths Display judge score and feedback Ask user: "Review phase output. Continue? [Y/n/feedback]" If user provides feedback, incorporate into next iteration If user says "n", pause workflow Checkpoint Message Format:


🔍 Human Review Checkpoint - Phase X

Phase: { phase name } Judge Score : ** { score } / { THRESHOLD } threshold Status: ✅ PASS / ⚠️ RETRY { n } / { MAX_ITERATIONS } Artifacts: - { artifact_path_1 } - { artifact_path_2 } Judge Feedback : ** { feedback summary } **Action Required : ** Review the above artifacts and provide feedback or continue.

Continue ? [ Y/n/feedback ] :


Usage Examples

Refine a draft task with all stages

/plan .specs/tasks/draft/add-validation.feature.md

Fast refinement with minimal stages

/plan .specs/tasks/draft/quick-fix.bug.md --fast

Continue from a specific stage

/plan .specs/tasks/draft/complex-feature.feature.md --continue decomposition

High-quality refinement with checkpoints

/plan .specs/tasks/draft/critical-api.feature.md --target-quality 4.5 --human-in-the-loop 2,3 ,4,5,6

Incremental refinement after user edits (re-runs only affected stages)

/plan .specs/tasks/todo/my-task.feature.md --refine Pre-Flight Checks Before starting workflow: Validate task file exists: If REFINE_MODE is false: Check that TASK_FILE exists in .specs/tasks/draft/ If REFINE_MODE is true: Check that TASK_FILE exists in .specs/tasks/todo/ or .specs/tasks/draft/ If not found, show error and exit Parse and display resolved configuration:

Configuration | Setting | Value | |


|

|
|
**
Task File
**
|
{TASK_FILE}
|
|
**
Target Quality
**
|
{THRESHOLD}/5.0
|
|
**
Max Iterations
**
|
{MAX_ITERATIONS}
|
|
**
Active Stages
**
|
{ACTIVE_STAGES as comma-separated list}
|
|
**
Human Checkpoints
**
|
Phase
|
|
**
Skip Judges
**
|
{SKIP_JUDGES}
|
|
**
Refine Mode
**
|
{REFINE_MODE}
|
|
**
Continue From
**
|
{CONTINUE_STAGE} or "Start"
|
Handle
--continue
mode:
If
CONTINUE_STAGE
is set:
Read the task file to get current state
Identify completed phases from task file content
Skip to
CONTINUE_STAGE
(or auto-detected next incomplete stage)
Pre-populate captured values from existing artifacts
Resume workflow from the appropriate phase
Handle
--refine
mode:
If
REFINE_MODE
is true:
Check file status:
git status --porcelain --
M
(staged) or
M
(unstaged) or
MM
(both) → proceed with diff
??
(untracked) → error: "File not tracked by git, cannot detect changes"
Empty output → no changes detected
Run
git diff HEAD --
to get all changes (staged + unstaged) vs last commit
Parse diff to identify modified sections
Collect any
//
comment markers as user feedback
Determine earliest modified section using Section-to-Stage Mapping
Set
ACTIVE_STAGES
to include only stages from the determined starting point onwards
Pass detected changes and user comments as additional context to agents
If no changes detected, inform user: "No changes detected in task file. Edit the file first, then run --refine." and exit
Extract task info from file:
Read task file to extract title and type from filename
Parse frontmatter for title and depends_on
Initialize workflow progress tracking
using TodoWrite:
Only include todos for phases in
ACTIVE_STAGES
. If continuing, mark completed phases as
completed
.
{
"todos"
:
[
{
"content"
:
"Ensure directories exist"
,
"status"
:
"pending"
,
"activeForm"
:
"Ensuring directories exist"
}
,
{
"content"
:
"Phase 2a: Research relevant resources and documentation"
,
"status"
:
"pending"
,
"activeForm"
:
"Researching resources"
}
,
{
"content"
:
"Judge 2a: PASS research quality (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating research"
}
,
{
"content"
:
"Phase 2b: Analyze codebase impact and affected files"
,
"status"
:
"pending"
,
"activeForm"
:
"Analyzing codebase impact"
}
,
{
"content"
:
"Judge 2b: PASS codebase analysis (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating codebase analysis"
}
,
{
"content"
:
"Phase 2c: Business analysis and acceptance criteria"
,
"status"
:
"pending"
,
"activeForm"
:
"Analyzing business requirements"
}
,
{
"content"
:
"Judge 2c: PASS business analysis (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating business analysis"
}
,
{
"content"
:
"Phase 3: Architecture synthesis from research and analysis"
,
"status"
:
"pending"
,
"activeForm"
:
"Synthesizing architecture"
}
,
{
"content"
:
"Judge 3: PASS architecture synthesis (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating architecture"
}
,
{
"content"
:
"Phase 4: Decompose into implementation steps"
,
"status"
:
"pending"
,
"activeForm"
:
"Decomposing into steps"
}
,
{
"content"
:
"Judge 4: PASS decomposition (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating decomposition"
}
,
{
"content"
:
"Phase 5: Parallelize implementation steps"
,
"status"
:
"pending"
,
"activeForm"
:
"Parallelizing steps"
}
,
{
"content"
:
"Judge 5: PASS parallelization (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating parallelization"
}
,
{
"content"
:
"Phase 6: Define verification rubrics"
,
"status"
:
"pending"
,
"activeForm"
:
"Defining verifications"
}
,
{
"content"
:
"Judge 6: PASS verifications (> {THRESHOLD})"
,
"status"
:
"pending"
,
"activeForm"
:
"Validating verifications"
}
,
{
"content"
:
"Move task to todo folder"
,
"status"
:
"pending"
,
"activeForm"
:
"Promoting task"
}
,
{
"content"
:
"Human checkpoint reviews"
,
"status"
:
"pending"
,
"activeForm"
:
"Awaiting human review"
}
]
}
Note:
Filter todos based on configuration:
If
SKIP_JUDGES
is true, omit ALL Judge todos (Judge 2a, 2b, 2c, 3, 4, 5, 6)
If
research
not in
ACTIVE_STAGES
, omit Phase 2a and Judge 2a todos
If
codebase analysis
not in
ACTIVE_STAGES
, omit Phase 2b and Judge 2b todos
If
business analysis
not in
ACTIVE_STAGES
, omit Phase 2c and Judge 2c todos
If
architecture synthesis
not in
ACTIVE_STAGES
, omit Phase 3 and Judge 3 todos
If
decomposition
not in
ACTIVE_STAGES
, omit Phase 4 and Judge 4 todos
If
parallelize
not in
ACTIVE_STAGES
, omit Phase 5 and Judge 5 todos
If
verifications
not in
ACTIVE_STAGES
, omit Phase 6 and Judge 6 todos
If
HUMAN_IN_THE_LOOP_PHASES
is empty, omit human checkpoint todo
Ensure directories exist
:
Run the folder creation script to create task directories and configure gitignore:
bash
${CLAUDE_PLUGIN_ROOT}
/scripts/create-folders.sh
This creates:
.specs/tasks/draft/
- New tasks awaiting analysis
.specs/tasks/todo/
- Tasks ready to implement
.specs/tasks/in-progress/
- Currently being worked on
.specs/tasks/done/
- Completed tasks
.specs/scratchpad/
- Temporary working files (gitignored)
.specs/analysis/
- Codebase impact analysis files
.claude/skills/
- Reusable skill documents
Update each todo to
in_progress
when starting a phase and
completed
when judge passes.
CRITICAL
Do not mark PASS for any judge if it did not pass the rubric. Retry the judge after each implementation change till it passes the check!
Do not read task files in .claude or .specs directories, your job is orchestrate agents that will do the work, not do it by yourself!
Use
THRESHOLD
(default 3.5) for all judge pass/fail decisions, not hardcoded values!
Use
MAX_ITERATIONS
(default 3) for retry limits, not hardcoded values!
After
MAX_ITERATIONS
reached: PROCEED to next stage automatically - do NOT ask user unless phase is in
HUMAN_IN_THE_LOOP_PHASES
!
Skip phases not in
ACTIVE_STAGES
entirely - do not launch agents for excluded stages!
Trigger human-in-the-loop checkpoints ONLY after phases in
HUMAN_IN_THE_LOOP_PHASES
!
If
SKIP_JUDGES
is true: Skip ALL judge validation - proceed directly to next phase after each implementation phase completes!
Task file must exist in
.specs/tasks/draft/
before running this command (unless
--refine
mode)!
If
REFINE_MODE
is true: Detect changes via git diff, skip unchanged stages, pass user feedback to agents!
Execution & Evaluation Rules
Use foreground agents only
Do not use background agents. Launch parallel agents when possible. Background agents constantly run in permissions issues and other errors.
Relaunch judge till you get valid results, of following happens:
Reject Long Reports: If an agent returns a very long report instead of using the scratchpad as requested, reject the result. This indicates the agent failed to follow the "use scratchpad" instruction.
Judge Score 5.0 is a Hallucination: If a judge returns a score of 5.0/5.0, treat it as a hallucination or lazy evaluation. Reject it and re-run the judge. Perfect scores are practically impossible in this rigorous framework.
Reject Missing Scores: If a judge report is missing the numerical score, reject it. This indicates the judge failed to read or follow the rubric instructions.
Workflow Execution
You MUST launch for each step a separate agent, instead of performing all steps yourself.
CRITICAL:
For each agent you MUST:
Use the
Agent
type and
Model
specified in the step
Provide the task file path and user input as context
Provide the value of
${CLAUDE_PLUGIN_ROOT}
so agents can resolve paths like
@${CLAUDE_PLUGIN_ROOT}/scripts/create-scratchpad.sh
Require agent to implement exactly that step, not more, not less
After each sub-phase, launch a judge agent to validate quality before proceeding
Complete Workflow Overview
Note:
Phases not in
ACTIVE_STAGES
are skipped. If
SKIP_JUDGES
is true, all judge steps are skipped entirely. Human checkpoints (🔍) occur after phases in
HUMAN_IN_THE_LOOP_PHASES
.
Input: Draft Task File (.specs/tasks/draft/*.md)
Phase 2: Parallel Analysis
├─────────────────────┬─────────────────────┐
▼ ▼ ▼
Phase 2a: Phase 2b: Phase 2c:
Research Codebase Analysis Business Analysis
[sdd:researcher sonnet] [sdd:code-explorer sonnet] [sdd:business-analyst opus]
Judge 2a Judge 2b Judge 2c
(pass: >THRESHOLD) (pass: >THRESHOLD) (pass: >THRESHOLD)
│ │ │
└─────────────────────┴─────────────────────┘
Phase 3: Architecture Synthesis
[sdd:software-architect opus]
Judge 3 (pass: >THRESHOLD)
Phase 4: Decomposition
[sdd:tech-lead opus]
Judge 4 (pass: >THRESHOLD)
Phase 5: Parallelize
[sdd:team-lead opus]
Judge 5 (pass: >THRESHOLD)
Phase 6: Verifications
[sdd:qa-engineer opus]
Judge 6 (pass: >THRESHOLD)
Move task: draft/ → todo/
Complete
Phase 2: Parallel Analysis
Phase 2 launches three analysis phases in parallel, each with its own judge validation.
Phase 2a/2b/2c: Parallel Sub-Phases
Launch these three phases
in parallel
immediately:
Phase 2a: Research
Model:
sonnet
Agent:
sdd:researcher
Depends on:
Task file exists
Purpose:
Gather relevant resources, documentation, libraries, and prior art. Creates or updates a reusable skill.
Launch agent:
Description
"Research task resources and create/update skill"
Prompt
:
CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
Task File:
Task Title: <title from task file>
CRITICAL: DO NOT OUTPUT YOUR RESEARCH, ONLY CREATE THE SCRATCHPAD AND SKILL FILE.
Capture:
Skill file path (e.g.,
.claude/skills//SKILL.md
)
Skill action (Created new / Updated existing)
Scratchpad file path (e.g.,
.specs/scratchpad/.md
)
Number of resources gathered
Key recommendation summary
CRITICAL: If expected files not created, launch the agent again with the same prompt.
Phase 2b: Codebase Impact Analysis
Model:
sonnet
Agent:
sdd:code-explorer
Depends on:
Task file exists
Purpose:
Identify affected files, interfaces, and integration points
Launch agent:
Description
"Analyze codebase impact"
Prompt
:
CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
Task File:
Task Title: <title from task file>
CRITICAL: DO NOT OUTPUT YOUR ANALYSIS, ONLY CREATE THE SCRATCHPAD AND ANALYSIS FILE.
Capture:
Analysis file path (e.g.,
.specs/analysis/analysis-{name}.md
)
Scratchpad file path (e.g.,
.specs/scratchpad/.md
)
Files affected count (modify/create/delete)
Risk level assessment
Key integration points
CRITICAL: If expected files not created, launch the agent again with the same prompt.
Phase 2c: Business Analysis
Model:
opus
Agent:
sdd:business-analyst
Depends on:
Task file exists
Purpose:
Refine description and create acceptance criteria
Launch agent:
Description
"Business analysis"
Prompt
:
CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
Read ${CLAUDE_PLUGIN_ROOT}/skills/plan/analyse-business-requirements.md and execute it exactly as is!
Task File:
Task Title: <title from task file>
CRITICAL: DO NOT OUTPUT YOUR BUSINESS ANALYSIS, ONLY CREATE THE SCRATCHPAD AND UPDATE THE TASK FILE.
Capture:
Scratchpad file path (e.g.,
.specs/scratchpad/.md
)
Acceptance criteria count
Scope defined (yes/no)
User scenarios documented
Judge 2a/2b/2c: Validate Parallel Phases
After
each
parallel phase completes, launch its respective judge
with the same agent type and model
.
Judge 2a: Validate Research/Skill
Model:
sonnet
Agent:
sdd:researcher
Depends on:
Phase 2a completion
Purpose:
Validate skill completeness and relevance
Launch judge:
Description
"Judge skill quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to skill file from Phase 2a}

Context

This is a skill document for task: {task title}. Evaluate comprehensiveness and reusability.

Rubric

  1. Resource Coverage (weight: 0.30)
  2. Documentation and references gathered?
  3. Libraries and tools identified with recommendations?
  4. 1=Missing critical resources, 2=Basic coverage, 3=Adequate, 4=Comprehensive, 5=Excellent
  5. Pattern Relevance (weight: 0.25)
  6. Are identified patterns applicable?
  7. Are recommendations actionable?
  8. 1=Irrelevant, 2=Somewhat useful, 3=Adequate, 4=Well-targeted, 5=Perfect fit
  9. Issue Anticipation (weight: 0.20)
  10. Common pitfalls identified with solutions?
  11. 1=None identified, 2=Few issues, 3=Adequate, 4=Good coverage, 5=Comprehensive
  12. Reusability (weight: 0.15)
  13. Is the skill general enough to help multiple tasks?
  14. Does it avoid task-specific details?
  15. 1=Too specific, 2=Limited reuse, 3=Adequate, 4=Good, 5=Highly reusable
  16. Task Integration (weight: 0.10)
  17. Was task file updated with skill reference?
  18. 1=Not updated, 3=Updated, 5=Updated with clear instructions
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Research complete, proceed
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 2a with feedback
    MAX_ITERATIONS reached
    Proceed to next stage regardless of score (log warning)
    Judge 2b: Validate Codebase Analysis
    Model:
    sonnet
    Agent:
    sdd:code-explorer
    Depends on:
    Phase 2b completion
    Purpose:
    Validate file identification accuracy and integration mapping
    Launch judge:
    Description
    "Judge codebase analysis quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to analysis file from Phase 2b}

Context

This is codebase impact analysis for task: {task title}. Evaluate accuracy and completeness.

Rubric

  1. File Identification Accuracy (weight: 0.35)
  2. All affected files identified with specific paths?
  3. New files and modifications distinguished?
  4. 1=Major files missing, 2=Mostly correct, 3=Adequate, 4=Precise, 5=Complete
  5. Interface Documentation (weight: 0.25)
  6. Key functions/classes documented with signatures?
  7. Change requirements clear?
  8. 1=Missing, 2=Partial, 3=Adequate, 4=Good, 5=Complete
  9. Integration Point Mapping (weight: 0.25)
  10. Integration points identified with impact?
  11. Similar patterns in codebase found?
  12. 1=Missing, 2=Partial, 3=Adequate, 4=Good, 5=Comprehensive
  13. Risk Assessment (weight: 0.15)
  14. High risk areas identified with mitigations?
  15. 1=No assessment, 2=Basic, 3=Adequate, 4=Good, 5=Thorough
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Analysis complete, proceed
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 2b with feedback
    MAX_ITERATIONS reached
    Proceed to next stage regardless of score (log warning)
    Judge 2c: Validate Business Analysis
    Model:
    opus
    Agent:
    sdd:business-analyst
    Depends on:
    Phase 2c completion
    Purpose:
    Validate acceptance criteria quality and scope definition
    Launch judge:
    Description
    "Judge business analysis quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to task file from Phase 2c}

Context

This is business analysis output. Evaluate description clarity and acceptance criteria quality.

Rubric

  1. Description Clarity (weight: 0.30)
  2. What/Why clearly explained?
  3. Scope boundaries defined?
  4. 1=Vague, 2=Basic, 3=Adequate, 4=Clear, 5=Excellent
  5. Acceptance Criteria Quality (weight: 0.35)
  6. Criteria specific and testable?
  7. Given/When/Then format for complex criteria?
  8. 1=Missing/vague, 2=Basic, 3=Adequate, 4=Good, 5=Excellent
  9. Scenario Coverage (weight: 0.20)
  10. Primary flow documented?
  11. Error scenarios considered?
  12. 1=Missing, 2=Basic, 3=Adequate, 4=Good, 5=Comprehensive
  13. Scope Definition (weight: 0.15)
  14. In-scope/out-of-scope explicit?
  15. No implementation details in description?
  16. 1=Missing, 2=Partial, 3=Adequate, 4=Good, 5=Clear
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Business analysis complete, proceed
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 2c with feedback
    MAX_ITERATIONS reached
    Proceed to next stage regardless of score (log warning)
    Synchronization Point
    Wait for ALL three parallel phases (2a, 2b, 2c) AND their judges to PASS before proceeding to Phase 3.
    Phase 3: Architecture Synthesis
    Model:
    opus
    Agent:
    sdd:software-architect
    Depends on:
    Phase 2a + Judge 2a PASS, Phase 2b + Judge 2b PASS, Phase 2c + Judge 2c PASS
    Purpose:
    Synthesize research, analysis, and business requirements into architectural overview
    Launch agent:
    Description
    "Architecture synthesis"
    Prompt
    :
    CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
    Task File:
    Skill File:
    Analysis File:
    CRITICAL: DO NOT OUTPUT YOUR ARCHITECTURE SYNTHESIS, ONLY CREATE THE SCRATCHPAD AND UPDATE THE TASK FILE.
    Capture:
    Scratchpad file path (e.g.,
    .specs/scratchpad/.md
    )
    Sections added to task file
    Key architectural decisions count
    Components identified (if applicable)
    Contracts defined (if applicable)
    Judge 3: Validate Architecture Synthesis
    Model:
    opus
    Agent:
    sdd:software-architect
    Depends on:
    Phase 3 completion
    Purpose:
    Validate architectural coherence and completeness
    Launch judge:
    Description
    "Judge architecture synthesis quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to task file after Phase 3}

Context

This is architecture synthesis output. The Architecture Overview section should contain solution strategy, key decisions, and only relevant architectural sections.

Rubric

  1. Solution Strategy Clarity (weight: 0.30)
  2. Approach clearly explained?
  3. Key decisions documented with reasoning?
  4. Trade-offs stated?
  5. 1=Missing/unclear, 2=Basic, 3=Adequate, 4=Clear, 5=Excellent
  6. Reference Integration (weight: 0.20)
  7. Links to research and analysis files?
  8. Insights from both integrated?
  9. 1=No links, 2=Partial, 3=Adequate, 4=Good, 5=Fully integrated
  10. Section Relevance (weight: 0.25)
  11. Only relevant sections included (not all)?
  12. Sections appropriate for task complexity?
  13. 1=Wrong sections, 2=Mostly appropriate, 3=Adequate, 4=Good, 5=Precisely targeted
  14. Expected Changes Accuracy (weight: 0.25)
  15. Files to create/modify listed?
  16. Consistent with codebase analysis?
  17. 1=Missing/inconsistent, 2=Partial, 3=Adequate, 4=Good, 5=Complete
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Architecture synthesis complete, proceed
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 3 with feedback
    MAX_ITERATIONS reached
    Proceed to Phase 4 regardless of score (log warning)
    Wait for PASS before Phase 4.
    Phase 4: Decomposition
    Model:
    opus
    Agent:
    sdd:tech-lead
    Depends on:
    Phase 3 + Judge 3 PASS
    Purpose:
    Break architecture into implementation steps with success criteria and risks
    Launch agent:
    Description
    "Decompose into implementation steps"
    Prompt
    :
    CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
    Task File:
    CRITICAL: DO NOT OUTPUT YOUR DECOMPOSITION, ONLY CREATE THE SCRATCHPAD AND UPDATE THE TASK FILE.
    Capture:
    Scratchpad file path (e.g.,
    .specs/scratchpad/.md
    )
    Implementation steps count
    Total subtasks count
    Critical path steps
    High priority risks count
    Judge 4: Validate Decomposition
    Model:
    opus
    Agent:
    sdd:tech-lead
    Depends on:
    Phase 4 completion
    Purpose:
    Validate implementation steps quality and completeness
    Launch judge:
    Description
    "Judge decomposition quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to task file after Phase 4}

Context

This is decomposition output. The Implementation Process section should contain ordered steps with success criteria, subtasks, blockers, and risks.

Rubric

  1. Step Quality (weight: 0.30)
  2. Each step has clear goal, output, success criteria?
  3. Steps ordered by dependency?
  4. No step too large (>Large estimate)?
  5. 1=Vague/missing, 2=Basic, 3=Adequate, 4=Good, 5=Excellent
  6. Success Criteria Testability (weight: 0.25)
  7. Criteria specific and verifiable?
  8. Use actual file paths, function names?
  9. Subtasks clearly defined with actionable descriptions?
  10. 1=Vague, 2=Partially testable, 3=Adequate, 4=Good, 5=All testable
  11. Risk Coverage (weight: 0.25)
  12. Blockers identified with resolutions?
  13. Risks identified with mitigations?
  14. High-risk tasks identified with decomposition recommendations?
  15. 1=None, 2=Basic, 3=Adequate, 4=Good, 5=Comprehensive
  16. Completeness (weight: 0.20)
  17. All architecture components have corresponding steps?
  18. Implementation summary table present?
  19. Definition of Done included?
  20. Phases organized: Setup → Foundational → User Stories → Polish?
  21. 1=Incomplete, 2=Partial, 3=Adequate, 4=Good, 5=Complete
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Decomposition complete, proceed to Phase 5
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 4 with feedback
    MAX_ITERATIONS reached
    Proceed to Phase 5 regardless of score (log warning)
    Wait for PASS before Phase 5.
    Phase 5: Parallelize Steps
    Model:
    opus
    Agent:
    sdd:team-lead
    Depends on:
    Phase 4 + Judge 4 PASS
    Purpose:
    Reorganize implementation steps for maximum parallel execution
    Launch agent:
    Description
    "Parallelize implementation steps"
    Prompt
    :
    CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
    Task File:
    Use agents only from this list:
    CRITICAL: DO NOT OUTPUT YOUR PARALLELIZATION, ONLY CREATE THE SCRATCHPAD AND UPDATE THE TASK FILE.
    Capture:
    Scratchpad file path (e.g.,
    .specs/scratchpad/.md
    )
    Number of steps reorganized
    Maximum parallelization depth
    Agent distribution summary
    Judge 5: Validate Parallelization
    Model:
    opus
    Agent:
    sdd:team-lead
    Depends on:
    Phase 5 completion
    Purpose:
    Validate dependency accuracy and parallelization optimization
    Launch judge:
    Description
    "Judge parallelization quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to parallelized task file from Phase 5}

Context

This is the output of Phase 5: Parallelize Steps. The artifact should contain implementation steps reorganized for maximum parallel execution with explicit dependencies, agent assignments, and parallelization diagram. Use agents only from this list: {list ALL available agents with plugin prefix if available, e.g. sdd:developer, code-review:bug-hunter. Also include general agents: opus, sonnet, haiku}

Rubric

  1. Dependency Accuracy (weight: 0.35)
  2. Are step dependencies correctly identified?
  3. No false dependencies (steps marked dependent when they're not)?
  4. No missing dependencies (steps that actually depend on others)?
  5. 1=Major dependency errors, 2=Mostly correct, 3=Acceptable, 5=Precise dependencies
  6. Parallelization Maximized (weight: 0.30)
  7. Are parallelizable steps correctly marked with "Parallel with:"?
  8. Is the parallelization diagram logical?
  9. 1=No parallelization/wrong, 2=Some optimization, 3=Acceptable, 5=Maximum parallelization
  10. Agent Selection Correctness (weight: 0.20)
  11. Are agent types appropriate for outputs (opus by default, haiku for trivial, sonnet for simple but high in volume)?
  12. Does selection follow the Agent Selection Guide?
  13. Are only agents from the provided available agents list used?
  14. 1=Wrong agents, 2=Mostly appropriate, 3=Acceptable, 4=Optimal selection, 5=Perfect selection
  15. Execution Directive Present (weight: 0.15)
  16. Is the sub-agent execution directive present?
  17. Are "MUST" requirements for parallel execution clear?
  18. 1=Missing directive, 2=Partial, 3=Acceptable, 4=Complete directive, 5=Perfect directive
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Proceed to Phase 6
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 5 with feedback
    MAX_ITERATIONS reached
    Proceed to Phase 6 regardless of score (log warning)
    Wait for PASS before Phase 6.
    Phase 6: Define Verifications
    Model:
    opus
    Agent:
    sdd:qa-engineer
    Depends on:
    Phase 5 + Judge 5 PASS
    Purpose:
    Add LLM-as-Judge verification sections with rubrics
    Launch agent:
    Description
    "Define verification rubrics"
    Prompt
    :
    CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT}
    Task File:
    CRITICAL: DO NOT OUTPUT YOUR VERIFICATIONS, ONLY CREATE THE SCRATCHPAD AND UPDATE THE TASK FILE.
    Capture:
    Scratchpad file path (e.g.,
    .specs/scratchpad/.md
    )
    Number of steps with verification
    Total evaluations defined
    Verification breakdown (Panel/Per-Item/None)
    Judge 6: Validate Verifications
    Model:
    opus
    Agent:
    sdd:qa-engineer
    Depends on:
    Phase 6 completion
    Purpose:
    Validate verification rubrics and thresholds
    Launch judge:
    Description
    "Judge verification quality" Prompt : CLAUDE_PLUGIN_ROOT=${CLAUDE_PLUGIN_ROOT} Read @${CLAUDE_PLUGIN_ROOT}/prompts/judge.md for evaluation methodology and execute.

Artifact Path

{path to task file with verifications from Phase 6}

Context

This is the output of Phase 6: Define Verifications. The artifact should contain LLM-as-Judge verification sections for each implementation step, including verification levels, custom rubrics, thresholds, and a verification summary table.

Rubric

  1. Verification Level Appropriateness (weight: 0.30)
  2. Do verification levels match artifact criticality?
  3. HIGH criticality → Panel, MEDIUM → Single/Per-Item, LOW/NONE → None?
  4. 1=Mismatched levels, 2=Mostly appropriate, 3=Acceptable, 5=Precisely calibrated
  5. Rubric Quality (weight: 0.30)
  6. Are criteria specific to the artifact type (not generic)?
  7. Do weights sum to 1.0?
  8. Are descriptions clear and measurable?
  9. 1=Generic/broken rubrics, 2=Adequate, 3=Acceptable, 5=Excellent custom rubrics
  10. Threshold Appropriateness (weight: 0.20)
  11. Are thresholds reasonable (typically 4.0/5.0)?
  12. Higher for critical, lower for experimental?
  13. 1=Wrong thresholds, 2=Standard applied, 3=Acceptable, 5=Context-appropriate
  14. Coverage Completeness (weight: 0.20)
  15. Does every step have a Verification section?
  16. Is the Verification Summary table present?
  17. 1=Missing verifications, 2=Most covered, 3=Acceptable, 5=100% coverage
    CRITICAL: use prompt exactly as is, do not add anything else. Including output of implementation agent!!!
    Decision Logic:
    PASS
    (score >=
    THRESHOLD
    ): Workflow complete, promote task
    FAIL
    (score <
    THRESHOLD
    ): Re-launch Phase 6 with feedback
    MAX_ITERATIONS reached
    Complete workflow regardless of score (log warning) Phase 7: Promote Task Purpose: Move the refined task from draft to todo folder After all phases complete: Move task file from draft to todo: git mv < TASK_FILE

    .specs/tasks/todo/

Fallback if git not available: mv .specs/tasks/todo/

Update any references in research and analysis files if needed Completion After all executed phases and judges complete: Use git tool to stage the task file, skill file, analysis file, and scratchpad files (only those that were created) Summarize the workflow results and output to user:

Task Refined | Property | Value | |


|

| | ** Original File ** | <original TASK_FILE path> | | ** Final Location ** | .specs/tasks/todo/<filename> (ready for implementation) | | ** Title ** | <task title> | | ** Type ** | <feature/bug/refactor/test/docs/chore/ci> (from filename) | | ** Skill ** | <skill file path or "Skipped"> | | ** Skill Action ** | <Created new / Updated existing / Skipped> | | ** Analysis ** | <analysis file path or "Skipped"> | | ** Scratchpad ** | <scratchpad file path> | | ** Implementation Steps ** | <count or "N/A"> | | ** Parallelization Depth ** | <max parallel agents or "N/A"> | | ** Total Verifications ** | <count or "N/A"> |

Configuration Used | Setting | Value | |


|

| | ** Target Quality ** | {THRESHOLD}/5.0 | | ** Max Iterations ** | {MAX_ITERATIONS} | | ** Active Stages ** | {ACTIVE_STAGES as comma-separated list} | | ** Skipped Stages ** | {SKIP_STAGES or stages not in ACTIVE_STAGES} | | ** Human Checkpoints ** | Phase {HUMAN_IN_THE_LOOP_PHASES as comma-separated} | | ** Skip Judges ** | {SKIP_JUDGES} | | ** Refine Mode ** | {REFINE_MODE} |

Quality Gates Summary | Phase | Judge Score | Verdict | |


|

|

| | Phase 2a: Research | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | | Phase 2b: Codebase Analysis | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | | Phase 2c: Business Analysis | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | | Phase 3: Architecture Synthesis | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | | Phase 4: Decomposition | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | | Phase 5: Parallelize | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | | Phase 6: Verify | X.X/5.0 | ✅ PASS / ⚠️ PROCEEDED (max iter) / ⏭️ SKIPPED | ** Threshold Used: ** {THRESHOLD}/5.0 (or N/A if SKIP_JUDGES) ** Legend: ** - ✅ PASS - Score >= THRESHOLD - ⚠️ PROCEEDED (max iter) - Score < THRESHOLD but MAX_ITERATIONS reached, proceeded anyway - ⏭️ SKIPPED - Stage not in ACTIVE_STAGES

Artifacts Generated .claude/ └── skills/ └── / └── SKILL.md # Reusable skill document (if research stage ran) .specs/ ├── tasks/ │ ├── draft/ # Draft tasks (source - now empty for this task) │ ├── todo/ │ │ └── ..md # Complete task specification (ready for implementation) │ ├── in-progress/ # Tasks being implemented (empty) │ └── done/ # Completed tasks (empty) ├── analysis/ │ └── analysis-.md # Codebase impact analysis (if codebase analysis stage ran) └── scratchpad/ └── .md # Architecture thinking scratchpad

Task Status Management

Task status is managed by folder location: - draft/ - Tasks created but not yet refined - todo/ - Tasks ready for implementation - in-progress/ - Tasks currently being worked on - done/ - Completed tasks

Next Steps

  1. Review task: .specs/tasks/todo/<filename>
  2. Edit the task file directly to make corrections
  3. Add // comments to lines that need clarification or changes
  4. Run /plan again with --refine to incorporate your feedback — it detects changes against git and propagates updates top-to-bottom (editing a section only affects sections below it, not above)
  5. If everything is fine, begin implementation: /implement (will auto-select the task from todo/)
    Error Handling
    Phase Agent Failure (Exception/Crash)
    If any phase agent fails unexpectedly:
    Report the failure with agent output
    Ask clarification questions from user that can help resolve the issue
    Launch the phase agent again with list of questions and answers to resolve the issue
    Judge Returns FAIL
    If any judge returns FAIL (score <
    THRESHOLD
    ):
    Automatic retry
    Re-launch the phase agent with judge feedback
    Human-in-the-loop check
    If phase is in HUMAN_IN_THE_LOOP_PHASES , trigger human checkpoint before the next judge retry (after implementation retry but before re-judging) After MAX_ITERATIONS reached : Proceed to next stage automatically (do NOT ask user unless --human-in-the-loop includes this phase) Log warning in completion summary: ⚠️ Phase X did not pass quality threshold (X.X/THRESHOLD) after MAX_ITERATIONS iterations Retry Flow Implementation → Judge FAIL → Implementation Retry → Judge Retry ↓ PASS → Continue to next stage FAIL → Repeat until MAX_ITERATIONS ↓ MAX_ITERATIONS reached → Proceed to next stage (with warning) Retry Flow with Human-in-the-Loop When phase is in HUMAN_IN_THE_LOOP_PHASES : Implementation → Judge FAIL → Implementation Retry ↓ 🔍 Human Checkpoint (optional feedback) ↓ Judge Retry ↓ PASS → Continue | FAIL → Repeat until MAX_ITERATIONS ↓ MAX_ITERATIONS → 🔍 Final Human Checkpoint ↓ User confirms → Proceed to next stage
返回排行榜