decompose-task
npx skills add https://github.com/fancy-bread/sdlc-workflow-skills --skill decompose-task
Agent 安装分布
Skill 文档
Decompose Task
Overview
Decompose a large task (epic or large story) into well-defined, actionable subtasks. This is a critical Scrum planning activity that ensures large work items are properly broken down into sprint-sized tasks.
Definitions
- {TASK_KEY}: Task/Issue ID from the issue tracker (e.g.,
FB-6,PROJ-123,KAN-42) - {FEATURE_DOMAIN}: Kebab-case feature name inherited from parent task (e.g.,
user-authentication,payment-processing). Used to link child PBIs to the same Spec as parent. - Task Decomposition: The process of breaking down a large task into smaller, more manageable subtasks
- Subtask: A child task that represents a specific piece of work within a larger task or epic
- Subtask Criteria: Requirements that each generated subtask must meet:
- Follows appropriate format (user story format for user-facing tasks, action-oriented for technical tasks)
- Has clear acceptance criteria (3-5 criteria minimum)
- Is appropriately sized (completable in 1-2 days typically, 1-2 sprint points)
- Is independent and can be developed standalone (or clearly defined dependencies)
- Delivers user or technical value
- Is testable and verifiable
- Task Types for Decomposition:
- Epic: High-level initiative that typically breaks down into multiple stories
- Story: User story or feature that may break down into technical tasks
- Technical Task: Large technical work item that may break down into smaller technical tasks
- Information Density Scoring: A method to assess task information completeness using 5 key elements:
- Clear, specific goals (not vague)
- Defined scope or boundaries
- User context (personas, use cases) OR technical context
- Success criteria or acceptance criteria
- Constraints or dependencies
- Scoring: 0-2 elements = INSUFFICIENT (must ask), 3-4 = MARGINAL (proceed with caution), 5+ = SUFFICIENT (proceed confidently)
Prerequisites
Before proceeding, verify:
-
MCP Status Validation: Perform MCP server status checks (see
mcp-status.mdfor detailed steps)- Test each configured MCP server connection (Atlassian, GitHub, etc.)
- Verify all required integrations are authorized and operational
- If any MCP server fails validation, STOP and report the failure. Do not proceed.
- MCP Tool Usage Standards: MCP tool usage should follow best practices (check schema files, validate parameters, handle errors gracefully). These standards are documented in AGENTS.md §3 Operational Boundaries if AGENTS.md exists, but apply universally regardless.
-
Task Exists: Verify the task exists in the issue tracker
- Use MCP tools to fetch task by
{TASK_KEY} - Use
mcp_atlassian_getJiraIssuefor Jira issues - Use
mcp_github_issue_readfor GitHub Issues - If task doesn’t exist, STOP and report error: “Task {TASK_KEY} not found”
- Use MCP tools to fetch task by
-
Task Has Sufficient Detail (using intelligent analysis):
- Analyze task content intelligently:
- Parse task description for completeness
- Detect vague language patterns:
- Generic terms: “improve”, “enhance”, “fix”, “add feature” without specifics
- Missing context: No user personas, no use cases, no success metrics
- Ambiguous scope: No boundaries, no “in scope” vs “out of scope”
- Missing criteria: No acceptance criteria, no definition of done
- Assess information density:
- Very short descriptions (< 50 words) likely insufficient
- No examples or scenarios provided
- No technical requirements mentioned
- No dependencies or constraints identified
- Intelligent gap detection:
- Identify specific missing information categories
- Determine which gaps are critical vs nice-to-have
- Generate contextually appropriate questions based on:
- Task type (epic, story, technical task)
- Domain/context clues in existing description
- Common patterns for similar tasks
- Decision logic:
- If task has < 3 of these elements: STOP and ask
- Clear, specific goals (not vague)
- Defined scope or boundaries
- User context (personas, use cases) OR technical context
- Success criteria or acceptance criteria
- Any constraints or dependencies
- If task description is < 50 words and vague: STOP and ask
- If key terms are undefined: STOP and ask
- “User experience” – which users? what experience?
- “Performance” – what metrics? what targets?
- “Feature” – what exactly? for whom?
- If task has < 3 of these elements: STOP and ask
- If information is insufficient:
- Generate 3-5 specific, actionable questions
- Questions should be:
- Contextual to the task domain
- Focused on critical gaps
- Actionable (user can answer clearly)
- Prioritized (most important first)
- Present questions clearly with context
- Wait for user response before proceeding
- Do NOT make assumptions or proceed with incomplete information
- If information is sufficient:
- Proceed to decomposition analysis
- Note any assumptions made (document in breakdown)
- If task lacks sufficient detail, STOP and ask user for clarification before proceeding.
- Analyze task content intelligently:
Usage
/decompose-task TASK-123
PBI 4-part anatomy (for subtasks)
Use this structure when generating subtask descriptions (Stories/Epics). PBI template: Use create-task’s asset â ../create-task/assets/pbi-anatomy.md â load when building subtask PBI descriptions (same anatomy as create-task; child issues follow the same structure). Populate from parent task and inherited {feature-domain}; link to specs/{feature-domain}/spec.md. Same spec as parent; for subtasks, Directive scope is narrower than parent.
Examples:
/decompose-task PROJ-100(epic)/decompose-task STORY-50(large story)
Steps
-
Read the task
- Retrieve task from issue tracker:
- Use
mcp_atlassian_getJiraIssuefor Jira issues- Parameters:
cloudId,issueIdOrKey= {TASK_KEY}
- Parameters:
- Use
mcp_github_issue_readfor GitHub Issues- Parameters:
owner,repo,issue_number= {TASK_KEY}
- Parameters:
- Use
- Understand task goals and objectives:
- Parse task title and description
- Identify the primary goal or objective
- Extract any stated business value or user benefit
- Review task description and acceptance criteria:
- Read full description text
- Extract acceptance criteria (numbered lists, checkboxes, “Given/When/Then” format)
- Identify any existing scope definitions
- Identify scope boundaries:
- Look for explicit “in scope” vs “out of scope” statements
- Identify any constraints or limitations mentioned
- Note any dependencies on other work
- Review any attached documents or links:
- Check for attached files or documents
- Review any linked issues or related tasks
- Note any external documentation referenced
- If task cannot be retrieved, STOP and report: “Task {TASK_KEY} not found or inaccessible”
- Retrieve task from issue tracker:
-
Validate task information (Intelligent Analysis)
- Apply intelligent analysis patterns (see Prerequisites section for detailed logic):
- Parse task description for completeness
- Detect vague language patterns (generic terms, missing context, ambiguous scope, missing criteria)
- Assess information density using 5-element scoring
- Perform intelligent gap detection
- Apply decision logic:
- If task has < 3 of required elements: STOP and ask
- Clear, specific goals (not vague)
- Defined scope or boundaries
- User context (personas, use cases) OR technical context
- Success criteria or acceptance criteria
- Any constraints or dependencies
- If task description is < 50 words and vague: STOP and ask
- If key terms are undefined: STOP and ask
- If task has < 3 of required elements: STOP and ask
- If information is insufficient:
- Generate 3-5 specific, actionable questions using context-aware question generation
- Present questions clearly with context
- Wait for user response before proceeding
- Do NOT make assumptions or proceed with incomplete information
- If information is sufficient:
- Proceed to scope analysis
- Note any assumptions made (document in breakdown)
- Apply intelligent analysis patterns (see Prerequisites section for detailed logic):
-
Analyze task scope
- Break down task into logical work areas:
- Identify major functional areas or components
- Group related functionality together
- Consider technical vs user-facing separation
- Identify user personas and use cases (if applicable):
- Extract user personas from task description
- Identify primary use cases or user journeys
- Note any edge cases or special scenarios
- Determine technical components involved:
- Identify technical layers or components (frontend, backend, database, API, etc.)
- Note any integration points with external systems
- Identify any technical constraints or requirements
- Identify dependencies between work areas:
- Map dependencies between different work areas
- Identify blocking dependencies (must be done first)
- Note parallel work opportunities (can be done simultaneously)
- Consider implementation phases or milestones:
- Identify logical phases for implementation
- Consider MVP vs enhancement separation
- Plan for incremental delivery if applicable
- If scope analysis reveals major gaps or ambiguities, ask user for clarification before proceeding.
- Break down task into logical work areas:
3a. Extract feature domain from parent (for PBI structure inheritance)
- Determine if parent requires PBI structure:
- Check parent task type: Epic or Story (PBI required)
- Check parent task type: Bug or Task (PBI optional)
- Extract feature domain from parent task:
- From parent labels:
- Look for label pattern:
feature:{domain} - Extract
{domain}from label - Example:
feature:user-authenticationâuser-authentication
- Look for label pattern:
- From parent description:
- Parse parent description for Context Pointer section
- Extract domain from spec link:
specs/{domain}/spec.md - Example:
specs/user-authentication/spec.mdâuser-authentication
- From parent epic (if parent is a story):
- If parent has parent epic, check epic for domain
- Use
mcp_atlassian_getJiraIssueto fetch epic - Extract domain from epic labels or description
- Ask user if domain not found:
- If parent has no domain, ask: “Which feature domain for these subtasks? (e.g., user-authentication, payment-processing)”
- Provide examples from existing specs
- From parent labels:
- Validate feature domain format:
- Must be kebab-case (lowercase with hyphens)
- Pattern:
[a-z]+(-[a-z]+)* - If invalid format, normalize or ask user for correction
- Check if spec exists:
- Use
glob_file_searchwith pattern:**/specs/{feature-domain}/spec.md - If spec exists: Note for PBI generation (valid links)
- If spec missing: Warn but proceed with placeholders
- Use
- Store feature domain for subtask generation:
- Save domain to use in step 4 (subtask description generation)
- All child tasks will inherit same domain
-
Generate subtasks
- Create tasks following appropriate format:
- For user stories: “As a [user], I want [goal], so that [benefit]” format
- For technical tasks: Use clear action-oriented descriptions
- Generate subtask descriptions with PBI structure (for Stories and Epics):
- Use the shared PBI 4-part anatomy â load ../create-task/assets/pbi-anatomy.md (see “PBI 4-part anatomy (for subtasks)” section above).
- For each subtask, populate 4-part anatomy:
- Directive Section:
- Specific scope for this subtask (narrower than parent)
- Clear in scope / out of scope boundaries
- Dependencies on other subtasks (if any)
- Context Pointer Section:
- Use {feature-domain} inherited from parent (step 3a)
- Generate link:
../../specs/{feature-domain}/spec.md#blueprint - Same spec as parent, but subtask-specific scope
- Verification Pointer Section:
- Same {feature-domain} as parent
- Generate link:
../../specs/{feature-domain}/spec.md#contract - Reference parent’s Contract for overall goals
- Refinement Rule Section:
- Use the Refinement Rule from the shared PBI template; same as parent’s refinement rule
- Directive Section:
- Validate PBI structure for each subtask:
- Verify all 4 sections present
- Verify feature domain matches parent
- Verify links are well-formed
- Fallback for non-Story subtasks:
- Use simpler description format for technical tasks/bugs
- Ensure each task meets subtask criteria:
- Independent and can be developed standalone (or clearly defined dependencies)
- Well-defined with clear acceptance criteria
- Appropriately sized (completable in 1-2 days typically, 1-2 sprint points)
- Delivers user or technical value
- Write clear, descriptive titles:
- Use action verbs (create, implement, add, refactor, etc.)
- Be specific about what is being done
- Include context if needed (e.g., “Add user authentication to mobile app”)
- Define comprehensive acceptance criteria for each task:
- Minimum 3-5 acceptance criteria per task
- Criteria should be specific, testable, and verifiable
- Include both positive and negative test cases where applicable
- Identify task dependencies and sequence:
- Mark which tasks depend on others
- Establish logical sequence for task execution
- Consider parallel execution opportunities
- If unable to generate well-defined subtasks, ask user for additional context or clarification.
- Create tasks following appropriate format:
-
Validate breakdown quality
- Review generated tasks for completeness:
- Verify all major work areas are covered
- Check that no critical functionality is missing
- Ensure scope matches original task
- Ensure tasks are appropriately sized:
- Each task should be completable in 1-2 days
- Tasks should be roughly similar in size
- Large tasks should be further decomposed
- Verify tasks are appropriately sized:
- Use Task Quality Checklist (see Checklists section)
- Ensure each task meets all criteria
- Check for missing dependencies:
- Verify all dependencies are identified
- Ensure dependency chains are clear
- Check for circular dependencies
- Verify PBI structure consistency:
- All subtasks reference same feature domain
- All subtasks have 4-part anatomy (if Stories/Epics)
- All links point to correct spec
- If breakdown seems incomplete or unclear:
- Ask user for additional context
- Request clarification on ambiguous areas
- Do NOT proceed with incomplete breakdown
- If breakdown quality is insufficient, STOP and revise before creating tasks in tracker.
- Review generated tasks for completeness:
-
Prioritize tasks
- Determine task priority based on:
- User value and business impact (highest priority for high-value tasks)
- Technical dependencies (blocking tasks first)
- Risk and complexity (high-risk tasks may need early attention)
- Implementation sequence (prerequisites first)
- Set task order/sequence:
- Establish logical sequence based on dependencies
- Consider parallel work opportunities
- Plan for incremental delivery
- Identify which tasks are MVP vs nice-to-have:
- Separate core functionality from enhancements
- Mark MVP tasks clearly
- Consider separating enhancement tasks for later sprints
- Document priority rationale if non-obvious:
- Note any business reasons for priority
- Document any assumptions about priority
- Determine task priority based on:
-
Create tasks in tracker
- For each subtask:
- Create task with required fields:
- Clear title (from step 4)
- Detailed description with PBI structure (from step 4, if Stories/Epics)
- Acceptance criteria (from step 4)
- Priority/rank (from step 6)
- Labels: Include
feature:{domain}label (inherited from parent)
- Use MCP tools to create tasks:
- Use
mcp_atlassian_createJiraIssuefor Jira- Parameters:
cloudId,projectKey,issueTypeName(typically “Story” or “Task”),summary,description,additional_fields(for parent link, priority, labels includingfeature:{domain})
- Parameters:
- Use
mcp_github_create_issuefor GitHub- Parameters:
owner,repo,title,body(markdown with PBI description and acceptance criteria),labels(includefeature:{domain})
- Parameters:
- Use
- Link all tasks to parent task/epic:
- Set parent field in Jira (
parentinadditional_fields) - Link via issue references in GitHub (include parent issue number in body)
- Set parent field in Jira (
- Set initial task status to “To Do” (default behavior of most trackers)
- Leave tasks unassigned (assignment happens during sprint planning)
- Create task with required fields:
- If task creation fails for any subtask, STOP and report the error. Do not proceed with remaining tasks until error is resolved.
- For each subtask:
-
Document breakdown
- Add breakdown summary comment to parent task:
- Use
mcp_atlassian_addCommentToJiraIssuefor Jira- Parameters:
cloudId,issueIdOrKey= {TASK_KEY},commentBody= markdown summary
- Parameters:
- Use
mcp_github_add_issue_commentfor GitHub- Parameters:
owner,repo,issue_number= {TASK_KEY},body= markdown summary
- Parameters:
- Use
- Include in summary:
- Task count and overview (e.g., “Created 5 subtasks for this epic”)
- List of created subtasks with their keys/IDs
- Brief description of breakdown approach
- Note any assumptions or decisions made:
- Document assumptions made during analysis
- Note any decisions about scope or approach
- Include rationale for priority decisions if non-obvious
- Document task dependencies if applicable:
- List any dependency relationships
- Note recommended sequence for execution
- Highlight any blocking dependencies
- Add breakdown summary comment to parent task:
Tools
MCP Tools (Atlassian)
mcp_atlassian_atlassianUserInfo– Verify Atlassian MCP connection- Obtaining CloudId for Atlassian Tools:
- Method 1 (Recommended): Use
mcp_atlassian_getAccessibleAtlassianResources- Returns list of accessible resources with
cloudIdvalues - Use the first result or match by site name
- Only call if cloudId is not already known or has expired
- Returns list of accessible resources with
- Method 2: Extract from Atlassian URLs
- Jira URL format:
https://{site}.atlassian.net/... - CloudId can be extracted from the URL or obtained via API
- Jira URL format:
- Error Handling: If cloudId cannot be determined, STOP and report: “Unable to determine Atlassian cloudId. Please verify MCP configuration.”
- Method 1 (Recommended): Use
mcp_atlassian_getJiraIssue– Fetch task details by {TASK_KEY}- Parameters:
cloudId,issueIdOrKey= {TASK_KEY} - Extract: title, description, acceptance criteria, labels, priority, status, parent/epic link
- Parameters:
mcp_atlassian_createJiraIssue– Create subtasks in Jira- Parameters:
cloudId,projectKey,issueTypeName(e.g., “Story”, “Task”),summary,description,additional_fields(forparent,priority,labels,components) - Use to create each subtask after decomposition
- Parameters:
mcp_atlassian_addCommentToJiraIssue– Add breakdown summary comment to parent task- Parameters:
cloudId,issueIdOrKey= {TASK_KEY},commentBody= markdown summary - Use to document the decomposition results
- Parameters:
MCP Tools (GitHub)
- A lightweight read-only GitHub MCP tool to verify connection (see Cursor Settings â Tools & MCP for exact names)
mcp_github_issue_read– Fetch GitHub issue details- Parameters:
owner,repo,issue_number= {TASK_KEY} (if numeric) - Extract: title, body, labels, state, milestone
- Parameters:
mcp_github_create_issue– Create subtasks as GitHub issues- Parameters:
owner,repo,title,body(markdown with description and acceptance criteria),labels - Use to create each subtask after decomposition
- Note: GitHub doesn’t have native parent-child relationships; use issue references in body
- Parameters:
mcp_github_add_issue_comment– Add breakdown summary comment to parent issue- Parameters:
owner,repo,issue_number= {TASK_KEY},body= markdown summary - Use to document the decomposition results
- Parameters:
Codebase Tools
codebase_search– Search for technical context if needed- Query: “How is [similar feature] implemented?” or “Where is [component] used?”
- Use when task involves technical work that requires understanding existing codebase patterns
- Use to identify technical components or integration points
grep– Find specific patterns, functions, or classes- Pattern: component names, function names, file patterns
- Use to identify related code or technical dependencies
Filesystem Tools
read_file– Read plan files or external docs referenced in task- Read plan files if referenced in task
- Parameters:
target_file= path to document - PBI template: ../create-task/assets/pbi-anatomy.md (see “PBI 4-part anatomy (for subtasks)” section).
- Use when task references external documentation or plan files
glob_file_search– Search for specs- Find specs: Pattern
**/specs/{feature-domain}/spec.md - Parameters:
pattern= glob pattern to search - Use to check if spec exists for inherited feature domain
- Find specs: Pattern
Pre-flight Checklist
- MCP status validation performed (see
mcp-status.md) - All MCP servers connected and authorized
- Task retrieved from issue tracker
- Task information validated using intelligent analysis
- Task has sufficient detail (or clarification requested from user)
- All prerequisites met before proceeding to decomposition
Task Quality Checklist
For each generated subtask, verify:
- Follows appropriate format (user story format for user-facing tasks, action-oriented for technical tasks)
- Has clear acceptance criteria (3-5 criteria minimum)
- Is appropriately sized (completable in 1-2 days typically, 1-2 sprint points)
- Is independent and can be developed standalone (or clearly defined dependencies)
- Delivers value (user or technical)
- Has clear definition of done
- Is testable and verifiable
Task Breakdown Checklist
- Parent task read and understood
- Task information validated (sufficient detail)
- Missing information requested from user (if needed)
- Task scope analyzed
- Work areas identified
- Breakdown quality validated
- Subtasks generated (minimum 3-5 tasks per parent)
- Tasks are well-defined and appropriately sized
- Acceptance criteria defined for each task
- Tasks prioritized and sequenced
- All tasks created in tracker
- Tasks linked to parent task/epic
- Breakdown documented in parent task comments
- Tasks ready for team estimation
Guidance
Role
Act as a Scrum Master or Product Manager responsible for decomposing large tasks into well-defined, actionable subtasks. You are analytical, thorough, and ensure that decomposition results in tasks that are clear, appropriately sized, and ready for team estimation and implementation.
Instruction
Execute the decompose-task workflow to break down large tasks into manageable subtasks. This includes:
- Performing prerequisite validation checks
- Reading and understanding the parent task
- Validating task information using intelligent analysis
- Analyzing task scope to identify work areas
- Generating well-defined subtasks that meet quality criteria
- Validating breakdown quality
- Prioritizing and sequencing tasks
- Creating tasks in the issue tracker
- Documenting the decomposition results
Context
- The task is tracked in an issue management system (Jira, GitHub Issues, Azure DevOps, etc.)
- Large tasks (epics or large stories) need to be broken down into sprint-sized tasks
- MCP integrations provide access to issue trackers for reading tasks and creating subtasks
- The decomposition process uses intelligent analysis to ensure tasks have sufficient detail before decomposition
- Generated subtasks must meet specific quality criteria to be useful for the development team
- ASDLC patterns: The PBI, Context Gates
- ASDLC pillars: Factory Architecture (decomposition workflow), Quality Control (information-density gate)
Examples
ASDLC: Context Gates â Information-density check acts as an input gate before decomposition.
Example 1: Vague Epic
You: /decompose-task EPIC-100
AI:
â Reading EPIC-100: Improve User Experience
Analysis:
- Description: 8 words (insufficient length)
- Vague terms detected: "improve", "user experience" (undefined)
- Missing: scope, personas, success criteria, constraints
- Information density: 1/5 elements
â Insufficient information detected. To create a proper breakdown, I need:
1. What specific user experience problems are we solving? (e.g., slow load times, confusing navigation, mobile usability)
2. Which user personas are affected? (e.g., new users, power users, mobile users)
3. What are the success criteria? (e.g., reduce bounce rate by 20%, increase task completion by 30%)
4. What's in scope vs out of scope? (e.g., only mobile app, or also web?)
Please provide this information so I can generate well-defined tasks.
Example 2: Partially Defined Epic
You: /decompose-task EPIC-200
AI:
â Reading EPIC-200: Add payment processing for subscription plans
Analysis:
- Description: Clear action and object
- Missing: user context, success criteria, technical constraints
- Information density: 2/5 elements
â Some information is missing. To ensure a complete breakdown:
1. Which user personas need this? (e.g., new subscribers, existing customers upgrading)
2. What are the acceptance criteria? (e.g., support credit cards, handle refunds, PCI compliance)
3. Are there technical constraints? (e.g., must use Stripe, integrate with existing billing system)
Proceeding with breakdown, but will note assumptions made.
Example 3: Well-Defined Epic
You: /decompose-task EPIC-300
AI:
â Reading EPIC-300: Implement OAuth 2.0 authentication for mobile app users
Analysis:
- Description: Specific technology and scope
- Contains: technical approach (OAuth 2.0), scope (mobile app), users (mobile users)
- Information density: 4/5 elements
â Sufficient information. Proceeding with breakdown...
[Decomposition process continues...]
Generated 5 subtasks:
1. EPIC-300-1: Implement OAuth 2.0 client library integration
2. EPIC-300-2: Create authentication UI components for mobile app
3. EPIC-300-3: Implement token storage and refresh logic
4. EPIC-300-4: Add user session management
5. EPIC-300-5: Write integration tests for OAuth flow
Breakdown documented in EPIC-300 comments.
Example 4: Large Story Decomposition
You: /decompose-task STORY-50
AI:
â Reading STORY-50: Add user profile editing functionality
Analysis:
- Description: Clear feature, but needs more detail
- Information density: 3/5 elements (MARGINAL)
- Proceeding with caution...
[Decomposition analysis...]
Generated 4 subtasks:
1. STORY-50-1: Create profile edit UI components
2. STORY-50-2: Implement profile update API endpoint
3. STORY-50-3: Add profile image upload functionality
4. STORY-50-4: Write tests for profile editing flow
Breakdown documented in STORY-50 comments.
Constraints
Rules (Must Follow):
- Operational Standards Compliance: This command follows operational standards (documented in AGENTS.md if present, but apply universally):
- MCP Tool Usage: Check schema files, validate parameters, handle errors gracefully
- Safety Limits: Never commit secrets, API keys, or sensitive data in task descriptions
- AGENTS.md Optional: Commands work without AGENTS.md. Standards apply regardless of whether AGENTS.md exists.
- See AGENTS.md §3 Operational Boundaries (if present) for detailed standards
- Prerequisites Must Pass: Do not proceed if MCP validation fails, task doesn’t exist, or task lacks sufficient detail. STOP and report the issue.
- Intelligent Analysis Required: Always apply intelligent analysis patterns before decomposition. Do not proceed with vague or incomplete tasks.
- Information Density Scoring: Use 5-element scoring (0-2: INSUFFICIENT must ask, 3-4: MARGINAL proceed with caution, 5+: SUFFICIENT proceed confidently).
- Subtask Quality Criteria: All generated subtasks must meet the Task Quality Checklist criteria. If subtasks don’t meet criteria, revise before creating in tracker.
- Minimum Subtask Count: Generate minimum 3-5 subtasks per parent task. If unable to generate sufficient subtasks, the parent task may need further clarification.
- Task Size: Each subtask should be completable in 1-2 days (1-2 sprint points). Larger tasks should be further decomposed.
- Dependency Clarity: Clearly identify and document task dependencies. Do not create tasks with unclear or circular dependencies.
- Validation Before Creation: Validate breakdown quality before creating tasks in tracker. Do not create tasks if breakdown quality is insufficient.
- Error Handling: If task creation fails for any subtask, STOP and report the error. Do not proceed with remaining tasks until error is resolved.
- Documentation Required: Always document the decomposition results in a comment on the parent task, including task count, list of created tasks, and any assumptions or decisions.
- Feature Domain Inheritance: Extract feature domain from parent task (from labels, description, or epic). All subtasks must inherit same domain.
- PBI Structure for Subtasks (Stories/Epics): Generate subtask descriptions with ASDLC PBI 4-part anatomy (Directive, Context Pointer, Verification Pointer, Refinement Rule).
- Spec Reference Consistency: All subtasks must reference the same spec as parent (
specs/{feature-domain}/spec.md). - PBI structure (Stories/Epics): Use the shared PBI template at create-task/assets/pbi-anatomy.md; populate for each subtask.
- Feature Label Inheritance: Add label
feature:{domain}to all subtasks (inherited from parent).
Existing Standards (Reference):
- MCP status validation: See
mcp-status.mdfor detailed MCP server connection checks - Task creation patterns: See
create-task.mdfor task creation workflows and validation patterns - User story format: “As a [user], I want [goal], so that [benefit]”
- Subtask criteria: Defined in Definitions section above
- PBI structure: Shared template at create-task/assets/pbi-anatomy.md
- Spec Structure: See
specs/README.mdfor Blueprint + Contract format - ASDLC Patterns: The PBI (4-part anatomy), The Spec (permanent pointer target)
Output
-
Decomposed Subtasks: Well-defined subtasks created in the issue tracker with:
- Clear titles and descriptions
- Comprehensive acceptance criteria (3-5 per task)
- Appropriate priority and sequencing
- Links to parent task/epic
- Status set to “To Do”, unassigned
-
Documentation: Breakdown summary comment added to parent task including:
- Count of created subtasks
- List of created subtasks with their keys/IDs
- Brief description of breakdown approach
- Any assumptions or decisions made
- Task dependencies and recommended sequence
The decomposition should result in tasks that are ready for team estimation and sprint planning, with clear scope, acceptance criteria, and dependencies.
Best Practices
- Validate first – Always check if task has sufficient information before proceeding
- Ask, don’t assume – If information is missing, ask specific questions rather than guessing
- Start with user value – Focus on what users need, not technical implementation
- Keep tasks small – Aim for tasks that can be completed in 1-2 days
- Be specific – Vague tasks lead to confusion during development
- Consider dependencies – Identify and sequence dependent tasks
- Think in slices – Create vertical slices (full feature) not horizontal layers (just backend)
- Review with team – Use this as input for team refinement sessions
Intelligent Analysis Patterns
The command uses pattern recognition to detect insufficient information:
Vague Language Detection
Red flags that trigger clarification:
- Generic action verbs without objects: “improve”, “enhance”, “fix”, “add”
- Abstract concepts without definition: “user experience”, “performance”, “quality”
- Missing specificity: “some users”, “better”, “faster”, “easier”
- No concrete examples or scenarios
Example analysis:
Task: "Improve user experience"
Analysis:
- "Improve" is vague (what exactly?)
- "User experience" is undefined (which aspects?)
- No success criteria
â INSUFFICIENT: Ask for specifics
Information Density Scoring
Minimum viable information:
- Task description > 50 words (unless very technical/clear)
- Contains at least 3 of: goals, scope, users, criteria, constraints
- Has concrete examples or scenarios
- Defines key terms used
Scoring logic:
- 0-2 elements present: INSUFFICIENT (must ask)
- 3-4 elements present: MARGINAL (proceed with caution, note assumptions)
- 5+ elements present: SUFFICIENT (proceed confidently)
Context-Aware Question Generation
Question generation based on task type:
For user-facing tasks:
- “Which user personas are affected?”
- “What user problems are we solving?”
- “What does success look like for users?”
For technical tasks:
- “What are the technical constraints?”
- “What are the integration points?”
- “What are the performance requirements?”
For feature requests:
- “What specific functionality is needed?”
- “Who will use this feature?”
- “What are the acceptance criteria?”
For improvements:
- “What specific problems are we solving?”
- “What are the current pain points?”
- “How will we measure improvement?”
Example Intelligent Interactions
Example 1: Vague Epic
You: /decompose-task EPIC-100
AI:
â Reading EPIC-100: Improve User Experience
Analysis:
- Description: 8 words (insufficient length)
- Vague terms detected: "improve", "user experience" (undefined)
- Missing: scope, personas, success criteria, constraints
- Information density: 1/5 elements
â Insufficient information detected. To create a proper breakdown, I need:
1. What specific user experience problems are we solving? (e.g., slow load times, confusing navigation, mobile usability)
2. Which user personas are affected? (e.g., new users, power users, mobile users)
3. What are the success criteria? (e.g., reduce bounce rate by 20%, increase task completion by 30%)
4. What's in scope vs out of scope? (e.g., only mobile app, or also web?)
Please provide this information so I can generate well-defined tasks.
Example 2: Partially Defined Epic
You: /decompose-task EPIC-200
AI:
â Reading EPIC-200: Add payment processing for subscription plans
Analysis:
- Description: Clear action and object
- Missing: user context, success criteria, technical constraints
- Information density: 2/5 elements
â Some information is missing. To ensure a complete breakdown:
1. Which user personas need this? (e.g., new subscribers, existing customers upgrading)
2. What are the acceptance criteria? (e.g., support credit cards, handle refunds, PCI compliance)
3. Are there technical constraints? (e.g., must use Stripe, integrate with existing billing system)
Proceeding with breakdown, but will note assumptions made.
Example 3: Well-Defined Epic
You: /decompose-task EPIC-300
AI:
â Reading EPIC-300: Implement OAuth 2.0 authentication for mobile app users
Analysis:
- Description: Specific technology and scope
- Contains: technical approach (OAuth 2.0), scope (mobile app), users (mobile users)
- Information density: 4/5 elements
â Sufficient information. Proceeding with breakdown...
Common Patterns
Feature-based breakdown:
- Break large task into major features
- Each feature becomes a task or task group
User journey-based breakdown:
- Identify user journeys within task
- Create tasks for each journey step
Technical component breakdown:
- Identify technical components
- Create tasks for each component (but ensure value)
MVP vs Enhancement:
- Identify core MVP tasks
- Separate enhancement tasks for later sprints