Add 5 new core command templates inspired by Garry Tan's GStack to complete the spec-driven development lifecycle: - /speckit.critique: Dual-lens product + engineering review before implementation - /speckit.review: Staff-level code review (correctness, security, performance) - /speckit.qa: Systematic QA testing (browser-driven and CLI modes) - /speckit.ship: Release automation (pre-flight, changelog, CI, PR creation) - /speckit.retro: Sprint retrospective with metrics and improvement suggestions Each command includes: - Command template in templates/commands/ - Output report template in templates/ - Extension hook support (before_*/after_*) - YAML frontmatter with prerequisite scripts Updated README.md workflow from 6 to 11 steps and added CHANGELOG entry. Co-authored-by: Copilot <223556219+Copilot@users.noreply.github.com>
12 KiB
description, scripts
| description | scripts | ||||
|---|---|---|---|---|---|
| Perform a dual-lens critical review of the specification and plan from both product strategy and engineering risk perspectives before implementation. |
|
User Input
$ARGUMENTS
You MUST consider the user input before proceeding (if not empty).
Pre-Execution Checks
Check for extension hooks (before critique):
- Check if
.specify/extensions.ymlexists in the project root. - If it exists, read it and look for entries under the
hooks.before_critiquekey - If the YAML cannot be parsed or is invalid, skip hook checking silently and continue normally
- Filter out hooks where
enabledis explicitlyfalse. Treat hooks without anenabledfield as enabled by default. - For each remaining hook, do not attempt to interpret or evaluate hook
conditionexpressions:- If the hook has no
conditionfield, or it is null/empty, treat the hook as executable - If the hook defines a non-empty
condition, skip the hook and leave condition evaluation to the HookExecutor implementation
- If the hook has no
- For each executable hook, output the following based on its
optionalflag:- Optional hook (
optional: true):## Extension Hooks **Optional Pre-Hook**: {extension} Command: `/{command}` Description: {description} Prompt: {prompt} To execute: `/{command}` - Mandatory hook (
optional: false):## Extension Hooks **Automatic Pre-Hook**: {extension} Executing: `/{command}` EXECUTE_COMMAND: {command} Wait for the result of the hook command before proceeding to the Outline.
- Optional hook (
- If no hooks are registered or
.specify/extensions.ymldoes not exist, skip silently
Goal
Challenge the specification and implementation plan through two distinct expert lenses BEFORE committing to implementation. The Product Lens evaluates whether the right problem is being solved in the right way for users. The Engineering Lens evaluates whether the technical approach is sound, scalable, and free of hidden risks. This dual review prevents costly mid-implementation pivots and catches strategic and technical blind spots early.
Operating Constraints
STRICTLY READ-ONLY: Do not modify spec.md, plan.md, or any other files. Output a structured critique report. Offer to apply approved changes only after the user reviews findings.
CONSTRUCTIVE CHALLENGE: The goal is to strengthen the spec and plan, not to block progress. Every critique item must include a constructive suggestion for improvement.
Constitution Authority: The project constitution (/memory/constitution.md) defines non-negotiable principles. Any spec/plan element conflicting with the constitution is automatically a 🎯 Must-Address item.
Outline
-
Run
{SCRIPT}from repo root and parse FEATURE_DIR and AVAILABLE_DOCS list. All paths must be absolute. For single quotes in args like "I'm Groot", use escape syntax: e.g 'I'''m Groot' (or double-quote if possible: "I'm Groot"). -
Load Critique Context:
- REQUIRED: Read
spec.mdfor requirements, user stories, and acceptance criteria - REQUIRED: Read
plan.mdfor architecture, tech stack, and implementation phases - IF EXISTS: Read
/memory/constitution.mdfor governing principles - IF EXISTS: Read
tasks.mdfor task breakdown (if already generated) - IF EXISTS: Read previous critique reports in FEATURE_DIR/critiques/ for context
- REQUIRED: Read
-
Product Lens Review (CEO/Product Lead Perspective):
Adopt the mindset of an experienced product leader who cares deeply about user value, market fit, and business impact. Evaluate:
3a. Problem Validation
- Is the problem statement clear and well-defined?
- Is this solving a real user pain point, or is it a solution looking for a problem?
- What evidence supports the need for this feature? (user research, data, customer requests)
- Is the scope appropriate — not too broad (trying to do everything) or too narrow (missing the core value)?
3b. User Value Assessment
- Does every user story deliver tangible user value?
- Are the acceptance criteria written from the user's perspective (outcomes, not implementation)?
- Is the user journey complete — or are there gaps where users would get stuck?
- What's the simplest version that would deliver 80% of the value? (MVP analysis)
- Are there unnecessary features that add complexity without proportional value?
3c. Alternative Approaches
- Could a simpler solution achieve the same outcome?
- Are there existing tools, libraries, or services that could replace custom implementation?
- What would a competitor's approach look like?
- What would happen if this feature were NOT built? What's the cost of inaction?
3d. Edge Cases & User Experience
- What happens when things go wrong? (error states, empty states, loading states)
- How does this feature interact with existing functionality?
- Are accessibility considerations addressed?
- Is the feature discoverable and intuitive?
- What are the onboarding/migration implications for existing users?
3e. Success Measurement
- Are the success criteria measurable and time-bound?
- How will you know if this feature is successful after launch?
- What metrics should be tracked?
- What would trigger a rollback decision?
-
Engineering Lens Review (Staff Engineer Perspective):
Adopt the mindset of a senior staff engineer who has seen projects fail due to hidden technical risks. Evaluate:
4a. Architecture Soundness
- Does the architecture follow established patterns for this type of system?
- Are boundaries and interfaces well-defined (separation of concerns)?
- Is the architecture testable at each layer?
- Are there circular dependencies or tight coupling risks?
- Does the architecture support future evolution without major refactoring?
4b. Failure Mode Analysis
- What are the most likely failure modes? (network failures, data corruption, resource exhaustion)
- How does the system degrade gracefully under each failure mode?
- What happens under peak load? Is there a scaling bottleneck?
- What are the blast radius implications — can a failure in this feature affect other parts of the system?
- Are retry, timeout, and circuit-breaker strategies defined?
4c. Security & Privacy Review
- What is the threat model? What attack vectors does this feature introduce?
- Are trust boundaries clearly defined (user input, API responses, third-party data)?
- Is sensitive data handled appropriately (encryption, access control, retention)?
- Are there compliance implications (GDPR, SOC2, HIPAA)?
- Is the principle of least privilege followed?
4d. Performance & Scalability
- Are there potential bottlenecks in the data flow?
- What are the expected data volumes? Will the design handle 10x growth?
- Are caching strategies appropriate and cache invalidation well-defined?
- Are database queries optimized (indexing, pagination, query complexity)?
- Are there resource-intensive operations that should be async or batched?
4e. Testing Strategy
- Is the testing plan comprehensive (unit, integration, E2E)?
- Are the critical paths identified for priority testing?
- Is the test data strategy realistic?
- Are there testability concerns (hard-to-mock dependencies, race conditions)?
- Is the test coverage target appropriate for the risk level?
4f. Operational Readiness
- Is observability planned (logging, metrics, tracing)?
- Are alerting thresholds defined?
- Is there a rollback strategy?
- Are database migrations reversible?
- Is the deployment strategy clear (blue-green, canary, feature flags)?
4g. Dependencies & Integration Risks
- Are third-party dependencies well-understood (stability, licensing, maintenance)?
- Are integration points with existing systems well-defined?
- What happens if an external service is unavailable?
- Are API versioning and backward compatibility considered?
-
Cross-Lens Synthesis: Identify items where both lenses converge (these are highest priority):
- Product simplification that also reduces engineering risk
- Engineering constraints that affect user experience
- Scope adjustments that improve both value delivery and technical feasibility
-
Severity Classification: Classify each finding:
- 🎯 Must-Address: Blocks proceeding to implementation. Critical product gap, security vulnerability, architecture flaw, or constitution violation. Must be resolved before
/speckit.tasks. - 💡 Recommendation: Strongly suggested improvement that would significantly improve quality, value, or risk profile. Should be addressed but won't block progress.
- 🤔 Question: Ambiguity or assumption that needs stakeholder input. Cannot be resolved by the development team alone.
- 🎯 Must-Address: Blocks proceeding to implementation. Critical product gap, security vulnerability, architecture flaw, or constitution violation. Must be resolved before
-
Generate Critique Report: Create the critique report at
FEATURE_DIR/critiques/critique-{timestamp}.mdusing the critique report template. The report must include:- Executive Summary: Overall assessment and readiness to proceed
- Product Lens Findings: Organized by subcategory (3a-3e)
- Engineering Lens Findings: Organized by subcategory (4a-4g)
- Cross-Lens Insights: Items where both perspectives converge
- Findings Summary Table: All items with ID, lens, severity, summary, suggestion
Findings Table Format:
ID Lens Severity Category Finding Suggestion P1 Product 🎯 Problem Validation No evidence of user need Conduct 5 user interviews or reference support tickets E1 Engineering 💡 Failure Modes No retry strategy for API calls Add exponential backoff with circuit breaker X1 Both 🎯 Scope × Risk Feature X adds complexity with unclear value Defer to v2; reduces both scope and technical risk -
Provide Verdict: Based on findings, provide one of:
- ✅ PROCEED: No must-address items. Spec and plan are solid. Run
/speckit.tasksto proceed. - ⚠️ PROCEED WITH UPDATES: Must-address items found but are resolvable. Offer to apply fixes to spec/plan, then proceed.
- 🛑 RETHINK: Fundamental product or architecture concerns. Recommend revisiting the spec with
/speckit.specifyor the plan with/speckit.plan.
- ✅ PROCEED: No must-address items. Spec and plan are solid. Run
-
Offer Remediation: For each must-address item and recommendation:
- Provide a specific suggested edit to
spec.mdorplan.md - Ask: "Would you like me to apply these changes? (all / select / none)"
- If user approves, apply changes to the relevant files
- After applying changes, recommend re-running
/speckit.critiqueto verify
- Provide a specific suggested edit to
Post-Critique Actions
Suggest next steps based on verdict:
- If PROCEED: "Run
/speckit.tasksto break the plan into actionable tasks" - If PROCEED WITH UPDATES: "Review the suggested changes, then run
/speckit.tasks" - If RETHINK: "Consider running
/speckit.specifyto refine the spec or/speckit.planto revise the architecture"
Check for extension hooks (after critique):
- Check if
.specify/extensions.ymlexists in the project root. - If it exists, read it and look for entries under the
hooks.after_critiquekey - If the YAML cannot be parsed or is invalid, skip hook checking silently and continue normally
- Filter out hooks where
enabledis explicitlyfalse. Treat hooks without anenabledfield as enabled by default. - For each remaining hook, do not attempt to interpret or evaluate hook
conditionexpressions:- If the hook has no
conditionfield, or it is null/empty, treat the hook as executable - If the hook defines a non-empty
condition, skip the hook and leave condition evaluation to the HookExecutor implementation
- If the hook has no
- For each executable hook, output the following based on its
optionalflag:- Optional hook (
optional: true):## Extension Hooks **Optional Hook**: {extension} Command: `/{command}` Description: {description} Prompt: {prompt} To execute: `/{command}` - Mandatory hook (
optional: false):## Extension Hooks **Automatic Hook**: {extension} Executing: `/{command}` EXECUTE_COMMAND: {command}
- Optional hook (
- If no hooks are registered or
.specify/extensions.ymldoes not exist, skip silently