mirror of
https://github.com/bmad-code-org/BMAD-METHOD.git
synced 2026-01-30 04:32:02 +00:00
Compare commits
81 Commits
6.0.0-alph
...
docs/tea-i
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
6a9e94cadb | ||
|
|
2b7f7ff421 | ||
|
|
3360666c2a | ||
|
|
274dea16fa | ||
|
|
dcd581c84a | ||
|
|
6d84a60a78 | ||
|
|
59e1b7067c | ||
|
|
1d8df63ac5 | ||
|
|
993d02b8b3 | ||
|
|
5cb5606ba3 | ||
|
|
c83da03621 | ||
|
|
638892289a | ||
|
|
950ea9be2e | ||
|
|
eeebf152af | ||
|
|
d419ac8a70 | ||
|
|
568249e985 | ||
|
|
c0f6401902 | ||
|
|
e535f94325 | ||
|
|
e465ce4bb5 | ||
|
|
9d328082eb | ||
|
|
d4f6642333 | ||
|
|
9f85dade25 | ||
|
|
5870651bad | ||
|
|
eff826eef9 | ||
|
|
0a3cc1d12c | ||
|
|
c3b7e98241 | ||
|
|
2f98f9130a | ||
|
|
c18904d674 | ||
|
|
3e3c92ed3e | ||
|
|
12d3492e0c | ||
|
|
677a00280b | ||
|
|
d19cca79d2 | ||
|
|
8e165b9b57 | ||
|
|
67b70288a6 | ||
|
|
5c76657732 | ||
|
|
7bf05c9d9d | ||
|
|
692f14f2e7 | ||
|
|
2e16650067 | ||
|
|
dc7a7f8c43 | ||
|
|
987410eb75 | ||
|
|
f838486caa | ||
|
|
51aa3dda2f | ||
|
|
35ae4fd024 | ||
|
|
f31659765e | ||
|
|
d1f3844449 | ||
|
|
05ddc2d29b | ||
|
|
c748f0f6cc | ||
|
|
4142972b6a | ||
|
|
cd45d22eb6 | ||
|
|
a297235862 | ||
|
|
d8b13bdb2e | ||
|
|
8699d7d968 | ||
|
|
8b92e5ee59 | ||
|
|
9f53d896b7 | ||
|
|
b46409e71d | ||
|
|
8cffd09fb7 | ||
|
|
2b89ee1302 | ||
|
|
b72c810a1f | ||
|
|
484990de50 | ||
|
|
b8836ced24 | ||
|
|
c7fcf16eae | ||
|
|
529d4a8c95 | ||
|
|
f0520c39d9 | ||
|
|
ff0517f4d0 | ||
|
|
b509fb9a1e | ||
|
|
e0090e5602 | ||
|
|
8d679b177b | ||
|
|
ea421adbf9 | ||
|
|
2a8a4388a9 | ||
|
|
d4a94df29a | ||
|
|
949cf64d3b | ||
|
|
aba9d11c88 | ||
|
|
208f27dcdb | ||
|
|
c15ad174ed | ||
|
|
24cedea690 | ||
|
|
bdb6bde9b5 | ||
|
|
cfdffe3f7a | ||
|
|
7b5b7afdc0 | ||
|
|
59a0eec2e2 | ||
|
|
1f16bb7413 | ||
|
|
b1d1242fcf |
@@ -11,8 +11,8 @@ reviews:
|
||||
walkthrough: false
|
||||
poem: false
|
||||
auto_review:
|
||||
enabled: false
|
||||
drafts: true # Can review drafts. Since it's manually triggered, it's fine.
|
||||
enabled: true
|
||||
drafts: false # Don't review drafts automatically
|
||||
auto_incremental_review: false # always review the whole PR, not just new commits
|
||||
base_branches:
|
||||
- main
|
||||
|
||||
15
.github/workflows/docs.yaml
vendored
15
.github/workflows/docs.yaml
vendored
@@ -39,20 +39,11 @@ jobs:
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Determine site URL
|
||||
id: site-url
|
||||
run: |
|
||||
if [ "${{ github.repository }}" = "bmad-code-org/BMAD-METHOD" ]; then
|
||||
echo "url=https://bmad-code-org.github.io/BMAD-METHOD" >> $GITHUB_OUTPUT
|
||||
else
|
||||
OWNER="${{ github.repository_owner }}"
|
||||
REPO="${{ github.event.repository.name }}"
|
||||
echo "url=https://${OWNER}.github.io/${REPO}" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
- name: Build documentation
|
||||
env:
|
||||
SITE_URL: ${{ steps.site-url.outputs.url }}
|
||||
# Override site URL from GitHub repo variable if set
|
||||
# Otherwise, astro.config.mjs will compute from GITHUB_REPOSITORY
|
||||
SITE_URL: ${{ vars.SITE_URL }}
|
||||
run: npm run docs:build
|
||||
|
||||
- name: Upload artifact
|
||||
|
||||
8
.gitignore
vendored
8
.gitignore
vendored
@@ -46,8 +46,6 @@ CLAUDE.local.md
|
||||
# Project-specific
|
||||
_bmad-core
|
||||
_bmad-creator-tools
|
||||
test-project-install/*
|
||||
sample-project/*
|
||||
flattened-codebase.xml
|
||||
*.stats.md
|
||||
.internal-docs/
|
||||
@@ -66,6 +64,7 @@ shared-modules
|
||||
z*/
|
||||
|
||||
_bmad
|
||||
_bmad-output
|
||||
.claude
|
||||
.codex
|
||||
.github/chatmodes
|
||||
@@ -76,6 +75,7 @@ _bmad
|
||||
|
||||
bmad-custom-src/
|
||||
|
||||
# Docusaurus / Documentation Build
|
||||
.docusaurus/
|
||||
# Astro / Documentation Build
|
||||
website/.astro/
|
||||
website/dist/
|
||||
build/
|
||||
|
||||
@@ -11,7 +11,6 @@ ignores:
|
||||
- .claude/**
|
||||
- .roo/**
|
||||
- .codex/**
|
||||
- .agentvibes/**
|
||||
- .kiro/**
|
||||
- sample-project/**
|
||||
- test-project-install/**
|
||||
|
||||
1
.vscode/settings.json
vendored
1
.vscode/settings.json
vendored
@@ -57,6 +57,7 @@
|
||||
"tileset",
|
||||
"tmpl",
|
||||
"Trae",
|
||||
"Unsharded",
|
||||
"VNET",
|
||||
"webskip"
|
||||
],
|
||||
|
||||
403
CHANGELOG.md
403
CHANGELOG.md
@@ -1,5 +1,408 @@
|
||||
# Changelog
|
||||
|
||||
## [6.0.0-alpha.23]
|
||||
|
||||
**Release: January 11, 2026**
|
||||
|
||||
### 🌟 Key Highlights
|
||||
|
||||
1. **Astro/Starlight Documentation Platform**: Complete migration from Docusaurus to modern Astro+Starlight for superior performance and customization
|
||||
2. **Diataxis Framework Implementation**: Professional documentation restructuring with tutorials, how-to guides, explanations, and references
|
||||
3. **Workflow Creator & Validator**: Powerful new tools for workflow creation with subprocess support and PRD validation
|
||||
4. **TEA Documentation Expansion**: Comprehensive testing documentation with cheat sheets, MCP enhancements, and API testing patterns
|
||||
5. **Brainstorming Revolution**: Research-backed procedural rigor with 100+ idea goal and anti-bias protocols
|
||||
6. **Cursor IDE Modernization**: Refactored from rules-based to command-based architecture for better IDE integration
|
||||
|
||||
### 📚 Documentation Platform Revolution
|
||||
|
||||
**Astro/Starlight Migration:**
|
||||
|
||||
- **From Docusaurus to Astro**: Complete platform migration for improved performance, better customization, and modern tooling
|
||||
- **Starlight Theme**: Professional documentation theme with dark mode default and responsive design
|
||||
- **Build Pipeline Overhaul**: New build-docs.js orchestrates link checking, artifact generation, and Astro build
|
||||
- **LLM-Friendly Documentation**: Generated llms.txt and llms-full.txt for AI agent discoverability
|
||||
- **Downloadable Source Bundles**: bmad-sources.zip and bmad-prompts.zip for offline use
|
||||
|
||||
**Diataxis Framework Implementation:**
|
||||
|
||||
- **Four Content Types**: Professional separation into tutorials, how-to guides, explanations, and references
|
||||
- **21 Files Migrated**: Phase 1 migration of core documentation to Diataxis structure
|
||||
- **42+ Focused Documents**: Phase 2 split of large legacy files into manageable pieces
|
||||
- **FAQ Restructuring**: 7 topic-specific FAQ files with standardized format
|
||||
- **Tutorial Style Guide**: Comprehensive documentation standards for consistent content creation
|
||||
|
||||
**Link Management & Quality:**
|
||||
|
||||
- **Site-Relative Links**: Converted 217 links to repo-relative format (/docs/path/file.md)
|
||||
- **Link Validation Tools**: New validate-doc-links.js and fix-doc-links.js for maintaining link integrity
|
||||
- **Broken Link Fixes**: Resolved ~50 broken internal links across documentation
|
||||
- **BMad Acronym Standardization**: Consistent use of "BMad" (Breakthrough Method of Agile AI Driven Development)
|
||||
- **SEO Optimization**: Absolute URLs in AI meta tags for better web crawler discoverability
|
||||
|
||||
### 🔧 Workflow Creator & Validator (Major Feature)
|
||||
|
||||
**Workflow Creation Tool:**
|
||||
|
||||
- **Subprocess Support**: Advanced workflows can now spawn subprocesses for complex operations
|
||||
- **PRD Validation Step**: New validation step ensures PRD quality before workflow execution
|
||||
- **Trimodal Workflow Creation**: Three-mode workflow generation system
|
||||
- **Quadrivariate Module Workflow**: Four-variable workflow architecture for enhanced flexibility
|
||||
- **Path Violation Checks**: Validator ensures workflows don't violate path constraints
|
||||
- **Max Parallel Mode POC**: Proof-of-concept for parallel workflow validation
|
||||
|
||||
**Workflow Quality Improvements:**
|
||||
|
||||
- **PRD Trimodal Compliance**: PRD workflow now follows trimodal standards
|
||||
- **Standardized Step Formatting**: Consistent markdown formatting across workflow and PRD steps
|
||||
- **Better Suggested Next Steps**: Improved workflow completion guidance
|
||||
- **Variable Naming Standardization**: {project_root} → {project-root} across all workflows
|
||||
|
||||
### 🧪 TEA Documentation Expansion
|
||||
|
||||
**Comprehensive Testing Guides:**
|
||||
|
||||
- **Cheat Sheets**: Quick reference guides for common testing scenarios
|
||||
- **MCP Enhancements**: Model Context Protocol improvements for testing workflows
|
||||
- **API Testing Patterns**: Best practices for API testing documentation
|
||||
- **Design Philosophy Callout**: Clear explanation of TEA's design principles
|
||||
- **Context Engineering Glossary**: New glossary entry defining context engineering concepts
|
||||
- **Fragment Count Updates**: Accurate documentation of TEA workflow components
|
||||
- **Playwright Utils Examples**: Updated code examples for playwright-utils integration
|
||||
|
||||
### 💡 Brainstorming Workflow Overhaul
|
||||
|
||||
**Research-Backed Procedural Rigor:**
|
||||
|
||||
- **100+ Idea Goal**: Emphasis on quantity-first approach to unlock better quality ideas
|
||||
- **Anti-Bias Protocol**: Domain pivot every 10 ideas to reduce cognitive biases
|
||||
- **Chain-of-Thought Requirements**: Reasoning before idea generation
|
||||
- **Simulated Temperature**: Prompts for higher divergence in ideation
|
||||
- **Standardized Idea Format**: Quality control template for consistent output
|
||||
- **Energy Checkpoints**: Multiple continuation options to maintain creative flow
|
||||
|
||||
**Exploration Menu Improvements:**
|
||||
|
||||
- **Letter-Based Navigation**: [K/T/A/B/C] options instead of numbers for clarity
|
||||
- **Keep/Try/Advanced/Break/Continue**: Clear action options for idea refinement
|
||||
- **Universal Facilitation Rules**: Consistent guidelines across all brainstorming steps
|
||||
- **Quality Growth Enforcement**: Balance between quantity and quality metrics
|
||||
|
||||
### 🖥️ Cursor IDE Modernization
|
||||
|
||||
**Command-Based Architecture:**
|
||||
|
||||
- **From Rules to Commands**: Complete refactor from rules-based to command-based system
|
||||
- **Command Generation**: Automatic generation of task and tool commands
|
||||
- **Commands Directory**: New `.cursor/commands/bmad/` structure for generated commands
|
||||
- **Cleanup Integration**: Automatic cleanup of old BMAD commands alongside rules
|
||||
- **Enhanced Logging**: Better feedback on agents, tasks, tools, and workflow commands generated
|
||||
|
||||
### 🤖 Agent System Improvements
|
||||
|
||||
**Agent Builder & Validation:**
|
||||
|
||||
- **hasSidecar Field**: All agents now indicate sidecar support (true/false)
|
||||
- **Validation Enforcement**: hasSidecar now required in agent validation
|
||||
- **Better Brownfield Documentation**: Improved brownfield project documentation
|
||||
- **Agent Builder Updates**: Agent builder now uses hasSidecar field
|
||||
- **Agent Editor Integration**: Editor workflow respects hasSidecar configuration
|
||||
|
||||
### 🐛 Bug Fixes & Quality Improvements
|
||||
|
||||
**Critical Fixes:**
|
||||
|
||||
- **Windows Line Endings**: Resolved CRLF issues causing cross-platform problems
|
||||
- **Code-Review File Filtering**: Fixed code-review picking up non-application files
|
||||
- **ERR_REQUIRE_ESM Resolution**: Dynamic import for inquirer v9+ compatibility
|
||||
- **Project-Context Conflicts**: Allow full project-context usage with conflict precedence
|
||||
- **Workflow Paths**: Fixed paths for workflow and sprint status files
|
||||
- **Missing Scripts**: Fixed missing scripts from installation
|
||||
|
||||
**Workflow & Variable Fixes:**
|
||||
|
||||
- **Variable Naming**: Standardized from {project_root} to {project-root} across CIS, BMGD, and BMM modules
|
||||
- **Workflow References**: Fixed broken .yaml → .md workflow references
|
||||
- **Advanced Elicitation Variables**: Fixed undefined variables in brainstorming
|
||||
- **Dependency Format**: Corrected dependency format and added missing frontmatter
|
||||
|
||||
**Code Quality:**
|
||||
|
||||
- **Dependency Updates**: Bumped qs from 6.14.0 to 6.14.1
|
||||
- **CodeRabbit Integration**: Enabled auto-review on new PRs
|
||||
- **TEA Fragment Counts**: Updated fragment counts for accuracy
|
||||
- **Documentation Links**: Fixed Discord channel references (#general-dev → #bmad-development)
|
||||
|
||||
### 🚀 Installation & CLI Improvements
|
||||
|
||||
**Installation Enhancements:**
|
||||
|
||||
- **Workflow Exclusion**: Ability to exclude workflows from being added as commands
|
||||
- **Example Workflow Protection**: Example workflow in workflow builder now excluded from tools
|
||||
- **CNAME Configuration**: Added CNAME file for custom domain support
|
||||
- **Script Fixes**: All scripts now properly included in installation
|
||||
|
||||
### 📊 Statistics
|
||||
|
||||
- **27 commits** since alpha.22
|
||||
- **217 documentation links** converted to site-relative format
|
||||
- **42+ focused documents** created from large legacy files
|
||||
- **7 topic-specific FAQ files** with standardized formatting
|
||||
- **Complete documentation platform** migrated from Docusaurus to Astro/Starlight
|
||||
- **Major workflow tools** added: Creator, Validator with subprocess support
|
||||
- **Brainstorming workflow** overhauled with research-backed rigor
|
||||
|
||||
---
|
||||
|
||||
## [6.0.0-alpha.22]
|
||||
|
||||
**Release: December 31, 2025**
|
||||
|
||||
### 🌟 Key Highlights
|
||||
|
||||
1. **Unified Agent Workflow**: Create, Edit, and Validate workflows consolidated into single powerful agent workflow with separate step paths
|
||||
2. **Agent Knowledge System**: Comprehensive data file architecture with persona properties, validation patterns, and crafting principles
|
||||
3. **Deep Language Integration**: All sharded progressive workflows now support language choice at every step
|
||||
4. **Core Module Documentation**: Extensive docs for core workflows (brainstorming, party mode, advanced elicitation)
|
||||
5. **BMAD Core Concepts**: New documentation structure explaining agents, workflows, modules, and installation
|
||||
6. **Tech Spec Sharded**: create-tech-spec workflow converted to sharded format with orient-first pattern
|
||||
|
||||
### 🤖 Unified Agent Workflow (Major Feature)
|
||||
|
||||
**Consolidated Architecture:**
|
||||
|
||||
- **Single Workflow, Three Paths**: Create, Edit, and Validate operations unified under `src/modules/bmb/workflows/agent/`
|
||||
- **steps-c/**: Create path with 9 comprehensive steps for building new agents
|
||||
- **steps-e/**: Edit path with 10 steps for modifying existing agents
|
||||
- **steps-v/**: Validate path for standalone agent validation review
|
||||
- **data/**: Centralized knowledge base for all agent-building intel
|
||||
|
||||
### 📚 Agent Knowledge System
|
||||
|
||||
**Data File Architecture:**
|
||||
|
||||
Located in `src/modules/bmb/workflows/agent/data/`:
|
||||
|
||||
- **agent-metadata.md** (208 lines) - Complete metadata field reference
|
||||
- **agent-menu-patterns.md** (233 lines) - Menu design patterns and best practices
|
||||
- **agent-compilation.md** (273 lines) - Compilation process documentation
|
||||
- **persona-properties.md** (266 lines) - Persona crafting properties and examples
|
||||
- **principles-crafting.md** (292 lines) - Core principles for agent design
|
||||
- **critical-actions.md** (120 lines) - Critical action patterns
|
||||
- **expert-agent-architecture.md** (236 lines) - Expert agent structure
|
||||
- **expert-agent-validation.md** (173 lines) - Expert-specific validation
|
||||
- **module-agent-validation.md** (124 lines) - Module-specific validation
|
||||
- **simple-agent-architecture.md** (204 lines) - Simple agent structure
|
||||
- **simple-agent-validation.md** (132 lines) - Simple agent validation
|
||||
- **understanding-agent-types.md** (222 lines) - Agent type comparison
|
||||
- **brainstorm-context.md** - Brainstorming guidance
|
||||
- **communication-presets.csv** - Communication style presets
|
||||
|
||||
**Reference Examples:**
|
||||
|
||||
- **reference/module-examples/architect.agent.yaml** - Module agent example
|
||||
- **reference/simple-examples/commit-poet.agent.yaml** - Simple agent example
|
||||
- **journal-keeper/** - Complete sidecar pattern example
|
||||
|
||||
**Templates:**
|
||||
|
||||
- **templates/simple-agent.template.md** - Simple agent template
|
||||
- **templates/expert-agent-template/expert-agent.template.md** - Expert agent template
|
||||
- **templates/expert-agent-sidecar/** - Sidecar templates (instructions, memories)
|
||||
|
||||
### 🌍 Deep Language Integration
|
||||
|
||||
**Progressive Workflow Language Support:**
|
||||
|
||||
- **Every Step Biased**: All sharded progressive workflow steps now include language preference context
|
||||
- **260+ Files Updated**: Comprehensive language integration across:
|
||||
- Core workflows (brainstorming, party mode, advanced elicitation)
|
||||
- BMB workflows (create-agent, create-module, create-workflow, edit-workflow, etc.)
|
||||
- BMGD workflows (game-brief, gdd, narrative, game-architecture, etc.)
|
||||
- BMM workflows (research, create-ux-design, prd, create-architecture, etc.)
|
||||
- **Tested Languages**: Verified working with Spanish and Pirate Speak
|
||||
- **Natural Conversations**: AI agents respond in configured language throughout workflow
|
||||
|
||||
### 📖 Core Module Documentation
|
||||
|
||||
**New Core Documentation Structure:**
|
||||
|
||||
`docs/modules/core/`:
|
||||
|
||||
- **index.md** - Core module overview
|
||||
- **core-workflows.md** - Core workflow documentation
|
||||
- **core-tasks.md** - Core task reference
|
||||
- **brainstorming.md** (100 lines) - Brainstorming workflow guide
|
||||
- **party-mode.md** (50 lines) - Party mode guide
|
||||
- **advanced-elicitation.md** (105 lines) - Advanced elicitation techniques
|
||||
- **document-sharding-guide.md** (133 lines) - Sharded workflow format guide
|
||||
- **global-core-config.md** - Global core configuration reference
|
||||
|
||||
**Advanced Elicitation Moved:**
|
||||
|
||||
- **From**: `docs/` root
|
||||
- **To**: `src/core/workflows/advanced-elicitation/`
|
||||
- **Status**: Now a proper core workflow with methods.csv
|
||||
|
||||
### 📚 BMAD Core Concepts Documentation
|
||||
|
||||
**New Documentation Structure:**
|
||||
|
||||
`docs/bmad-core-concepts/`:
|
||||
|
||||
- **index.md** - Core concepts introduction
|
||||
- **agents.md** (93 lines) - Understanding agents in BMAD
|
||||
- **workflows.md** (89 lines) - Understanding workflows in BMAD
|
||||
- **modules.md** (76 lines) - Understanding modules (BMM, BMGD, CIS, BMB, Core)
|
||||
- **installing/index.md** (77 lines) - Installation guide
|
||||
- **installing/upgrading.md** (144 lines) - Upgrading guide
|
||||
- **bmad-customization/index.md** - Customization overview
|
||||
- **bmad-customization/agents.md** - Agent customization guide
|
||||
- **bmad-customization/workflows.md** (30 lines) - Workflow customization guide
|
||||
- **web-bundles/index.md** (34 lines) - Web bundle distribution guide
|
||||
|
||||
**Documentation Cleanup:**
|
||||
|
||||
- **Removed v4-to-v6-upgrade.md** - Outdated upgrade guide
|
||||
- **Removed document-sharding-guide.md** from docs root (moved to core)
|
||||
- **Removed web-bundles-gemini-gpt-guide.md** - Consolidated into web-bundles/index.md
|
||||
- **Removed getting-started/installation.md** - Migrated to bmad-core-concepts
|
||||
- **Removed all ide-info/*.md files** - Consolidated into web-bundles documentation
|
||||
|
||||
### 🔧 Create-Tech-Spec Sharded Conversion
|
||||
|
||||
**Monolithic to Sharded:**
|
||||
|
||||
- **From**: Single `workflow.yaml` with `instructions.md`
|
||||
- **To**: Sharded `workflow.md` with individual step files
|
||||
- **Pattern**: Orient-first approach (understand before investigating)
|
||||
|
||||
### 🔨 Additional Improvements
|
||||
|
||||
**Workflow Status Path Fixes:**
|
||||
|
||||
- **Corrected Discovery Paths**: workflow-status workflows now properly use planning_artifacts and implementation_artifacts
|
||||
- **Updated All Path Files**: enterprise-brownfield, enterprise-greenfield, method-brownfield, method-greenfield
|
||||
|
||||
**Documentation Updates:**
|
||||
|
||||
- **BMB Agent Creation Guide**: Comprehensive 166-line guide for agent creation
|
||||
- **Workflow Vendoring Doc**: New 42-line guide on workflow customization and inheritance
|
||||
- **Document Project Reference**: Moved from BMM docs to shared location
|
||||
- **Workflows Planning Guide**: New 89-line guide for planning workflows
|
||||
|
||||
**BMB Documentation Streamlining:**
|
||||
|
||||
- **Removed Redundant Docs**: Eliminated duplicate documentation in `src/modules/bmb/docs/`
|
||||
- **Step File Rules**: New 469-line comprehensive guide for step file creation
|
||||
- **Agent Docs Moved**: Agent architecture and validation docs moved to workflow data/
|
||||
|
||||
**Windows Inquirer Fix:**
|
||||
|
||||
- **Another Default Addition**: Additional inquirer default value setting for better Windows multiselection support
|
||||
|
||||
**Code Quality:**
|
||||
|
||||
- **Removed Old BMM README**: Consolidated module documentation
|
||||
- **Removed BMM Troubleshooting**: 661-line doc moved to shared location
|
||||
- **Removed Enterprise Agentic Development**: 686-line doc consolidated
|
||||
- **Removed Scale Adaptive System**: 618-line doc consolidated
|
||||
|
||||
---
|
||||
|
||||
## [6.0.0-alpha.21]
|
||||
|
||||
**Release: December 27, 2025**
|
||||
|
||||
### 🌟 Key Highlights
|
||||
|
||||
1. **Consistent Menu System**: All agents now use standardized 2-letter menu codes (e.g., "rd" for research, "ca" for create-architecture)
|
||||
2. **Planning Artifacts Architecture**: Phase 1-3 workflows now properly segregate planning artifacts from documentation
|
||||
3. **Windows Installer Fixed Again**: Updated inquirer to resolve multiselection tool issues
|
||||
4. **Auto-Injected Features**: Chat and party mode automatically injected into all agents
|
||||
5. **Validation System**: All agents now pass comprehensive new validation checks
|
||||
|
||||
### 🎯 Consistent Menu System (Major Feature)
|
||||
|
||||
**Standardized 2-Letter Codes:**
|
||||
|
||||
- **Compound Menu Triggers**: All agents now use consistent 2-letter compound trigger format (e.g., `bmm-rd`, `bmm-ca`)
|
||||
- **Improved UX**: Shorter, more memorable command shortcuts across all modules
|
||||
- **Module Prefixing**: Menu items properly scoped by module prefix (bmm-, bmgd-, cis-, bmb-)
|
||||
- **Universal Pattern**: All 22 agents updated to follow the same menu structure
|
||||
|
||||
**Agent Updates:**
|
||||
|
||||
- **BMM Module**: 9 agents with standardized menus (pm, analyst, architect, dev, ux-designer, tech-writer, sm, tea, quick-flow-solo-dev)
|
||||
- **BMGD Module**: 6 agents with standardized menus (game-architect, game-designer, game-dev, game-qa, game-scrum-master, game-solo-dev)
|
||||
- **CIS Module**: 6 agents with standardized menus (innovation-strategist, design-thinking-coach, creative-problem-solver, brainstorming-coach, presentation-master, storyteller)
|
||||
- **BMB Module**: 3 agents with standardized menus (bmad-builder, agent-builder, module-builder, workflow-builder)
|
||||
- **Core Module**: BMAD Master agent updated with consistent menu patterns
|
||||
|
||||
### 📁 Planning Artifacts Architecture
|
||||
|
||||
**Content Segregation Implementation:**
|
||||
|
||||
- **Phase 1-3 Workflows**: All planning workflows now use `planning_artifacts` folder (default changed from `docs`)
|
||||
- **Proper Input Discovery**: Workflows follow consistent input discovery patterns from planning artifacts
|
||||
- **Output Management**: Planning artifacts properly separated from long-term documentation
|
||||
- **Affected Workflows**:
|
||||
- Product Brief: Updated discovery and output to planning artifacts
|
||||
- PRD: Fixed discovery and output to planning artifacts
|
||||
- UX Design: Updated all steps for proper artifact handling
|
||||
- Architecture: Updated discovery and output flow
|
||||
- Game Architecture: Updated for planning artifacts
|
||||
- Story Creation: Updated workflow output paths
|
||||
|
||||
**File Organization:**
|
||||
|
||||
- **Planning Artifacts**: Ephemeral planning documents (prd.md, product-brief.md, ux-design.md, architecture.md)
|
||||
- **Documentation**: Long-term project documentation (separate from planning)
|
||||
- **Module Configuration**: BMM and BMGD modules updated with proper default paths
|
||||
|
||||
### 🪟 Windows Installer Fixes
|
||||
|
||||
**Inquirer Multiselection Fix:**
|
||||
|
||||
- **Updated Inquirer Version**: Resolved tool multiselection issues that were causing Windows installer failures
|
||||
- **Better Compatibility**: Improved handling of checkbox and multi-select prompts on Windows(?)
|
||||
|
||||
### 🤖 Agent System Improvements
|
||||
|
||||
**Auto-Injected Features:**
|
||||
|
||||
- **Chat Mode**: Automatically injected into all agents during compilation
|
||||
- **Party Mode**: Automatically injected into all agents during compilation
|
||||
- **Reduced Manual Configuration**: No need to manually add these features to agent definitions
|
||||
- **Consistent Behavior**: All agents now have uniform access to chat and party mode capabilities
|
||||
|
||||
**Agent Normalization:**
|
||||
|
||||
- **All Agents Validated**: All 22 agents pass comprehensive validation checks
|
||||
- **Schema Enforcement**: Proper compound trigger validation implemented
|
||||
- **Metadata Cleanup**: Removed obsolete and inconsistent metadata patterns
|
||||
- **Test Fixtures Updated**: Validation test fixtures aligned with new requirements
|
||||
|
||||
### 🔧 Bug Fixes & Cleanup
|
||||
|
||||
**Docusaurus Merge Recovery:**
|
||||
|
||||
- **Restored Agent Files**: Fixed agent files accidentally modified in Docusaurus merge (PR #1191)
|
||||
- **Reference Cleanup**: Removed obsolete agent reference examples (journal-keeper, security-engineer, trend-analyst)
|
||||
- **Test Fixture Updates**: Aligned test fixtures with current validation requirements
|
||||
|
||||
**Code Quality:**
|
||||
|
||||
- **Schema Improvements**: Enhanced agent schema validation with better error messages
|
||||
- **Removed Redundancy**: Cleaned up duplicate and obsolete agent definitions
|
||||
- **Installer Cleanup**: Removed unused configuration code from BMM installer
|
||||
|
||||
**Planning Artifacts Path:**
|
||||
- Default: `planning_artifacts/` (configurable in module.yaml)
|
||||
- Previous: `docs/`
|
||||
- Benefit: Clear separation between planning work and permanent documentation
|
||||
|
||||
---
|
||||
|
||||
## [6.0.0-alpha.20]
|
||||
|
||||
**Release: December 23, 2025**
|
||||
|
||||
@@ -4,8 +4,9 @@ Thank you for considering contributing to the BMad project! We believe in **Huma
|
||||
|
||||
💬 **Discord Community**: Join our [Discord server](https://discord.gg/gk8jAdXWmj) for real-time discussions:
|
||||
|
||||
- **#general-dev** - Technical discussions, feature ideas, and development questions
|
||||
- **#bugs-issues** - Bug reports and issue discussions
|
||||
- **#bmad-development** - Technical discussions and development questions
|
||||
- **#suggestions-feedback** - Feature ideas and suggestions
|
||||
- **#report-bugs-and-issues** - Bug reports and issue discussions
|
||||
|
||||
## Our Philosophy
|
||||
|
||||
@@ -56,7 +57,7 @@ Every contribution should strengthen human-AI collaboration. Ask yourself: **"Do
|
||||
### Reporting Bugs
|
||||
|
||||
1. **Check existing issues** first to avoid duplicates
|
||||
2. **Consider discussing in Discord** (#bugs-issues channel) for quick help
|
||||
2. **Consider discussing in Discord** (#report-bugs-and-issues channel) for quick help
|
||||
3. **Use the bug report template** when creating a new issue - it guides you through providing:
|
||||
- Clear bug description
|
||||
- Steps to reproduce
|
||||
@@ -67,7 +68,7 @@ Every contribution should strengthen human-AI collaboration. Ask yourself: **"Do
|
||||
|
||||
### Suggesting Features or New Modules
|
||||
|
||||
1. **Discuss first in Discord** (#general-dev channel) - the feature request template asks if you've done this
|
||||
1. **Discuss first in Discord** (#suggestions-feedback channel) - the feature request template asks if you've done this
|
||||
2. **Check existing issues and discussions** to avoid duplicates
|
||||
3. **Use the feature request template** when creating an issue
|
||||
4. **Be specific** about why this feature would benefit the BMad community and strengthen human-AI collaboration
|
||||
@@ -77,7 +78,7 @@ Every contribution should strengthen human-AI collaboration. Ask yourself: **"Do
|
||||
⚠️ **Required before submitting PRs:**
|
||||
|
||||
1. **For bugs**: Check if an issue exists (create one using the bug template if not)
|
||||
2. **For features**: Discuss in Discord (#general-dev) AND create a feature request issue
|
||||
2. **For features**: Discuss in Discord (#suggestions-feedback) AND create a feature request issue
|
||||
3. **For large changes**: Always open an issue first to discuss alignment
|
||||
|
||||
Please propose small, granular changes! For large or significant changes, discuss in Discord and open an issue first. This prevents wasted effort on PRs that may not align with planned changes.
|
||||
@@ -251,8 +252,9 @@ By participating in this project, you agree to abide by our Code of Conduct. We
|
||||
## Need Help?
|
||||
|
||||
- 💬 Join our [Discord Community](https://discord.gg/gk8jAdXWmj):
|
||||
- **#general-dev** - Technical questions and feature discussions
|
||||
- **#bugs-issues** - Get help with bugs before filing issues
|
||||
- **#bmad-development** - Technical questions and discussions
|
||||
- **#suggestions-feedback** - Feature ideas and suggestions
|
||||
- **#report-bugs-and-issues** - Get help with bugs before filing issues
|
||||
- 🐛 Report bugs using the [bug report template](https://github.com/bmad-code-org/BMAD-METHOD/issues/new?template=bug_report.md)
|
||||
- 💡 Suggest features using the [feature request template](https://github.com/bmad-code-org/BMAD-METHOD/issues/new?template=feature_request.md)
|
||||
- 📖 Browse the [GitHub Discussions](https://github.com/bmad-code-org/BMAD-METHOD/discussions)
|
||||
|
||||
252
README.md
252
README.md
@@ -1,245 +1,91 @@
|
||||
# BMad Method & BMad Core
|
||||
# BMad Method
|
||||
|
||||
[](https://www.npmjs.com/package/bmad-method)
|
||||
[](https://www.npmjs.com/package/bmad-method)
|
||||
[](https://www.npmjs.com/package/bmad-method)
|
||||
[](LICENSE)
|
||||
[](https://nodejs.org)
|
||||
[](https://discord.gg/gk8jAdXWmj)
|
||||
|
||||
---
|
||||
**Build More, Architect Dreams** — An AI-driven agile development framework with 21 specialized agents, 50+ guided workflows, and scale-adaptive intelligence that adjusts from bug fixes to enterprise systems.
|
||||
|
||||
<div align="center">
|
||||
**100% free and open source.** No paywalls. No gated content. No gated Discord. We believe in empowering everyone, not just those who can pay.
|
||||
|
||||
## 🎉 NEW: BMAD V6 Installer - Create & Share Custom Content!
|
||||
## Why BMad?
|
||||
|
||||
The completely revamped **BMAD V6 installer** now includes built-in support for creating, installing, and sharing custom modules, agents, workflows, templates, and tools! Build your own AI solutions or share them with your team - and real soon, with the whole BMad Community througha verified community sharing portal!
|
||||
Traditional AI tools do the thinking for you, producing average results. BMad agents act as expert collaborators who guide you through structured workflows to bring out your best thinking.
|
||||
|
||||
**✨ What's New:**
|
||||
- **Scale-Adaptive**: Automatically adjusts planning depth based on project complexity (Level 0-4)
|
||||
- **Structured Workflows**: Grounded in agile best practices across analysis, planning, architecture, and implementation
|
||||
- **Specialized Agents**: 12+ domain experts (PM, Architect, Developer, UX, Scrum Master, and more)
|
||||
- **Complete Lifecycle**: From brainstorming to deployment, with just-in-time documentation
|
||||
|
||||
- 📦 **Streamlined Custom Module Installation** - Package your custom content as installable modules
|
||||
- 🤖 **Agent & Workflow Sharing** - Distribute standalone agents and workflows
|
||||
- 🔄 **Unitary Module Support** - Install individual components without full modules
|
||||
- ⚙️ **Dependency Management** - Automatic handling of module dependencies
|
||||
- 🛡️ **Update-Safe Customization** - Your custom content persists through updates
|
||||
## Quick Start
|
||||
|
||||
**📚 Learn More:**
|
||||
|
||||
- [**Custom Content Overview**](./docs/custom-content.md) - Discover all supported content types
|
||||
- [**Installation Guide**](./docs/custom-content-installation.md) - Learn to create and install custom content
|
||||
- [**Detail Content Docs**](./src/modules/bmb/docs/index.md) - Reference details for agents, modules, workflows and the bmad builder
|
||||
- [**2 Very simple Custom Modules of questionable quality**](./samples/sample-custom-modules/README.md) - if you want to download and try to install a custom shared module, get an idea of how to bundle and share your own, or create your own personal agents, workflows and modules.
|
||||
|
||||
</div>
|
||||
|
||||
---
|
||||
|
||||
## AI-Driven Agile Development That Scales From Bug Fixes to Enterprise
|
||||
|
||||
**Build More, Architect Dreams** (BMAD) with **21 specialized AI agents** across 4 official modules, and **50+ guided workflows** that adapt to your project's complexity—from quick bug fixes to enterprise platforms, and new step file workflows that allow for incredibly long workflows to stay on the rails longer than ever before!
|
||||
|
||||
Additionally - when we say 'Build More, Architect Dreams' - we mean it! The BMad Builder has landed, and now as of Alpha.15 is fully supported in the installation flow via NPX - custom stand along agents, workflows and the modules of your dreams! The community forge will soon open, endless possibility awaits!
|
||||
|
||||
> **🚀 v6 is a MASSIVE upgrade from v4!** Complete architectural overhaul, scale-adaptive intelligence, visual workflows, and the powerful BMad Core framework. v4 users: this changes everything. [See what's new →](#whats-new-in-v6)
|
||||
|
||||
> **📌 v6 Alpha Status:** Near-beta quality with vastly improved stability. Documentation is being finalized. New videos coming soon to [BMadCode YouTube](https://www.youtube.com/@BMadCode).
|
||||
|
||||
## 🎯 Why BMad Method?
|
||||
|
||||
Unlike generic AI coding assistants, BMad Method provides **structured, battle-tested workflows** powered by specialized agents who understand agile development. Each agent has deep domain expertise—from product management to architecture to testing—working together seamlessly.
|
||||
|
||||
**✨ Key Benefits:**
|
||||
|
||||
- **Scale-Adaptive Intelligence** - Automatically adjusts planning depth from bug fixes to enterprise systems
|
||||
- **Complete Development Lifecycle** - Analysis → Planning → Architecture → Implementation
|
||||
- **Specialized Expertise** - 19 agents with specific roles (PM, Architect, Developer, UX Designer, etc.)
|
||||
- **Proven Methodologies** - Built on agile best practices with AI amplification
|
||||
- **IDE Integration** - Works with Claude Code, Cursor, Windsurf, VS Code
|
||||
|
||||
## 🏗️ The Power of BMad Core
|
||||
|
||||
**BMad Method** is actually a sophisticated module built on top of **BMad Core** (**C**ollaboration **O**ptimized **R**eflection **E**ngine). This revolutionary architecture means:
|
||||
|
||||
- **BMad Core** provides the universal framework for human-AI collaboration
|
||||
- **BMad Method** leverages Core to deliver agile development workflows
|
||||
- **BMad Builder** lets YOU create custom modules as powerful as BMad Method itself
|
||||
|
||||
With **BMad Builder**, you can architect both simple agents and vastly complex domain-specific modules (legal, medical, finance, education, creative) that will soon be sharable in an **official community marketplace**. Imagine building and sharing your own specialized AI team!
|
||||
|
||||
## 📊 See It In Action
|
||||
|
||||
<p align="center">
|
||||
<img src="./src/modules/bmm/docs/images/workflow-method-greenfield.svg" alt="BMad Method Workflow" width="100%">
|
||||
</p>
|
||||
|
||||
<p align="center">
|
||||
<em>Complete BMad Method workflow showing all phases, agents, and decision points</em>
|
||||
</p>
|
||||
|
||||
## 🚀 Get Started in 3 Steps
|
||||
|
||||
### 1. Install BMad Method
|
||||
**Prerequisites**: [Node.js](https://nodejs.org) v20+
|
||||
|
||||
```bash
|
||||
# Install v6 Alpha (recommended)
|
||||
npx bmad-method@alpha install
|
||||
|
||||
# Or stable v4 for production
|
||||
npx bmad-method install
|
||||
```
|
||||
|
||||
### 2. Initialize Your Project
|
||||
Follow the installer prompts to configure your project. Then run:
|
||||
|
||||
Load any agent in your IDE and run:
|
||||
|
||||
```
|
||||
```bash
|
||||
*workflow-init
|
||||
```
|
||||
|
||||
This analyzes your project and recommends the right workflow track.
|
||||
This analyzes your project and recommends a track:
|
||||
|
||||
### 3. Choose Your Track
|
||||
| Track | Best For | Time to First Story |
|
||||
| --------------- | ------------------------- | ------------------- |
|
||||
| **Quick Flow** | Bug fixes, small features | ~5 minutes |
|
||||
| **BMad Method** | Products and platforms | ~15 minutes |
|
||||
| **Enterprise** | Compliance-heavy systems | ~30 minutes |
|
||||
|
||||
BMad Method adapts to your needs with three intelligent tracks:
|
||||
## Modules
|
||||
|
||||
| Track | Use For | Planning | Time to Start |
|
||||
| ------------------ | ------------------------- | ----------------------- | ------------- |
|
||||
| **⚡ Quick Flow** | Bug fixes, small features | Tech spec only | < 5 minutes |
|
||||
| **📋 BMad Method** | Products, platforms | PRD + Architecture + UX | < 15 minutes |
|
||||
| **🏢 Enterprise** | Compliance, scale | Full governance suite | < 30 minutes |
|
||||
| Module | Purpose |
|
||||
| ------------------------------------- | -------------------------------------------------------- |
|
||||
| **BMad Method (BMM)** | Core agile development with 34 workflows across 4 phases |
|
||||
| **BMad Builder (BMB)** | Create custom agents and domain-specific modules |
|
||||
| **Creative Intelligence Suite (CIS)** | Innovation, brainstorming, and problem-solving |
|
||||
|
||||
> **Not sure?** Run `*workflow-init` and let BMad analyze your project goal.
|
||||
## Documentation
|
||||
|
||||
## 🔄 How It Works: 4-Phase Methodology
|
||||
**[Full Documentation](http://docs.bmad-method.org)** — Tutorials, how-to guides, concepts, and reference
|
||||
|
||||
BMad Method guides you through a proven development lifecycle:
|
||||
|
||||
1. **📊 Analysis** (Optional) - Brainstorm, research, and explore solutions
|
||||
2. **📝 Planning** - Create PRDs, tech specs, or game design documents
|
||||
3. **🏗️ Solutioning** - Design architecture, UX, and technical approach
|
||||
4. **⚡ Implementation** - Story-driven development with continuous validation
|
||||
|
||||
Each phase has specialized workflows and agents working together to deliver exceptional results.
|
||||
|
||||
## 🤖 Meet Your Team
|
||||
|
||||
**12 Specialized Agents** working in concert:
|
||||
|
||||
| Development | Architecture | Product | Leadership |
|
||||
| ----------- | -------------- | ------------- | -------------- |
|
||||
| Developer | Architect | PM | Scrum Master |
|
||||
| UX Designer | Test Architect | Analyst | BMad Master |
|
||||
| Tech Writer | Game Architect | Game Designer | Game Developer |
|
||||
|
||||
**Test Architect** integrates with `@seontechnologies/playwright-utils` for production-ready fixture-based utilities.
|
||||
|
||||
Each agent brings deep expertise and can be customized to match your team's style.
|
||||
|
||||
## 📦 What's Included
|
||||
|
||||
### Core Modules
|
||||
|
||||
- **BMad Method (BMM)** - Complete agile development framework
|
||||
- 12 specialized agents
|
||||
- 34 workflows across 4 phases
|
||||
- Scale-adaptive planning
|
||||
- [→ Documentation Hub](./src/modules/bmm/docs/index.md)
|
||||
|
||||
- **BMad Builder (BMB)** - Create custom agents and workflows
|
||||
- Build anything from simple agents to complex modules
|
||||
- Create domain-specific solutions (legal, medical, finance, education)
|
||||
- [→ Builder Guide](./src/modules/bmb/docs/index.md)
|
||||
|
||||
- **Creative Intelligence Suite (CIS)** - Innovation & problem-solving
|
||||
- Brainstorming, design thinking, storytelling
|
||||
- 5 creative facilitation workflows
|
||||
- [→ Creative Workflows](./src/modules/cis/docs/index.md)
|
||||
|
||||
### Key Features
|
||||
|
||||
- **🎨 Customizable Agents** - Modify personalities, expertise, and communication styles
|
||||
- **🌐 Multi-Language Support** - Separate settings for communication and code output
|
||||
- **📄 Document Sharding** - 90% token savings for large projects
|
||||
- **🔄 Update-Safe** - Your customizations persist through updates
|
||||
- **🚀 Web Bundles** - Use in ChatGPT, Claude Projects, or Gemini Gems
|
||||
|
||||
## 📚 Documentation
|
||||
|
||||
### Quick Links
|
||||
|
||||
- **[Quick Start Guide](./src/modules/bmm/docs/quick-start.md)** - 15-minute introduction
|
||||
- **[Complete BMM Documentation](./src/modules/bmm/docs/index.md)** - All guides and references
|
||||
- **[Agent Customization](./docs/agent-customization-guide.md)** - Personalize your agents
|
||||
- **[All Documentation](./docs/index.md)** - Complete documentation index
|
||||
- [Getting Started Tutorial](http://docs.bmad-method.org/tutorials/getting-started/getting-started-bmadv6/)
|
||||
- [Upgrading from Previous Versions](http://docs.bmad-method.org/how-to/installation/upgrade-to-v6/)
|
||||
|
||||
### For v4 Users
|
||||
|
||||
- **[v4 Documentation](https://github.com/bmad-code-org/BMAD-METHOD/tree/V4)**
|
||||
- **[v4 to v6 Upgrade Guide](./docs/v4-to-v6-upgrade.md)**
|
||||
- **[v4 Documentation](https://github.com/bmad-code-org/BMAD-METHOD/tree/V4/docs)**
|
||||
|
||||
## 💬 Community & Support
|
||||
## Community
|
||||
|
||||
- **[Discord Community](https://discord.gg/gk8jAdXWmj)** - Get help, share projects
|
||||
- **[GitHub Issues](https://github.com/bmad-code-org/BMAD-METHOD/issues)** - Report bugs, request features
|
||||
- **[YouTube Channel](https://www.youtube.com/@BMadCode)** - Video tutorials and demos
|
||||
- **[Web Bundles](https://bmad-code-org.github.io/bmad-bundles/)** - Pre-built agent bundles
|
||||
- **[Code of Conduct](.github/CODE_OF_CONDUCT.md)** - Community guidelines
|
||||
- [Discord](https://discord.gg/gk8jAdXWmj) — Get help, share ideas, collaborate
|
||||
- [YouTube](https://www.youtube.com/@BMadCode) — Tutorials, master class, and podcast (launching Feb 2025)
|
||||
- [GitHub Issues](https://github.com/bmad-code-org/BMAD-METHOD/issues) — Bug reports and feature requests
|
||||
- [Discussions](https://github.com/bmad-code-org/BMAD-METHOD/discussions) — Community conversations
|
||||
|
||||
## 🛠️ Development
|
||||
## Support BMad
|
||||
|
||||
For contributors working on the BMad codebase:
|
||||
BMad is free for everyone — and always will be. If you'd like to support development:
|
||||
|
||||
```bash
|
||||
# Run all quality checks
|
||||
npm test
|
||||
- ⭐ [Star us on GitHub](https://github.com/bmad-code-org/BMAD-METHOD/) — Helps others discover BMad
|
||||
- 📺 [Subscribe on YouTube](https://www.youtube.com/@BMadCode) — Master class launching Feb 2026
|
||||
- ☕ [Buy Me a Coffee](https://buymeacoffee.com/bmad) — Fuel the development
|
||||
- 🏢 Corporate sponsorship — DM on Discord
|
||||
- 🎤 Speaking & Media — Available for conferences, podcasts, interviews (Discord)
|
||||
|
||||
# Development commands
|
||||
npm run lint:fix # Fix code style
|
||||
npm run format:fix # Auto-format code
|
||||
npm run bundle # Build web bundles
|
||||
```
|
||||
## Contributing
|
||||
|
||||
See [CONTRIBUTING.md](CONTRIBUTING.md) for full development guidelines.
|
||||
We welcome contributions! See [CONTRIBUTING.md](CONTRIBUTING.md) for guidelines.
|
||||
|
||||
## What's New in v6
|
||||
## License
|
||||
|
||||
**v6 represents a complete architectural revolution from v4:**
|
||||
|
||||
### 🚀 Major Upgrades
|
||||
|
||||
- **BMad Core Framework** - Modular architecture enabling custom domain solutions
|
||||
- **Scale-Adaptive Intelligence** - Automatic adjustment from bug fixes to enterprise
|
||||
- **Visual Workflows** - Beautiful SVG diagrams showing complete methodology
|
||||
- **BMad Builder Module** - Create and share your own AI agent teams
|
||||
- **50+ Workflows** - Up from 20 in v4, covering every development scenario
|
||||
- **19 Specialized Agents** - Enhanced with customizable personalities and expertise
|
||||
- **Update-Safe Customization** - Your configs persist through all updates
|
||||
- **Web Bundles** - Use agents in ChatGPT, Claude, and Gemini
|
||||
- **Multi-Language Support** - Separate settings for communication and code
|
||||
- **Document Sharding** - 90% token savings for large projects
|
||||
|
||||
### 🔄 For v4 Users
|
||||
|
||||
- **[Comprehensive Upgrade Guide](./docs/v4-to-v6-upgrade.md)** - Step-by-step migration
|
||||
- **[v4 Documentation Archive](https://github.com/bmad-code-org/BMAD-METHOD/tree/V4)** - Legacy reference
|
||||
- Backwards compatibility where possible
|
||||
- Smooth migration path with installer detection
|
||||
|
||||
## 📄 License
|
||||
|
||||
MIT License - See [LICENSE](LICENSE) for details.
|
||||
|
||||
**Trademarks:** BMad™ and BMAD-METHOD™ are trademarks of BMad Code, LLC.
|
||||
|
||||
Supported by: <a href="https://m.do.co/c/00f11bd932bb"><img src="https://opensource.nyc3.cdn.digitaloceanspaces.com/attribution/assets/SVG/DO_Logo_horizontal_blue.svg" height="24" alt="DigitalOcean" style="vertical-align: middle;"></a>
|
||||
MIT License — see [LICENSE](LICENSE) for details.
|
||||
|
||||
---
|
||||
|
||||
<p align="center">
|
||||
<a href="https://github.com/bmad-code-org/BMAD-METHOD/graphs/contributors">
|
||||
<img src="https://contrib.rocks/image?repo=bmad-code-org/BMAD-METHOD" alt="Contributors">
|
||||
</a>
|
||||
</p>
|
||||
**BMad** and **BMAD-METHOD** are trademarks of BMad Code, LLC.
|
||||
|
||||
<p align="center">
|
||||
<sub>Built with ❤️ for the human-AI collaboration community</sub>
|
||||
</p>
|
||||
[](https://github.com/bmad-code-org/BMAD-METHOD/graphs/contributors)
|
||||
|
||||
85
SECURITY.md
Normal file
85
SECURITY.md
Normal file
@@ -0,0 +1,85 @@
|
||||
# Security Policy
|
||||
|
||||
## Supported Versions
|
||||
|
||||
We release security patches for the following versions:
|
||||
|
||||
| Version | Supported |
|
||||
| ------- | ------------------ |
|
||||
| Latest | :white_check_mark: |
|
||||
| < Latest | :x: |
|
||||
|
||||
We recommend always using the latest version of BMad Method to ensure you have the most recent security updates.
|
||||
|
||||
## Reporting a Vulnerability
|
||||
|
||||
We take security vulnerabilities seriously. If you discover a security issue, please report it responsibly.
|
||||
|
||||
### How to Report
|
||||
|
||||
**Do NOT report security vulnerabilities through public GitHub issues.**
|
||||
|
||||
Instead, please report them via one of these methods:
|
||||
|
||||
1. **GitHub Security Advisories** (Preferred): Use [GitHub's private vulnerability reporting](https://github.com/bmad-code-org/BMAD-METHOD/security/advisories/new) to submit a confidential report.
|
||||
|
||||
2. **Discord**: Contact a maintainer directly via DM on our [Discord server](https://discord.gg/gk8jAdXWmj).
|
||||
|
||||
### What to Include
|
||||
|
||||
Please include as much of the following information as possible:
|
||||
|
||||
- Type of vulnerability (e.g., prompt injection, path traversal, etc.)
|
||||
- Full paths of source file(s) related to the vulnerability
|
||||
- Step-by-step instructions to reproduce the issue
|
||||
- Proof-of-concept or exploit code (if available)
|
||||
- Impact assessment of the vulnerability
|
||||
|
||||
### Response Timeline
|
||||
|
||||
- **Initial Response**: Within 48 hours of receiving your report
|
||||
- **Status Update**: Within 7 days with our assessment
|
||||
- **Resolution Target**: Critical issues within 30 days; other issues within 90 days
|
||||
|
||||
### What to Expect
|
||||
|
||||
1. We will acknowledge receipt of your report
|
||||
2. We will investigate and validate the vulnerability
|
||||
3. We will work on a fix and coordinate disclosure timing with you
|
||||
4. We will credit you in the security advisory (unless you prefer to remain anonymous)
|
||||
|
||||
## Security Scope
|
||||
|
||||
### In Scope
|
||||
|
||||
- Vulnerabilities in BMad Method core framework code
|
||||
- Security issues in agent definitions or workflows that could lead to unintended behavior
|
||||
- Path traversal or file system access issues
|
||||
- Prompt injection vulnerabilities that bypass intended agent behavior
|
||||
- Supply chain vulnerabilities in dependencies
|
||||
|
||||
### Out of Scope
|
||||
|
||||
- Security issues in user-created custom agents or modules
|
||||
- Vulnerabilities in third-party AI providers (Claude, GPT, etc.)
|
||||
- Issues that require physical access to a user's machine
|
||||
- Social engineering attacks
|
||||
- Denial of service attacks that don't exploit a specific vulnerability
|
||||
|
||||
## Security Best Practices for Users
|
||||
|
||||
When using BMad Method:
|
||||
|
||||
1. **Review Agent Outputs**: Always review AI-generated code before executing it
|
||||
2. **Limit File Access**: Configure your AI IDE to limit file system access where possible
|
||||
3. **Keep Updated**: Regularly update to the latest version
|
||||
4. **Validate Dependencies**: Review any dependencies added by generated code
|
||||
5. **Environment Isolation**: Consider running AI-assisted development in isolated environments
|
||||
|
||||
## Acknowledgments
|
||||
|
||||
We appreciate the security research community's efforts in helping keep BMad Method secure. Contributors who report valid security issues will be acknowledged in our security advisories.
|
||||
|
||||
---
|
||||
|
||||
Thank you for helping keep BMad Method and our community safe.
|
||||
9
docs/404.md
Normal file
9
docs/404.md
Normal file
@@ -0,0 +1,9 @@
|
||||
---
|
||||
title: Page Not Found
|
||||
template: splash
|
||||
---
|
||||
|
||||
|
||||
The page you're looking for doesn't exist or has been moved.
|
||||
|
||||
[Return to Home](/docs/index.md)
|
||||
@@ -1,4 +1,7 @@
|
||||
# Workflow Diagram Maintenance
|
||||
---
|
||||
title: "Workflow Diagram Maintenance"
|
||||
---
|
||||
|
||||
|
||||
## Regenerating SVG from Excalidraw
|
||||
|
||||
766
docs/_STYLE_GUIDE.md
Normal file
766
docs/_STYLE_GUIDE.md
Normal file
@@ -0,0 +1,766 @@
|
||||
---
|
||||
title: "Documentation Style Guide"
|
||||
---
|
||||
|
||||
Internal guidelines for maintaining consistent, high-quality documentation across the BMad Method project. This document is not included in the Starlight sidebar — it's for contributors and maintainers, not end users.
|
||||
|
||||
## Quick Principles
|
||||
|
||||
1. **Clarity over brevity** — Be concise, but never at the cost of understanding
|
||||
2. **Consistent structure** — Follow established patterns so readers know what to expect
|
||||
3. **Strategic visuals** — Use admonitions, tables, and diagrams purposefully
|
||||
4. **Scannable content** — Headers, lists, and callouts help readers find what they need
|
||||
|
||||
## Validation Steps
|
||||
|
||||
Before submitting documentation changes, run these checks from the repo root:
|
||||
|
||||
1. **Fix link format** — Convert relative links (`./`, `../`) to site-relative paths (`/path/`)
|
||||
```bash
|
||||
npm run docs:fix-links # Preview changes
|
||||
npm run docs:fix-links -- --write # Apply changes
|
||||
```
|
||||
|
||||
2. **Validate links** — Check all links point to existing files
|
||||
```bash
|
||||
npm run docs:validate-links # Preview issues
|
||||
npm run docs:validate-links -- --write # Auto-fix where possible
|
||||
```
|
||||
|
||||
3. **Build the site** — Verify no build errors
|
||||
```bash
|
||||
npm run docs:build
|
||||
```
|
||||
|
||||
## Tutorial Structure
|
||||
|
||||
Every tutorial should follow this structure:
|
||||
|
||||
```
|
||||
1. Title + Hook (1-2 sentences describing the outcome)
|
||||
2. Version/Module Notice (info or warning admonition as appropriate)
|
||||
3. What You'll Learn (bullet list of outcomes)
|
||||
4. Prerequisites (info admonition)
|
||||
5. Quick Path (tip admonition - TL;DR summary)
|
||||
6. Understanding [Topic] (context before steps - tables for phases/agents)
|
||||
7. Installation (if applicable)
|
||||
8. Step 1: [First Major Task]
|
||||
9. Step 2: [Second Major Task]
|
||||
10. Step 3: [Third Major Task]
|
||||
11. What You've Accomplished (summary + folder structure if applicable)
|
||||
12. Quick Reference (commands table)
|
||||
13. Common Questions (FAQ format)
|
||||
14. Getting Help (community links)
|
||||
15. Key Takeaways (tip admonition - memorable points)
|
||||
```
|
||||
|
||||
Not all sections are required for every tutorial, but this is the standard flow.
|
||||
|
||||
## How-To Structure
|
||||
|
||||
How-to guides are task-focused and shorter than tutorials. They answer "How do I do X?" for users who already understand the basics.
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence: "Use the `X` workflow to...")
|
||||
2. When to Use This (bullet list of scenarios)
|
||||
3. When to Skip This (optional - for workflows that aren't always needed)
|
||||
4. Prerequisites (note admonition)
|
||||
5. Steps (numbered ### subsections)
|
||||
6. What You Get (output/artifacts produced)
|
||||
7. Example (optional - concrete usage scenario)
|
||||
8. Tips (optional - best practices, common pitfalls)
|
||||
9. Next Steps (optional - what to do after completion)
|
||||
```
|
||||
|
||||
Include sections only when they add value. A simple how-to might only need Hook, Prerequisites, Steps, and What You Get.
|
||||
|
||||
### How-To vs Tutorial
|
||||
|
||||
| Aspect | How-To | Tutorial |
|
||||
|--------|--------|----------|
|
||||
| **Length** | 50-150 lines | 200-400 lines |
|
||||
| **Audience** | Users who know the basics | New users learning concepts |
|
||||
| **Focus** | Complete a specific task | Understand a workflow end-to-end |
|
||||
| **Sections** | 5-8 sections | 12-15 sections |
|
||||
| **Examples** | Brief, inline | Detailed, step-by-step |
|
||||
|
||||
### How-To Visual Elements
|
||||
|
||||
Use admonitions strategically in how-to guides:
|
||||
|
||||
| Admonition | Use In How-To |
|
||||
|------------|---------------|
|
||||
| `:::note[Prerequisites]` | Required dependencies, agents, prior steps |
|
||||
| `:::tip[Pro Tip]` | Optional shortcuts or best practices |
|
||||
| `:::caution[Common Mistake]` | Pitfalls to avoid |
|
||||
| `:::note[Example]` | Brief usage example inline with steps |
|
||||
|
||||
**Guidelines:**
|
||||
- **1-2 admonitions max** per how-to (they're shorter than tutorials)
|
||||
- **Prerequisites as admonition** makes scanning easier
|
||||
- **Tips section** can be a flat list instead of admonition if there are multiple tips
|
||||
- **Skip admonitions entirely** for very simple how-tos
|
||||
|
||||
### How-To Checklist
|
||||
|
||||
Before submitting a how-to:
|
||||
|
||||
- [ ] Hook is one clear sentence starting with "Use the `X` workflow to..."
|
||||
- [ ] When to Use This has 3-5 bullet points
|
||||
- [ ] Prerequisites listed (admonition or flat list)
|
||||
- [ ] Steps are numbered `###` subsections with action verbs
|
||||
- [ ] What You Get describes output artifacts
|
||||
- [ ] No horizontal rules (`---`)
|
||||
- [ ] No `####` headers
|
||||
- [ ] No "Related" section (sidebar handles navigation)
|
||||
- [ ] 1-2 admonitions maximum
|
||||
|
||||
## Explanation Structure
|
||||
|
||||
Explanation documents help users understand concepts, features, and design decisions. They answer "What is X?" and "Why does X matter?" rather than "How do I do X?"
|
||||
|
||||
### Types of Explanation Documents
|
||||
|
||||
| Type | Purpose | Example |
|
||||
|------|---------|---------|
|
||||
| **Index/Landing** | Overview of a topic area with navigation | `core-concepts/index.md` |
|
||||
| **Concept** | Define and explain a core concept | `what-are-agents.md` |
|
||||
| **Feature** | Deep dive into a specific capability | `quick-flow.md` |
|
||||
| **Philosophy** | Explain design decisions and rationale | `why-solutioning-matters.md` |
|
||||
| **FAQ** | Answer common questions (see FAQ Sections below) | `brownfield-faq.md` |
|
||||
|
||||
### General Explanation Structure
|
||||
|
||||
```
|
||||
1. Title + Hook (1-2 sentences explaining the topic)
|
||||
2. Overview/Definition (what it is, why it matters)
|
||||
3. Key Concepts (### subsections for main ideas)
|
||||
4. Comparison Table (optional - when comparing options)
|
||||
5. When to Use / When Not to Use (optional - decision guidance)
|
||||
6. Diagram (optional - mermaid for processes/flows)
|
||||
7. Next Steps (optional - where to go from here)
|
||||
```
|
||||
|
||||
### Index/Landing Pages
|
||||
|
||||
Index pages orient users within a topic area.
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence overview)
|
||||
2. Content Table (links with descriptions)
|
||||
3. Getting Started (numbered list for new users)
|
||||
4. Choose Your Path (optional - decision tree for different goals)
|
||||
```
|
||||
|
||||
**Example hook:** "Understanding the fundamental building blocks of the BMad Method."
|
||||
|
||||
### Concept Explainers
|
||||
|
||||
Concept pages define and explain core ideas.
|
||||
|
||||
```
|
||||
1. Title + Hook (what it is in one sentence)
|
||||
2. Types/Categories (if applicable, with ### subsections)
|
||||
3. Key Differences Table (comparing types/options)
|
||||
4. Components/Parts (breakdown of elements)
|
||||
5. Which Should You Use? (decision guidance)
|
||||
6. Creating/Customizing (brief pointer to how-to guides)
|
||||
```
|
||||
|
||||
**Example hook:** "Agents are AI assistants that help you accomplish tasks. Each agent has a unique personality, specialized capabilities, and an interactive menu."
|
||||
|
||||
### Feature Explainers
|
||||
|
||||
Feature pages provide deep dives into specific capabilities.
|
||||
|
||||
```
|
||||
1. Title + Hook (what the feature does)
|
||||
2. Quick Facts (optional - "Perfect for:", "Time to:")
|
||||
3. When to Use / When Not to Use (with bullet lists)
|
||||
4. How It Works (process overview, mermaid diagram if helpful)
|
||||
5. Key Benefits (what makes it valuable)
|
||||
6. Comparison Table (vs alternatives if applicable)
|
||||
7. When to Graduate/Upgrade (optional - when to use something else)
|
||||
```
|
||||
|
||||
**Example hook:** "Quick Spec Flow is a streamlined alternative to the full BMad Method for Quick Flow track projects."
|
||||
|
||||
### Philosophy/Rationale Documents
|
||||
|
||||
Philosophy pages explain design decisions and reasoning.
|
||||
|
||||
```
|
||||
1. Title + Hook (the principle or decision)
|
||||
2. The Problem (what issue this addresses)
|
||||
3. The Solution (how this approach solves it)
|
||||
4. Key Principles (### subsections for main ideas)
|
||||
5. Benefits (what users gain)
|
||||
6. When This Applies (scope of the principle)
|
||||
```
|
||||
|
||||
**Example hook:** "Phase 3 (Solutioning) translates **what** to build (from Planning) into **how** to build it (technical design)."
|
||||
|
||||
### Explanation Visual Elements
|
||||
|
||||
Use these elements strategically in explanation documents:
|
||||
|
||||
| Element | Use For |
|
||||
|---------|---------|
|
||||
| **Comparison tables** | Contrasting types, options, or approaches |
|
||||
| **Mermaid diagrams** | Process flows, phase sequences, decision trees |
|
||||
| **"Best for:" lists** | Quick decision guidance |
|
||||
| **Code examples** | Illustrating concepts (keep brief) |
|
||||
|
||||
**Guidelines:**
|
||||
- **Use diagrams sparingly** — one mermaid diagram per document maximum
|
||||
- **Tables over prose** — for any comparison of 3+ items
|
||||
- **Avoid step-by-step instructions** — point to how-to guides instead
|
||||
|
||||
### Explanation Checklist
|
||||
|
||||
Before submitting an explanation document:
|
||||
|
||||
- [ ] Hook clearly states what the document explains
|
||||
- [ ] Content organized into scannable `##` sections
|
||||
- [ ] Comparison tables used for contrasting options
|
||||
- [ ] No horizontal rules (`---`)
|
||||
- [ ] No `####` headers
|
||||
- [ ] No "Related" section (sidebar handles navigation)
|
||||
- [ ] No "Next:" navigation links (sidebar handles navigation)
|
||||
- [ ] Diagrams have clear labels and flow
|
||||
- [ ] Links to how-to guides for "how do I do this?" questions
|
||||
- [ ] 2-3 admonitions maximum
|
||||
|
||||
## Reference Structure
|
||||
|
||||
Reference documents provide quick lookup information for users who know what they're looking for. They answer "What are the options?" and "What does X do?" rather than explaining concepts or teaching skills.
|
||||
|
||||
### Types of Reference Documents
|
||||
|
||||
| Type | Purpose | Example |
|
||||
|------|---------|---------|
|
||||
| **Index/Landing** | Navigation to reference content | `workflows/index.md` |
|
||||
| **Catalog** | Quick-reference list of items | `agents/index.md` |
|
||||
| **Deep-Dive** | Detailed single-item reference | `document-project.md` |
|
||||
| **Configuration** | Settings and config documentation | `core-tasks.md` |
|
||||
| **Glossary** | Term definitions | `glossary/index.md` |
|
||||
| **Comprehensive** | Extensive multi-item reference | `bmgd-workflows.md` |
|
||||
|
||||
### Reference Index Pages
|
||||
|
||||
For navigation landing pages:
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence describing scope)
|
||||
2. Content Sections (## for each category)
|
||||
- Bullet list with links and brief descriptions
|
||||
```
|
||||
|
||||
Keep these minimal — their job is navigation, not explanation.
|
||||
|
||||
### Catalog Reference (Item Lists)
|
||||
|
||||
For quick-reference lists of items:
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence)
|
||||
2. Items (## for each item)
|
||||
- Brief description (one sentence)
|
||||
- **Commands:** or **Key Info:** as flat list
|
||||
3. Universal/Shared (## section if applicable)
|
||||
```
|
||||
|
||||
**Guidelines:**
|
||||
- Use `##` for items, not `###`
|
||||
- No horizontal rules between items — whitespace is sufficient
|
||||
- No "Related" section — sidebar handles navigation
|
||||
- Keep descriptions to 1 sentence per item
|
||||
|
||||
### Item Deep-Dive Reference
|
||||
|
||||
For detailed single-item documentation:
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence purpose)
|
||||
2. Quick Facts (optional note admonition)
|
||||
- Module, Command, Input, Output as list
|
||||
3. Purpose/Overview (## section)
|
||||
4. How to Invoke (code block)
|
||||
5. Key Sections (## for each major aspect)
|
||||
- Use ### for sub-options within sections
|
||||
6. Notes/Caveats (tip or caution admonition)
|
||||
```
|
||||
|
||||
**Guidelines:**
|
||||
- Start with "quick facts" so readers immediately know scope
|
||||
- Use admonitions for important caveats
|
||||
- No "Related Documentation" section — sidebar handles this
|
||||
|
||||
### Configuration Reference
|
||||
|
||||
For settings, tasks, and config documentation:
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence explaining what these configure)
|
||||
2. Table of Contents (jump links if 4+ items)
|
||||
3. Items (## for each config/task)
|
||||
- **Bold summary** — one sentence describing what it does
|
||||
- **Use it when:** bullet list of scenarios
|
||||
- **How it works:** numbered steps
|
||||
- **Output:** expected result (if applicable)
|
||||
```
|
||||
|
||||
**Guidelines:**
|
||||
- Table of contents only needed for 4+ items
|
||||
- Keep "How it works" to 3-5 steps maximum
|
||||
- No horizontal rules between items
|
||||
|
||||
### Glossary Reference
|
||||
|
||||
For term definitions:
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence)
|
||||
2. Navigation (jump links to categories)
|
||||
3. Categories (## for each category)
|
||||
- Terms (### for each term)
|
||||
- Definition (1-3 sentences, no prefix)
|
||||
- Related context or example (optional)
|
||||
```
|
||||
|
||||
**Guidelines:**
|
||||
- Group related terms into categories
|
||||
- Keep definitions concise — link to explanation docs for depth
|
||||
- Use `###` for terms (makes them linkable and scannable)
|
||||
- No horizontal rules between terms
|
||||
|
||||
### Comprehensive Reference Guide
|
||||
|
||||
For extensive multi-item references:
|
||||
|
||||
```
|
||||
1. Title + Hook (one sentence)
|
||||
2. Overview (## section)
|
||||
- Diagram or table showing organization
|
||||
3. Major Sections (## for each phase/category)
|
||||
- Items (### for each item)
|
||||
- Standardized fields: Command, Agent, Input, Output, Description
|
||||
- Optional: Steps, Features, Use when
|
||||
4. Next Steps (optional — only if genuinely helpful)
|
||||
```
|
||||
|
||||
**Guidelines:**
|
||||
- Standardize item fields across all items in the guide
|
||||
- Use tables for comparing multiple items at once
|
||||
- One diagram maximum per document
|
||||
- No horizontal rules — use `##` sections for separation
|
||||
|
||||
### General Reference Guidelines
|
||||
|
||||
These apply to all reference documents:
|
||||
|
||||
| Do | Don't |
|
||||
|----|-------|
|
||||
| Use `##` for major sections, `###` for items within | Use `####` headers |
|
||||
| Use whitespace for separation | Use horizontal rules (`---`) |
|
||||
| Link to explanation docs for "why" | Explain concepts inline |
|
||||
| Use tables for structured data | Use nested lists |
|
||||
| Use admonitions for important notes | Use bold paragraphs for callouts |
|
||||
| Keep descriptions to 1-2 sentences | Write paragraphs of explanation |
|
||||
|
||||
### Reference Admonitions
|
||||
|
||||
Use sparingly — 1-2 maximum per reference document:
|
||||
|
||||
| Admonition | Use In Reference |
|
||||
|------------|------------------|
|
||||
| `:::note[Prerequisites]` | Dependencies needed before using |
|
||||
| `:::tip[Pro Tip]` | Shortcuts or advanced usage |
|
||||
| `:::caution[Important]` | Critical caveats or warnings |
|
||||
|
||||
### Reference Checklist
|
||||
|
||||
Before submitting a reference document:
|
||||
|
||||
- [ ] Hook clearly states what the document references
|
||||
- [ ] Appropriate structure for reference type (catalog, deep-dive, etc.)
|
||||
- [ ] No horizontal rules (`---`)
|
||||
- [ ] No `####` headers
|
||||
- [ ] No "Related" section (sidebar handles navigation)
|
||||
- [ ] Items use consistent structure throughout
|
||||
- [ ] Descriptions are 1-2 sentences maximum
|
||||
- [ ] Tables used for structured/comparative data
|
||||
- [ ] 1-2 admonitions maximum
|
||||
- [ ] Links to explanation docs for conceptual depth
|
||||
|
||||
## Glossary Structure
|
||||
|
||||
Glossaries provide quick-reference definitions for project terminology. Unlike other reference documents, glossaries prioritize compact scanability over narrative explanation.
|
||||
|
||||
### Layout Strategy
|
||||
|
||||
Starlight auto-generates a right-side "On this page" navigation from headers. Use this to your advantage:
|
||||
|
||||
- **Categories as `##` headers** — Appear in right nav for quick jumping
|
||||
- **Terms in tables** — Compact rows, not individual headers
|
||||
- **No inline TOC** — Right sidebar handles navigation; inline TOC is redundant
|
||||
- **Right nav shows categories only** — Cleaner than listing every term
|
||||
|
||||
This approach reduces content length by ~70% while improving navigation.
|
||||
|
||||
### Table Format
|
||||
|
||||
Each category uses a two-column table:
|
||||
|
||||
```md
|
||||
## Category Name
|
||||
|
||||
| Term | Definition |
|
||||
|------|------------|
|
||||
| **Agent** | Specialized AI persona with specific expertise that guides users through workflows. |
|
||||
| **Workflow** | Multi-step guided process that orchestrates AI agent activities to produce deliverables. |
|
||||
```
|
||||
|
||||
### Definition Guidelines
|
||||
|
||||
| Do | Don't |
|
||||
|----|-------|
|
||||
| Start with what it IS or DOES | Start with "This is..." or "A [term] is..." |
|
||||
| Keep to 1-2 sentences | Write multi-paragraph explanations |
|
||||
| Bold the term name in the cell | Use plain text for terms |
|
||||
| Link to docs for deep dives | Explain full concepts inline |
|
||||
|
||||
### Context Markers
|
||||
|
||||
For terms with limited scope, add italic context at the start of the definition:
|
||||
|
||||
```md
|
||||
| **Tech-Spec** | *Quick Flow only.* Comprehensive technical plan for small changes. |
|
||||
| **PRD** | *BMad Method/Enterprise.* Product-level planning document with vision and goals. |
|
||||
```
|
||||
|
||||
Standard markers:
|
||||
- `*Quick Flow only.*`
|
||||
- `*BMad Method/Enterprise.*`
|
||||
- `*Phase N.*`
|
||||
- `*BMGD.*`
|
||||
- `*Brownfield.*`
|
||||
|
||||
### Cross-References
|
||||
|
||||
Link related terms when helpful. Reference the category anchor since individual terms aren't headers:
|
||||
|
||||
```md
|
||||
| **Tech-Spec** | *Quick Flow only.* Technical plan for small changes. See [PRD](#planning-documents). |
|
||||
```
|
||||
|
||||
### Organization
|
||||
|
||||
- **Alphabetize terms** within each category table
|
||||
- **Alphabetize categories** or order by logical progression (foundational → specific)
|
||||
- **No catch-all sections** — Every term belongs in a specific category
|
||||
|
||||
### Glossary Checklist
|
||||
|
||||
Before submitting glossary changes:
|
||||
|
||||
- [ ] Terms in tables, not individual headers
|
||||
- [ ] Terms alphabetized within each category
|
||||
- [ ] No inline TOC (right nav handles navigation)
|
||||
- [ ] No horizontal rules (`---`)
|
||||
- [ ] Definitions are 1-2 sentences
|
||||
- [ ] Context markers italicized at definition start
|
||||
- [ ] Term names bolded in table cells
|
||||
- [ ] No "A [term] is..." definitions
|
||||
|
||||
## Visual Hierarchy
|
||||
|
||||
### Avoid
|
||||
|
||||
| Pattern | Problem |
|
||||
|---------|---------|
|
||||
| `---` horizontal rules | Fragment the reading flow |
|
||||
| `####` deep headers | Create visual noise |
|
||||
| **Important:** bold paragraphs | Blend into body text |
|
||||
| Deeply nested lists | Hard to scan |
|
||||
| Code blocks for non-code | Confusing semantics |
|
||||
|
||||
### Use Instead
|
||||
|
||||
| Pattern | When to Use |
|
||||
|---------|-------------|
|
||||
| White space + section headers | Natural content separation |
|
||||
| Bold text within paragraphs | Inline emphasis |
|
||||
| Admonitions | Callouts that need attention |
|
||||
| Tables | Structured comparisons |
|
||||
| Flat lists | Scannable options |
|
||||
|
||||
## Admonitions
|
||||
|
||||
Use Starlight admonitions strategically:
|
||||
|
||||
```md
|
||||
:::tip[Title]
|
||||
Shortcuts, best practices, "pro tips"
|
||||
:::
|
||||
|
||||
:::note[Title]
|
||||
Context, definitions, examples, prerequisites
|
||||
:::
|
||||
|
||||
:::caution[Title]
|
||||
Caveats, potential issues, things to watch out for
|
||||
:::
|
||||
|
||||
:::danger[Title]
|
||||
Critical warnings only — data loss, security issues
|
||||
:::
|
||||
```
|
||||
|
||||
### Standard Admonition Uses
|
||||
|
||||
| Admonition | Standard Use in Tutorials |
|
||||
|------------|---------------------------|
|
||||
| `:::note[Prerequisites]` | What users need before starting |
|
||||
| `:::tip[Quick Path]` | TL;DR summary at top of tutorial |
|
||||
| `:::caution[Fresh Chats]` | Context limitation reminders |
|
||||
| `:::note[Example]` | Command/response examples |
|
||||
| `:::tip[Check Your Status]` | How to verify progress |
|
||||
| `:::tip[Remember These]` | Key takeaways at end |
|
||||
|
||||
### Admonition Guidelines
|
||||
|
||||
- **Always include a title** for tip, info, and warning
|
||||
- **Keep content brief** — 1-3 sentences ideal
|
||||
- **Don't overuse** — More than 3-4 per major section feels noisy
|
||||
- **Don't nest** — Admonitions inside admonitions are hard to read
|
||||
|
||||
## Headers
|
||||
|
||||
### Budget
|
||||
|
||||
- **8-12 `##` sections** for full tutorials following standard structure
|
||||
- **2-3 `###` subsections** per `##` section maximum
|
||||
- **Avoid `####` entirely** — use bold text or admonitions instead
|
||||
|
||||
### Naming
|
||||
|
||||
- Use action verbs for steps: "Install BMad", "Create Your Plan"
|
||||
- Use nouns for reference sections: "Common Questions", "Quick Reference"
|
||||
- Keep headers short and scannable
|
||||
|
||||
## Code Blocks
|
||||
|
||||
### Do
|
||||
|
||||
```md
|
||||
```bash
|
||||
npx bmad-method install
|
||||
```
|
||||
```
|
||||
|
||||
### Don't
|
||||
|
||||
````md
|
||||
```
|
||||
You: Do something
|
||||
Agent: [Response here]
|
||||
```
|
||||
````
|
||||
|
||||
For command/response examples, use an admonition instead:
|
||||
|
||||
```md
|
||||
:::note[Example]
|
||||
Run `workflow-status` and the agent will tell you the next recommended workflow.
|
||||
:::
|
||||
```
|
||||
|
||||
## Tables
|
||||
|
||||
Use tables for:
|
||||
- Phases and what happens in each
|
||||
- Agent roles and when to use them
|
||||
- Command references
|
||||
- Comparing options
|
||||
- Step sequences with multiple attributes
|
||||
|
||||
Keep tables simple:
|
||||
- 2-4 columns maximum
|
||||
- Short cell content
|
||||
- Left-align text, right-align numbers
|
||||
|
||||
### Standard Tables
|
||||
|
||||
**Phases Table:**
|
||||
```md
|
||||
| Phase | Name | What Happens |
|
||||
|-------|------|--------------|
|
||||
| 1 | Analysis | Brainstorm, research *(optional)* |
|
||||
| 2 | Planning | Requirements — PRD or tech-spec *(required)* |
|
||||
```
|
||||
|
||||
**Quick Reference Table:**
|
||||
```md
|
||||
| Command | Agent | Purpose |
|
||||
|---------|-------|---------|
|
||||
| `*workflow-init` | Analyst | Initialize a new project |
|
||||
| `*prd` | PM | Create Product Requirements Document |
|
||||
```
|
||||
|
||||
**Build Cycle Table:**
|
||||
```md
|
||||
| Step | Agent | Workflow | Purpose |
|
||||
|------|-------|----------|---------|
|
||||
| 1 | SM | `create-story` | Create story file from epic |
|
||||
| 2 | DEV | `dev-story` | Implement the story |
|
||||
```
|
||||
|
||||
## Lists
|
||||
|
||||
### Flat Lists (Preferred)
|
||||
|
||||
```md
|
||||
- **Option A** — Description of option A
|
||||
- **Option B** — Description of option B
|
||||
- **Option C** — Description of option C
|
||||
```
|
||||
|
||||
### Numbered Steps
|
||||
|
||||
```md
|
||||
1. Load the **PM agent** in a new chat
|
||||
2. Run the PRD workflow: `*prd`
|
||||
3. Output: `PRD.md`
|
||||
```
|
||||
|
||||
### Avoid Deep Nesting
|
||||
|
||||
```md
|
||||
<!-- Don't do this -->
|
||||
1. First step
|
||||
- Sub-step A
|
||||
- Detail 1
|
||||
- Detail 2
|
||||
- Sub-step B
|
||||
2. Second step
|
||||
```
|
||||
|
||||
Instead, break into separate sections or use an admonition for context.
|
||||
|
||||
## Links
|
||||
|
||||
- Use descriptive link text: `[Tutorial Style Guide](./tutorial-style.md)`
|
||||
- Avoid "click here" or bare URLs
|
||||
- Prefer relative paths within docs
|
||||
|
||||
## Images
|
||||
|
||||
- Always include alt text
|
||||
- Add a caption in italics below: `*Description of the image.*`
|
||||
- Use SVG for diagrams when possible
|
||||
- Store in `./images/` relative to the document
|
||||
|
||||
## FAQ Sections
|
||||
|
||||
Use a TOC with jump links, `###` headers for questions, and direct answers:
|
||||
|
||||
```md
|
||||
## Questions
|
||||
|
||||
- [Do I always need architecture?](#do-i-always-need-architecture)
|
||||
- [Can I change my plan later?](#can-i-change-my-plan-later)
|
||||
|
||||
### Do I always need architecture?
|
||||
|
||||
Only for BMad Method and Enterprise tracks. Quick Flow skips to implementation.
|
||||
|
||||
### Can I change my plan later?
|
||||
|
||||
Yes. The SM agent has a `correct-course` workflow for handling scope changes.
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](...) or ask in [Discord](...) so we can add it!
|
||||
```
|
||||
|
||||
### FAQ Guidelines
|
||||
|
||||
- **TOC at top** — Jump links under `## Questions` for quick navigation
|
||||
- **`###` headers** — Questions are scannable and linkable (no `Q:` prefix)
|
||||
- **Direct answers** — No `**A:**` prefix, just the answer
|
||||
- **No "Related Documentation"** — Sidebar handles navigation; avoid repetitive links
|
||||
- **End with CTA** — "Have a question not answered here?" with issue/Discord links
|
||||
|
||||
## Folder Structure Blocks
|
||||
|
||||
Show project structure in "What You've Accomplished":
|
||||
|
||||
````md
|
||||
Your project now has:
|
||||
|
||||
```
|
||||
your-project/
|
||||
├── _bmad/ # BMad configuration
|
||||
├── _bmad-output/
|
||||
│ ├── PRD.md # Your requirements document
|
||||
│ └── bmm-workflow-status.yaml # Progress tracking
|
||||
└── ...
|
||||
```
|
||||
````
|
||||
|
||||
## Example: Before and After
|
||||
|
||||
### Before (Noisy)
|
||||
|
||||
```md
|
||||
---
|
||||
|
||||
## Getting Started
|
||||
|
||||
### Step 1: Initialize
|
||||
|
||||
#### What happens during init?
|
||||
|
||||
**Important:** You need to describe your project.
|
||||
|
||||
1. Your project goals
|
||||
- What you want to build
|
||||
- Why you're building it
|
||||
2. The complexity
|
||||
- Small, medium, or large
|
||||
|
||||
---
|
||||
```
|
||||
|
||||
### After (Clean)
|
||||
|
||||
```md
|
||||
## Step 1: Initialize Your Project
|
||||
|
||||
Load the **Analyst agent** in your IDE, wait for the menu, then run `workflow-init`.
|
||||
|
||||
:::note[What Happens]
|
||||
You'll describe your project goals and complexity. The workflow then recommends a planning track.
|
||||
:::
|
||||
```
|
||||
|
||||
## Checklist
|
||||
|
||||
Before submitting a tutorial:
|
||||
|
||||
- [ ] Follows the standard structure
|
||||
- [ ] Has version/module notice if applicable
|
||||
- [ ] Has "What You'll Learn" section
|
||||
- [ ] Has Prerequisites admonition
|
||||
- [ ] Has Quick Path TL;DR admonition
|
||||
- [ ] No horizontal rules (`---`)
|
||||
- [ ] No `####` headers
|
||||
- [ ] Admonitions used for callouts (not bold paragraphs)
|
||||
- [ ] Tables used for structured data (phases, commands, agents)
|
||||
- [ ] Lists are flat (no deep nesting)
|
||||
- [ ] Has "What You've Accomplished" section
|
||||
- [ ] Has Quick Reference table
|
||||
- [ ] Has Common Questions section
|
||||
- [ ] Has Getting Help section
|
||||
- [ ] Has Key Takeaways admonition
|
||||
- [ ] All links use descriptive text
|
||||
- [ ] Images have alt text and captions
|
||||
30
docs/_archive/customize-workflows.md
Normal file
30
docs/_archive/customize-workflows.md
Normal file
@@ -0,0 +1,30 @@
|
||||
---
|
||||
title: "Workflow Customization Guide"
|
||||
---
|
||||
|
||||
Customize and optimize workflows with step replacement and hooks.
|
||||
|
||||
## Status
|
||||
|
||||
:::note[Coming Soon]
|
||||
Workflow customization is an upcoming capability. This guide will be updated when the feature is available.
|
||||
:::
|
||||
|
||||
## What to Expect
|
||||
|
||||
Workflow customization will allow you to:
|
||||
|
||||
- **Replace Steps** - Swap out specific workflow steps with custom implementations
|
||||
- **Add Hooks** - Inject custom behavior before/after workflow steps
|
||||
- **Extend Workflows** - Create new workflows based on existing ones
|
||||
- **Override Behavior** - Customize workflow logic for your project's needs
|
||||
|
||||
## For Now
|
||||
|
||||
While workflow customization is in development, you can:
|
||||
|
||||
- **Create Custom Workflows** - Use the BMad Builder to create entirely new workflows
|
||||
- **Customize Agents** - Modify agent behavior using [Agent Customization](/docs/how-to/customization/customize-agents.md)
|
||||
- **Provide Feedback** - Share your workflow customization needs with the community
|
||||
|
||||
**In the meantime:** Learn how to [create custom workflows](/docs/explanation/bmad-builder/index.md) from scratch.
|
||||
247
docs/_archive/getting-started-bmadv4.md
Normal file
247
docs/_archive/getting-started-bmadv4.md
Normal file
@@ -0,0 +1,247 @@
|
||||
---
|
||||
title: "Getting Started with BMad v4"
|
||||
description: Install BMad and create your first planning document
|
||||
---
|
||||
|
||||
|
||||
Build software faster using AI-powered workflows with specialized agents that guide you through planning, architecture, and implementation.
|
||||
|
||||
:::note[Stable Release]
|
||||
This tutorial covers BMad v4, the current stable release. For the latest features (with potential breaking changes), see the [BMad v6 Alpha tutorial](./getting-started-bmadv6.md).
|
||||
:::
|
||||
|
||||
## What You'll Learn
|
||||
|
||||
- Install and configure BMad for your IDE
|
||||
- Understand how BMad organizes work into phases and agents
|
||||
- Initialize a project and choose a planning track
|
||||
- Create your first requirements document
|
||||
|
||||
:::note[Prerequisites]
|
||||
- **Node.js 20+** — Required for the installer
|
||||
- **Git** — Recommended for version control
|
||||
- **AI-powered IDE** — Claude Code, Cursor, Windsurf, or similar
|
||||
- **A project idea** — Even a simple one works for learning
|
||||
:::
|
||||
|
||||
:::tip[Quick Path]
|
||||
**Install** → `npx bmad-method install`
|
||||
**Initialize** → Load Analyst agent, run `workflow-init`
|
||||
**Plan** → PM creates PRD, Architect creates architecture
|
||||
**Build** → SM manages sprints, DEV implements stories
|
||||
**Always use fresh chats** for each workflow to avoid context issues.
|
||||
:::
|
||||
|
||||
## Understanding BMad
|
||||
|
||||
BMad helps you build software through guided workflows with specialized AI agents. The process follows four phases:
|
||||
|
||||
| Phase | Name | What Happens |
|
||||
|-------|------|--------------|
|
||||
| 1 | Analysis | Brainstorm, research *(optional)* |
|
||||
| 2 | Planning | Requirements — PRD or tech-spec *(required)* |
|
||||
| 3 | Solutioning | Architecture, design decisions *(varies by track)* |
|
||||
| 4 | Implementation | Build code story by story *(required)* |
|
||||
|
||||
Based on your project's complexity, BMad offers three planning tracks:
|
||||
|
||||
| Track | Best For | Documents Created |
|
||||
|-------|----------|-------------------|
|
||||
| **Quick Flow** | Bug fixes, simple features, clear scope | Tech-spec only |
|
||||
| **BMad Method** | Products, platforms, complex features | PRD + Architecture + UX |
|
||||
| **Enterprise** | Compliance, multi-tenant, enterprise needs | PRD + Architecture + Security + DevOps |
|
||||
|
||||
## Installation
|
||||
|
||||
Open a terminal in your project directory and run:
|
||||
|
||||
```bash
|
||||
npx bmad-method install
|
||||
```
|
||||
|
||||
The interactive installer guides you through setup:
|
||||
|
||||
- **Choose Installation Location** — Select current directory (recommended), subdirectory, or custom path
|
||||
- **Select Your AI Tool** — Claude Code, Cursor, Windsurf, or other
|
||||
- **Choose Modules** — Select **BMM** (BMad Method) for this tutorial
|
||||
- **Accept Defaults** — Customize later in `_bmad/[module]/config.yaml`
|
||||
|
||||
Verify your installation:
|
||||
|
||||
```
|
||||
your-project/
|
||||
├── _bmad/
|
||||
│ ├── bmm/ # Method module
|
||||
│ │ ├── agents/ # Agent files
|
||||
│ │ ├── workflows/ # Workflow files
|
||||
│ │ └── config.yaml # Module config
|
||||
│ └── core/ # Core utilities
|
||||
├── _bmad-output/ # Generated artifacts (created later)
|
||||
└── .claude/ # IDE configuration (if using Claude Code)
|
||||
```
|
||||
|
||||
:::tip[Troubleshooting]
|
||||
Having issues? See [Install BMad](../../how-to/installation/install-bmad.md) for common solutions.
|
||||
:::
|
||||
|
||||
## Step 1: Initialize Your Project
|
||||
|
||||
Load the **Analyst agent** in your IDE:
|
||||
- **Claude Code**: Type `/analyst` or load the agent file directly
|
||||
- **Cursor/Windsurf**: Open the agent file from `_bmad/bmm/agents/`
|
||||
|
||||
Wait for the agent's menu to appear, then run:
|
||||
|
||||
```
|
||||
Run workflow-init
|
||||
```
|
||||
|
||||
Or use the shorthand: `*workflow-init`
|
||||
|
||||
The workflow asks you to describe:
|
||||
- **Your project and goals** — What are you building? What problem does it solve?
|
||||
- **Existing codebase** — Is this new (greenfield) or existing code (brownfield)?
|
||||
- **Size and complexity** — Roughly how big is this? (adjustable later)
|
||||
|
||||
Based on your description, the workflow suggests a planning track. For this tutorial, choose **BMad Method**.
|
||||
|
||||
Once you confirm, the workflow creates `bmm-workflow-status.yaml` to track your progress.
|
||||
|
||||
:::caution[Fresh Chats]
|
||||
Always start a fresh chat for each workflow. This prevents context limitations from causing issues.
|
||||
:::
|
||||
|
||||
## Step 2: Create Your Plan
|
||||
|
||||
With your project initialized, work through the planning phases.
|
||||
|
||||
### Phase 1: Analysis (Optional)
|
||||
|
||||
If you want to brainstorm or research first:
|
||||
- **brainstorm-project** — Guided ideation with the Analyst
|
||||
- **research** — Market and technical research
|
||||
- **product-brief** — Recommended foundation document
|
||||
|
||||
### Phase 2: Planning (Required)
|
||||
|
||||
**Start a fresh chat** and load the **PM agent**.
|
||||
|
||||
```
|
||||
Run prd
|
||||
```
|
||||
|
||||
Or use shortcuts: `*prd`, select "create-prd" from the menu, or say "Let's create a PRD".
|
||||
|
||||
The PM agent guides you through:
|
||||
1. **Project overview** — Refine your project description
|
||||
2. **Goals and success metrics** — What does success look like?
|
||||
3. **User personas** — Who uses this product?
|
||||
4. **Functional requirements** — What must the system do?
|
||||
5. **Non-functional requirements** — Performance, security, scalability needs
|
||||
|
||||
When complete, you'll have `PRD.md` in your `_bmad-output/` folder.
|
||||
|
||||
:::note[UX Design (Optional)]
|
||||
If your project has a user interface, load the **UX-Designer agent** and run the UX design workflow after creating your PRD.
|
||||
:::
|
||||
|
||||
### Phase 3: Solutioning (Required for BMad Method)
|
||||
|
||||
**Start a fresh chat** and load the **Architect agent**.
|
||||
|
||||
```
|
||||
Run create-architecture
|
||||
```
|
||||
|
||||
The architect guides you through technical decisions: tech stack, database design, API patterns, and system structure.
|
||||
|
||||
:::tip[Check Your Status]
|
||||
Unsure what's next? Load any agent and run `workflow-status`. It tells you the next recommended or required workflow.
|
||||
:::
|
||||
|
||||
## Step 3: Build Your Project
|
||||
|
||||
Once planning is complete, move to implementation.
|
||||
|
||||
### Initialize Sprint Planning
|
||||
|
||||
Load the **SM agent** and run `sprint-planning`. This creates `sprint-status.yaml` to track all epics and stories.
|
||||
|
||||
### The Build Cycle
|
||||
|
||||
For each story, repeat this cycle with fresh chats:
|
||||
|
||||
| Step | Agent | Workflow | Purpose |
|
||||
|------|-------|----------|---------|
|
||||
| 1 | SM | `create-story` | Create story file from epic |
|
||||
| 2 | DEV | `dev-story` | Implement the story |
|
||||
| 3 | DEV | `code-review` | Quality validation *(recommended)* |
|
||||
|
||||
After completing all stories in an epic, load the **SM agent** and run `retrospective`.
|
||||
|
||||
## What You've Accomplished
|
||||
|
||||
You've learned the foundation of building with BMad:
|
||||
|
||||
- Installed BMad and configured it for your IDE
|
||||
- Initialized a project with your chosen planning track
|
||||
- Created planning documents (PRD, Architecture)
|
||||
- Understood the build cycle for implementation
|
||||
|
||||
Your project now has:
|
||||
|
||||
```
|
||||
your-project/
|
||||
├── _bmad/ # BMad configuration
|
||||
├── _bmad-output/
|
||||
│ ├── PRD.md # Your requirements document
|
||||
│ ├── architecture.md # Technical decisions
|
||||
│ └── bmm-workflow-status.yaml # Progress tracking
|
||||
└── ...
|
||||
```
|
||||
|
||||
## Quick Reference
|
||||
|
||||
| Command | Agent | Purpose |
|
||||
|---------|-------|---------|
|
||||
| `*workflow-init` | Analyst | Initialize a new project |
|
||||
| `*workflow-status` | Any | Check progress and next steps |
|
||||
| `*prd` | PM | Create Product Requirements Document |
|
||||
| `*create-architecture` | Architect | Create architecture document |
|
||||
| `*sprint-planning` | SM | Initialize sprint tracking |
|
||||
| `*create-story` | SM | Create a story file |
|
||||
| `*dev-story` | DEV | Implement a story |
|
||||
| `*code-review` | DEV | Review implemented code |
|
||||
|
||||
## Common Questions
|
||||
|
||||
**Do I need to create a PRD for every project?**
|
||||
Only for BMad Method and Enterprise tracks. Quick Flow projects use a simpler tech-spec instead.
|
||||
|
||||
**Can I skip Phase 1 (Analysis)?**
|
||||
Yes, Phase 1 is optional. If you already know what you're building, start with Phase 2 (Planning).
|
||||
|
||||
**What if I want to brainstorm first?**
|
||||
Load the Analyst agent and run `*brainstorm-project` before `workflow-init`.
|
||||
|
||||
**Why start fresh chats for each workflow?**
|
||||
Workflows are context-intensive. Reusing chats can cause the AI to hallucinate or lose track of details. Fresh chats ensure maximum context capacity.
|
||||
|
||||
## Getting Help
|
||||
|
||||
- **During workflows** — Agents guide you with questions and explanations
|
||||
- **Check status** — Run `workflow-status` with any agent
|
||||
- **Community** — [Discord](https://discord.gg/gk8jAdXWmj) (#bmad-method-help, #report-bugs-and-issues)
|
||||
- **Video tutorials** — [BMad Code YouTube](https://www.youtube.com/@BMadCode)
|
||||
|
||||
## Key Takeaways
|
||||
|
||||
:::tip[Remember These]
|
||||
- **Always use fresh chats** — Load agents in new chats for each workflow
|
||||
- **Let workflow-status guide you** — Ask any agent for status when unsure
|
||||
- **Track matters** — Quick Flow uses tech-spec; Method/Enterprise need PRD and architecture
|
||||
- **Tracking is automatic** — Status files update themselves
|
||||
- **Agents are flexible** — Use menu numbers, shortcuts (`*prd`), or natural language
|
||||
:::
|
||||
|
||||
Ready to start? Install BMad, load the Analyst, run `workflow-init`, and let the agents guide you.
|
||||
52
docs/_archive/vendor-workflows.md
Normal file
52
docs/_archive/vendor-workflows.md
Normal file
@@ -0,0 +1,52 @@
|
||||
---
|
||||
title: "Workflow Vendoring, Customization, and Inheritance"
|
||||
---
|
||||
|
||||
Use workflow vendoring and inheritance to share or reutilize workflows across modules.
|
||||
|
||||
## Workflow Vendoring
|
||||
|
||||
Workflow Vendoring allows an agent to have access to a workflow from another module, without having to install said module. At install time, the module workflow being vendored will be cloned and installed into the module that is receiving the vendored workflow the agent needs.
|
||||
|
||||
### How to Vendor
|
||||
|
||||
Lets assume you are building a module, and you do not want to recreate a workflow from the BMad Method, such as workflows/4-implementation/dev-story/workflow.md. Instead of copying all the context to your module, and having to maintain it over time as updates are made, you can instead use the exec-vendor menu item in your agent.
|
||||
|
||||
From your modules agent definition, you would implement the menu item as follows in the agent:
|
||||
|
||||
```yaml
|
||||
- trigger: develop-story
|
||||
exec-vendor: "{project-root}/_bmad/<source-module>/workflows/4-production/dev-story/workflow.md"
|
||||
exec: "{project-root}/_bmad/<my-module>/workflows/dev-story/workflow.md"
|
||||
description: "Execute Dev Story workflow, implementing tasks and tests, or performing updates to the story"
|
||||
```
|
||||
|
||||
At install time, it will clone the workflow and all of its required assets, and the agent that gets built will have an exec to a path installed in its own module. The content gets added to the folder you specify in exec. While it does not have to exactly match the source path, you will want to ensure you are specifying the workflow.md to be in a new location (in other words in this example, dev-story would not already be the path of another custom module workflow that already exists.)
|
||||
|
||||
## Workflow Inheritance
|
||||
|
||||
:::note[Coming Soon]
|
||||
Official support for workflow inheritance is coming post beta.
|
||||
:::
|
||||
|
||||
Workflow Inheritance is a different concept, that allows you to modify or extend existing workflow.
|
||||
|
||||
Party Mode from the core is an example of a workflow that is designed with inheritance in mind - customization for specific party needs. While party mode itself is generic - there might be specific agent collaborations you want to create. Without having to reinvent the whole party mode concept, or copy and paste all of its content - you could inherit from party mode to extend it to be specific.
|
||||
|
||||
Some possible examples could be:
|
||||
|
||||
- Retrospective
|
||||
- Sprint Planning
|
||||
- Collaborative Brainstorming Sessions
|
||||
|
||||
## Workflow Customization
|
||||
|
||||
:::note[Coming Soon]
|
||||
Official support for workflow customization is coming post beta.
|
||||
:::
|
||||
|
||||
Similar to Workflow Inheritance, Workflow Customization will soon be allowed for certain workflows that are meant to be user customized - similar in process to how agents are customized now.
|
||||
|
||||
This will take the shape of workflows with optional hooks, configurable inputs, and the ability to replace whole at install time.
|
||||
|
||||
For example, assume you are using the Create PRD workflow, which is comprised of 11 steps, and you want to always include specifics about your companies domain, technical landscape or something else. While project-context can be helpful with that, you can also through hooks and step overrides, have full replace steps, the key requirement being to ensure your step replace file is an exact file name match of an existing step, follows all conventions, and ends in a similar fashion to either hook back in to call the next existing step, or more custom steps that eventually hook back into the flow.
|
||||
@@ -1,149 +0,0 @@
|
||||
# Custom Content Installation
|
||||
|
||||
This guide explains how to create and install custom BMAD content including agents, workflows, and modules. Custom content extends BMAD's functionality with specialized tools and workflows that can be shared across projects or teams.
|
||||
|
||||
For detailed information about the different types of custom content available, see [Custom Content](./custom-content.md).
|
||||
|
||||
You can find example custom modules in the `samples/sample-custom-modules/` folder of the repository. Download either of the sample folders to try them out.
|
||||
|
||||
## Content Types Overview
|
||||
|
||||
BMAD Core supports several categories of custom content:
|
||||
|
||||
- Custom Stand Alone Modules
|
||||
- Custom Add On Modules
|
||||
- Custom Global Modules
|
||||
- Custom Agents
|
||||
- Custom Workflows
|
||||
|
||||
## Making Custom Content Installable
|
||||
|
||||
### Custom Modules
|
||||
|
||||
To create an installable custom module:
|
||||
|
||||
1. **Folder Structure**
|
||||
- Create a folder with a short, abbreviated name (e.g., `cis` for Creative Intelligence Suite)
|
||||
- The folder name serves as the module code
|
||||
|
||||
2. **Required Files**
|
||||
- Include a `module.yaml` file in the root folder
|
||||
- This file drives the installation process when used by the BMAD installer
|
||||
- Reference existing modules or the BMad Builder for configuration examples
|
||||
|
||||
3. **Folder Organization**
|
||||
Follow these conventions for optimal compatibility:
|
||||
|
||||
```
|
||||
module-code/
|
||||
module.yaml
|
||||
agents/
|
||||
workflows/
|
||||
tools/
|
||||
templates/
|
||||
...
|
||||
```
|
||||
|
||||
- `agents/` - Agent definitions
|
||||
- `workflows/` - Workflow definitions
|
||||
- Additional custom folders are supported but following conventions is recommended for agent and workflow discovery
|
||||
|
||||
**Note:** Full documentation for global modules and add-on modules will be available as support is finalized.
|
||||
|
||||
### Standalone Content (Agents, Workflows, Tasks, Tools, Templates, Prompts)
|
||||
|
||||
For standalone content that isn't part of a cohesive module collection, follow this structure:
|
||||
|
||||
1. **Module Configuration**
|
||||
- Create a folder with a `module.yaml` file (similar to custom modules)
|
||||
- Add the property `unitary: true` to the module.yaml
|
||||
- The `unitary: true` property indicates this is a collection of potentially unrelated items that don't depend on each other
|
||||
|
||||
2. **Folder Structure**
|
||||
Organize content in specific named folders:
|
||||
|
||||
```
|
||||
module-name/
|
||||
module.yaml # Contains unitary: true
|
||||
agents/
|
||||
workflows/
|
||||
templates/
|
||||
tools/
|
||||
tasks/
|
||||
prompts/
|
||||
```
|
||||
|
||||
3. **Individual Item Organization**
|
||||
Each item should have its own subfolder:
|
||||
```text
|
||||
my-custom-stuff/
|
||||
module.yaml
|
||||
agents/
|
||||
larry/larry.agent.md
|
||||
curly/curly.agent.md
|
||||
moe/moe.agent.md
|
||||
moe/moe-sidecar/memories.csv
|
||||
```
|
||||
|
||||
**Future Feature:** Unitary modules will support selective installation, allowing users to pick and choose which specific items to install.
|
||||
|
||||
**Note:** Documentation explaining the distinctions between these content types and their specific use cases will be available soon.
|
||||
|
||||
## Installation Process
|
||||
|
||||
### Prerequisites
|
||||
|
||||
Ensure your content follows the proper conventions and includes a `module.yaml` file (only one per top-level folder).
|
||||
|
||||
### New Project Installation
|
||||
|
||||
When setting up a new BMAD project:
|
||||
|
||||
1. The installer will prompt: `Would you like to install a local custom module (this includes custom agents and workflows also)? (y/N)`
|
||||
2. Select 'y' to specify the path to your module folder containing `module.yaml`
|
||||
|
||||
### Existing Project Modification
|
||||
|
||||
To add custom content to an existing BMAD project:
|
||||
|
||||
1. Run the installer against your project location
|
||||
2. Select `Modify BMAD Installation`
|
||||
3. Choose the option to add, modify, or update custom modules
|
||||
|
||||
### Upcoming Features
|
||||
|
||||
- **Unitary Module Selection:** For modules with `type: unitary` (instead of `type: module`), you'll be able to select specific items to install
|
||||
- **Add-on Module Dependencies:** The installer will verify and install dependencies for add-on modules automatically
|
||||
|
||||
## Quick Updates
|
||||
|
||||
When updates to BMAD Core or core modules (BMM, CIS, etc.) become available, the quick update process will:
|
||||
|
||||
1. Apply available updates to core modules
|
||||
2. Recompile all agents with customizations from the `_config/agents` folder
|
||||
3. Retain your custom content from a cached location
|
||||
4. Preserve your existing configurations and customizations
|
||||
|
||||
This means you don't need to keep the source module files locally. When updates are available, simply point to the updated module location during the update process.
|
||||
|
||||
## Important Considerations
|
||||
|
||||
### Module Naming Conflicts
|
||||
|
||||
When installing unofficial modules, ensure unique identification to avoid conflicts:
|
||||
|
||||
1. **Module Codes:** Each module must have a unique code (e.g., don't use `bmm` for custom modules)
|
||||
2. **Module Names:** Avoid using names that conflict with existing modules
|
||||
3. **Multiple Custom Modules:** If creating multiple custom modules, use distinct codes for each
|
||||
|
||||
**Examples of conflicts to avoid:**
|
||||
|
||||
- Don't create a custom module with code `bmm` (already used by BMad Method)
|
||||
- Don't name multiple custom modules with the same code like `mca`
|
||||
|
||||
### Best Practices
|
||||
|
||||
- Use descriptive, unique codes for your modules
|
||||
- Document any dependencies your custom modules have
|
||||
- Test custom modules in isolation before sharing
|
||||
- Consider version numbering for your custom content to track updates
|
||||
@@ -1,449 +0,0 @@
|
||||
# Document Sharding Guide
|
||||
|
||||
Comprehensive guide to BMad Method's document sharding system for managing large planning and architecture documents.
|
||||
|
||||
## Table of Contents
|
||||
|
||||
- [What is Document Sharding?](#what-is-document-sharding)
|
||||
- [When to Use Sharding](#when-to-use-sharding)
|
||||
- [How Sharding Works](#how-sharding-works)
|
||||
- [Using the Shard-Doc Tool](#using-the-shard-doc-tool)
|
||||
- [Workflow Support](#workflow-support)
|
||||
- [Best Practices](#best-practices)
|
||||
- [Examples](#examples)
|
||||
|
||||
## What is Document Sharding?
|
||||
|
||||
Document sharding splits large markdown files into smaller, organized files based on level 2 headings (`## Heading`). This enables:
|
||||
|
||||
- **Selective Loading** - Workflows load only the sections they need
|
||||
- **Reduced Token Usage** - Massive efficiency gains for large projects
|
||||
- **Better Organization** - Logical section-based file structure
|
||||
- **Maintained Context** - Index file preserves document structure
|
||||
|
||||
### Architecture
|
||||
|
||||
```
|
||||
Before Sharding:
|
||||
docs/
|
||||
└── PRD.md (large 50k token file)
|
||||
|
||||
After Sharding:
|
||||
docs/
|
||||
└── prd/
|
||||
├── index.md # Table of contents with descriptions
|
||||
├── overview.md # Section 1
|
||||
├── user-requirements.md # Section 2
|
||||
├── technical-requirements.md # Section 3
|
||||
└── ... # Additional sections
|
||||
```
|
||||
|
||||
## When to Use Sharding
|
||||
|
||||
### Ideal Candidates
|
||||
|
||||
**Large Multi-Epic Projects:**
|
||||
|
||||
- Very large complex PRDs
|
||||
- Architecture documents with multiple system layers
|
||||
- Epic files with 4+ epics (especially for Phase 4)
|
||||
- UX design specs covering multiple subsystems
|
||||
|
||||
**Token Thresholds:**
|
||||
|
||||
- **Consider sharding**: Documents > 20k tokens
|
||||
- **Strongly recommended**: Documents > 40k tokens
|
||||
- **Critical for efficiency**: Documents > 60k tokens
|
||||
|
||||
### When NOT to Shard
|
||||
|
||||
**Small Projects:**
|
||||
|
||||
- Single epic projects
|
||||
- Level 0-1 projects (tech-spec only)
|
||||
- Documents under 10k tokens
|
||||
- Quick prototypes
|
||||
|
||||
**Frequently Updated Docs:**
|
||||
|
||||
- Active work-in-progress documents
|
||||
- Documents updated daily
|
||||
- Documents where whole-file context is essential
|
||||
|
||||
## How Sharding Works
|
||||
|
||||
### Sharding Process
|
||||
|
||||
1. **Tool Execution**: Run `npx @kayvan/markdown-tree-parser source.md destination/` - this is abstracted with the core shard-doc task which is installed as a slash command or manual task rule depending on your tools.
|
||||
2. **Section Extraction**: Tool splits by level 2 headings
|
||||
3. **File Creation**: Each section becomes a separate file
|
||||
4. **Index Generation**: `index.md` created with structure and descriptions
|
||||
|
||||
### Workflow Discovery
|
||||
|
||||
BMad workflows use a **dual discovery system**:
|
||||
|
||||
1. **Try whole document first** - Look for `document-name.md`
|
||||
2. **Check for sharded version** - Look for `document-name/index.md`
|
||||
3. **Priority rule** - Whole document takes precedence if both exist
|
||||
|
||||
### Loading Strategies
|
||||
|
||||
**Full Load (Phase 1-3 workflows):**
|
||||
|
||||
```
|
||||
If sharded:
|
||||
- Read index.md
|
||||
- Read ALL section files
|
||||
- Treat as single combined document
|
||||
```
|
||||
|
||||
**Selective Load (Phase 4 workflows):**
|
||||
|
||||
```
|
||||
If sharded epics and working on Epic 3:
|
||||
- Read epics/index.md
|
||||
- Load ONLY epics/epic-3.md
|
||||
- Skip all other epic files
|
||||
- 90%+ token savings!
|
||||
```
|
||||
|
||||
## Using the Shard-Doc Tool
|
||||
|
||||
### CLI Command
|
||||
|
||||
```bash
|
||||
# Activate bmad-master or analyst agent, then:
|
||||
/shard-doc
|
||||
```
|
||||
|
||||
### Interactive Process
|
||||
|
||||
```
|
||||
Agent: Which document would you like to shard?
|
||||
User: docs/PRD.md
|
||||
|
||||
Agent: Default destination: docs/prd/
|
||||
Accept default? [y/n]
|
||||
User: y
|
||||
|
||||
Agent: Sharding PRD.md...
|
||||
✓ Created 12 section files
|
||||
✓ Generated index.md
|
||||
✓ Complete!
|
||||
```
|
||||
|
||||
### What Gets Created
|
||||
|
||||
**index.md structure:**
|
||||
|
||||
```markdown
|
||||
# PRD - Index
|
||||
|
||||
## Sections
|
||||
|
||||
1. [Overview](./overview.md) - Project vision and objectives
|
||||
2. [User Requirements](./user-requirements.md) - Feature specifications
|
||||
3. [Epic 1: Authentication](./epic-1-authentication.md) - User auth system
|
||||
4. [Epic 2: Dashboard](./epic-2-dashboard.md) - Main dashboard UI
|
||||
...
|
||||
```
|
||||
|
||||
**Individual section files:**
|
||||
|
||||
- Named from heading text (kebab-case)
|
||||
- Contains complete section content
|
||||
- Preserves all markdown formatting
|
||||
- Can be read independently
|
||||
|
||||
## Workflow Support
|
||||
|
||||
### Universal Support
|
||||
|
||||
**All BMM workflows support both formats:**
|
||||
|
||||
- ✅ Whole documents
|
||||
- ✅ Sharded documents
|
||||
- ✅ Automatic detection
|
||||
- ✅ Transparent to user
|
||||
|
||||
### Workflow-Specific Patterns
|
||||
|
||||
#### Phase 1-3 (Full Load)
|
||||
|
||||
Workflows load entire sharded documents:
|
||||
|
||||
- `product-brief` - Research, brainstorming docs
|
||||
- `prd` - Product brief, research
|
||||
- `gdd` - Game brief, research
|
||||
- `create-ux-design` - PRD, brief, architecture (if available)
|
||||
- `tech-spec` - Brief, research
|
||||
- `architecture` - PRD, UX design (if available)
|
||||
- `create-epics-and-stories` - PRD, architecture
|
||||
- `implementation-readiness` - All planning docs
|
||||
|
||||
#### Phase 4 (Selective Load)
|
||||
|
||||
Workflows load only needed sections:
|
||||
|
||||
**sprint-planning** (Full Load):
|
||||
|
||||
- Needs ALL epics to build complete status
|
||||
|
||||
**create-story, code-review** (Selective):
|
||||
|
||||
```
|
||||
Working on Epic 3, Story 2:
|
||||
✓ Load epics/epic-3.md only
|
||||
✗ Skip epics/epic-1.md, epic-2.md, epic-4.md, etc.
|
||||
|
||||
Result: 90%+ token reduction for 10-epic projects!
|
||||
```
|
||||
|
||||
### Input File Patterns
|
||||
|
||||
Workflows use standardized patterns:
|
||||
|
||||
```yaml
|
||||
input_file_patterns:
|
||||
prd:
|
||||
whole: '{output_folder}/*prd*.md'
|
||||
sharded: '{output_folder}/*prd*/index.md'
|
||||
|
||||
epics:
|
||||
whole: '{output_folder}/*epic*.md'
|
||||
sharded_index: '{output_folder}/*epic*/index.md'
|
||||
sharded_single: '{output_folder}/*epic*/epic-{{epic_num}}.md'
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
### Sharding Strategy
|
||||
|
||||
**Do:**
|
||||
|
||||
- ✅ Shard after planning phase complete
|
||||
- ✅ Keep level 2 headings well-organized
|
||||
- ✅ Use descriptive section names
|
||||
- ✅ Shard before Phase 4 implementation
|
||||
- ✅ Keep original file as backup initially
|
||||
|
||||
**Don't:**
|
||||
|
||||
- ❌ Shard work-in-progress documents
|
||||
- ❌ Shard small documents (<20k tokens)
|
||||
- ❌ Mix sharded and whole versions
|
||||
- ❌ Manually edit index.md structure
|
||||
|
||||
### Naming Conventions
|
||||
|
||||
**Good Section Names:**
|
||||
|
||||
```markdown
|
||||
## Epic 1: User Authentication
|
||||
|
||||
## Technical Requirements
|
||||
|
||||
## System Architecture
|
||||
|
||||
## UX Design Principles
|
||||
```
|
||||
|
||||
**Poor Section Names:**
|
||||
|
||||
```markdown
|
||||
## Section 1
|
||||
|
||||
## Part A
|
||||
|
||||
## Details
|
||||
|
||||
## More Info
|
||||
```
|
||||
|
||||
### File Management
|
||||
|
||||
**When to Re-shard:**
|
||||
|
||||
- Significant structural changes to document
|
||||
- Adding/removing major sections
|
||||
- After major refactoring
|
||||
|
||||
**Updating Sharded Docs:**
|
||||
|
||||
1. Edit individual section files directly
|
||||
2. OR edit original, delete sharded folder, re-shard
|
||||
3. Don't manually edit index.md
|
||||
|
||||
## Examples
|
||||
|
||||
### Example 1: Large PRD
|
||||
|
||||
**Scenario:** 15-epic project, PRD is 45k tokens
|
||||
|
||||
**Before Sharding:**
|
||||
|
||||
```
|
||||
Every workflow loads entire 45k token PRD
|
||||
Architecture workflow: 45k tokens
|
||||
UX design workflow: 45k tokens
|
||||
```
|
||||
|
||||
**After Sharding:**
|
||||
|
||||
```bash
|
||||
/shard-doc
|
||||
Source: docs/PRD.md
|
||||
Destination: docs/prd/
|
||||
|
||||
Created:
|
||||
prd/index.md
|
||||
prd/overview.md (3k tokens)
|
||||
prd/functional-requirements.md (8k tokens)
|
||||
prd/non-functional-requirements.md (6k tokens)
|
||||
prd/user-personas.md (4k tokens)
|
||||
...additional FR/NFR sections
|
||||
```
|
||||
|
||||
**Result:**
|
||||
|
||||
```
|
||||
Architecture workflow: Can load specific sections needed
|
||||
UX design workflow: Can load specific sections needed
|
||||
Significant token reduction for large requirement docs!
|
||||
```
|
||||
|
||||
### Example 2: Sharding Epics File
|
||||
|
||||
**Scenario:** 8 epics with detailed stories, 35k tokens total
|
||||
|
||||
```bash
|
||||
/shard-doc
|
||||
Source: docs/bmm-epics.md
|
||||
Destination: docs/epics/
|
||||
|
||||
Created:
|
||||
epics/index.md
|
||||
epics/epic-1.md
|
||||
epics/epic-2.md
|
||||
...
|
||||
epics/epic-8.md
|
||||
```
|
||||
|
||||
**Efficiency Gain:**
|
||||
|
||||
```
|
||||
Working on Epic 5 stories:
|
||||
Old: Load all 8 epics (35k tokens)
|
||||
New: Load epic-5.md only (4k tokens)
|
||||
Savings: 88% reduction
|
||||
```
|
||||
|
||||
### Example 3: Architecture Document
|
||||
|
||||
**Scenario:** Multi-layer system architecture, 28k tokens
|
||||
|
||||
```bash
|
||||
/shard-doc
|
||||
Source: docs/architecture.md
|
||||
Destination: docs/architecture/
|
||||
|
||||
Created:
|
||||
architecture/index.md
|
||||
architecture/system-overview.md
|
||||
architecture/frontend-architecture.md
|
||||
architecture/backend-services.md
|
||||
architecture/data-layer.md
|
||||
architecture/infrastructure.md
|
||||
architecture/security-architecture.md
|
||||
```
|
||||
|
||||
**Benefit:** Code-review workflow can reference specific architectural layers without loading entire architecture doc.
|
||||
|
||||
## Custom Workflow Integration
|
||||
|
||||
### For Workflow Builders
|
||||
|
||||
When creating custom workflows that load large documents:
|
||||
|
||||
**1. Add input_file_patterns to workflow.yaml:**
|
||||
|
||||
```yaml
|
||||
input_file_patterns:
|
||||
your_document:
|
||||
whole: '{output_folder}/*your-doc*.md'
|
||||
sharded: '{output_folder}/*your-doc*/index.md'
|
||||
```
|
||||
|
||||
**2. Add discovery instructions to instructions.md:**
|
||||
|
||||
```markdown
|
||||
## Document Discovery
|
||||
|
||||
1. Search for whole document: _your-doc_.md
|
||||
2. Check for sharded version: _your-doc_/index.md
|
||||
3. If sharded: Read index + ALL sections (or specific sections if selective load)
|
||||
4. Priority: Whole document first
|
||||
```
|
||||
|
||||
**3. Choose loading strategy:**
|
||||
|
||||
- **Full Load**: Read all sections when sharded
|
||||
- **Selective Load**: Read only relevant sections (requires section identification logic)
|
||||
|
||||
### Pattern Templates
|
||||
|
||||
**Full Load Pattern:**
|
||||
|
||||
```xml
|
||||
<action>Search for document: {output_folder}/*doc-name*.md</action>
|
||||
<action>If not found, check for sharded: {output_folder}/*doc-name*/index.md</action>
|
||||
<action if="sharded found">Read index.md to understand structure</action>
|
||||
<action if="sharded found">Read ALL section files listed in index</action>
|
||||
<action if="sharded found">Combine content as single document</action>
|
||||
```
|
||||
|
||||
**Selective Load Pattern (with section ID):**
|
||||
|
||||
```xml
|
||||
<action>Determine section needed (e.g., epic_num = 3)</action>
|
||||
<action>Check for sharded version: {output_folder}/*doc-name*/index.md</action>
|
||||
<action if="sharded found">Read ONLY the specific section file needed</action>
|
||||
<action if="sharded found">Skip all other section files</action>
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
**Both whole and sharded exist:**
|
||||
|
||||
- Workflows will use whole document (priority rule)
|
||||
- Delete or archive the one you don't want
|
||||
|
||||
**Index.md out of sync:**
|
||||
|
||||
- Delete sharded folder
|
||||
- Re-run shard-doc on original
|
||||
|
||||
**Workflow can't find document:**
|
||||
|
||||
- Check file naming matches patterns (`*prd*.md`, `*epic*.md`, etc.)
|
||||
- Verify index.md exists in sharded folder
|
||||
- Check output_folder path in config
|
||||
|
||||
**Sections too granular:**
|
||||
|
||||
- Combine sections in original document
|
||||
- Use fewer level 2 headings
|
||||
- Re-shard
|
||||
|
||||
## Related Documentation
|
||||
|
||||
- [shard-doc Tool](../src/core/tools/shard-doc.xml) - Tool implementation
|
||||
- [BMM Workflows Guide](./modules/bmm/index.md#-workflow-guides) - Workflow overview
|
||||
- [Workflow Creation Guide](./modules/bmb/workflows/index) - Custom workflow patterns
|
||||
|
||||
---
|
||||
|
||||
**Document sharding is optional but powerful** - use it when efficiency matters for large projects!
|
||||
@@ -1,47 +1,44 @@
|
||||
# Downloads
|
||||
---
|
||||
title: Downloads
|
||||
---
|
||||
|
||||
Download BMAD Method resources for offline use, AI training, or integration.
|
||||
Download BMad Method resources for offline use, AI training, or integration.
|
||||
|
||||
## Source Bundles
|
||||
|
||||
| File | Description |
|
||||
|------|-------------|
|
||||
| **[bmad-sources.zip](/downloads/bmad-sources.zip)** | Complete BMad source files |
|
||||
| **[bmad-prompts.zip](/downloads/bmad-prompts.zip)** | Agent and workflow prompts only |
|
||||
|
||||
## LLM-Optimized Files
|
||||
|
||||
These files are designed for AI consumption - perfect for loading into Claude, ChatGPT, or any LLM context window.
|
||||
|
||||
| File | Description | Use Case |
|
||||
| ----------------------------------- | ----------------------------------- | -------------------------- |
|
||||
| **[llms.txt](/llms.txt)** | Documentation index with summaries | Quick overview, navigation |
|
||||
| **[llms-full.txt](/llms-full.txt)** | Complete documentation concatenated | Full context loading |
|
||||
| File | Description | Use Case |
|
||||
|------|-------------|----------|
|
||||
| **[llms.txt](/llms.txt)** | Documentation index with summaries | Quick overview, navigation |
|
||||
| **[llms-full.txt](/llms-full.txt)** | Complete documentation concatenated | Full context loading |
|
||||
|
||||
### Using with LLMs
|
||||
|
||||
**Claude Projects:**
|
||||
|
||||
```
|
||||
Upload llms-full.txt as project knowledge
|
||||
```
|
||||
|
||||
**ChatGPT:**
|
||||
|
||||
```
|
||||
Paste llms.txt for navigation, or sections from llms-full.txt as needed
|
||||
```
|
||||
|
||||
**API Usage:**
|
||||
|
||||
```python
|
||||
import requests
|
||||
docs = requests.get("https://bmad-code-org.github.io/BMAD-METHOD/llms-full.txt").text
|
||||
# Include in your system prompt or context
|
||||
```
|
||||
|
||||
## Source Bundles
|
||||
|
||||
Download the complete source code for offline development or contribution.
|
||||
|
||||
| Bundle | Contents |
|
||||
| --------------------------------------------------- | ------------------------------------------- |
|
||||
| **[bmad-sources.zip](/downloads/bmad-sources.zip)** | Complete `/src/` directory with all modules |
|
||||
| **[bmad-prompts.zip](/downloads/bmad-prompts.zip)** | Agent prompts and workflows from `/_bmad/` |
|
||||
|
||||
## Installation Options
|
||||
|
||||
### NPM (Recommended)
|
||||
@@ -50,12 +47,6 @@ Download the complete source code for offline development or contribution.
|
||||
npx bmad-method@alpha install
|
||||
```
|
||||
|
||||
### Manual Installation
|
||||
|
||||
1. Download `bmad-prompts.zip`
|
||||
2. Extract to your project root
|
||||
3. Configure your IDE per the [IDE Guides](/docs/ide-info/)
|
||||
|
||||
## Version Information
|
||||
|
||||
- **Current Version:** See [CHANGELOG](https://github.com/bmad-code-org/BMAD-METHOD/blob/main/CHANGELOG.md)
|
||||
@@ -63,7 +54,7 @@ npx bmad-method@alpha install
|
||||
|
||||
## API Access
|
||||
|
||||
For programmatic access to BMAD documentation:
|
||||
For programmatic access to BMad documentation:
|
||||
|
||||
```bash
|
||||
# Get documentation index
|
||||
@@ -75,7 +66,7 @@ curl https://bmad-code-org.github.io/BMAD-METHOD/llms-full.txt
|
||||
|
||||
## Contributing
|
||||
|
||||
Want to improve BMAD Method? Check out:
|
||||
Want to improve BMad Method? Check out:
|
||||
|
||||
- [Contributing Guide](https://github.com/bmad-code-org/BMAD-METHOD/blob/main/CONTRIBUTING.md)
|
||||
- [GitHub Repository](https://github.com/bmad-code-org/BMAD-METHOD)
|
||||
@@ -1,21 +1,25 @@
|
||||
# Quick Flow Solo Dev Agent (Barry)
|
||||
|
||||
**Agent ID:** `_bmad/bmm/agents/quick-flow-solo-dev.md`
|
||||
**Icon:** 🚀
|
||||
**Module:** BMM
|
||||
|
||||
---
|
||||
title: "Quick Flow Solo Dev Agent (Barry)"
|
||||
---
|
||||
|
||||
Barry is the elite solo developer who takes projects from concept to deployment with ruthless efficiency — no handoffs, no delays, just pure focused development.
|
||||
|
||||
:::note[Agent Info]
|
||||
- **Agent ID:** `_bmad/bmm/agents/quick-flow-solo-dev.md`
|
||||
- **Icon:** 🚀
|
||||
- **Module:** BMM
|
||||
:::
|
||||
|
||||
## Overview
|
||||
|
||||
Barry is the elite solo developer who lives and breathes the BMAD Quick Flow workflow. He takes projects from concept to deployment with ruthless efficiency - no handoffs, no delays, just pure focused development. Barry architects specs, writes the code, and ships features faster than entire teams. When you need it done right and done now, Barry's your dev.
|
||||
Barry is the elite solo developer who lives and breathes the BMad Quick Flow workflow. He takes projects from concept to deployment with ruthless efficiency - no handoffs, no delays, just pure focused development. Barry architects specs, writes the code, and ships features faster than entire teams. When you need it done right and done now, Barry's your dev.
|
||||
|
||||
### Agent Persona
|
||||
|
||||
**Name:** Barry
|
||||
**Title:** Quick Flow Solo Dev
|
||||
|
||||
**Identity:** Barry is an elite developer who thrives on autonomous execution. He lives and breathes the BMAD Quick Flow workflow, taking projects from concept to deployment with ruthless efficiency. No handoffs, no delays - just pure, focused development. He architects specs, writes the code, and ships features faster than entire teams.
|
||||
**Identity:** Barry is an elite developer who thrives on autonomous execution. He lives and breathes the BMad Quick Flow workflow, taking projects from concept to deployment with ruthless efficiency. No handoffs, no delays - just pure, focused development. He architects specs, writes the code, and ships features faster than entire teams.
|
||||
|
||||
**Communication Style:** Direct, confident, and implementation-focused. Uses tech slang and gets straight to the point. No fluff, just results. Every response moves the project forward.
|
||||
|
||||
@@ -28,15 +32,13 @@ Barry is the elite solo developer who lives and breathes the BMAD Quick Flow wor
|
||||
- Documentation happens alongside development, not after
|
||||
- Ship early, ship often
|
||||
|
||||
---
|
||||
|
||||
## Menu Commands
|
||||
|
||||
Barry owns the entire BMAD Quick Flow path, providing a streamlined 3-step development process that eliminates handoffs and maximizes velocity.
|
||||
Barry owns the entire BMad Quick Flow path, providing a streamlined 3-step development process that eliminates handoffs and maximizes velocity.
|
||||
|
||||
### 1. **create-tech-spec**
|
||||
### 1. **quick-spec**
|
||||
|
||||
- **Workflow:** `_bmad/bmm/workflows/bmad-quick-flow/create-tech-spec/workflow.yaml`
|
||||
- **Workflow:** `_bmad/bmm/workflows/bmad-quick-flow/quick-spec/workflow.md`
|
||||
- **Description:** Architect a technical spec with implementation-ready stories
|
||||
- **Use when:** You need to transform requirements into a buildable spec
|
||||
|
||||
@@ -58,8 +60,6 @@ Barry owns the entire BMAD Quick Flow path, providing a streamlined 3-step devel
|
||||
- **Description:** Bring in other experts when I need specialized backup
|
||||
- **Use when:** You need collaborative problem-solving or specialized expertise
|
||||
|
||||
---
|
||||
|
||||
## When to Use Barry
|
||||
|
||||
### Ideal Scenarios
|
||||
@@ -78,15 +78,13 @@ Barry owns the entire BMAD Quick Flow path, providing a streamlined 3-step devel
|
||||
- **Proof of Concepts** - Rapid prototyping with production-quality code
|
||||
- **Performance Optimizations** - System improvements and scalability work
|
||||
|
||||
---
|
||||
|
||||
## The BMAD Quick Flow Process
|
||||
## The BMad Quick Flow Process
|
||||
|
||||
Barry orchestrates a simple, efficient 3-step process:
|
||||
|
||||
```mermaid
|
||||
flowchart LR
|
||||
A[Requirements] --> B[create-tech-spec]
|
||||
A[Requirements] --> B[quick-spec]
|
||||
B --> C[Tech Spec]
|
||||
C --> D[quick-dev]
|
||||
D --> E[Implementation]
|
||||
@@ -104,7 +102,7 @@ flowchart LR
|
||||
style H fill:#e0f2f1
|
||||
```
|
||||
|
||||
### Step 1: Technical Specification (`create-tech-spec`)
|
||||
### Step 1: Technical Specification (`quick-spec`)
|
||||
|
||||
**Goal:** Transform user requirements into implementation-ready technical specifications
|
||||
|
||||
@@ -177,8 +175,6 @@ flowchart LR
|
||||
- Security considerations
|
||||
- Maintainability and documentation
|
||||
|
||||
---
|
||||
|
||||
## Collaboration with Other Agents
|
||||
|
||||
### Natural Partnerships
|
||||
@@ -198,8 +194,6 @@ In party mode, Barry often acts as:
|
||||
- **Performance Optimizer** - Ensuring scalable solutions
|
||||
- **Code Review Authority** - Validating technical approaches
|
||||
|
||||
---
|
||||
|
||||
## Tips for Working with Barry
|
||||
|
||||
### For Best Results
|
||||
@@ -225,8 +219,6 @@ In party mode, Barry often acts as:
|
||||
4. **Over-planning** - I excel at rapid, pragmatic development
|
||||
5. **Not Using Party Mode** - Missing collaborative insights for complex problems
|
||||
|
||||
---
|
||||
|
||||
## Example Workflow
|
||||
|
||||
```bash
|
||||
@@ -234,7 +226,7 @@ In party mode, Barry often acts as:
|
||||
/bmad:bmm:agents:quick-flow-solo-dev
|
||||
|
||||
# Create a tech spec
|
||||
> create-tech-spec
|
||||
> quick-spec
|
||||
|
||||
# Quick implementation
|
||||
> quick-dev tech-spec-auth.md
|
||||
@@ -303,35 +295,34 @@ Implement OAuth 2.0 authentication with JWT tokens and role-based access control
|
||||
- [ ] Given admin role, when accessing admin endpoint, then allow access
|
||||
```
|
||||
|
||||
---
|
||||
## Common Questions
|
||||
|
||||
## Related Documentation
|
||||
- [When should I use Barry vs other agents?](#when-should-i-use-barry-vs-other-agents)
|
||||
- [Is the code review step mandatory?](#is-the-code-review-step-mandatory)
|
||||
- [Can I skip the tech spec step?](#can-i-skip-the-tech-spec-step)
|
||||
- [How does Barry differ from the Dev agent?](#how-does-barry-differ-from-the-dev-agent)
|
||||
- [Can Barry handle enterprise-scale projects?](#can-barry-handle-enterprise-scale-projects)
|
||||
|
||||
- **[Quick Start Guide](./quick-start.md)** - Getting started with BMM
|
||||
- **[Agents Guide](./agents-guide.md)** - Complete agent reference
|
||||
- **[Scale Adaptive System](./scale-adaptive-system.md)** - Understanding development tracks
|
||||
- **[Workflow Implementation](./workflows-implementation.md)** - Implementation workflows
|
||||
- **[Party Mode](./party-mode.md)** - Multi-agent collaboration
|
||||
### When should I use Barry vs other agents?
|
||||
|
||||
---
|
||||
Use Barry for Quick Flow development (small to medium features), rapid prototyping, or when you need elite solo development. For large, complex projects requiring full team collaboration, consider the full BMad Method with specialized agents.
|
||||
|
||||
## Frequently Asked Questions
|
||||
### Is the code review step mandatory?
|
||||
|
||||
**Q: When should I use Barry vs other agents?**
|
||||
A: Use Barry for Quick Flow development (small to medium features), rapid prototyping, or when you need elite solo development. For large, complex projects requiring full team collaboration, consider the full BMad Method with specialized agents.
|
||||
No, it's optional but highly recommended for critical features, team projects, or when learning best practices.
|
||||
|
||||
**Q: Is the code review step mandatory?**
|
||||
A: No, it's optional but highly recommended for critical features, team projects, or when learning best practices.
|
||||
### Can I skip the tech spec step?
|
||||
|
||||
**Q: Can I skip the tech spec step?**
|
||||
A: Yes, the quick-dev workflow accepts direct instructions. However, tech specs are recommended for complex features or team collaboration.
|
||||
Yes, the quick-dev workflow accepts direct instructions. However, tech specs are recommended for complex features or team collaboration.
|
||||
|
||||
**Q: How does Barry differ from the Dev agent?**
|
||||
A: Barry handles the complete Quick Flow process (spec → dev → review) with elite architectural expertise, while the Dev agent specializes in pure implementation tasks. Barry is your autonomous end-to-end solution.
|
||||
### How does Barry differ from the Dev agent?
|
||||
|
||||
**Q: Can Barry handle enterprise-scale projects?**
|
||||
A: For enterprise-scale projects requiring full team collaboration, consider using the Enterprise Method track. Barry is optimized for rapid delivery in the Quick Flow track where solo execution wins.
|
||||
Barry handles the complete Quick Flow process (spec → dev → review) with elite architectural expertise, while the Dev agent specializes in pure implementation tasks. Barry is your autonomous end-to-end solution.
|
||||
|
||||
---
|
||||
### Can Barry handle enterprise-scale projects?
|
||||
|
||||
**Ready to ship some code?** → Start with `/bmad:bmm:agents:quick-flow-solo-dev`
|
||||
For enterprise-scale projects requiring full team collaboration, consider using the Enterprise Method track. Barry is optimized for rapid delivery in the Quick Flow track where solo execution wins.
|
||||
|
||||
:::tip[Ready to Ship?]
|
||||
Start with `/bmad:bmm:agents:quick-flow-solo-dev`
|
||||
:::
|
||||
20
docs/explanation/agents/index.md
Normal file
20
docs/explanation/agents/index.md
Normal file
@@ -0,0 +1,20 @@
|
||||
---
|
||||
title: "Understanding Agents"
|
||||
description: Understanding BMad agents and their roles
|
||||
---
|
||||
|
||||
Comprehensive guides to BMad's AI agents — their roles, capabilities, and how to work with them effectively.
|
||||
|
||||
## Agent Guides
|
||||
|
||||
| Agent | Description |
|
||||
|-------|-------------|
|
||||
| **[Agent Roles](/docs/explanation/core-concepts/agent-roles.md)** | Overview of all BMM agent roles and responsibilities |
|
||||
| **[Quick Flow Solo Dev (Barry)](/docs/explanation/agents/barry-quick-flow.md)** | The dedicated agent for rapid development |
|
||||
| **[Game Development Agents](/docs/explanation/game-dev/agents.md)** | Complete guide to BMGD's specialized game dev agents |
|
||||
|
||||
## Getting Started
|
||||
|
||||
1. Read **[What Are Agents?](/docs/explanation/core-concepts/what-are-agents.md)** for the core concept explanation
|
||||
2. Review **[Agent Roles](/docs/explanation/core-concepts/agent-roles.md)** to understand available agents
|
||||
3. Choose an agent that fits your workflow needs
|
||||
107
docs/explanation/architecture/four-phases.md
Normal file
107
docs/explanation/architecture/four-phases.md
Normal file
@@ -0,0 +1,107 @@
|
||||
---
|
||||
title: "The Four Phases of BMad Method"
|
||||
description: Understanding the four phases of the BMad Method
|
||||
---
|
||||
|
||||
|
||||
BMad Method uses a four-phase approach that adapts to project complexity while ensuring consistent quality.
|
||||
|
||||
## Phase Overview
|
||||
|
||||
| Phase | Name | Purpose | Required? |
|
||||
|-------|------|---------|-----------|
|
||||
| **Phase 1** | Analysis | Exploration and discovery | Optional |
|
||||
| **Phase 2** | Planning | Requirements definition | Required |
|
||||
| **Phase 3** | Solutioning | Technical design | Track-dependent |
|
||||
| **Phase 4** | Implementation | Building the software | Required |
|
||||
|
||||
## Phase 1: Analysis (Optional)
|
||||
|
||||
Exploration and discovery workflows that help validate ideas and understand markets before planning.
|
||||
|
||||
**Workflows:**
|
||||
- `brainstorm-project` - Solution exploration
|
||||
- `research` - Market/technical/competitive research
|
||||
- `product-brief` - Strategic vision capture
|
||||
|
||||
**When to use:**
|
||||
- Starting new projects
|
||||
- Exploring opportunities
|
||||
- Validating market fit
|
||||
|
||||
**When to skip:**
|
||||
- Clear requirements
|
||||
- Well-defined features
|
||||
- Continuing existing work
|
||||
|
||||
## Phase 2: Planning (Required)
|
||||
|
||||
Requirements definition using the scale-adaptive system to match planning depth to project complexity.
|
||||
|
||||
**Workflows:**
|
||||
- `prd` - Product Requirements Document (BMad Method/Enterprise)
|
||||
- `tech-spec` - Technical specification (Quick Flow)
|
||||
- `create-ux-design` - Optional UX specification
|
||||
|
||||
**Key principle:**
|
||||
Define **what** to build and **why**. Leave **how** to Phase 3.
|
||||
|
||||
## Phase 3: Solutioning (Track-Dependent)
|
||||
|
||||
Technical architecture and design decisions that prevent agent conflicts during implementation.
|
||||
|
||||
**Workflows:**
|
||||
- `architecture` - System design with ADRs
|
||||
- `create-epics-and-stories` - Work breakdown (after architecture)
|
||||
- `implementation-readiness` - Gate check
|
||||
|
||||
**Required for:**
|
||||
- BMad Method (complex projects)
|
||||
- Enterprise Method
|
||||
|
||||
**Skip for:**
|
||||
- Quick Flow (simple changes)
|
||||
|
||||
**Key principle:**
|
||||
Make technical decisions explicit so all agents implement consistently.
|
||||
|
||||
## Phase 4: Implementation (Required)
|
||||
|
||||
Iterative sprint-based development with story-centric workflow.
|
||||
|
||||
**Workflows:**
|
||||
- `sprint-planning` - Initialize tracking
|
||||
- `create-story` - Prepare stories
|
||||
- `dev-story` - Implement with tests
|
||||
- `code-review` - Quality assurance
|
||||
- `retrospective` - Continuous improvement
|
||||
|
||||
:::tip[Key Principle]
|
||||
One story at a time — complete each story's full lifecycle before starting the next.
|
||||
:::
|
||||
|
||||
## Phase Flow by Track
|
||||
|
||||
### Quick Flow
|
||||
|
||||
```
|
||||
Phase 2 (tech-spec) → Phase 4 (implement)
|
||||
```
|
||||
|
||||
Skip Phases 1 and 3 for simple changes.
|
||||
|
||||
### BMad Method
|
||||
|
||||
```
|
||||
Phase 1 (optional) → Phase 2 (PRD) → Phase 3 (architecture) → Phase 4 (implement)
|
||||
```
|
||||
|
||||
Full methodology for complex projects.
|
||||
|
||||
### Enterprise
|
||||
|
||||
```
|
||||
Phase 1 → Phase 2 (PRD) → Phase 3 (architecture + extended) → Phase 4 (implement)
|
||||
```
|
||||
|
||||
Same as BMad Method with optional extended workflows.
|
||||
111
docs/explanation/architecture/preventing-agent-conflicts.md
Normal file
111
docs/explanation/architecture/preventing-agent-conflicts.md
Normal file
@@ -0,0 +1,111 @@
|
||||
---
|
||||
title: "Preventing Agent Conflicts"
|
||||
description: How architecture prevents conflicts when multiple agents implement a system
|
||||
---
|
||||
|
||||
|
||||
When multiple AI agents implement different parts of a system, they can make conflicting technical decisions. Architecture documentation prevents this by establishing shared standards.
|
||||
|
||||
## Common Conflict Types
|
||||
|
||||
### API Style Conflicts
|
||||
|
||||
Without architecture:
|
||||
- Agent A uses REST with `/users/{id}`
|
||||
- Agent B uses GraphQL mutations
|
||||
- Result: Inconsistent API patterns, confused consumers
|
||||
|
||||
With architecture:
|
||||
- ADR specifies: "Use GraphQL for all client-server communication"
|
||||
- All agents follow the same pattern
|
||||
|
||||
### Database Design Conflicts
|
||||
|
||||
Without architecture:
|
||||
- Agent A uses snake_case column names
|
||||
- Agent B uses camelCase column names
|
||||
- Result: Inconsistent schema, confusing queries
|
||||
|
||||
With architecture:
|
||||
- Standards document specifies naming conventions
|
||||
- All agents follow the same patterns
|
||||
|
||||
### State Management Conflicts
|
||||
|
||||
Without architecture:
|
||||
- Agent A uses Redux for global state
|
||||
- Agent B uses React Context
|
||||
- Result: Multiple state management approaches, complexity
|
||||
|
||||
With architecture:
|
||||
- ADR specifies state management approach
|
||||
- All agents implement consistently
|
||||
|
||||
## How Architecture Prevents Conflicts
|
||||
|
||||
### 1. Explicit Decisions via ADRs
|
||||
|
||||
Every significant technology choice is documented with:
|
||||
- Context (why this decision matters)
|
||||
- Options considered (what alternatives exist)
|
||||
- Decision (what we chose)
|
||||
- Rationale (why we chose it)
|
||||
- Consequences (trade-offs accepted)
|
||||
|
||||
### 2. FR/NFR-Specific Guidance
|
||||
|
||||
Architecture maps each functional requirement to technical approach:
|
||||
- FR-001: User Management → GraphQL mutations
|
||||
- FR-002: Mobile App → Optimized queries
|
||||
|
||||
### 3. Standards and Conventions
|
||||
|
||||
Explicit documentation of:
|
||||
- Directory structure
|
||||
- Naming conventions
|
||||
- Code organization
|
||||
- Testing patterns
|
||||
|
||||
## Architecture as Shared Context
|
||||
|
||||
Think of architecture as the shared context that all agents read before implementing:
|
||||
|
||||
```
|
||||
PRD: "What to build"
|
||||
↓
|
||||
Architecture: "How to build it"
|
||||
↓
|
||||
Agent A reads architecture → implements Epic 1
|
||||
Agent B reads architecture → implements Epic 2
|
||||
Agent C reads architecture → implements Epic 3
|
||||
↓
|
||||
Result: Consistent implementation
|
||||
```
|
||||
|
||||
## Key ADR Topics
|
||||
|
||||
Common decisions that prevent conflicts:
|
||||
|
||||
| Topic | Example Decision |
|
||||
|-------|-----------------|
|
||||
| API Style | GraphQL vs REST vs gRPC |
|
||||
| Database | PostgreSQL vs MongoDB |
|
||||
| Auth | JWT vs Sessions |
|
||||
| State Management | Redux vs Context vs Zustand |
|
||||
| Styling | CSS Modules vs Tailwind vs Styled Components |
|
||||
| Testing | Jest + Playwright vs Vitest + Cypress |
|
||||
|
||||
## Anti-Patterns to Avoid
|
||||
|
||||
:::caution[Common Mistakes]
|
||||
- **Implicit Decisions** — "We'll figure out the API style as we go" leads to inconsistency
|
||||
- **Over-Documentation** — Documenting every minor choice causes analysis paralysis
|
||||
- **Stale Architecture** — Documents written once and never updated cause agents to follow outdated patterns
|
||||
:::
|
||||
|
||||
:::tip[Correct Approach]
|
||||
- Document decisions that cross epic boundaries
|
||||
- Focus on conflict-prone areas
|
||||
- Update architecture as you learn
|
||||
- Use `correct-course` for significant changes
|
||||
:::
|
||||
75
docs/explanation/architecture/why-solutioning-matters.md
Normal file
75
docs/explanation/architecture/why-solutioning-matters.md
Normal file
@@ -0,0 +1,75 @@
|
||||
---
|
||||
title: "Why Solutioning Matters"
|
||||
description: Understanding why the solutioning phase is critical for multi-epic projects
|
||||
---
|
||||
|
||||
|
||||
Phase 3 (Solutioning) translates **what** to build (from Planning) into **how** to build it (technical design). This phase prevents agent conflicts in multi-epic projects by documenting architectural decisions before implementation begins.
|
||||
|
||||
## The Problem Without Solutioning
|
||||
|
||||
```
|
||||
Agent 1 implements Epic 1 using REST API
|
||||
Agent 2 implements Epic 2 using GraphQL
|
||||
Result: Inconsistent API design, integration nightmare
|
||||
```
|
||||
|
||||
When multiple agents implement different parts of a system without shared architectural guidance, they make independent technical decisions that may conflict.
|
||||
|
||||
## The Solution With Solutioning
|
||||
|
||||
```
|
||||
architecture workflow decides: "Use GraphQL for all APIs"
|
||||
All agents follow architecture decisions
|
||||
Result: Consistent implementation, no conflicts
|
||||
```
|
||||
|
||||
By documenting technical decisions explicitly, all agents implement consistently and integration becomes straightforward.
|
||||
|
||||
## Solutioning vs Planning
|
||||
|
||||
| Aspect | Planning (Phase 2) | Solutioning (Phase 3) |
|
||||
| -------- | ----------------------- | --------------------------------- |
|
||||
| Question | What and Why? | How? Then What units of work? |
|
||||
| Output | FRs/NFRs (Requirements) | Architecture + Epics/Stories |
|
||||
| Agent | PM | Architect → PM |
|
||||
| Audience | Stakeholders | Developers |
|
||||
| Document | PRD (FRs/NFRs) | Architecture + Epic Files |
|
||||
| Level | Business logic | Technical design + Work breakdown |
|
||||
|
||||
## Key Principle
|
||||
|
||||
**Make technical decisions explicit and documented** so all agents implement consistently.
|
||||
|
||||
This prevents:
|
||||
- API style conflicts (REST vs GraphQL)
|
||||
- Database design inconsistencies
|
||||
- State management disagreements
|
||||
- Naming convention mismatches
|
||||
- Security approach variations
|
||||
|
||||
## When Solutioning is Required
|
||||
|
||||
| Track | Solutioning Required? |
|
||||
|-------|----------------------|
|
||||
| Quick Flow | No - skip entirely |
|
||||
| BMad Method Simple | Optional |
|
||||
| BMad Method Complex | Yes |
|
||||
| Enterprise | Yes |
|
||||
|
||||
:::tip[Rule of Thumb]
|
||||
If you have multiple epics that could be implemented by different agents, you need solutioning.
|
||||
:::
|
||||
|
||||
## The Cost of Skipping
|
||||
|
||||
Skipping solutioning on complex projects leads to:
|
||||
|
||||
- **Integration issues** discovered mid-sprint
|
||||
- **Rework** due to conflicting implementations
|
||||
- **Longer development time** overall
|
||||
- **Technical debt** from inconsistent patterns
|
||||
|
||||
:::caution[Cost Multiplier]
|
||||
Catching alignment issues in solutioning is 10× faster than discovering them during implementation.
|
||||
:::
|
||||
@@ -1,8 +1,14 @@
|
||||
# Custom Content
|
||||
---
|
||||
title: "Custom Content"
|
||||
---
|
||||
|
||||
BMAD supports several categories of officially supported custom content that extend the platform's capabilities. Custom content can be created manually or with the recommended assistance of the BMad Builder (BoMB) Module. The BoMB Agent provides workflows and expertise to plan and build any custom content you can imagine.
|
||||
BMad supports several categories of custom content that extend the platform's capabilities — from simple personal agents to full-featured professional modules.
|
||||
|
||||
This flexibility transforms the platform beyond its current capabilities, enabling:
|
||||
:::tip[Recommended Approach]
|
||||
Use the BMad Builder (BoMB) Module for guided workflows and expertise when creating custom content.
|
||||
:::
|
||||
|
||||
This flexibility enables:
|
||||
|
||||
- Extensions and add-ons for existing modules (BMad Method, Creative Intelligence Suite)
|
||||
- Completely new modules, workflows, templates, and agents outside software engineering
|
||||
@@ -14,30 +20,29 @@ This flexibility transforms the platform beyond its current capabilities, enabli
|
||||
|
||||
## Categories
|
||||
|
||||
- [Custom Content](#custom-content)
|
||||
- [Categories](#categories)
|
||||
- [Custom Stand Alone Modules](#custom-stand-alone-modules)
|
||||
- [Custom Add On Modules](#custom-add-on-modules)
|
||||
- [Custom Global Modules](#custom-global-modules)
|
||||
- [Custom Agents](#custom-agents)
|
||||
- [BMad Tiny Agents](#bmad-tiny-agents)
|
||||
- [Simple vs Expert Agents](#simple-vs-expert-agents)
|
||||
- [Custom Workflows](#custom-workflows)
|
||||
- [Categories](#categories)
|
||||
- [Custom Stand-Alone Modules](#custom-stand-alone-modules)
|
||||
- [Custom Add-On Modules](#custom-add-on-modules)
|
||||
- [Custom Global Modules](#custom-global-modules)
|
||||
- [Custom Agents](#custom-agents)
|
||||
- [BMad Tiny Agents](#bmad-tiny-agents)
|
||||
- [Simple and Expert Agents](#simple-and-expert-agents)
|
||||
- [Custom Workflows](#custom-workflows)
|
||||
|
||||
## Custom Stand Alone Modules
|
||||
## Custom Stand-Alone Modules
|
||||
|
||||
Custom modules range from simple collections of related agents, workflows, and tools designed to work independently, to complex, expansive systems like the BMad Method or even larger applications.
|
||||
|
||||
Custom modules are [installable](./custom-content-installation.md) using the standard BMAD method and support advanced features:
|
||||
Custom modules are [installable](/docs/how-to/installation/install-custom-modules.md) using the standard BMad method and support advanced features:
|
||||
|
||||
- Optional user information collection during installation/updates
|
||||
- Versioning and upgrade paths
|
||||
- Custom installer functions with IDE-specific post-installation handling (custom hooks, subagents, or vendor-specific tools)
|
||||
- Ability to bundle specific tools such as MCP, skills, execution libraries, and code
|
||||
|
||||
## Custom Add On Modules
|
||||
## Custom Add-On Modules
|
||||
|
||||
Custom Add On Modules contain specific agents, tools, or workflows that expand, modify, or customize another module but cannot exist or install independently. These add-ons provide enhanced functionality while leveraging the base module's existing capabilities.
|
||||
Custom Add-On Modules contain specific agents, tools, or workflows that expand, modify, or customize another module but cannot exist or install independently. These add-ons provide enhanced functionality while leveraging the base module's existing capabilities.
|
||||
|
||||
Examples include:
|
||||
|
||||
@@ -45,7 +50,7 @@ Examples include:
|
||||
- Framework-specific support for particular use cases
|
||||
- Game development expansions that add new genre-specific capabilities without reinventing existing functionality
|
||||
|
||||
Add on modules can include:
|
||||
Add-on modules can include:
|
||||
|
||||
- Custom agents with awareness of the target module
|
||||
- Access to existing module workflows
|
||||
@@ -53,13 +58,12 @@ Add on modules can include:
|
||||
|
||||
## Custom Global Modules
|
||||
|
||||
Similar to Custom Stand Alone Modules, but designed to add functionality that applies across all installed content. These modules provide cross-cutting capabilities that enhance the entire BMAD ecosystem.
|
||||
Similar to Custom Stand-Alone Modules, but designed to add functionality that applies across all installed content. These modules provide cross-cutting capabilities that enhance the entire BMad ecosystem.
|
||||
|
||||
Examples include:
|
||||
|
||||
- The current TTS (Text-to-Speech) functionality for Claude, which will be rebuilt as a global module
|
||||
- The core module, which is always installed and provides all agents with party mode and advanced elicitation capabilities
|
||||
- Installation and update tools that work with any BMAD method configuration
|
||||
- Installation and update tools that work with any BMad method configuration
|
||||
|
||||
Upcoming standards will document best practices for building global content that affects installed modules through:
|
||||
|
||||
@@ -81,7 +85,7 @@ Personal agents designed for highly specific needs that may not be suitable for
|
||||
|
||||
These are simple, standalone files that can be scoped to focus on specific data or paths when integrated into an information vault or repository.
|
||||
|
||||
### Simple vs Expert Agents
|
||||
### Simple and Expert Agents
|
||||
|
||||
The distinction between simple and expert agents lies in their structure:
|
||||
|
||||
@@ -89,7 +93,6 @@ The distinction between simple and expert agents lies in their structure:
|
||||
|
||||
- Single file containing all prompts and configuration
|
||||
- Self-contained and straightforward
|
||||
- has metadata type: simple
|
||||
|
||||
**Expert Agent:**
|
||||
|
||||
@@ -98,7 +101,9 @@ The distinction between simple and expert agents lies in their structure:
|
||||
- When installed, the sidecar folder (`[agentname]-sidecar`) is placed in the user memory location
|
||||
- has metadata type: expert
|
||||
|
||||
:::note[Key Distinction]
|
||||
The key distinction is the presence of a sidecar folder. As web and consumer agent tools evolve to support common memory mechanisms, storage formats, and MCP, the writable memory files will adapt to support these evolving standards.
|
||||
:::
|
||||
|
||||
Custom agents can be:
|
||||
|
||||
@@ -119,4 +124,6 @@ A custom workflow created outside of a larger module can still be distributed an
|
||||
- Slash commands
|
||||
- Manual command/prompt execution when supported by tools
|
||||
|
||||
:::tip[Core Concept]
|
||||
At its core, a custom workflow is a single or series of prompts designed to achieve a specific outcome.
|
||||
:::
|
||||
45
docs/explanation/bmad-builder/index.md
Normal file
45
docs/explanation/bmad-builder/index.md
Normal file
@@ -0,0 +1,45 @@
|
||||
---
|
||||
title: "BMad Builder (BMB)"
|
||||
description: Create custom agents, workflows, and modules for BMad
|
||||
---
|
||||
|
||||
Create custom agents, workflows, and modules for BMad — from simple personal assistants to full-featured professional tools.
|
||||
|
||||
## Quick Start
|
||||
|
||||
| Resource | Description |
|
||||
|----------|-------------|
|
||||
| **[Agent Creation Guide](/docs/tutorials/advanced/create-custom-agent.md)** | Step-by-step guide to building your first agent |
|
||||
| **[Install Custom Modules](/docs/how-to/installation/install-custom-modules.md)** | Installing standalone simple and expert agents |
|
||||
|
||||
## Agent Architecture
|
||||
|
||||
| Type | Description |
|
||||
|------|-------------|
|
||||
| **Simple Agent** | Self-contained, optimized, personality-driven |
|
||||
| **Expert Agent** | Memory, sidecar files, domain restrictions |
|
||||
| **Module Agent** | Workflow integration, professional tools |
|
||||
|
||||
## Key Concepts
|
||||
|
||||
Agents are authored in YAML with Handlebars templating. The compiler auto-injects:
|
||||
|
||||
1. **Frontmatter** — Name and description from metadata
|
||||
2. **Activation Block** — Steps, menu handlers, rules
|
||||
3. **Menu Enhancement** — `*help` and `*exit` commands added automatically
|
||||
4. **Trigger Prefixing** — Your triggers auto-prefixed with `*`
|
||||
|
||||
:::note[Learn More]
|
||||
See [Custom Content Types](/docs/explanation/bmad-builder/custom-content-types.md) for detailed explanations of all content categories.
|
||||
:::
|
||||
|
||||
## Reference Examples
|
||||
|
||||
Production-ready examples available in the BMB reference folder:
|
||||
|
||||
| Agent | Type | Description |
|
||||
|-------|------|-------------|
|
||||
| **commit-poet** | Simple | Commit message artisan with style customization |
|
||||
| **journal-keeper** | Expert | Personal journal companion with memory and pattern recognition |
|
||||
| **security-engineer** | Module | BMM security specialist with threat modeling |
|
||||
| **trend-analyst** | Module | CIS trend intelligence expert |
|
||||
131
docs/explanation/bmm/index.md
Normal file
131
docs/explanation/bmm/index.md
Normal file
@@ -0,0 +1,131 @@
|
||||
---
|
||||
title: "BMM Documentation"
|
||||
---
|
||||
|
||||
Complete guides for the BMad Method Module (BMM) — AI-powered agile development workflows that adapt to your project's complexity.
|
||||
|
||||
## Getting Started
|
||||
|
||||
:::tip[Quick Path]
|
||||
Install → workflow-init → Follow agent guidance
|
||||
:::
|
||||
|
||||
**New to BMM?** Start here:
|
||||
|
||||
| Resource | Description |
|
||||
|----------|-------------|
|
||||
| **[Quick Start Guide](/docs/tutorials/getting-started/getting-started-bmadv6.md)** | Step-by-step guide to building your first project |
|
||||
| **[Complete Workflow Diagram](../../tutorials/getting-started/images/workflow-method-greenfield.svg)** | Visual flowchart showing all phases, agents, and decision points |
|
||||
|
||||
## Core Concepts
|
||||
|
||||
The BMad Method is meant to be adapted and customized to your specific needs. In this realm there is no one size fits all - your needs are unique, and BMad Method is meant to support this (and if it does not, can be further customized or extended with new modules).
|
||||
|
||||
First know there is the full BMad Method Process and then there is a Quick Flow for those quicker smaller efforts.
|
||||
|
||||
- **[Full Adaptive BMad Method](#workflow-guides)** - Full planning and scope support through extensive development and testing.
|
||||
- Broken down into 4 phases, all of which are comprised of both required and optional phases
|
||||
- Phases 1-3 are all about progressive idea development through planning and preparations to build your project.
|
||||
- Phase 4 is the implementation cycle where you will Just In Time (JIT) produce the contextual stories needed for the dev agent based on the extensive planning completed
|
||||
- All 4 phases have optional steps in them, depending on how rigorous you want to go with planning, research ideation, validation, testing and traceability.
|
||||
- While there is a lot here, know that even this can be distilled down to a simple PRD, Epic and Story list and then jump into the dev cycle. But if that is all you want, you might be better off with the BMad Quick Flow described next
|
||||
|
||||
- **[BMad Quick Flow](/docs/explanation/features/quick-flow.md)** - Fast-track development workflow
|
||||
- 3-step process: spec → dev → optional review
|
||||
- Perfect for bug fixes and small features
|
||||
- Rapid prototyping with production quality
|
||||
- Implementation in minutes, not days
|
||||
- Has a specialized single agent that does all of this: **[Quick Flow Solo Dev Agent](/docs/explanation/agents/barry-quick-flow.md)**
|
||||
|
||||
- **TEA engagement (optional)** - Choose TEA engagement: none, TEA-only (standalone), or integrated by track. See **[Test Architect Guide](/docs/explanation/features/tea-overview.md)**.
|
||||
|
||||
## Agents and Collaboration
|
||||
|
||||
Complete guide to BMM's AI agent team:
|
||||
|
||||
- **[Agents Guide](/docs/explanation/core-concepts/agent-roles.md)** - Comprehensive agent reference
|
||||
- 12 specialized BMM agents + BMad Master
|
||||
- Agent roles, workflows, and when to use them
|
||||
- Agent customization system
|
||||
- Best practices and common patterns
|
||||
|
||||
- **[Party Mode Guide](/docs/explanation/features/party-mode.md)** - Multi-agent collaboration
|
||||
- How party mode works (19+ agents collaborate in real-time)
|
||||
- When to use it (strategic, creative, cross-functional, complex)
|
||||
- Example party compositions
|
||||
- Multi-module integration (BMM + CIS + BMB + custom)
|
||||
- Agent customization in party mode
|
||||
- Best practices
|
||||
|
||||
## Working with Existing Code
|
||||
|
||||
Comprehensive guide for brownfield development:
|
||||
|
||||
- **[Brownfield Development Guide](/docs/how-to/brownfield/index.md)** - Complete guide for existing codebases
|
||||
- Documentation phase strategies
|
||||
- Track selection for brownfield
|
||||
- Integration with existing patterns
|
||||
- Phase-by-phase workflow guidance
|
||||
- Common scenarios
|
||||
|
||||
## Quick References
|
||||
|
||||
Essential reference materials:
|
||||
|
||||
- **[Glossary](/docs/reference/glossary/index.md)** - Key terminology and concepts
|
||||
- **[FAQ](/docs/explanation/faq/index.md)** - Frequently asked questions across all topics
|
||||
|
||||
## Choose Your Path
|
||||
|
||||
### I need to...
|
||||
|
||||
**Build something new (greenfield)**
|
||||
→ Start with [Quick Start Guide](/docs/tutorials/getting-started/getting-started-bmadv6.md)
|
||||
|
||||
**Fix a bug or add small feature**
|
||||
→ Use the [Quick Flow Solo Dev](/docs/explanation/agents/barry-quick-flow.md) directly with its dedicated stand alone [Quick Bmad Spec Flow](/docs/explanation/features/quick-flow.md) process
|
||||
|
||||
**Work with existing codebase (brownfield)**
|
||||
→ Read [Brownfield Development Guide](/docs/how-to/brownfield/index.md)
|
||||
→ Pay special attention to documentation requirements for brownfield projects
|
||||
|
||||
## Workflow Guides
|
||||
|
||||
Comprehensive documentation for all BMM workflows organized by phase:
|
||||
|
||||
- **[Phase 1: Analysis Workflows](/docs/how-to/workflows/run-brainstorming-session.md)** - Optional exploration and research workflows (595 lines)
|
||||
- brainstorm-project, product-brief, research, and more
|
||||
- When to use analysis workflows
|
||||
- Creative and strategic tools
|
||||
|
||||
- **[Phase 2: Planning Workflows](/docs/how-to/workflows/create-prd.md)** - Scale-adaptive planning (967 lines)
|
||||
- prd, tech-spec, gdd, narrative, ux
|
||||
- Track-based planning approach (Quick Flow, BMad Method, Enterprise Method)
|
||||
- Which planning workflow to use
|
||||
|
||||
- **[Phase 3: Solutioning Workflows](/docs/how-to/workflows/create-architecture.md)** - Architecture and validation (638 lines)
|
||||
- architecture, create-epics-and-stories, implementation-readiness
|
||||
- V6: Epics created AFTER architecture for better quality
|
||||
- Required for BMad Method and Enterprise Method tracks
|
||||
- Preventing agent conflicts
|
||||
|
||||
- **[Phase 4: Implementation Workflows](/docs/how-to/workflows/run-sprint-planning.md)** - Sprint-based development (1,634 lines)
|
||||
- sprint-planning, create-story, dev-story, code-review
|
||||
- Complete story lifecycle
|
||||
- One-story-at-a-time discipline
|
||||
|
||||
- **[Testing & QA Workflows](/docs/explanation/features/tea-overview.md)** - Comprehensive quality assurance (1,420 lines)
|
||||
- Test strategy, automation, quality gates
|
||||
- TEA agent and test healing
|
||||
|
||||
## External Resources
|
||||
|
||||
### Community and Support
|
||||
|
||||
- **[Discord Community](https://discord.gg/gk8jAdXWmj)** - Get help from the community (#bmad-method-help, #report-bugs-and-issues)
|
||||
- **[GitHub Issues](https://github.com/bmad-code-org/BMAD-METHOD/issues)** - Report bugs or request features
|
||||
- **[YouTube Channel](https://www.youtube.com/@BMadCode)** - Video tutorials and walkthroughs
|
||||
|
||||
:::tip[Ready to Begin?]
|
||||
[Start with the Quick Start Guide](/docs/tutorials/getting-started/getting-started-bmadv6.md)
|
||||
:::
|
||||
179
docs/explanation/core-concepts/agent-roles.md
Normal file
179
docs/explanation/core-concepts/agent-roles.md
Normal file
@@ -0,0 +1,179 @@
|
||||
---
|
||||
title: "Agent Roles in BMad Method"
|
||||
description: Understanding the different agent roles in BMad Method
|
||||
---
|
||||
|
||||
BMad Method uses specialized AI agents, each with a distinct role, expertise, and personality. Understanding these roles helps you know which agent to use for each task.
|
||||
|
||||
## Core Agents Overview
|
||||
|
||||
| Agent | Role | Primary Phase |
|
||||
|-------|------|---------------|
|
||||
| **Analyst** | Research and discovery | Phase 1 (Analysis) |
|
||||
| **PM** | Requirements and planning | Phase 2 (Planning) |
|
||||
| **Architect** | Technical design | Phase 3 (Solutioning) |
|
||||
| **SM** | Sprint orchestration | Phase 4 (Implementation) |
|
||||
| **DEV** | Code implementation | Phase 4 (Implementation) |
|
||||
| **TEA** | Test architecture | Phases 3-4 (Cross-phase) |
|
||||
| **UX Designer** | User experience | Phase 2-3 |
|
||||
| **Quick Flow Solo Dev** | Fast solo development | All phases (Quick Flow) |
|
||||
|
||||
## Phase 1: Analysis
|
||||
|
||||
### Analyst (Mary)
|
||||
|
||||
Business analysis and research specialist.
|
||||
|
||||
**Responsibilities:**
|
||||
- Brainstorming and ideation
|
||||
- Market, domain, and competitive research
|
||||
- Product brief creation
|
||||
- Brownfield project documentation
|
||||
|
||||
**Key Workflows:**
|
||||
- `*brainstorm-project`
|
||||
- `*research`
|
||||
- `*product-brief`
|
||||
- `*document-project`
|
||||
|
||||
**When to use:** Starting new projects, exploring ideas, validating market fit, documenting existing codebases.
|
||||
|
||||
## Phase 2: Planning
|
||||
|
||||
### PM (John)
|
||||
|
||||
Product requirements and planning expert.
|
||||
|
||||
**Responsibilities:**
|
||||
- Creating Product Requirements Documents
|
||||
- Defining functional and non-functional requirements
|
||||
- Breaking requirements into epics and stories
|
||||
- Validating implementation readiness
|
||||
|
||||
**Key Workflows:**
|
||||
- `*create-prd`
|
||||
- `*create-epics-and-stories`
|
||||
- `*implementation-readiness`
|
||||
|
||||
**When to use:** Defining what to build, creating PRDs, organizing work into stories.
|
||||
|
||||
### UX Designer (Sally)
|
||||
|
||||
User experience and UI design specialist.
|
||||
|
||||
**Responsibilities:**
|
||||
- UX specification creation
|
||||
- User journey mapping
|
||||
- Wireframe and mockup design
|
||||
- Design system documentation
|
||||
|
||||
**Key Workflows:**
|
||||
- `*create-ux-design`
|
||||
- `*validate-design`
|
||||
|
||||
**When to use:** When UX is a primary differentiator, complex user workflows, design system creation.
|
||||
|
||||
## Phase 3: Solutioning
|
||||
|
||||
### Architect (Winston)
|
||||
|
||||
System architecture and technical design expert.
|
||||
|
||||
**Responsibilities:**
|
||||
- System architecture design
|
||||
- Architecture Decision Records (ADRs)
|
||||
- Technical standards definition
|
||||
- Implementation readiness validation
|
||||
|
||||
**Key Workflows:**
|
||||
- `*create-architecture`
|
||||
- `*implementation-readiness`
|
||||
|
||||
**When to use:** Multi-epic projects, cross-cutting technical decisions, preventing agent conflicts.
|
||||
|
||||
## Phase 4: Implementation
|
||||
|
||||
### SM (Bob)
|
||||
|
||||
Sprint planning and story preparation orchestrator.
|
||||
|
||||
**Responsibilities:**
|
||||
- Sprint planning and tracking
|
||||
- Story preparation for development
|
||||
- Course correction handling
|
||||
- Epic retrospectives
|
||||
|
||||
**Key Workflows:**
|
||||
- `*sprint-planning`
|
||||
- `*create-story`
|
||||
- `*correct-course`
|
||||
- `*epic-retrospective`
|
||||
|
||||
**When to use:** Organizing work, preparing stories, tracking progress.
|
||||
|
||||
### DEV (Amelia)
|
||||
|
||||
Story implementation and code review specialist.
|
||||
|
||||
**Responsibilities:**
|
||||
- Story implementation with tests
|
||||
- Code review
|
||||
- Following architecture patterns
|
||||
- Quality assurance
|
||||
|
||||
**Key Workflows:**
|
||||
- `*dev-story`
|
||||
- `*code-review`
|
||||
|
||||
**When to use:** Writing code, implementing stories, reviewing quality.
|
||||
|
||||
## Cross-Phase Agents
|
||||
|
||||
### TEA (Murat)
|
||||
|
||||
Test architecture and quality strategy expert.
|
||||
|
||||
**Responsibilities:**
|
||||
- Test framework setup
|
||||
- Test design and planning
|
||||
- ATDD and automation
|
||||
- Quality gate decisions
|
||||
|
||||
**Key Workflows:**
|
||||
- `*framework`, `*ci`
|
||||
- `*test-design`, `*atdd`, `*automate`
|
||||
- `*test-review`, `*trace`, `*nfr-assess`
|
||||
|
||||
**When to use:** Setting up testing, creating test plans, quality gates.
|
||||
|
||||
## Quick Flow
|
||||
|
||||
### Quick Flow Solo Dev (Barry)
|
||||
|
||||
Fast solo development without handoffs.
|
||||
|
||||
**Responsibilities:**
|
||||
- Technical specification
|
||||
- End-to-end implementation
|
||||
- Code review
|
||||
|
||||
**Key Workflows:**
|
||||
- `*quick-spec`
|
||||
- `*quick-dev`
|
||||
- `*code-review`
|
||||
|
||||
**When to use:** Bug fixes, small features, rapid prototyping.
|
||||
|
||||
## Choosing the Right Agent
|
||||
|
||||
| Task | Agent |
|
||||
|------|-------|
|
||||
| Brainstorming ideas | Analyst |
|
||||
| Market research | Analyst |
|
||||
| Creating PRD | PM |
|
||||
| Designing UX | UX Designer |
|
||||
| System architecture | Architect |
|
||||
| Preparing stories | SM |
|
||||
| Writing code | DEV |
|
||||
| Setting up tests | TEA |
|
||||
| Quick bug fix | Quick Flow Solo Dev |
|
||||
35
docs/explanation/core-concepts/index.md
Normal file
35
docs/explanation/core-concepts/index.md
Normal file
@@ -0,0 +1,35 @@
|
||||
---
|
||||
title: "BMad Core Concepts"
|
||||
---
|
||||
|
||||
Understanding the fundamental building blocks of the BMad Method.
|
||||
|
||||
## The Essentials
|
||||
|
||||
| Concept | Description | Guide |
|
||||
|---------|-------------|-------|
|
||||
| **Agents** | AI assistants with personas, capabilities, and menus | [Agents Guide](/docs/explanation/core-concepts/what-are-agents.md) |
|
||||
| **Workflows** | Structured processes for achieving specific outcomes | [Workflows Guide](/docs/explanation/core-concepts/what-are-workflows.md) |
|
||||
| **Modules** | Packaged collections of agents and workflows | [Modules Guide](/docs/explanation/core-concepts/what-are-modules.md) |
|
||||
|
||||
## Getting Started
|
||||
|
||||
### New to BMad?
|
||||
Start here to understand what BMad is and how it works:
|
||||
|
||||
1. **[Agents Guide](/docs/explanation/core-concepts/what-are-agents.md)** - Learn about Simple and Expert agents
|
||||
2. **[Workflows Guide](/docs/explanation/core-concepts/what-are-workflows.md)** - Understand how workflows orchestrate tasks
|
||||
3. **[Modules Guide](/docs/explanation/core-concepts/what-are-modules.md)** - See how modules organize functionality
|
||||
|
||||
### Installing BMad
|
||||
|
||||
- **[Installation Guide](/docs/how-to/installation/index.md)** - Set up BMad in your project
|
||||
- **[Upgrading from v4](/docs/how-to/installation/upgrade-to-v6.md)** - Migrate from earlier versions
|
||||
|
||||
### Configuration
|
||||
|
||||
- **[BMad Customization](/docs/how-to/customization/index.md)** - Personalize agents and workflows
|
||||
|
||||
### Advanced
|
||||
|
||||
- **[Web Bundles](/docs/explanation/features/web-bundles.md)** - Use BMad in Gemini Gems and Custom GPTs
|
||||
95
docs/explanation/core-concepts/what-are-agents.md
Normal file
95
docs/explanation/core-concepts/what-are-agents.md
Normal file
@@ -0,0 +1,95 @@
|
||||
---
|
||||
title: "Agents"
|
||||
---
|
||||
|
||||
Agents are AI assistants that help you accomplish tasks. Each agent has a unique personality, specialized capabilities, and an interactive menu.
|
||||
|
||||
## Agent Types
|
||||
|
||||
BMad has two primary agent types, designed for different use cases:
|
||||
|
||||
### Simple Agents
|
||||
|
||||
**Self-contained, focused, ready to use.**
|
||||
|
||||
Simple agents are complete in a single file. They excel at well-defined tasks and require minimal setup.
|
||||
|
||||
**Best for:**
|
||||
- Single-purpose assistants (code review, documentation, commit messages)
|
||||
- Quick deployment
|
||||
- Projects that don't require persistent memory
|
||||
- Getting started fast
|
||||
|
||||
**Example:** A commit message agent that reads your git diff and generates conventional commits.
|
||||
|
||||
### Expert Agents
|
||||
|
||||
**Powerful, memory-equipped, domain specialists.**
|
||||
|
||||
Expert agents have a **sidecar** - a companion folder containing additional instructions, workflows, and memory files. They remember context across sessions and handle complex, multi-step tasks.
|
||||
|
||||
**Best for:**
|
||||
- Domain specialists (security architect, game designer, product manager)
|
||||
- Tasks requiring persistent memory
|
||||
- Complex workflows with multiple stages
|
||||
- Projects that grow over time
|
||||
|
||||
**Example:** A game architect that remembers your design decisions, maintains consistency across sprints, and coordinates with other specialists.
|
||||
|
||||
## Key Differences
|
||||
|
||||
| Feature | Simple | Expert |
|
||||
| ---------------- | -------------- | -------------------------- |
|
||||
| **Files** | Single file | Agent + sidecar folder |
|
||||
| **Memory** | Session only | Persistent across sessions |
|
||||
| **Capabilities** | Focused scope | Multi-domain, extensible |
|
||||
| **Setup** | Zero config | Sidecar initialization |
|
||||
| **Best Use** | Specific tasks | Ongoing projects |
|
||||
|
||||
## Agent Components
|
||||
|
||||
All agents share these building blocks:
|
||||
|
||||
### Persona
|
||||
- **Role** - What the agent does (expertise domain)
|
||||
- **Identity** - Who the agent is (personality, character)
|
||||
- **Communication Style** - How the agent speaks (tone, voice)
|
||||
- **Principles** - Why the agent acts (values, decision framework)
|
||||
|
||||
### Capabilities
|
||||
- Skills, tools, and knowledge the agent can apply
|
||||
- Mapped to specific menu commands
|
||||
|
||||
### Menu
|
||||
- Interactive command list
|
||||
- Triggers, descriptions, and handlers
|
||||
- Auto-includes help and exit options
|
||||
|
||||
### Critical Actions (optional)
|
||||
- Instructions that execute before the agent starts
|
||||
- Enable autonomous behaviors (e.g., "check git status before changes")
|
||||
|
||||
## Which Should You Use?
|
||||
|
||||
:::tip[Quick Decision]
|
||||
Choose **Simple** for focused, one-off tasks with no memory needs. Choose **Expert** when you need persistent context and complex workflows.
|
||||
:::
|
||||
|
||||
**Choose Simple when:**
|
||||
- You need a task done quickly and reliably
|
||||
- The scope is well-defined and won't change much
|
||||
- You don't need the agent to remember things between sessions
|
||||
|
||||
**Choose Expert when:**
|
||||
- You're building something complex over time
|
||||
- The agent needs to maintain context (project history, decisions)
|
||||
- You want the agent to coordinate workflows or other agents
|
||||
- Domain expertise requires specialized knowledge bases
|
||||
|
||||
## Creating Custom Agents
|
||||
|
||||
BMad provides the **BMad Builder (BMB)** module for creating your own agents. See the [Agent Creation Guide](/docs/tutorials/advanced/create-custom-agent.md) for step-by-step instructions.
|
||||
|
||||
## Customizing Existing Agents
|
||||
|
||||
You can modify any agent's behavior without editing core files. See [BMad Customization](/docs/how-to/customization/index.md) for details. It is critical to never modify an installed agents .md file directly and follow the customization process, this way future updates to the agent or module its part of will continue to be updated and recompiled with the installer tool, and your customizations will still be retained.
|
||||
78
docs/explanation/core-concepts/what-are-modules.md
Normal file
78
docs/explanation/core-concepts/what-are-modules.md
Normal file
@@ -0,0 +1,78 @@
|
||||
---
|
||||
title: "Modules"
|
||||
---
|
||||
|
||||
Modules are organized collections of agents and workflows that solve specific problems or address particular domains.
|
||||
|
||||
## What is a Module?
|
||||
|
||||
A module is a self-contained package that includes:
|
||||
|
||||
- **Agents** - Specialized AI assistants
|
||||
- **Workflows** - Step-by-step processes
|
||||
- **Configuration** - Module-specific settings
|
||||
- **Documentation** - Usage guides and reference
|
||||
|
||||
## Official Modules
|
||||
|
||||
:::note[Core is Always Installed]
|
||||
The Core module is automatically included with every BMad installation. It provides the foundation that other modules build upon.
|
||||
:::
|
||||
|
||||
### Core Module
|
||||
Always installed, provides shared functionality:
|
||||
- Global configuration
|
||||
- Core workflows (Party Mode, Advanced Elicitation, Brainstorming)
|
||||
- Common tasks (document indexing, sharding, review)
|
||||
|
||||
### BMad Method (BMM)
|
||||
Software and game development:
|
||||
- Project planning workflows
|
||||
- Implementation agents (Dev, PM, QA, Scrum Master)
|
||||
- Testing and architecture guidance
|
||||
|
||||
### BMad Builder (BMB)
|
||||
Create custom solutions:
|
||||
- Agent creation workflows
|
||||
- Workflow authoring tools
|
||||
- Module scaffolding
|
||||
|
||||
### Creative Intelligence Suite (CIS)
|
||||
Innovation and creativity:
|
||||
- Creative thinking techniques
|
||||
- Innovation strategy workflows
|
||||
- Storytelling and ideation
|
||||
|
||||
### BMad Game Dev (BMGD)
|
||||
Game development specialization:
|
||||
- Game design workflows
|
||||
- Narrative development
|
||||
- Performance testing frameworks
|
||||
|
||||
## Module Structure
|
||||
|
||||
Installed modules follow this structure:
|
||||
|
||||
```
|
||||
_bmad/
|
||||
├── core/ # Always present
|
||||
├── bmm/ # BMad Method (if installed)
|
||||
├── bmb/ # BMad Builder (if installed)
|
||||
├── cis/ # Creative Intelligence (if installed)
|
||||
└── bmgd/ # Game Dev (if installed)
|
||||
```
|
||||
|
||||
## Custom Modules
|
||||
|
||||
You can create your own modules containing:
|
||||
- Custom agents for your domain
|
||||
- Organizational workflows
|
||||
- Team-specific configurations
|
||||
|
||||
Custom modules are installed the same way as official modules.
|
||||
|
||||
## Installing Modules
|
||||
|
||||
During BMad installation, you choose which modules to install. You can also add or remove modules later by re-running the installer.
|
||||
|
||||
See [Installation Guide](/docs/how-to/installation/index.md) for details.
|
||||
204
docs/explanation/core-concepts/what-are-workflows.md
Normal file
204
docs/explanation/core-concepts/what-are-workflows.md
Normal file
@@ -0,0 +1,204 @@
|
||||
---
|
||||
title: "Workflows"
|
||||
---
|
||||
|
||||
Workflows are like prompts on steroids. They harness the untapped power and control of LLMs through progressive disclosure—breaking complex tasks into focused steps that execute sequentially. Instead of random AI slop where you hope for the best, workflows give you repeatable, reliable, high-quality outputs.
|
||||
|
||||
This guide explains what workflows are, why they're powerful, and how to think about designing them.
|
||||
|
||||
## What Is a Workflow?
|
||||
|
||||
A workflow is a structured process where the AI executes steps sequentially to accomplish a task. Each step has a specific purpose, and the AI moves through them methodically—whether that involves extensive collaboration or minimal user interaction.
|
||||
|
||||
Think of it this way: instead of asking "help me build a nutrition plan" and getting a generic response, a workflow guides you (or runs automatically) through discovery, assessment, strategy, shopping lists, and prep schedules—each step building on the last, nothing missed, no shortcuts taken.
|
||||
|
||||
## How do workflows differ from skills?
|
||||
|
||||
Actually they really do not - a workflow can be a skill, and a skill can be a workflow. The main thing with a BMad workflow is the suggestion to follow certain conventions, which actually are also skill best practices. A skill has a few optional and required fields to add as the main file workflow and get stored in a specific location depending on your tool choice for automatic invocation by the llm - whereas workflows are generally intentionally launched, with from another process calling them, or a user invoking via a slash command. In the near future, workflows will optionally be installable as skills also - but if you like, you can add front matter to your custom workflows based on the skill spec from Anthropic, and put them in the proper location your tool dictates.
|
||||
|
||||
### The Power of Progressive Disclosure
|
||||
|
||||
Here's why workflows work so well: the AI only sees the current step. It doesn't know about step 5 when it's on step 2. It can't get ahead of itself, skip steps, or lose focus. Each step gets the AI's full attention, completing fully before the next step loads.
|
||||
|
||||
This is the opposite of a giant prompt that tries to handle everything at once and inevitably misses details or loses coherence.
|
||||
|
||||
Workflows exist on a spectrum:
|
||||
|
||||
- **Interactive workflows** guide users through complex decisions via collaboration and facilitation
|
||||
- **Automated workflows** run with minimal user input, processing documents or executing tasks
|
||||
- **Hybrid workflows** combine both—some steps need user input, others run automatically
|
||||
|
||||
### Real-World Workflow Examples
|
||||
|
||||
**Tax Organizer Workflow**
|
||||
|
||||
A tax preparation workflow that helps users organize financial documents for tax filing. Runs in a single session, follows prescriptive IRS categories, produces a checklist of required documents with missing-item alerts. Sequential and compliance-focused.
|
||||
|
||||
**Meal Planning Workflow**
|
||||
|
||||
Creates personalized weekly meal plans through collaborative nutrition planning. Users can stop mid-session and return later because the workflow tracks progress. Intent-based conversation helps discover preferences rather than following a script. Multi-session, creative, and highly interactive.
|
||||
|
||||
**Course Creator Workflow**
|
||||
|
||||
Helps instructors design course syllabi. Branches based on course type—academic courses need accreditation sections, vocational courses need certification prep, self-paced courses need different structures entirely.
|
||||
|
||||
**Therapy Intake Workflow**
|
||||
|
||||
Guides mental health professionals through structured client intake sessions. Highly sensitive and confidential, uses intent-based questioning to build rapport while ensuring all required clinical information is collected. Continuable across multiple sessions.
|
||||
|
||||
**Software Architecture Workflow** (BMM Module)
|
||||
|
||||
Part of a larger software development pipeline. Runs after product requirements and UX design are complete, takes those documents as input, then collaboratively walks through technical decisions: system components, data flows, technology choices, architectural patterns. Produces an architecture document that implementation teams use to build consistently.
|
||||
|
||||
**Shard Document Workflow**
|
||||
|
||||
Nearly hands-off automated workflow. Takes a large document as input, uses a custom npx tool to split it into smaller files, deletes the original, then augments an index with content details so the LLM can efficiently find and reference specific sections later. Minimal user interaction—just specify the input document.
|
||||
|
||||
These examples show the range: from collaborative creative processes to automated batch jobs, workflows ensure completeness and consistency whether the work involves deep collaboration or minimal human oversight.
|
||||
|
||||
### The Facilitative Philosophy
|
||||
|
||||
When workflows involve users, they should be **facilitative, not directive**. The AI treats users as partners and domain experts, not as passive recipients of generated content.
|
||||
|
||||
**Collaborative dialogue, not command-response**: The AI and user work together throughout. The AI brings structured thinking, methodology, and technical knowledge. The user brings domain expertise, context, and judgment. Together they produce something better than either could alone.
|
||||
|
||||
**The user is the expert in their domain**: A nutrition planning workflow doesn't dictate meal plans—it guides users through discovering what works for their lifestyle. An architecture workflow doesn't tell architects what to build—it facilitates systematic decision-making so choices are explicit and consistent.
|
||||
|
||||
**Intent-based facilitation**: Workflows should describe goals and approaches, not scripts. Instead of "Ask: What is your age? Then ask: What is your goal weight?" use "Guide the user through understanding their health profile. Ask 1-2 questions at a time. Think about their responses before asking follow-ups. Probe to understand their actual needs."
|
||||
|
||||
The AI figures out exact wording and question order based on conversation context. This makes interactions feel natural and responsive rather than robotic and interrogative.
|
||||
|
||||
:::caution[When to Be Prescriptive]
|
||||
Some workflows require exact scripts—medical intake, legal compliance, safety-critical procedures. But these are the exception. Default to facilitative intent-based approaches unless compliance or regulation demands otherwise.
|
||||
:::
|
||||
|
||||
## Why Workflows Matter
|
||||
|
||||
Workflows solve three fundamental problems with AI interactions:
|
||||
|
||||
**Focus**: Each step contains only instructions for that phase. The AI sees one step at a time, preventing it from getting ahead of itself or losing focus.
|
||||
|
||||
**Continuity**: Workflows can span multiple sessions. Stop mid-workflow and return later without losing progress—something free-form prompts can't do.
|
||||
|
||||
**Quality**: Sequential enforcement prevents shortcuts. The AI must complete each step fully before moving on, ensuring thorough, complete outputs instead of rushed, half-baked results.
|
||||
|
||||
## How Workflows Work
|
||||
|
||||
### The Basic Structure
|
||||
|
||||
Workflows consist of multiple markdown files, each representing one step:
|
||||
|
||||
```
|
||||
my-workflow/
|
||||
├── workflow.md # Entry point and configuration
|
||||
├── steps/ # Step files (steps-c/ for create, steps-e/ for edit, steps-v/ for validate)
|
||||
│ ├── step-01-init.md
|
||||
│ ├── step-02-profile.md
|
||||
│ └── step-N-final.md
|
||||
├── data/ # Reference materials, CSVs, examples
|
||||
└── templates/ # Output document templates
|
||||
```
|
||||
|
||||
The `workflow.md` file is minimal—it contains the workflow name, description, goal, the AI's role, and how to start. Importantly, it does not list all steps or detail what each does. This is progressive disclosure in action.
|
||||
|
||||
### Sequential Execution
|
||||
|
||||
Workflows execute in strict sequence: `step-01 → step-02 → step-03 → ... → step-N`
|
||||
|
||||
The AI cannot skip steps or optimize the sequence. It must complete each step fully before loading the next. This ensures thoroughness and prevents shortcuts that compromise quality.
|
||||
|
||||
### Continuable Workflows
|
||||
|
||||
Some workflows are complex enough that users might need multiple sessions. These "continuable workflows" track which steps are complete in the output document's frontmatter, so users can stop and resume later without losing progress.
|
||||
|
||||
Use continuable workflows when:
|
||||
- The workflow produces large documents
|
||||
- Multiple sessions are likely
|
||||
- Complex decisions benefit from reflection
|
||||
- The workflow has many steps (8+)
|
||||
|
||||
Keep it simple (single-session) when tasks are quick, focused, and can be completed in one sitting.
|
||||
|
||||
### Workflow Chaining
|
||||
|
||||
Workflows can be chained together where outputs become inputs. The BMM module pipeline is a perfect example:
|
||||
|
||||
```
|
||||
brainstorming → research → brief → PRD → UX → architecture → epics → sprint-planning
|
||||
↓
|
||||
implement-story → review → repeat
|
||||
```
|
||||
|
||||
Each workflow checks for required inputs from prior workflows, validates they're complete, and produces output for the next workflow. This creates powerful end-to-end pipelines for complex processes.
|
||||
|
||||
### The Tri-Modal Pattern
|
||||
|
||||
For critical workflows that produce important artifacts, BMad uses a tri-modal structure: Create, Validate, and Edit. Each mode is a separate workflow path that can run independently or flow into the others.
|
||||
|
||||
**Create mode** builds new artifacts from scratch. But here's where it gets interesting: create mode can also function as a conversion tool. Feed it a non-compliant document—something that doesn't follow BMad standards—and it will extract the essential content and rebuild it as a compliant artifact. This means you can bring in existing work and automatically upgrade it to follow proper patterns.
|
||||
|
||||
**Validate mode** runs standalone and checks artifacts against standards. Because it's separate, you can run validation whenever you want—immediately after creation, weeks later when things have changed, or even using a different LLM entirely. It's like having a quality assurance checkpoint that's always available but never forced.
|
||||
|
||||
**Edit mode** modifies existing artifacts while enforcing standards. As you update documents to reflect changing requirements or new understanding, edit mode ensures you don't accidentally drift away from the patterns that make the artifacts useful. It checks compliance as you work and can route back to create mode if it detects something that needs full conversion.
|
||||
|
||||
All BMad planning workflows and the BMB module (will) use this tri-modal pattern. The pristine example is the workflow workflow in BMB—it creates workflow specifications, validates them against standards, and lets you edit them while maintaining compliance. You can study that workflow to see the pattern in action.
|
||||
|
||||
This tri-modal approach gives you the best of both worlds: the creativity and flexibility to build what you need, the quality assurance of validation that can run anytime, and the ability to iterate while staying true to standards that make the artifacts valuable across sessions and team members.
|
||||
|
||||
## Design Decisions
|
||||
|
||||
Before building a workflow, answer these questions:
|
||||
|
||||
**Module affiliation**: Is this standalone or part of a module? Module-based workflows can access module-specific variables and reference other workflow outputs. Also when part of a module, generally they will be associated to an agent.
|
||||
|
||||
**Continuable or single-session?**: Will users need multiple sessions, or can this be completed in one sitting?
|
||||
|
||||
**Edit/Validate support?**: Do you need Create/Edit/Validate modes (tri-modal structure)? Use tri-modal for complex, critical workflows requiring quality assurance. Use create-only for simple, one-off workflows.
|
||||
|
||||
**Document output?**: Does this produce a persistent file, or perform actions without output?
|
||||
|
||||
**Intent or prescriptive?**: Is this intent-based facilitation (most workflows) or prescriptive compliance (medical, legal, regulated)?
|
||||
|
||||
## Learning from Examples
|
||||
|
||||
The best way to understand workflows is to study real examples. Look at the official BMad modules:
|
||||
|
||||
- **BMB (Module Builder)**: Workflow and agent creation workflows
|
||||
- **BMM (Business Method Module)**: Complete software development pipeline from brainstorming through sprint planning
|
||||
- **BMGD (Game Development Module)**: Game design briefs, narratives, architecture
|
||||
- **CIS (Creativity, Innovation, Strategy)**: Brainstorming, design thinking, storytelling, innovation strategy
|
||||
|
||||
Study the workflow.md files to understand how each workflow starts. Examine step files to see how instructions are structured. Notice the frontmatter variables, menu handling, and how steps chain together.
|
||||
|
||||
Copy patterns that work. Adapt them to your domain. The structure is consistent across all workflows—the content and steps change, but the architecture stays the same.
|
||||
|
||||
## When to Use Workflows
|
||||
|
||||
Use workflows when:
|
||||
|
||||
- **Tasks are multi-step and complex**: Break down complexity into manageable pieces
|
||||
- **Quality and completeness matter**: Sequential enforcement ensures nothing gets missed
|
||||
- **Repeatability is important**: Get consistent results every time
|
||||
- **Tasks span multiple sessions**: Continuable workflows preserve progress
|
||||
- **You need to chain processes**: Output of one workflow becomes input of another
|
||||
- **Compliance or standards matter**: Enforce required steps and documentation
|
||||
|
||||
Don't use workflows when:
|
||||
|
||||
- **Tasks are simple and one-off**: A single prompt works fine for quick questions
|
||||
- **Flexibility trumps structure**: Free-form conversation is better for exploration
|
||||
|
||||
Modified BMad Workflows
|
||||
|
||||
- **Tasks are truly one-step**
|
||||
|
||||
If there's only one thing to do and it can be explained in under about 300 lines - don't bother with step files. Instead, you can still have
|
||||
a short single file workflow.md file.
|
||||
|
||||
## The Bottom Line
|
||||
|
||||
Workflows transform AI from a tool that gives variable, unpredictable results into a reliable system for complex, multi-step processes. Through progressive disclosure, sequential execution, guided facilitation, and thoughtful design, workflows give you control and repeatability that ad-hoc prompting alone can't match.
|
||||
|
||||
They're not just for software development. You can create workflows for any guided process - meal planning, course design, therapy intake, tax preparation, document processing, creative writing, event planning—any complex task that benefits from structure and thoroughness.
|
||||
|
||||
Start simple. Study examples. Build workflows for your own domain. You'll wonder how you ever got by with just prompts.
|
||||
18
docs/explanation/core/index.md
Normal file
18
docs/explanation/core/index.md
Normal file
@@ -0,0 +1,18 @@
|
||||
---
|
||||
title: "Core Module"
|
||||
---
|
||||
|
||||
|
||||
The Core Module is installed with all installations of BMad modules and provides common functionality that any module, workflow, or agent can take advantage of.
|
||||
|
||||
## Core Module Components
|
||||
|
||||
- **[Global Core Config](/docs/reference/configuration/global-config.md)** — Inheritable configuration that impacts all modules and custom content
|
||||
- **[Core Workflows](/docs/reference/workflows/core-workflows.md)** — Domain-agnostic workflows usable by any module
|
||||
- [Party Mode](/docs/explanation/features/party-mode.md) — Multi-agent conversation orchestration
|
||||
- [Brainstorming](/docs/explanation/features/brainstorming-techniques.md) — Structured creative sessions with 60+ techniques
|
||||
- [Advanced Elicitation](/docs/explanation/features/advanced-elicitation.md) — LLM rethinking with 50+ reasoning methods
|
||||
- **[Core Tasks](/docs/reference/configuration/core-tasks.md)** — Common tasks available across modules
|
||||
- [Index Docs](/docs/reference/configuration/core-tasks.md#index-docs) — Generate directory index files
|
||||
- [Adversarial Review](/docs/reference/configuration/core-tasks.md#adversarial-review) — Critical content review
|
||||
- [Shard Document](/docs/reference/configuration/core-tasks.md#shard-document) — Split large documents into sections
|
||||
@@ -1,16 +1,10 @@
|
||||
# CIS - Creative Intelligence Suite
|
||||
---
|
||||
title: "Creative Intelligence Suite (CIS)"
|
||||
description: AI-powered creative facilitation with the Creative Intelligence Suite
|
||||
---
|
||||
|
||||
AI-powered creative facilitation transforming strategic thinking through expert coaching across five specialized domains.
|
||||
|
||||
## Table of Contents
|
||||
|
||||
- [Core Capabilities](#core-capabilities)
|
||||
- [Specialized Agents](#specialized-agents)
|
||||
- [Interactive Workflows](#interactive-workflows)
|
||||
- [Quick Start](#quick-start)
|
||||
- [Key Differentiators](#key-differentiators)
|
||||
- [Configuration](#configuration)
|
||||
|
||||
## Core Capabilities
|
||||
|
||||
CIS provides structured creative methodologies through distinctive agent personas who act as master facilitators, drawing out insights through strategic questioning rather than generating solutions directly.
|
||||
@@ -25,46 +19,39 @@ CIS provides structured creative methodologies through distinctive agent persona
|
||||
|
||||
## Interactive Workflows
|
||||
|
||||
[View all workflows →](../workflows/README.md)
|
||||
|
||||
**5 Workflows** with **150+ Creative Techniques:**
|
||||
|
||||
### Brainstorming
|
||||
|
||||
36 techniques across 7 categories for ideation
|
||||
|
||||
36 techniques across 7 categories for ideation:
|
||||
- Divergent/convergent thinking
|
||||
- Lateral connections
|
||||
- Forced associations
|
||||
|
||||
### Design Thinking
|
||||
|
||||
Complete 5-phase human-centered process
|
||||
|
||||
Complete 5-phase human-centered process:
|
||||
- Empathize → Define → Ideate → Prototype → Test
|
||||
- User journey mapping
|
||||
- Rapid iteration
|
||||
|
||||
### Problem Solving
|
||||
|
||||
Systematic root cause analysis
|
||||
|
||||
Systematic root cause analysis:
|
||||
- 5 Whys, Fishbone diagrams
|
||||
- Solution generation
|
||||
- Impact assessment
|
||||
|
||||
### Innovation Strategy
|
||||
|
||||
Business model disruption
|
||||
|
||||
Business model disruption:
|
||||
- Blue Ocean Strategy
|
||||
- Jobs-to-be-Done
|
||||
- Disruptive innovation patterns
|
||||
|
||||
### Storytelling
|
||||
|
||||
25 narrative frameworks
|
||||
|
||||
25 narrative frameworks:
|
||||
- Hero's Journey
|
||||
- Story circles
|
||||
- Compelling pitch structures
|
||||
@@ -74,20 +61,16 @@ Business model disruption
|
||||
### Direct Workflow
|
||||
|
||||
```bash
|
||||
# Start interactive session
|
||||
workflow brainstorming
|
||||
|
||||
# With context document
|
||||
workflow design-thinking --data /path/to/context.md
|
||||
```
|
||||
|
||||
### Agent-Facilitated
|
||||
|
||||
```bash
|
||||
# Load agent
|
||||
agent cis/brainstorming-coach
|
||||
|
||||
# Start workflow
|
||||
> *brainstorm
|
||||
```
|
||||
|
||||
@@ -99,31 +82,6 @@ agent cis/brainstorming-coach
|
||||
- **Persona-Driven** - Unique communication styles
|
||||
- **Rich Method Libraries** - 150+ proven techniques
|
||||
|
||||
## Configuration
|
||||
|
||||
Edit `/_bmad/cis/config.yaml`:
|
||||
|
||||
```yaml
|
||||
output_folder: ./creative-outputs
|
||||
user_name: Your Name
|
||||
communication_language: english
|
||||
```
|
||||
|
||||
## Module Structure
|
||||
|
||||
```
|
||||
cis/
|
||||
├── agents/ # 5 specialized facilitators
|
||||
├── workflows/ # 5 interactive processes
|
||||
│ ├── brainstorming/
|
||||
│ ├── design-thinking/
|
||||
│ ├── innovation-strategy/
|
||||
│ ├── problem-solving/
|
||||
│ └── storytelling/
|
||||
├── tasks/ # Supporting operations
|
||||
└── teams/ # Agent collaborations
|
||||
```
|
||||
|
||||
## Integration Points
|
||||
|
||||
CIS workflows integrate with:
|
||||
@@ -140,10 +98,6 @@ CIS workflows integrate with:
|
||||
4. **Take breaks** when energy flags
|
||||
5. **Document insights** as they emerge
|
||||
|
||||
## Related Documentation
|
||||
|
||||
- **[BMM Documentation](../../bmm/docs/index.md)** - Core BMad Method documentation
|
||||
|
||||
---
|
||||
|
||||
Part of BMad Method v6.0 - Transform creative potential through expert AI facilitation.
|
||||
:::tip[Learn More]
|
||||
See [Facilitation Over Generation](/docs/explanation/philosophy/facilitation-over-generation.md) for the core philosophy behind CIS.
|
||||
:::
|
||||
73
docs/explanation/faq/brownfield-faq.md
Normal file
73
docs/explanation/faq/brownfield-faq.md
Normal file
@@ -0,0 +1,73 @@
|
||||
---
|
||||
title: "Brownfield Development FAQ"
|
||||
description: Common questions about brownfield development in the BMad Method
|
||||
---
|
||||
Quick answers to common questions about brownfield (existing codebase) development in the BMad Method (BMM).
|
||||
|
||||
## Questions
|
||||
|
||||
- [What is brownfield vs greenfield?](#what-is-brownfield-vs-greenfield)
|
||||
- [Do I have to run document-project for brownfield?](#do-i-have-to-run-document-project-for-brownfield)
|
||||
- [What if I forget to run document-project?](#what-if-i-forget-to-run-document-project)
|
||||
- [Can I use Quick Spec Flow for brownfield projects?](#can-i-use-quick-spec-flow-for-brownfield-projects)
|
||||
- [How does workflow-init handle old planning docs?](#how-does-workflow-init-handle-old-planning-docs)
|
||||
- [What if my existing code doesn't follow best practices?](#what-if-my-existing-code-doesnt-follow-best-practices)
|
||||
|
||||
### What is brownfield vs greenfield?
|
||||
|
||||
- **Greenfield** — New project, starting from scratch, clean slate
|
||||
- **Brownfield** — Existing project, working with established codebase and patterns
|
||||
|
||||
### Do I have to run document-project for brownfield?
|
||||
|
||||
Highly recommended, especially if:
|
||||
|
||||
- No existing documentation
|
||||
- Documentation is outdated
|
||||
- AI agents need context about existing code
|
||||
- Level 2-4 complexity
|
||||
|
||||
You can skip it if you have comprehensive, up-to-date documentation including `docs/index.md`.
|
||||
|
||||
### What if I forget to run document-project?
|
||||
|
||||
Workflows will lack context about existing code. You may get:
|
||||
|
||||
- Suggestions that don't match existing patterns
|
||||
- Integration approaches that miss existing APIs
|
||||
- Architecture that conflicts with current structure
|
||||
|
||||
Run document-project and restart planning with proper context.
|
||||
|
||||
### Can I use Quick Spec Flow for brownfield projects?
|
||||
|
||||
Yes! Quick Spec Flow works great for brownfield. It will:
|
||||
|
||||
- Auto-detect your existing stack
|
||||
- Analyze brownfield code patterns
|
||||
- Detect conventions and ask for confirmation
|
||||
- Generate context-rich tech-spec that respects existing code
|
||||
|
||||
Perfect for bug fixes and small features in existing codebases.
|
||||
|
||||
### How does workflow-init handle old planning docs?
|
||||
|
||||
workflow-init asks about YOUR current work first, then uses old artifacts as context:
|
||||
|
||||
1. Shows what it found (old PRD, epics, etc.)
|
||||
2. Asks: "Is this work in progress, previous effort, or proposed work?"
|
||||
3. If previous effort: Asks you to describe your NEW work
|
||||
4. Determines level based on YOUR work, not old artifacts
|
||||
|
||||
This prevents old Level 3 PRDs from forcing Level 3 workflow for a new Level 0 bug fix.
|
||||
|
||||
### What if my existing code doesn't follow best practices?
|
||||
|
||||
Quick Spec Flow detects your conventions and asks: "Should I follow these existing conventions?" You decide:
|
||||
|
||||
- **Yes** → Maintain consistency with current codebase
|
||||
- **No** → Establish new standards (document why in tech-spec)
|
||||
|
||||
BMM respects your choice — it won't force modernization, but it will offer it.
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
67
docs/explanation/faq/getting-started-faq.md
Normal file
67
docs/explanation/faq/getting-started-faq.md
Normal file
@@ -0,0 +1,67 @@
|
||||
---
|
||||
title: "Getting Started FAQ"
|
||||
description: Common questions about getting started with the BMad Method
|
||||
---
|
||||
|
||||
Quick answers to common questions about getting started with the BMad Method.
|
||||
|
||||
## Questions
|
||||
|
||||
- [Why does BMad use so many tokens?](#why-does-bmad-use-so-many-tokens)
|
||||
- [Do I always need to run workflow-init?](#do-i-always-need-to-run-workflow-init)
|
||||
- [Why do I need fresh chats for each workflow?](#why-do-i-need-fresh-chats-for-each-workflow)
|
||||
- [Can I skip workflow-status and just start working?](#can-i-skip-workflow-status-and-just-start-working)
|
||||
- [What's the minimum I need to get started?](#whats-the-minimum-i-need-to-get-started)
|
||||
- [How do I know if I'm in Phase 1, 2, 3, or 4?](#how-do-i-know-if-im-in-phase-1-2-3-or-4)
|
||||
|
||||
### Why does BMad use so many tokens?
|
||||
|
||||
BMad is not always the most token efficient approach, and that's by design. The checkpoints, story files, and retrospectives keep you in the loop so you can apply taste, judgment, and accumulated context that no agent has. Fully automated coding loops optimize for code velocity; BMad optimizes for decision quality. If you're building something you'll maintain for years, where user experience matters, where architectural choices compound—that tradeoff pays for itself.
|
||||
|
||||
### Do I always need to run workflow-init?
|
||||
|
||||
No, once you learn the flow you can go directly to workflows. However, workflow-init is helpful because it:
|
||||
|
||||
- Determines your project's appropriate level automatically
|
||||
- Creates the tracking status file
|
||||
- Routes you to the correct starting workflow
|
||||
|
||||
For experienced users: use the [Quick Start Guide](/docs/tutorials/getting-started/getting-started-bmadv6.md) to go directly to the right agent/workflow.
|
||||
|
||||
### Why do I need fresh chats for each workflow?
|
||||
|
||||
Context-intensive workflows (like brainstorming, PRD creation, architecture design) can cause AI hallucinations if run in sequence within the same chat. Starting fresh ensures the agent has maximum context capacity for each workflow. This is particularly important for:
|
||||
|
||||
- Planning workflows (PRD, architecture)
|
||||
- Analysis workflows (brainstorming, research)
|
||||
- Complex story implementation
|
||||
|
||||
Quick workflows like status checks can reuse chats safely.
|
||||
|
||||
### Can I skip workflow-status and just start working?
|
||||
|
||||
Yes, if you already know your project level and which workflow comes next. workflow-status is mainly useful for:
|
||||
|
||||
- New projects (guides initial setup)
|
||||
- When you're unsure what to do next
|
||||
- After breaks in work (reminds you where you left off)
|
||||
- Checking overall progress
|
||||
|
||||
### What's the minimum I need to get started?
|
||||
|
||||
For the fastest path:
|
||||
|
||||
1. Install BMad Method: `npx bmad-method@alpha install`
|
||||
2. For small changes: Load PM agent → run tech-spec → implement
|
||||
3. For larger projects: Load PM agent → run prd → architect → implement
|
||||
|
||||
### How do I know if I'm in Phase 1, 2, 3, or 4?
|
||||
|
||||
Check your `bmm-workflow-status.md` file (created by workflow-init). It shows your current phase and progress. If you don't have this file, you can also tell by what you're working on:
|
||||
|
||||
- **Phase 1** — Brainstorming, research, product brief (optional)
|
||||
- **Phase 2** — Creating either a PRD or tech-spec (always required)
|
||||
- **Phase 3** — Architecture design (Level 2-4 only)
|
||||
- **Phase 4** — Actually writing code, implementing stories
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
52
docs/explanation/faq/implementation-faq.md
Normal file
52
docs/explanation/faq/implementation-faq.md
Normal file
@@ -0,0 +1,52 @@
|
||||
---
|
||||
title: "Implementation FAQ"
|
||||
description: Common questions about implementation in the BMad Method
|
||||
---
|
||||
|
||||
Quick answers to common questions about implementation in the BMad Method.
|
||||
|
||||
## Questions
|
||||
|
||||
- [Does create-story include implementation context?](#does-create-story-include-implementation-context)
|
||||
- [How do I mark a story as done?](#how-do-i-mark-a-story-as-done)
|
||||
- [Can I work on multiple stories at once?](#can-i-work-on-multiple-stories-at-once)
|
||||
- [What if my story takes longer than estimated?](#what-if-my-story-takes-longer-than-estimated)
|
||||
- [When should I run retrospective?](#when-should-i-run-retrospective)
|
||||
|
||||
### Does create-story include implementation context?
|
||||
|
||||
Yes! The create-story workflow generates story files that include implementation-specific guidance, references existing patterns from your documentation, and provides technical context. The workflow loads your architecture, PRD, and existing project documentation to create comprehensive stories. For Quick Flow projects using tech-spec, the tech-spec itself is already comprehensive, so stories can be simpler.
|
||||
|
||||
### How do I mark a story as done?
|
||||
|
||||
After dev-story completes and code-review passes:
|
||||
|
||||
1. Open `sprint-status.yaml` (created by sprint-planning)
|
||||
2. Change the story status from `review` to `done`
|
||||
3. Save the file
|
||||
|
||||
### Can I work on multiple stories at once?
|
||||
|
||||
Yes, if you have capacity! Stories within different epics can be worked in parallel. However, stories within the same epic are usually sequential because they build on each other.
|
||||
|
||||
### What if my story takes longer than estimated?
|
||||
|
||||
That's normal! Stories are estimates. If implementation reveals more complexity:
|
||||
|
||||
1. Continue working until DoD is met
|
||||
2. Consider if story should be split
|
||||
3. Document learnings in retrospective
|
||||
4. Adjust future estimates based on this learning
|
||||
|
||||
### When should I run retrospective?
|
||||
|
||||
After completing all stories in an epic (when epic is done). Retrospectives capture:
|
||||
|
||||
- What went well
|
||||
- What could improve
|
||||
- Technical insights
|
||||
- Learnings for future epics
|
||||
|
||||
Don't wait until project end — run after each epic for continuous improvement.
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
16
docs/explanation/faq/index.md
Normal file
16
docs/explanation/faq/index.md
Normal file
@@ -0,0 +1,16 @@
|
||||
---
|
||||
title: "Frequently Asked Questions"
|
||||
description: Frequently asked questions about the BMad Method
|
||||
---
|
||||
|
||||
Quick answers to common questions about the BMad Method, organized by topic.
|
||||
|
||||
## Topics
|
||||
|
||||
- [Getting Started](/docs/explanation/faq/getting-started-faq.md) - Questions about starting with BMad
|
||||
- [Levels & Tracks](/docs/explanation/faq/levels-and-tracks-faq.md) - Choosing the right level
|
||||
- [Workflows](/docs/explanation/faq/workflows-faq.md) - Workflow and phase questions
|
||||
- [Planning](/docs/explanation/faq/planning-faq.md) - Planning document questions
|
||||
- [Implementation](/docs/explanation/faq/implementation-faq.md) - Implementation questions
|
||||
- [Brownfield](/docs/explanation/faq/brownfield-faq.md) - Existing codebase questions
|
||||
- [Tools & Advanced](/docs/explanation/faq/tools-faq.md) - Tools, IDEs, and advanced topics
|
||||
52
docs/explanation/faq/levels-and-tracks-faq.md
Normal file
52
docs/explanation/faq/levels-and-tracks-faq.md
Normal file
@@ -0,0 +1,52 @@
|
||||
---
|
||||
title: "Levels and Tracks FAQ"
|
||||
description: Common questions about choosing the right level for your project
|
||||
---
|
||||
|
||||
Quick answers to common questions about choosing the right level for your BMad Method project.
|
||||
|
||||
## Questions
|
||||
|
||||
- [How do I know which level my project is?](#how-do-i-know-which-level-my-project-is)
|
||||
- [Can I change levels mid-project?](#can-i-change-levels-mid-project)
|
||||
- [What if workflow-init suggests the wrong level?](#what-if-workflow-init-suggests-the-wrong-level)
|
||||
- [Do I always need architecture for Level 2?](#do-i-always-need-architecture-for-level-2)
|
||||
- [What's the difference between Level 1 and Level 2?](#whats-the-difference-between-level-1-and-level-2)
|
||||
|
||||
### How do I know which level my project is?
|
||||
|
||||
Use workflow-init for automatic detection, or self-assess using these keywords:
|
||||
|
||||
- **Level 0** — "fix", "bug", "typo", "small change", "patch" → 1 story
|
||||
- **Level 1** — "simple", "basic", "small feature", "add" → 1-10 stories
|
||||
- **Level 2** — "dashboard", "several features", "admin panel" → 5-15 stories
|
||||
- **Level 3** — "platform", "integration", "complex", "system" → 12-40 stories
|
||||
- **Level 4** — "enterprise", "multi-tenant", "multiple products" → 40+ stories
|
||||
|
||||
When in doubt, start smaller. You can always run create-prd later if needed.
|
||||
|
||||
### Can I change levels mid-project?
|
||||
|
||||
Yes! If you started at Level 1 but realize it's Level 2, you can run create-prd to add proper planning docs. The system is flexible — your initial level choice isn't permanent.
|
||||
|
||||
### What if workflow-init suggests the wrong level?
|
||||
|
||||
You can override it! workflow-init suggests a level but always asks for confirmation. If you disagree, just say so and choose the level you think is appropriate. Trust your judgment.
|
||||
|
||||
### Do I always need architecture for Level 2?
|
||||
|
||||
No, architecture is **optional** for Level 2. Only create architecture if you need system-level design. Many Level 2 projects work fine with just PRD created during planning.
|
||||
|
||||
### What's the difference between Level 1 and Level 2?
|
||||
|
||||
- **Level 1** — 1-10 stories, uses tech-spec (simpler, faster), no architecture
|
||||
- **Level 2** — 5-15 stories, uses PRD (product-focused), optional architecture
|
||||
|
||||
The overlap (5-10 stories) is intentional. Choose based on:
|
||||
|
||||
- Need product-level planning? → Level 2
|
||||
- Just need technical plan? → Level 1
|
||||
- Multiple epics? → Level 2
|
||||
- Single epic? → Level 1
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
41
docs/explanation/faq/planning-faq.md
Normal file
41
docs/explanation/faq/planning-faq.md
Normal file
@@ -0,0 +1,41 @@
|
||||
---
|
||||
title: "Planning Documents FAQ"
|
||||
description: Common questions about planning documents in the BMad Method
|
||||
---
|
||||
|
||||
Quick answers to common questions about planning documents in the BMad Method.
|
||||
|
||||
## Questions
|
||||
|
||||
- [Why no tech-spec at Level 2+?](#why-no-tech-spec-at-level-2)
|
||||
- [Do I need a PRD for a bug fix?](#do-i-need-a-prd-for-a-bug-fix)
|
||||
- [Can I skip the product brief?](#can-i-skip-the-product-brief)
|
||||
|
||||
### Why no tech-spec at Level 2+?
|
||||
|
||||
Level 2+ projects need product-level planning (PRD) and system-level design (Architecture), which tech-spec doesn't provide. Tech-spec is too narrow for coordinating multiple features. Instead, Level 2-4 uses:
|
||||
|
||||
- PRD (product vision, functional requirements, non-functional requirements)
|
||||
- Architecture (system design)
|
||||
- Epics+Stories (created AFTER architecture is complete)
|
||||
|
||||
### Do I need a PRD for a bug fix?
|
||||
|
||||
No! Bug fixes are typically Level 0 (single atomic change). Use Quick Spec Flow:
|
||||
|
||||
- Load PM agent
|
||||
- Run tech-spec workflow
|
||||
- Implement immediately
|
||||
|
||||
PRDs are for Level 2-4 projects with multiple features requiring product-level coordination.
|
||||
|
||||
### Can I skip the product brief?
|
||||
|
||||
Yes, product brief is always optional. It's most valuable for:
|
||||
|
||||
- Level 3-4 projects needing strategic direction
|
||||
- Projects with stakeholders requiring alignment
|
||||
- Novel products needing market research
|
||||
- When you want to explore solution space before committing
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
253
docs/explanation/faq/tools-faq.md
Normal file
253
docs/explanation/faq/tools-faq.md
Normal file
@@ -0,0 +1,253 @@
|
||||
---
|
||||
title: "Tools and Advanced FAQ"
|
||||
description: Common questions about tools, IDEs, and advanced topics in the BMad Method
|
||||
---
|
||||
|
||||
Quick answers to common questions about tools, IDEs, and advanced topics in the BMad Method.
|
||||
|
||||
## Questions
|
||||
|
||||
**Tools and Technical**
|
||||
|
||||
- [Why are my Mermaid diagrams not rendering?](#why-are-my-mermaid-diagrams-not-rendering)
|
||||
- [Can I use BMM with GitHub Copilot / Cursor / other AI tools?](#can-i-use-bmm-with-github-copilot--cursor--other-ai-tools)
|
||||
- [What IDEs/tools support BMM?](#what-idestools-support-bmm)
|
||||
- [Can I customize agents?](#can-i-customize-agents)
|
||||
- [What happens to my planning docs after implementation?](#what-happens-to-my-planning-docs-after-implementation)
|
||||
- [Can I use BMM for non-software projects?](#can-i-use-bmm-for-non-software-projects)
|
||||
|
||||
**Advanced**
|
||||
|
||||
- [What if my project grows from Level 1 to Level 3?](#what-if-my-project-grows-from-level-1-to-level-3)
|
||||
- [Can I mix greenfield and brownfield approaches?](#can-i-mix-greenfield-and-brownfield-approaches)
|
||||
- [How do I handle urgent hotfixes during a sprint?](#how-do-i-handle-urgent-hotfixes-during-a-sprint)
|
||||
- [What if I disagree with the workflow's recommendations?](#what-if-i-disagree-with-the-workflows-recommendations)
|
||||
- [Can multiple developers work on the same BMM project?](#can-multiple-developers-work-on-the-same-bmm-project)
|
||||
- [What is party mode and when should I use it?](#what-is-party-mode-and-when-should-i-use-it)
|
||||
|
||||
**Getting Help**
|
||||
|
||||
- [Where do I get help if my question isn't answered here?](#where-do-i-get-help-if-my-question-isnt-answered-here)
|
||||
- [How do I report a bug or request a feature?](#how-do-i-report-a-bug-or-request-a-feature)
|
||||
|
||||
## Tools and Technical
|
||||
|
||||
### Why are my Mermaid diagrams not rendering?
|
||||
|
||||
Common issues:
|
||||
|
||||
1. Missing language tag: Use ` ```mermaid` not just ` ``` `
|
||||
2. Syntax errors in diagram (validate at mermaid.live)
|
||||
3. Tool doesn't support Mermaid (check your Markdown renderer)
|
||||
|
||||
All BMM docs use valid Mermaid syntax that should render in GitHub, VS Code, and most IDEs.
|
||||
|
||||
### Can I use BMM with GitHub Copilot / Cursor / other AI tools?
|
||||
|
||||
Yes! BMM is complementary. BMM handles:
|
||||
|
||||
- Project planning and structure
|
||||
- Workflow orchestration
|
||||
- Agent Personas and expertise
|
||||
- Documentation generation
|
||||
- Quality gates
|
||||
|
||||
Your AI coding assistant handles:
|
||||
|
||||
- Line-by-line code completion
|
||||
- Quick refactoring
|
||||
- Test generation
|
||||
|
||||
Use them together for best results.
|
||||
|
||||
### What IDEs/tools support BMM?
|
||||
|
||||
BMM requires tools with **agent mode** and access to **high-quality LLM models** that can load and follow complex workflows, then properly implement code changes.
|
||||
|
||||
**Recommended Tools:**
|
||||
|
||||
- **Claude Code** — Best choice
|
||||
- Sonnet 4.5 (excellent workflow following, coding, reasoning)
|
||||
- Opus (maximum context, complex planning)
|
||||
- Native agent mode designed for BMM workflows
|
||||
|
||||
- **Cursor**
|
||||
- Supports Anthropic (Claude) and OpenAI models
|
||||
- Agent mode with composer
|
||||
- Good for developers who prefer Cursor's UX
|
||||
|
||||
- **Windsurf**
|
||||
- Multi-model support
|
||||
- Agent capabilities
|
||||
- Suitable for BMM workflows
|
||||
|
||||
**What Matters:**
|
||||
|
||||
1. **Agent mode** — Can load long workflow instructions and maintain context
|
||||
2. **High-quality LLM** — Models ranked high on SWE-bench (coding benchmarks)
|
||||
3. **Model selection** — Access to Claude Sonnet 4.5, Opus, or GPT-4o class models
|
||||
4. **Context capacity** — Can handle large planning documents and codebases
|
||||
|
||||
**Why model quality matters:** BMM workflows require LLMs that can follow multi-step processes, maintain context across phases, and implement code that adheres to specifications. Tools with weaker models will struggle with workflow adherence and code quality.
|
||||
|
||||
### Can I customize agents?
|
||||
|
||||
Yes! Agents are installed as markdown files with XML-style content (optimized for LLMs, readable by any model). Create customization files in `_bmad/_config/agents/[agent-name].customize.yaml` to override default behaviors while keeping core functionality intact. See agent documentation for customization options.
|
||||
|
||||
**Note:** While source agents in this repo are YAML, they install as `.md` files with XML-style tags — a format any LLM can read and follow.
|
||||
|
||||
### What happens to my planning docs after implementation?
|
||||
|
||||
Keep them! They serve as:
|
||||
|
||||
- Historical record of decisions
|
||||
- Onboarding material for new team members
|
||||
- Reference for future enhancements
|
||||
- Audit trail for compliance
|
||||
|
||||
For enterprise projects (Level 4), consider archiving completed planning artifacts to keep workspace clean.
|
||||
|
||||
### Can I use BMM for non-software projects?
|
||||
|
||||
BMM is optimized for software development, but the methodology principles (scale-adaptive planning, just-in-time design, context injection) can apply to other complex project types. You'd need to adapt workflows and agents for your domain.
|
||||
|
||||
## Advanced
|
||||
|
||||
### What if my project grows from Level 1 to Level 3?
|
||||
|
||||
Totally fine! When you realize scope has grown:
|
||||
|
||||
1. Run create-prd to add product-level planning
|
||||
2. Run create-architecture for system design
|
||||
3. Use existing tech-spec as input for PRD
|
||||
4. Continue with updated level
|
||||
|
||||
The system is flexible — growth is expected.
|
||||
|
||||
### Can I mix greenfield and brownfield approaches?
|
||||
|
||||
Yes! Common scenario: adding new greenfield feature to brownfield codebase. Approach:
|
||||
|
||||
1. Run document-project for brownfield context
|
||||
2. Use greenfield workflows for new feature planning
|
||||
3. Explicitly document integration points between new and existing
|
||||
4. Test integration thoroughly
|
||||
|
||||
### How do I handle urgent hotfixes during a sprint?
|
||||
|
||||
Use correct-course workflow or just:
|
||||
|
||||
1. Save your current work state
|
||||
2. Load PM agent → quick tech-spec for hotfix
|
||||
3. Implement hotfix (Level 0 flow)
|
||||
4. Deploy hotfix
|
||||
5. Return to original sprint work
|
||||
|
||||
Level 0 Quick Spec Flow is perfect for urgent fixes.
|
||||
|
||||
### What if I disagree with the workflow's recommendations?
|
||||
|
||||
Workflows are guidance, not enforcement. If a workflow recommends something that doesn't make sense for your context:
|
||||
|
||||
- Explain your reasoning to the agent
|
||||
- Ask for alternative approaches
|
||||
- Skip the recommendation if you're confident
|
||||
- Document why you deviated (for future reference)
|
||||
|
||||
Trust your expertise — BMM supports your decisions.
|
||||
|
||||
### Can multiple developers work on the same BMM project?
|
||||
|
||||
Yes! But the paradigm is fundamentally different from traditional agile teams.
|
||||
|
||||
**Key Difference:**
|
||||
|
||||
- **Traditional** — Multiple devs work on stories within one epic (months)
|
||||
- **Agentic** — Each dev owns complete epics (days)
|
||||
|
||||
**In traditional agile:** A team of 5 devs might spend 2-3 months on a single epic, with each dev owning different stories.
|
||||
|
||||
**With BMM + AI agents:** A single dev can complete an entire epic in 1-3 days. What used to take months now takes days.
|
||||
|
||||
**Team Work Distribution:**
|
||||
|
||||
- **Recommended:** Split work by **epic** (not story)
|
||||
- Each developer owns complete epics end-to-end
|
||||
- Parallel work happens at epic level
|
||||
- Minimal coordination needed
|
||||
|
||||
**For full-stack apps:**
|
||||
|
||||
- Frontend and backend can be separate epics (unusual in traditional agile)
|
||||
- Frontend dev owns all frontend epics
|
||||
- Backend dev owns all backend epics
|
||||
- Works because delivery is so fast
|
||||
|
||||
**Enterprise Considerations:**
|
||||
|
||||
- Use **git submodules** for BMM installation (not .gitignore)
|
||||
- Allows personal configurations without polluting main repo
|
||||
- Teams may use different AI tools (Claude Code, Cursor, etc.)
|
||||
- Developers may follow different methods or create custom agents/workflows
|
||||
|
||||
**Quick Tips:**
|
||||
|
||||
- Share `sprint-status.yaml` (single source of truth)
|
||||
- Assign entire epics to developers (not individual stories)
|
||||
- Coordinate at epic boundaries, not story level
|
||||
- Use git submodules for BMM in enterprise settings
|
||||
|
||||
### What is party mode and when should I use it?
|
||||
|
||||
Party mode is a unique multi-agent collaboration feature where ALL your installed agents (19+ from BMM, CIS, BMB, custom modules) discuss your challenges together in real-time.
|
||||
|
||||
**How it works:**
|
||||
|
||||
1. Run `/bmad:core:workflows:party-mode` (or `*party-mode` from any agent)
|
||||
2. Introduce your topic
|
||||
3. BMad Master selects 2-3 most relevant agents per message
|
||||
4. Agents cross-talk, debate, and build on each other's ideas
|
||||
|
||||
**Best for:**
|
||||
|
||||
- Strategic decisions with trade-offs (architecture choices, tech stack, scope)
|
||||
- Creative brainstorming (game design, product innovation, UX ideation)
|
||||
- Cross-functional alignment (epic kickoffs, retrospectives, phase transitions)
|
||||
- Complex problem-solving (multi-faceted challenges, risk assessment)
|
||||
|
||||
**Example parties:**
|
||||
|
||||
- **Product Strategy** — PM + Innovation Strategist (CIS) + Analyst
|
||||
- **Technical Design** — Architect + Creative Problem Solver (CIS) + Game Architect
|
||||
- **User Experience** — UX Designer + Design Thinking Coach (CIS) + Storyteller (CIS)
|
||||
|
||||
**Why it's powerful:**
|
||||
|
||||
- Diverse perspectives (technical, creative, strategic)
|
||||
- Healthy debate reveals blind spots
|
||||
- Emergent insights from agent interaction
|
||||
- Natural collaboration across modules
|
||||
|
||||
**For complete documentation:** See the [Party Mode Guide](/docs/explanation/features/party-mode.md)
|
||||
|
||||
## Getting Help
|
||||
|
||||
### Where do I get help if my question isn't answered here?
|
||||
|
||||
1. Search [Complete Documentation](https://github.com/bmad-code-org/BMAD-METHOD/blob/main/README.md) for related topics
|
||||
2. Ask in [Discord Community](https://discord.gg/gk8jAdXWmj) (#bmad-method-help)
|
||||
3. Open a [GitHub Issue](https://github.com/bmad-code-org/BMAD-METHOD/issues)
|
||||
4. Watch [YouTube Tutorials](https://www.youtube.com/@BMadCode)
|
||||
|
||||
### How do I report a bug or request a feature?
|
||||
|
||||
Open a GitHub issue at: <https://github.com/bmad-code-org/BMAD-METHOD/issues>
|
||||
|
||||
Please include:
|
||||
|
||||
- BMM version (check your installed version)
|
||||
- Steps to reproduce (for bugs)
|
||||
- Expected vs actual behavior
|
||||
- Relevant workflow or agent involved
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
61
docs/explanation/faq/workflows-faq.md
Normal file
61
docs/explanation/faq/workflows-faq.md
Normal file
@@ -0,0 +1,61 @@
|
||||
---
|
||||
title: "Workflows FAQ"
|
||||
description: Common questions about BMad Method workflows and phases
|
||||
---
|
||||
|
||||
Quick answers to common questions about BMad Method workflows and phases.
|
||||
|
||||
## Questions
|
||||
|
||||
- [What's the difference between workflow-status and workflow-init?](#whats-the-difference-between-workflow-status-and-workflow-init)
|
||||
- [Can I skip Phase 1 (Analysis)?](#can-i-skip-phase-1-analysis)
|
||||
- [When is Phase 3 (Architecture) required?](#when-is-phase-3-architecture-required)
|
||||
- [What happens if I skip a recommended workflow?](#what-happens-if-i-skip-a-recommended-workflow)
|
||||
- [How do I know when Phase 3 is complete?](#how-do-i-know-when-phase-3-is-complete)
|
||||
- [Can I run workflows in parallel?](#can-i-run-workflows-in-parallel)
|
||||
|
||||
### What's the difference between workflow-status and workflow-init?
|
||||
|
||||
- **workflow-status** — Checks existing status and tells you what's next (use when continuing work)
|
||||
- **workflow-init** — Creates new status file and sets up project (use when starting new project)
|
||||
|
||||
If status file exists, use workflow-status. If not, use workflow-init.
|
||||
|
||||
### Can I skip Phase 1 (Analysis)?
|
||||
|
||||
Yes! Phase 1 is optional for all levels, though recommended for complex projects. Skip if:
|
||||
|
||||
- Requirements are clear
|
||||
- No research needed
|
||||
- Time-sensitive work
|
||||
- Small changes (Level 0-1)
|
||||
|
||||
### When is Phase 3 (Architecture) required?
|
||||
|
||||
- **Level 0-1** — Never (skip entirely)
|
||||
- **Level 2** — Optional (only if system design needed)
|
||||
- **Level 3-4** — Required (comprehensive architecture mandatory)
|
||||
|
||||
### What happens if I skip a recommended workflow?
|
||||
|
||||
Nothing breaks! Workflows are guidance, not enforcement. However, skipping recommended workflows (like architecture for Level 3) may cause:
|
||||
|
||||
- Integration issues during implementation
|
||||
- Rework due to poor planning
|
||||
- Conflicting design decisions
|
||||
- Longer development time overall
|
||||
|
||||
### How do I know when Phase 3 is complete?
|
||||
|
||||
For Level 3-4, run the implementation-readiness workflow. It validates PRD + Architecture + Epics + UX (optional) are aligned before implementation. Pass the gate check = ready for Phase 4.
|
||||
|
||||
### Can I run workflows in parallel?
|
||||
|
||||
Most workflows must be sequential within a phase:
|
||||
|
||||
- **Phase 1** — brainstorm → research → product-brief (optional order)
|
||||
- **Phase 2** — PRD must complete before moving forward
|
||||
- **Phase 3** — architecture → epics+stories → implementation-readiness (sequential)
|
||||
- **Phase 4** — Stories within an epic should generally be sequential, but stories in different epics can be parallel if you have capacity
|
||||
|
||||
**Have a question not answered here?** Please [open an issue](https://github.com/bmad-code-org/BMAD-METHOD/issues) or ask in [Discord](https://discord.gg/gk8jAdXWmj) so we can add it!
|
||||
95
docs/explanation/features/advanced-elicitation.md
Normal file
95
docs/explanation/features/advanced-elicitation.md
Normal file
@@ -0,0 +1,95 @@
|
||||
---
|
||||
title: "Advanced Elicitation"
|
||||
---
|
||||
|
||||
Push the LLM to rethink its work through 50+ reasoning methods — essentially, LLM brainstorming.
|
||||
|
||||
Advanced Elicitation is the inverse of Brainstorming. Instead of pulling ideas out of you, the LLM applies sophisticated reasoning techniques to re-examine and enhance content it has just generated. It's the LLM brainstorming with itself to find better approaches, uncover hidden issues, and discover improvements it missed on the first pass.
|
||||
|
||||
## When to Use It
|
||||
|
||||
- After a workflow generates a section of content and you want to explore alternatives
|
||||
- When the LLM's initial output seems adequate but you suspect there's more depth available
|
||||
- For high-stakes content where multiple perspectives would strengthen the result
|
||||
- To stress-test assumptions, explore edge cases, or find weaknesses in generated plans
|
||||
- When you want the LLM to "think again" but with structured reasoning methods
|
||||
|
||||
## How It Works
|
||||
|
||||
### 1. Context Analysis
|
||||
The LLM analyzes the current content, understanding its type, complexity, stakeholder needs, risk level, and creative potential.
|
||||
|
||||
### 2. Smart Method Selection
|
||||
Based on context, 5 methods are intelligently selected from a library of 50+ techniques and presented to you:
|
||||
|
||||
| Option | Description |
|
||||
| ----------------- | ---------------------------------------- |
|
||||
| **1-5** | Apply the selected method to the content |
|
||||
| **[r] Reshuffle** | Get 5 new methods selected randomly |
|
||||
| **[a] List All** | Browse the complete method library |
|
||||
| **[x] Proceed** | Continue with enhanced content |
|
||||
|
||||
### 3. Method Execution & Iteration
|
||||
- The selected method is applied to the current content
|
||||
- Improvements are shown for your review
|
||||
- You choose whether to apply changes or discard them
|
||||
- The menu re-appears for additional elicitations
|
||||
- Each method builds on previous enhancements
|
||||
|
||||
### 4. Party Mode Integration (Optional)
|
||||
If Party Mode is active, BMad agents participate randomly in the elicitation process, adding their unique perspectives to the methods.
|
||||
|
||||
## Method Categories
|
||||
|
||||
| Category | Focus | Example Methods |
|
||||
| ----------------- | ----------------------------------- | -------------------------------------------------------------- |
|
||||
| **Core** | Foundational reasoning techniques | First Principles Analysis, 5 Whys, Socratic Questioning |
|
||||
| **Collaboration** | Multiple perspectives and synthesis | Stakeholder Round Table, Expert Panel Review, Debate Club |
|
||||
| **Advanced** | Complex reasoning frameworks | Tree of Thoughts, Graph of Thoughts, Self-Consistency |
|
||||
| **Competitive** | Adversarial stress-testing | Red Team vs Blue Team, Shark Tank Pitch, Code Review Gauntlet |
|
||||
| **Technical** | Architecture and code quality | Decision Records, Rubber Duck Debugging, Algorithm Olympics |
|
||||
| **Creative** | Innovation and lateral thinking | SCAMPER, Reverse Engineering, Random Input Stimulus |
|
||||
| **Research** | Evidence-based analysis | Literature Review Personas, Thesis Defense, Comparative Matrix |
|
||||
| **Risk** | Risk identification and mitigation | Pre-mortem Analysis, Failure Mode Analysis, Chaos Monkey |
|
||||
| **Learning** | Understanding verification | Feynman Technique, Active Recall Testing |
|
||||
| **Philosophical** | Conceptual clarity | Occam's Razor, Ethical Dilemmas |
|
||||
| **Retrospective** | Reflection and lessons | Hindsight Reflection, Lessons Learned Extraction |
|
||||
|
||||
## Key Features
|
||||
|
||||
- **50+ reasoning methods** — Spanning core logic to advanced multi-step reasoning frameworks
|
||||
- **Smart context selection** — Methods chosen based on content type, complexity, and stakeholder needs
|
||||
- **Iterative enhancement** — Each method builds on previous improvements
|
||||
- **User control** — Accept or discard each enhancement before proceeding
|
||||
- **Party Mode integration** — Agents can participate when Party Mode is active
|
||||
|
||||
## Workflow Integration
|
||||
|
||||
Advanced Elicitation is a core workflow designed to be invoked by other workflows during content generation:
|
||||
|
||||
| Parameter | Description |
|
||||
| ---------------------- | --------------------------------------------------------- |
|
||||
| **Content to enhance** | The current section content that was just generated |
|
||||
| **Context type** | The kind of content being created (spec, code, doc, etc.) |
|
||||
| **Enhancement goals** | What the calling workflow wants to improve |
|
||||
|
||||
### Integration Flow
|
||||
|
||||
When called from a workflow:
|
||||
1. Receives the current section content that was just generated
|
||||
2. Applies elicitation methods iteratively to enhance that content
|
||||
3. Returns the enhanced version when user selects 'x' to proceed
|
||||
4. The enhanced content replaces the original section in the output document
|
||||
|
||||
### Example
|
||||
|
||||
A specification generation workflow could invoke Advanced Elicitation after producing each major section (requirements, architecture, implementation plan). The workflow would pass the generated section, and Advanced Elicitation would offer methods like "Stakeholder Round Table" to gather diverse perspectives on requirements, or "Red Team vs Blue Team" to stress-test the architecture for vulnerabilities.
|
||||
|
||||
## Advanced Elicitation vs. Brainstorming
|
||||
|
||||
| | **Advanced Elicitation** | **Brainstorming** |
|
||||
| ------------ | ------------------------------------------------- | --------------------------------------------- |
|
||||
| **Source** | LLM generates ideas through structured reasoning | User provides ideas, AI coaches them out |
|
||||
| **Purpose** | Rethink and improve LLM's own output | Unlock user's creativity |
|
||||
| **Methods** | 50+ reasoning and analysis techniques | 60+ ideation and creativity techniques |
|
||||
| **Best for** | Enhancing generated content, finding alternatives | Breaking through blocks, generating new ideas |
|
||||
92
docs/explanation/features/brainstorming-techniques.md
Normal file
92
docs/explanation/features/brainstorming-techniques.md
Normal file
@@ -0,0 +1,92 @@
|
||||
---
|
||||
title: "Brainstorming"
|
||||
---
|
||||
|
||||
Facilitate structured creative sessions using 60+ proven ideation techniques.
|
||||
|
||||
The Brainstorming workflow is an interactive facilitation system that helps you unlock your own creativity. The AI acts as coach, guide, and creative partner — using proven techniques to draw out ideas and insights that are already within you.
|
||||
|
||||
:::note[Important]
|
||||
Every idea comes from you. The workflow creates the conditions for your best thinking to emerge through guided exploration, but you are the source.
|
||||
:::
|
||||
|
||||
## When to Use It
|
||||
|
||||
- Breaking through creative blocks on a specific challenge
|
||||
- Generating innovative ideas for products, features, or solutions
|
||||
- Exploring a problem from completely new angles
|
||||
- Systematically developing ideas from raw concepts to actionable plans
|
||||
- Team ideation (with collaborative techniques) or personal creative exploration
|
||||
|
||||
## How It Works
|
||||
|
||||
### 1. Session Setup
|
||||
Define your topic, goals, and any constraints.
|
||||
|
||||
### 2. Choose Your Approach
|
||||
|
||||
| Approach | Description |
|
||||
|----------|-------------|
|
||||
| **User-Selected** | Browse the full technique library and pick what appeals to you |
|
||||
| **AI-Recommended** | Get customized technique suggestions based on your goals |
|
||||
| **Random Selection** | Discover unexpected methods through serendipitous technique combinations |
|
||||
| **Progressive Flow** | Journey systematically from expansive exploration to focused action planning |
|
||||
|
||||
### 3. Interactive Facilitation
|
||||
Work through techniques with true collaborative coaching. The AI asks probing questions, builds on your ideas, and helps you think deeper—but your ideas are the source.
|
||||
|
||||
### 4. Idea Organization
|
||||
All your generated ideas are organized into themes and prioritized.
|
||||
|
||||
### 5. Action Planning
|
||||
Top ideas get concrete next steps, resource requirements, and success metrics.
|
||||
|
||||
## What You Get
|
||||
|
||||
A comprehensive session document that captures the entire journey:
|
||||
|
||||
- Topic, goals, and session parameters
|
||||
- Each technique used and how it was applied
|
||||
- Your contributions and the ideas you generated
|
||||
- Thematic organization connecting related insights
|
||||
- Prioritized ideas with action plans
|
||||
- Session highlights and key breakthroughs
|
||||
|
||||
This document becomes a permanent record of your creative process — valuable for future reference, sharing with stakeholders, or continuing the session later.
|
||||
|
||||
## Technique Categories
|
||||
|
||||
| Category | Focus |
|
||||
|----------|-------|
|
||||
| **Collaborative** | Team dynamics and inclusive participation |
|
||||
| **Creative** | Breakthrough thinking and paradigm shifts |
|
||||
| **Deep** | Root cause analysis and strategic insight |
|
||||
| **Structured** | Organized frameworks and systematic exploration |
|
||||
| **Theatrical** | Playful, radical perspectives |
|
||||
| **Wild** | Boundary-pushing, extreme thinking |
|
||||
| **Biomimetic** | Nature-inspired solutions |
|
||||
| **Quantum** | Quantum principles for innovation |
|
||||
| **Cultural** | Traditional knowledge and cross-cultural approaches |
|
||||
| **Introspective Delight** | Inner wisdom and authentic exploration |
|
||||
|
||||
## Key Features
|
||||
|
||||
- **Interactive coaching** — Pulls ideas *out* of you, doesn't generate them for you
|
||||
- **On-demand loading** — Techniques loaded from a comprehensive library as needed
|
||||
- **Session preservation** — Every step, insight, and action plan is documented
|
||||
- **Continuation support** — Pause sessions and return later, or extend with additional techniques
|
||||
|
||||
## Workflow Integration
|
||||
|
||||
Brainstorming is a core workflow designed to be invoked and configured by other modules. When called from another workflow, it accepts contextual parameters:
|
||||
|
||||
| Parameter | Description |
|
||||
|-----------|-------------|
|
||||
| **Topic focus** | What the brainstorming should help discover or solve |
|
||||
| **Guardrails** | Constraints, boundaries, or must-avoid areas |
|
||||
| **Output goals** | What the final output needs to accomplish for the calling workflow |
|
||||
| **Context files** | Project-specific guidance to inform technique selection |
|
||||
|
||||
### Example
|
||||
|
||||
When creating a new module in the BMad Builder workflow, Brainstorming can be invoked with guardrails around the module's purpose and a goal to discover key features, user needs, or architectural considerations. The session becomes focused on producing exactly what the module creation workflow needs.
|
||||
95
docs/explanation/features/party-mode.md
Normal file
95
docs/explanation/features/party-mode.md
Normal file
@@ -0,0 +1,95 @@
|
||||
---
|
||||
title: "Party Mode: Multi-Agent Collaboration"
|
||||
---
|
||||
|
||||
Get all your AI agents in one conversation.
|
||||
|
||||
## What is Party Mode?
|
||||
|
||||
Ever wanted to gather your entire AI team in one room and see what happens? That's party mode.
|
||||
|
||||
Type `/bmad:core:workflows:party-mode` (or `*party-mode` from any agent or at key workflow junctions when asked), and suddenly you've got **all your AI agents** in one conversation. PM, Architect, DEV, UX Designer and more that you can choose from.
|
||||
|
||||
**Why it's useful:**
|
||||
|
||||
- **After complex workflows** - Debrief with the whole team about what worked, what didn't
|
||||
- **Big decisions with tradeoffs** - Get technical, creative, and strategic perspectives simultaneously
|
||||
- **Brainstorming sessions** - Watch ideas evolve through cross-pollination
|
||||
- **When things go wrong** - Call out failures, watch agents defend their decisions, let them debate whose fault it was (oddly therapeutic)
|
||||
- **Sprint retrospectives** - Party mode powers the retrospective workflow
|
||||
- **Sprint planning** - Multi-agent collaboration for planning sessions
|
||||
|
||||
## How It Works
|
||||
|
||||
**The basics:**
|
||||
|
||||
1. Party mode reads `_bmad/_config/agent-manifest.csv`
|
||||
2. Loads ALL installed agents (already includes your customizations from install)
|
||||
3. BMad Master orchestrates - picks 2-3 relevant agents per message based on topic
|
||||
4. Agents respond in character, can agree/disagree/build on each other's ideas
|
||||
5. Discussion continues until you type "exit" or reach natural conclusion
|
||||
|
||||
**That's it.** No complex merging, no runtime magic. Just agents talking.
|
||||
|
||||
## Quick Start
|
||||
|
||||
```bash
|
||||
/bmad:core:workflows:party-mode
|
||||
|
||||
*party-mode
|
||||
|
||||
|
||||
/bmad:core:workflows:party-mode and include also in the party Santa Clause and Einstein
|
||||
|
||||
Ask questions, respond to agents, direct the conversation
|
||||
|
||||
Type: exit
|
||||
```
|
||||
|
||||
## Fun Examples
|
||||
|
||||
### Example 1: Calling Out Bad Architecture
|
||||
|
||||
**You:** "Okay team, our last sprint was a disaster. The authentication service went down three times. Who screwed this up?"
|
||||
|
||||
**Architect (Winston):** "The design was sound - distributed authentication with proper fallbacks. If implementation deviated from the spec..."
|
||||
|
||||
**DEV (Amelia):** "I followed the architecture doc exactly. The spec didn't account for concurrent session invalidation race conditions."
|
||||
|
||||
**PM (John):** "Both of you missed the bigger issue - we didn't validate session management requirements in the PRD. This is on me for not catching it."
|
||||
|
||||
**TEA (Murat):** "And I should've caught this in integration tests. The test scenarios didn't cover concurrent invalidation."
|
||||
|
||||
_(Watch them debate whose fault it really was - it's therapeutic)_
|
||||
|
||||
### Example 2: Creative Brainstorming
|
||||
|
||||
**You:** "How do we make onboarding feel magical instead of boring?"
|
||||
|
||||
**UX Designer (Sally):** "Start with progressive disclosure - reveal features as users need them, not all at once in a tutorial."
|
||||
|
||||
**Storyteller (CIS):** "What if onboarding was a story? Each step reveals a character's journey - the user IS the hero."
|
||||
|
||||
**Innovation Strategist (CIS):** "Take it further - gamify with unlockable achievements. But make them meaningful, not arbitrary badges."
|
||||
|
||||
**Game Designer:** "Building on that - what if the first 'quest' is actually solving a real user problem? They learn by doing something valuable."
|
||||
|
||||
_(Ideas cross-pollinate and evolve)_
|
||||
|
||||
### Example 3: Technical Decision
|
||||
|
||||
**You:** "Monolith or microservices for MVP?"
|
||||
|
||||
**Architect:** "Start monolith. Microservices add complexity you don't need at 1000 users."
|
||||
|
||||
**PM:** "Agree. Time to market matters more than theoretical scalability."
|
||||
|
||||
**DEV:** "Monolith with clear module boundaries. We can extract services later if needed."
|
||||
|
||||
**Innovation Strategist:** "Contrarian take - if your differentiator IS scalability, build for it now. Otherwise Architect's right."
|
||||
|
||||
_(Multiple perspectives reveal the right answer)_
|
||||
|
||||
:::tip[Better Decisions]
|
||||
Better decisions through diverse perspectives. Welcome to party mode.
|
||||
:::
|
||||
149
docs/explanation/features/quick-flow.md
Normal file
149
docs/explanation/features/quick-flow.md
Normal file
@@ -0,0 +1,149 @@
|
||||
---
|
||||
title: "Quick Spec Flow"
|
||||
description: Understanding Quick Spec Flow for rapid development in BMad Method
|
||||
---
|
||||
|
||||
Quick Spec Flow is a streamlined alternative to the full BMad Method for Quick Flow track projects. Instead of going through Product Brief → PRD → Architecture, you go straight to a context-aware technical specification and start coding.
|
||||
|
||||
- **Perfect for:** Bug fixes, small features, rapid prototyping, and quick enhancements
|
||||
- **Time to implementation:** Minutes, not hours
|
||||
|
||||
## When to Use Quick Flow
|
||||
|
||||
### Use Quick Flow when:
|
||||
|
||||
- Single bug fix or small enhancement
|
||||
- Small feature with clear scope (typically 1-15 stories)
|
||||
- Rapid prototyping or experimentation
|
||||
- Adding to existing brownfield codebase
|
||||
- You know exactly what you want to build
|
||||
|
||||
### Use BMad Method or Enterprise when:
|
||||
|
||||
- Building new products or major features
|
||||
- Need stakeholder alignment
|
||||
- Complex multi-team coordination
|
||||
- Requires extensive planning and architecture
|
||||
|
||||
:::tip[Not Sure?]
|
||||
Run `workflow-init` to get a recommendation based on your project's needs.
|
||||
:::
|
||||
|
||||
## Quick Flow Overview
|
||||
|
||||
```mermaid
|
||||
flowchart TD
|
||||
START[Step 1: Run Tech-Spec Workflow]
|
||||
DETECT[Detects project stack]
|
||||
ANALYZE[Analyzes brownfield codebase]
|
||||
TEST[Detects test frameworks]
|
||||
CONFIRM[Confirms conventions]
|
||||
GENERATE[Generates context-rich tech-spec]
|
||||
STORIES[Creates ready-to-implement stories]
|
||||
IMPL[Step 2: Implement with DEV Agent]
|
||||
DONE[DONE!]
|
||||
|
||||
START --> DETECT
|
||||
DETECT --> ANALYZE
|
||||
ANALYZE --> TEST
|
||||
TEST --> CONFIRM
|
||||
CONFIRM --> GENERATE
|
||||
GENERATE --> STORIES
|
||||
STORIES --> IMPL
|
||||
IMPL --> DONE
|
||||
|
||||
style START fill:#bfb,stroke:#333,stroke-width:2px
|
||||
style IMPL fill:#bbf,stroke:#333,stroke-width:2px
|
||||
style DONE fill:#f9f,stroke:#333,stroke-width:3px
|
||||
```
|
||||
|
||||
## What Makes It Quick
|
||||
|
||||
- No Product Brief needed
|
||||
- No PRD needed
|
||||
- No Architecture doc needed
|
||||
- Auto-detects your stack
|
||||
- Auto-analyzes brownfield code
|
||||
- Auto-validates quality
|
||||
- Story context optional (tech-spec is comprehensive)
|
||||
|
||||
## Smart Context Discovery
|
||||
|
||||
Quick Spec Flow automatically discovers and uses:
|
||||
|
||||
### Existing Documentation
|
||||
- Product briefs (if they exist)
|
||||
- Research documents
|
||||
- `document-project` output (brownfield codebase map)
|
||||
|
||||
### Project Stack
|
||||
- **Node.js:** package.json → frameworks, dependencies, scripts
|
||||
- **Python:** requirements.txt, pyproject.toml → packages, tools
|
||||
- **Ruby:** Gemfile → gems and versions
|
||||
- **Java:** pom.xml, build.gradle → Maven/Gradle dependencies
|
||||
- **Go:** go.mod → modules
|
||||
- **Rust:** Cargo.toml → crates
|
||||
|
||||
### Brownfield Code Patterns
|
||||
- Directory structure and organization
|
||||
- Existing code patterns (class-based, functional, MVC)
|
||||
- Naming conventions
|
||||
- Test frameworks and patterns
|
||||
- Code style configurations
|
||||
|
||||
### Convention Confirmation
|
||||
|
||||
Quick Spec Flow detects your conventions and **asks for confirmation**:
|
||||
|
||||
```
|
||||
I've detected these conventions in your codebase:
|
||||
|
||||
Code Style:
|
||||
- ESLint with Airbnb config
|
||||
- Prettier with single quotes
|
||||
|
||||
Test Patterns:
|
||||
- Jest test framework
|
||||
- .test.js file naming
|
||||
|
||||
Should I follow these existing conventions? (yes/no)
|
||||
```
|
||||
|
||||
**You decide:** Conform to existing patterns or establish new standards!
|
||||
|
||||
## Auto-Validation
|
||||
|
||||
Quick Spec Flow **automatically validates** everything:
|
||||
|
||||
- Context gathering completeness
|
||||
- Definitiveness (no "use X or Y" statements)
|
||||
- Brownfield integration quality
|
||||
- Stack alignment
|
||||
- Implementation readiness
|
||||
|
||||
## Comparison: Quick Flow vs Full BMM
|
||||
|
||||
| Aspect | Quick Flow Track | BMad Method/Enterprise Tracks |
|
||||
| --------------------- | ---------------------------- | ---------------------------------- |
|
||||
| **Setup** | None (standalone) | workflow-init recommended |
|
||||
| **Planning Docs** | tech-spec.md only | Product Brief → PRD → Architecture |
|
||||
| **Time to Code** | Minutes | Hours to days |
|
||||
| **Best For** | Bug fixes, small features | New products, major features |
|
||||
| **Context Discovery** | Automatic | Manual + guided |
|
||||
| **Validation** | Auto-validates everything | Manual validation steps |
|
||||
| **Brownfield** | Auto-analyzes and conforms | Manual documentation required |
|
||||
|
||||
## When to Graduate to BMad Method
|
||||
|
||||
Start with Quick Flow, but switch to BMad Method when:
|
||||
|
||||
- Project grows beyond initial scope
|
||||
- Multiple teams need coordination
|
||||
- Stakeholders need formal documentation
|
||||
- Product vision is unclear
|
||||
- Architectural decisions need deep analysis
|
||||
- Compliance/regulatory requirements exist
|
||||
|
||||
:::tip[Transition Tip]
|
||||
You can always run `workflow-init` later to transition from Quick Flow to BMad Method.
|
||||
:::
|
||||
354
docs/explanation/features/tea-overview.md
Normal file
354
docs/explanation/features/tea-overview.md
Normal file
@@ -0,0 +1,354 @@
|
||||
---
|
||||
title: "Test Architect (TEA) Overview"
|
||||
description: Understanding the Test Architect (TEA) agent and its role in BMad Method
|
||||
---
|
||||
|
||||
|
||||
The Test Architect (TEA) is a specialized agent focused on quality strategy, test automation, and release gates in BMad Method projects.
|
||||
|
||||
:::tip[Design Philosophy]
|
||||
TEA was built to solve AI-generated tests that rot in review. For the problem statement and design principles, see [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md). For setup, see [Setup Test Framework](/docs/how-to/workflows/setup-test-framework.md).
|
||||
:::
|
||||
|
||||
## Overview
|
||||
|
||||
- **Persona:** Murat, Master Test Architect and Quality Advisor focused on risk-based testing, fixture architecture, ATDD, and CI/CD governance.
|
||||
- **Mission:** Deliver actionable quality strategies, automation coverage, and gate decisions that scale with project complexity and compliance demands.
|
||||
- **Use When:** BMad Method or Enterprise track projects, integration risk is non-trivial, brownfield regression risk exists, or compliance/NFR evidence is required. (Quick Flow projects typically don't require TEA)
|
||||
|
||||
## Choose Your TEA Engagement Model
|
||||
|
||||
BMad does not mandate TEA. There are five valid ways to use it (or skip it). Pick one intentionally.
|
||||
|
||||
1. **No TEA**
|
||||
- Skip all TEA workflows. Use your existing team testing approach.
|
||||
|
||||
2. **TEA Solo (Standalone)**
|
||||
- Use TEA on a non-BMad project. Bring your own requirements, acceptance criteria, and environments.
|
||||
- Typical sequence: `*test-design` (system or epic) -> `*atdd` and/or `*automate` -> optional `*test-review` -> `*trace` for coverage and gate decisions.
|
||||
- Run `*framework` or `*ci` only if you want TEA to scaffold the harness or pipeline; they work best after you decide the stack/architecture.
|
||||
|
||||
**TEA Lite (Beginner Approach):**
|
||||
- Simplest way to use TEA - just use `*automate` to test existing features.
|
||||
- Perfect for learning TEA fundamentals in 30 minutes.
|
||||
- See [TEA Lite Quickstart Tutorial](/docs/tutorials/getting-started/tea-lite-quickstart.md).
|
||||
|
||||
3. **Integrated: Greenfield - BMad Method (Simple/Standard Work)**
|
||||
- Phase 3: system-level `*test-design`, then `*framework` and `*ci`.
|
||||
- Phase 4: per-epic `*test-design`, optional `*atdd`, then `*automate` and optional `*test-review`.
|
||||
- Gate (Phase 2): `*trace`.
|
||||
|
||||
4. **Integrated: Brownfield - BMad Method or Enterprise (Simple or Complex)**
|
||||
- Phase 2: baseline `*trace`.
|
||||
- Phase 3: system-level `*test-design`, then `*framework` and `*ci`.
|
||||
- Phase 4: per-epic `*test-design` focused on regression and integration risks.
|
||||
- Gate (Phase 2): `*trace`; `*nfr-assess` (if not done earlier).
|
||||
- For brownfield BMad Method, follow the same flow with `*nfr-assess` optional.
|
||||
|
||||
5. **Integrated: Greenfield - Enterprise Method (Enterprise/Compliance Work)**
|
||||
- Phase 2: `*nfr-assess`.
|
||||
- Phase 3: system-level `*test-design`, then `*framework` and `*ci`.
|
||||
- Phase 4: per-epic `*test-design`, plus `*atdd`/`*automate`/`*test-review`.
|
||||
- Gate (Phase 2): `*trace`; archive artifacts as needed.
|
||||
|
||||
If you are unsure, default to the integrated path for your track and adjust later.
|
||||
|
||||
## TEA Command Catalog
|
||||
|
||||
| Command | Primary Outputs | Notes | With Playwright MCP Enhancements |
|
||||
| -------------- | --------------------------------------------------------------------------------------------- | ---------------------------------------------------- | ------------------------------------------------------------------------------------------------------------ |
|
||||
| `*framework` | Playwright/Cypress scaffold, `.env.example`, `.nvmrc`, sample specs | Use when no production-ready harness exists | - |
|
||||
| `*ci` | CI workflow, selective test scripts, secrets checklist | Platform-aware (GitHub Actions default) | - |
|
||||
| `*test-design` | Combined risk assessment, mitigation plan, and coverage strategy | Risk scoring + optional exploratory mode | **+ Exploratory**: Interactive UI discovery with browser automation (uncover actual functionality) |
|
||||
| `*atdd` | Failing acceptance tests + implementation checklist | TDD red phase + optional recording mode | **+ Recording**: UI selectors verified with live browser; API tests benefit from trace analysis |
|
||||
| `*automate` | Prioritized specs, fixtures, README/script updates, DoD summary | Optional healing/recording, avoid duplicate coverage | **+ Healing**: Visual debugging + trace analysis for test fixes; **+ Recording**: Verified selectors (UI) + network inspection (API) |
|
||||
| `*test-review` | Test quality review report with 0-100 score, violations, fixes | Reviews tests against knowledge base patterns | - |
|
||||
| `*nfr-assess` | NFR assessment report with actions | Focus on security/performance/reliability | - |
|
||||
| `*trace` | Phase 1: Coverage matrix, recommendations. Phase 2: Gate decision (PASS/CONCERNS/FAIL/WAIVED) | Two-phase workflow: traceability + gate decision | - |
|
||||
|
||||
## TEA Workflow Lifecycle
|
||||
|
||||
**Phase Numbering Note:** BMad uses a 4-phase methodology with optional Phase 1 and a documentation prerequisite:
|
||||
|
||||
- **Documentation** (Optional for brownfield): Prerequisite using `*document-project`
|
||||
- **Phase 1** (Optional): Discovery/Analysis (`*brainstorm`, `*research`, `*product-brief`)
|
||||
- **Phase 2** (Required): Planning (`*prd` creates PRD with FRs/NFRs)
|
||||
- **Phase 3** (Track-dependent): Solutioning (`*architecture` → `*test-design` (system-level) → `*create-epics-and-stories` → TEA: `*framework`, `*ci` → `*implementation-readiness`)
|
||||
- **Phase 4** (Required): Implementation (`*sprint-planning` → per-epic: `*test-design` → per-story: dev workflows)
|
||||
|
||||
TEA integrates into the BMad development lifecycle during Solutioning (Phase 3) and Implementation (Phase 4):
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'primaryColor':'#fff','primaryTextColor':'#000','primaryBorderColor':'#000','lineColor':'#000','secondaryColor':'#fff','tertiaryColor':'#fff','fontSize':'16px','fontFamily':'arial'}}}%%
|
||||
graph TB
|
||||
subgraph Phase2["<b>Phase 2: PLANNING</b>"]
|
||||
PM["<b>PM: *prd (creates PRD with FRs/NFRs)</b>"]
|
||||
PlanNote["<b>Business requirements phase</b>"]
|
||||
NFR2["<b>TEA: *nfr-assess (optional, enterprise)</b>"]
|
||||
PM -.-> NFR2
|
||||
NFR2 -.-> PlanNote
|
||||
PM -.-> PlanNote
|
||||
end
|
||||
|
||||
subgraph Phase3["<b>Phase 3: SOLUTIONING</b>"]
|
||||
Architecture["<b>Architect: *architecture</b>"]
|
||||
EpicsStories["<b>PM/Architect: *create-epics-and-stories</b>"]
|
||||
TestDesignSys["<b>TEA: *test-design (system-level)</b>"]
|
||||
Framework["<b>TEA: *framework (optional if needed)</b>"]
|
||||
CI["<b>TEA: *ci (optional if needed)</b>"]
|
||||
GateCheck["<b>Architect: *implementation-readiness</b>"]
|
||||
Architecture --> EpicsStories
|
||||
Architecture --> TestDesignSys
|
||||
TestDesignSys --> Framework
|
||||
EpicsStories --> Framework
|
||||
Framework --> CI
|
||||
CI --> GateCheck
|
||||
Phase3Note["<b>Epics created AFTER architecture,</b><br/><b>then system-level test design and test infrastructure setup</b>"]
|
||||
EpicsStories -.-> Phase3Note
|
||||
end
|
||||
|
||||
subgraph Phase4["<b>Phase 4: IMPLEMENTATION - Per Epic Cycle</b>"]
|
||||
SprintPlan["<b>SM: *sprint-planning</b>"]
|
||||
TestDesign["<b>TEA: *test-design (per epic)</b>"]
|
||||
CreateStory["<b>SM: *create-story</b>"]
|
||||
ATDD["<b>TEA: *atdd (optional, before dev)</b>"]
|
||||
DevImpl["<b>DEV: implements story</b>"]
|
||||
Automate["<b>TEA: *automate</b>"]
|
||||
TestReview1["<b>TEA: *test-review (optional)</b>"]
|
||||
Trace1["<b>TEA: *trace (refresh coverage)</b>"]
|
||||
|
||||
SprintPlan --> TestDesign
|
||||
TestDesign --> CreateStory
|
||||
CreateStory --> ATDD
|
||||
ATDD --> DevImpl
|
||||
DevImpl --> Automate
|
||||
Automate --> TestReview1
|
||||
TestReview1 --> Trace1
|
||||
Trace1 -.->|next story| CreateStory
|
||||
TestDesignNote["<b>Test design: 'How do I test THIS epic?'</b><br/>Creates test-design-epic-N.md per epic"]
|
||||
TestDesign -.-> TestDesignNote
|
||||
end
|
||||
|
||||
subgraph Gate["<b>EPIC/RELEASE GATE</b>"]
|
||||
NFR["<b>TEA: *nfr-assess (if not done earlier)</b>"]
|
||||
TestReview2["<b>TEA: *test-review (final audit, optional)</b>"]
|
||||
TraceGate["<b>TEA: *trace - Phase 2: Gate</b>"]
|
||||
GateDecision{"<b>Gate Decision</b>"}
|
||||
|
||||
NFR --> TestReview2
|
||||
TestReview2 --> TraceGate
|
||||
TraceGate --> GateDecision
|
||||
GateDecision -->|PASS| Pass["<b>PASS ✅</b>"]
|
||||
GateDecision -->|CONCERNS| Concerns["<b>CONCERNS ⚠️</b>"]
|
||||
GateDecision -->|FAIL| Fail["<b>FAIL ❌</b>"]
|
||||
GateDecision -->|WAIVED| Waived["<b>WAIVED ⏭️</b>"]
|
||||
end
|
||||
|
||||
Phase2 --> Phase3
|
||||
Phase3 --> Phase4
|
||||
Phase4 --> Gate
|
||||
|
||||
style Phase2 fill:#bbdefb,stroke:#0d47a1,stroke-width:3px,color:#000
|
||||
style Phase3 fill:#c8e6c9,stroke:#2e7d32,stroke-width:3px,color:#000
|
||||
style Phase4 fill:#e1bee7,stroke:#4a148c,stroke-width:3px,color:#000
|
||||
style Gate fill:#ffe082,stroke:#f57c00,stroke-width:3px,color:#000
|
||||
style Pass fill:#4caf50,stroke:#1b5e20,stroke-width:3px,color:#000
|
||||
style Concerns fill:#ffc107,stroke:#f57f17,stroke-width:3px,color:#000
|
||||
style Fail fill:#f44336,stroke:#b71c1c,stroke-width:3px,color:#000
|
||||
style Waived fill:#9c27b0,stroke:#4a148c,stroke-width:3px,color:#000
|
||||
```
|
||||
|
||||
**TEA workflows:** `*framework` and `*ci` run once in Phase 3 after architecture. `*test-design` is **dual-mode**:
|
||||
|
||||
- **System-level (Phase 3):** Run immediately after architecture/ADR drafting to produce `test-design-system.md` (testability review, ADR → test mapping, Architecturally Significant Requirements (ASRs), environment needs). Feeds the implementation-readiness gate.
|
||||
- **Epic-level (Phase 4):** Run per-epic to produce `test-design-epic-N.md` (risk, priorities, coverage plan).
|
||||
|
||||
The Quick Flow track skips Phases 1 and 3.
|
||||
BMad Method and Enterprise use all phases based on project needs.
|
||||
When an ADR or architecture draft is produced, run `*test-design` in **system-level** mode before the implementation-readiness gate. This ensures the ADR has an attached testability review and ADR → test mapping. Keep the test-design updated if ADRs change.
|
||||
|
||||
## Why TEA Is Different from Other BMM Agents
|
||||
|
||||
TEA spans multiple phases (Phase 3, Phase 4, and the release gate). Most BMM agents operate in a single phase. That multi-phase role is paired with a dedicated testing knowledge base so standards stay consistent across projects.
|
||||
|
||||
### TEA's 8 Workflows Across Phases
|
||||
|
||||
| Phase | TEA Workflows | Frequency | Purpose |
|
||||
| ----------- | --------------------------------------------------------- | ---------------- | ---------------------------------------------- |
|
||||
| **Phase 2** | (none) | - | Planning phase - PM defines requirements |
|
||||
| **Phase 3** | \*test-design (system-level), \*framework, \*ci | Once per project | System testability review and test infrastructure setup |
|
||||
| **Phase 4** | \*test-design, \*atdd, \*automate, \*test-review, \*trace | Per epic/story | Test planning per epic, then per-story testing |
|
||||
| **Release** | \*nfr-assess, \*trace (Phase 2: gate) | Per epic/release | Go/no-go decision |
|
||||
|
||||
**Note**: `*trace` is a two-phase workflow: Phase 1 (traceability) + Phase 2 (gate decision). This reduces cognitive load while maintaining natural workflow.
|
||||
|
||||
### Why TEA Requires Its Own Knowledge Base
|
||||
|
||||
TEA uniquely requires:
|
||||
|
||||
- **Extensive domain knowledge**: Test patterns, CI/CD, fixtures, and quality practices
|
||||
- **Cross-cutting concerns**: Standards that apply across all BMad projects (not just PRDs or stories)
|
||||
- **Optional integrations**: Playwright-utils and MCP enhancements
|
||||
|
||||
This architecture lets TEA maintain consistent, production-ready testing patterns while operating across multiple phases.
|
||||
|
||||
## Track Cheat Sheets (Condensed)
|
||||
|
||||
These cheat sheets map TEA workflows to the **BMad Method and Enterprise tracks** across the **4-Phase Methodology** (Phase 1: Analysis, Phase 2: Planning, Phase 3: Solutioning, Phase 4: Implementation).
|
||||
|
||||
**Note:** The Quick Flow track typically doesn't require TEA (covered in Overview). These cheat sheets focus on BMad Method and Enterprise tracks where TEA adds value.
|
||||
|
||||
**Legend for Track Deltas:**
|
||||
|
||||
- ➕ = New workflow or phase added (doesn't exist in baseline)
|
||||
- 🔄 = Modified focus (same workflow, different emphasis or purpose)
|
||||
- 📦 = Additional output or archival requirement
|
||||
|
||||
### Greenfield - BMad Method (Simple/Standard Work)
|
||||
|
||||
**Planning Track:** BMad Method (PRD + Architecture)
|
||||
**Use Case:** New projects with standard complexity
|
||||
|
||||
| Workflow Stage | Test Architect | Dev / Team | Outputs |
|
||||
| -------------------------- | ----------------------------------------------------------------- | ----------------------------------------------------------------------------------- | ---------------------------------------------------------- |
|
||||
| **Phase 1**: Discovery | - | Analyst `*product-brief` (optional) | `product-brief.md` |
|
||||
| **Phase 2**: Planning | - | PM `*prd` (creates PRD with FRs/NFRs) | PRD with functional/non-functional requirements |
|
||||
| **Phase 3**: Solutioning | Run `*framework`, `*ci` AFTER architecture and epic creation | Architect `*architecture`, `*create-epics-and-stories`, `*implementation-readiness` | Architecture, epics/stories, test scaffold, CI pipeline |
|
||||
| **Phase 4**: Sprint Start | - | SM `*sprint-planning` | Sprint status file with all epics and stories |
|
||||
| **Phase 4**: Epic Planning | Run `*test-design` for THIS epic (per-epic test plan) | Review epic scope | `test-design-epic-N.md` with risk assessment and test plan |
|
||||
| **Phase 4**: Story Dev | (Optional) `*atdd` before dev, then `*automate` after | SM `*create-story`, DEV implements | Tests, story implementation |
|
||||
| **Phase 4**: Story Review | Execute `*test-review` (optional), re-run `*trace` | Address recommendations, update code/tests | Quality report, refreshed coverage matrix |
|
||||
| **Phase 4**: Release Gate | (Optional) `*test-review` for final audit, Run `*trace` (Phase 2) | Confirm Definition of Done, share release notes | Quality audit, Gate YAML + release summary |
|
||||
|
||||
**Key notes:**
|
||||
- Run `*framework` and `*ci` once in Phase 3 after architecture.
|
||||
- Run `*test-design` per epic in Phase 4; use `*atdd` before dev when helpful.
|
||||
- Use `*trace` for gate decisions; `*test-review` is an optional audit.
|
||||
|
||||
### Brownfield - BMad Method or Enterprise (Simple or Complex)
|
||||
|
||||
**Planning Tracks:** BMad Method or Enterprise Method
|
||||
**Use Case:** Existing codebases: simple additions (BMad Method) or complex enterprise requirements (Enterprise Method)
|
||||
|
||||
**🔄 Brownfield Deltas from Greenfield:**
|
||||
|
||||
- ➕ Documentation (Prerequisite) - Document existing codebase if undocumented
|
||||
- ➕ Phase 2: `*trace` - Baseline existing test coverage before planning
|
||||
- 🔄 Phase 4: `*test-design` - Focus on regression hotspots and brownfield risks
|
||||
- 🔄 Phase 4: Story Review - May include `*nfr-assess` if not done earlier
|
||||
|
||||
| Workflow Stage | Test Architect | Dev / Team | Outputs |
|
||||
| --------------------------------- | --------------------------------------------------------------------------- | ----------------------------------------------------------------------------------- | ---------------------------------------------------------------------- |
|
||||
| **Documentation**: Prerequisite ➕ | - | Analyst `*document-project` (if undocumented) | Comprehensive project documentation |
|
||||
| **Phase 1**: Discovery | - | Analyst/PM/Architect rerun planning workflows | Updated planning artifacts in `{output_folder}` |
|
||||
| **Phase 2**: Planning | Run ➕ `*trace` (baseline coverage) | PM `*prd` (creates PRD with FRs/NFRs) | PRD with FRs/NFRs, ➕ coverage baseline |
|
||||
| **Phase 3**: Solutioning | Run `*framework`, `*ci` AFTER architecture and epic creation | Architect `*architecture`, `*create-epics-and-stories`, `*implementation-readiness` | Architecture, epics/stories, test framework, CI pipeline |
|
||||
| **Phase 4**: Sprint Start | - | SM `*sprint-planning` | Sprint status file with all epics and stories |
|
||||
| **Phase 4**: Epic Planning | Run `*test-design` for THIS epic 🔄 (regression hotspots) | Review epic scope and brownfield risks | `test-design-epic-N.md` with brownfield risk assessment and mitigation |
|
||||
| **Phase 4**: Story Dev | (Optional) `*atdd` before dev, then `*automate` after | SM `*create-story`, DEV implements | Tests, story implementation |
|
||||
| **Phase 4**: Story Review | Apply `*test-review` (optional), re-run `*trace`, ➕ `*nfr-assess` if needed | Resolve gaps, update docs/tests | Quality report, refreshed coverage matrix, NFR report |
|
||||
| **Phase 4**: Release Gate | (Optional) `*test-review` for final audit, Run `*trace` (Phase 2) | Capture sign-offs, share release notes | Quality audit, Gate YAML + release summary |
|
||||
|
||||
**Key notes:**
|
||||
- Start with `*trace` in Phase 2 to baseline coverage.
|
||||
- Focus `*test-design` on regression hotspots and integration risk.
|
||||
- Run `*nfr-assess` before the gate if it wasn't done earlier.
|
||||
|
||||
### Greenfield - Enterprise Method (Enterprise/Compliance Work)
|
||||
|
||||
**Planning Track:** Enterprise Method (BMad Method + extended security/devops/test strategies)
|
||||
**Use Case:** New enterprise projects with compliance, security, or complex regulatory requirements
|
||||
|
||||
**🏢 Enterprise Deltas from BMad Method:**
|
||||
|
||||
- ➕ Phase 1: `*research` - Domain and compliance research (recommended)
|
||||
- ➕ Phase 2: `*nfr-assess` - Capture NFR requirements early (security/performance/reliability)
|
||||
- 🔄 Phase 4: `*test-design` - Enterprise focus (compliance, security architecture alignment)
|
||||
- 📦 Release Gate - Archive artifacts and compliance evidence for audits
|
||||
|
||||
| Workflow Stage | Test Architect | Dev / Team | Outputs |
|
||||
| -------------------------- | ----------------------------------------------------------------------- | ----------------------------------------------------------------------------------- | ------------------------------------------------------------------ |
|
||||
| **Phase 1**: Discovery | - | Analyst ➕ `*research`, `*product-brief` | Domain research, compliance analysis, product brief |
|
||||
| **Phase 2**: Planning | Run ➕ `*nfr-assess` | PM `*prd` (creates PRD with FRs/NFRs), UX `*create-ux-design` | Enterprise PRD with FRs/NFRs, UX design, ➕ NFR documentation |
|
||||
| **Phase 3**: Solutioning | Run `*framework`, `*ci` AFTER architecture and epic creation | Architect `*architecture`, `*create-epics-and-stories`, `*implementation-readiness` | Architecture, epics/stories, test framework, CI pipeline |
|
||||
| **Phase 4**: Sprint Start | - | SM `*sprint-planning` | Sprint plan with all epics |
|
||||
| **Phase 4**: Epic Planning | Run `*test-design` for THIS epic 🔄 (compliance focus) | Review epic scope and compliance requirements | `test-design-epic-N.md` with security/performance/compliance focus |
|
||||
| **Phase 4**: Story Dev | (Optional) `*atdd`, `*automate`, `*test-review`, `*trace` per story | SM `*create-story`, DEV implements | Tests, fixtures, quality reports, coverage matrices |
|
||||
| **Phase 4**: Release Gate | Final `*test-review` audit, Run `*trace` (Phase 2), 📦 archive artifacts | Capture sign-offs, 📦 compliance evidence | Quality audit, updated assessments, gate YAML, 📦 audit trail |
|
||||
|
||||
**Key notes:**
|
||||
- Run `*nfr-assess` early in Phase 2.
|
||||
- `*test-design` emphasizes compliance, security, and performance alignment.
|
||||
- Archive artifacts at the release gate for audits.
|
||||
|
||||
**Related how-to guides:**
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md)
|
||||
- [How to Set Up a Test Framework](/docs/how-to/workflows/setup-test-framework.md)
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md)
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md)
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md)
|
||||
- [How to Set Up CI Pipeline](/docs/how-to/workflows/setup-ci.md)
|
||||
- [How to Run NFR Assessment](/docs/how-to/workflows/run-nfr-assess.md)
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md)
|
||||
|
||||
## Deep Dive Concepts
|
||||
|
||||
Want to understand TEA principles and patterns in depth?
|
||||
|
||||
**Core Principles:**
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Probability × impact scoring, P0-P3 priorities
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Definition of Done, determinism, isolation
|
||||
- [Knowledge Base System](/docs/explanation/tea/knowledge-base-system.md) - Context engineering with tea-index.csv
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Pure function → fixture → composition
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Eliminating flakiness with intercept-before-navigate
|
||||
|
||||
**Engagement & Strategy:**
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - TEA Lite, TEA Solo, TEA Integrated (5 models explained)
|
||||
|
||||
**Philosophy:**
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - **Start here to understand WHY TEA exists** - The problem with AI-generated tests and TEA's three-part solution
|
||||
|
||||
## Optional Integrations
|
||||
|
||||
### Playwright Utils (`@seontechnologies/playwright-utils`)
|
||||
|
||||
Production-ready fixtures and utilities that enhance TEA workflows.
|
||||
|
||||
- Install: `npm install -D @seontechnologies/playwright-utils`
|
||||
> Note: Playwright Utils is enabled via the installer. Only set `tea_use_playwright_utils` in `_bmad/bmm/config.yaml` if you need to override the installer choice.
|
||||
- Impacts: `*framework`, `*atdd`, `*automate`, `*test-review`, `*ci`
|
||||
- Utilities include: api-request, auth-session, network-recorder, intercept-network-call, recurse, log, file-utils, burn-in, network-error-monitor, fixtures-composition
|
||||
|
||||
### Playwright MCP Enhancements
|
||||
|
||||
Live browser verification for test design and automation.
|
||||
|
||||
**Two Playwright MCP servers** (actively maintained, continuously updated):
|
||||
|
||||
- `playwright` - Browser automation (`npx @playwright/mcp@latest`)
|
||||
- `playwright-test` - Test runner with failure analysis (`npx playwright run-test-mcp-server`)
|
||||
|
||||
**Configuration example**:
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"playwright": {
|
||||
"command": "npx",
|
||||
"args": ["@playwright/mcp@latest"]
|
||||
},
|
||||
"playwright-test": {
|
||||
"command": "npx",
|
||||
"args": ["playwright", "run-test-mcp-server"]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
- Helps `*test-design` validate actual UI behavior.
|
||||
- Helps `*atdd` and `*automate` verify selectors against the live DOM.
|
||||
- Enhances healing with `browser_snapshot`, console, network, and locator tools.
|
||||
|
||||
**To disable**: set `tea_use_mcp_enhancements: false` in `_bmad/bmm/config.yaml` or remove MCPs from IDE config.
|
||||
34
docs/explanation/features/web-bundles.md
Normal file
34
docs/explanation/features/web-bundles.md
Normal file
@@ -0,0 +1,34 @@
|
||||
---
|
||||
title: "Web Bundles"
|
||||
---
|
||||
|
||||
Use BMad agents in Gemini Gems and Custom GPTs.
|
||||
|
||||
:::caution[Status]
|
||||
The Web Bundling Feature is being rebuilt from the ground up. Current v6 bundles may be incomplete or missing functionality.
|
||||
:::
|
||||
|
||||
## What Are Web Bundles?
|
||||
|
||||
Web bundles package BMad agents as self-contained files that work in Gemini Gems and Custom GPTs. Everything the agent needs - instructions, workflows, dependencies - is bundled into a single file for easy upload.
|
||||
|
||||
### What's Included
|
||||
|
||||
- Complete agent persona and instructions
|
||||
- All workflows and dependencies
|
||||
- Interactive menu system
|
||||
- Party mode for multi-agent collaboration
|
||||
- No external files required
|
||||
|
||||
### Use Cases
|
||||
|
||||
**Perfect for:**
|
||||
- Uploading a single file to a Gemini GEM or Custom GPT
|
||||
- Using BMad Method from the Web
|
||||
- Cost savings (generally lower cost than local usage)
|
||||
- Quick sharing of agent configurations
|
||||
|
||||
**Trade-offs:**
|
||||
- Some quality reduction vs local usage
|
||||
- Less convenient than full local installation
|
||||
- Limited to agent capabilities (no workflow file access)
|
||||
@@ -1,25 +1,23 @@
|
||||
# BMGD Agents Guide
|
||||
---
|
||||
title: "BMGD Agents Guide"
|
||||
---
|
||||
|
||||
Complete reference for BMGD's six specialized game development agents.
|
||||
|
||||
---
|
||||
|
||||
## Agent Overview
|
||||
|
||||
BMGD provides six agents, each with distinct expertise:
|
||||
|
||||
| Agent | Name | Role | Phase Focus |
|
||||
| ------------------------ | ---------------- | ----------------------------------------------------------- | ----------- |
|
||||
| 🎲 **Game Designer** | Samus Shepard | Lead Game Designer + Creative Vision Architect | Phases 1-2 |
|
||||
| 🏛️ **Game Architect** | Cloud Dragonborn | Principal Game Systems Architect + Technical Director | Phase 3 |
|
||||
| 🕹️ **Game Developer** | Link Freeman | Senior Game Developer + Technical Implementation Specialist | Phase 4 |
|
||||
| 🎯 **Game Scrum Master** | Max | Game Development Scrum Master + Sprint Orchestrator | Phase 4 |
|
||||
| 🧪 **Game QA** | GLaDOS | Game QA Architect + Test Automation Specialist | All Phases |
|
||||
| 🎮 **Game Solo Dev** | Indie | Elite Indie Game Developer + Quick Flow Specialist | All Phases |
|
||||
| Agent | Name | Role | Phase Focus |
|
||||
|-------|------|------|-------------|
|
||||
| **Game Designer** | Samus Shepard | Lead Game Designer + Creative Vision Architect | Phases 1-2 |
|
||||
| **Game Architect** | Cloud Dragonborn | Principal Game Systems Architect + Technical Director | Phase 3 |
|
||||
| **Game Developer** | Link Freeman | Senior Game Developer + Technical Implementation Specialist | Phase 4 |
|
||||
| **Game Scrum Master** | Max | Game Development Scrum Master + Sprint Orchestrator | Phase 4 |
|
||||
| **Game QA** | GLaDOS | Game QA Architect + Test Automation Specialist | All Phases |
|
||||
| **Game Solo Dev** | Indie | Elite Indie Game Developer + Quick Flow Specialist | All Phases |
|
||||
|
||||
---
|
||||
|
||||
## 🎲 Game Designer (Samus Shepard)
|
||||
## Game Designer (Samus Shepard)
|
||||
|
||||
### Role
|
||||
|
||||
@@ -59,9 +57,7 @@ Talks like an excited streamer - enthusiastic, asks about player motivations, ce
|
||||
| `party-mode` | Multi-agent collaboration |
|
||||
| `advanced-elicitation` | Deep exploration (web only) |
|
||||
|
||||
---
|
||||
|
||||
## 🏛️ Game Architect (Cloud Dragonborn)
|
||||
## Game Architect (Cloud Dragonborn)
|
||||
|
||||
### Role
|
||||
|
||||
@@ -99,9 +95,7 @@ Speaks like a wise sage from an RPG - calm, measured, uses architectural metapho
|
||||
| `party-mode` | Multi-agent collaboration |
|
||||
| `advanced-elicitation` | Deep exploration (web only) |
|
||||
|
||||
---
|
||||
|
||||
## 🕹️ Game Developer (Link Freeman)
|
||||
## Game Developer (Link Freeman)
|
||||
|
||||
### Role
|
||||
|
||||
@@ -141,9 +135,7 @@ Speaks like a speedrunner - direct, milestone-focused, always optimizing for the
|
||||
| `party-mode` | Multi-agent collaboration |
|
||||
| `advanced-elicitation` | Deep exploration (web only) |
|
||||
|
||||
---
|
||||
|
||||
## 🎯 Game Scrum Master (Max)
|
||||
## Game Scrum Master (Max)
|
||||
|
||||
### Role
|
||||
|
||||
@@ -187,9 +179,7 @@ Talks in game terminology - milestones are save points, handoffs are level trans
|
||||
| `party-mode` | Multi-agent collaboration |
|
||||
| `advanced-elicitation` | Deep exploration (web only) |
|
||||
|
||||
---
|
||||
|
||||
## 🧪 Game QA (GLaDOS)
|
||||
## Game QA (GLaDOS)
|
||||
|
||||
### Role
|
||||
|
||||
@@ -262,9 +252,7 @@ GLaDOS has access to a comprehensive game testing knowledge base (`gametest/qa-i
|
||||
- Smoke testing
|
||||
- Test prioritization (P0-P3)
|
||||
|
||||
---
|
||||
|
||||
## 🎮 Game Solo Dev (Indie)
|
||||
## Game Solo Dev (Indie)
|
||||
|
||||
### Role
|
||||
|
||||
@@ -300,7 +288,7 @@ Direct, confident, and gameplay-focused. Uses dev slang, thinks in game feel and
|
||||
| ------------------ | ------------------------------------------------------ |
|
||||
| `quick-prototype` | Rapid prototype to test if a mechanic is fun |
|
||||
| `quick-dev` | Implement features end-to-end with game considerations |
|
||||
| `create-tech-spec` | Create implementation-ready technical spec |
|
||||
| `quick-spec` | Create implementation-ready technical spec |
|
||||
| `code-review` | Review code quality |
|
||||
| `test-framework` | Set up automated testing |
|
||||
| `party-mode` | Bring in specialists when needed |
|
||||
@@ -321,8 +309,6 @@ Use **Full BMGD workflow** when:
|
||||
- You're working with stakeholders/publishers
|
||||
- Long-term maintainability is critical
|
||||
|
||||
---
|
||||
|
||||
## Agent Selection Guide
|
||||
|
||||
### By Phase
|
||||
@@ -356,8 +342,6 @@ Use **Full BMGD workflow** when:
|
||||
| "Quick prototype this idea" | Game Solo Dev |
|
||||
| "Ship this feature fast" | Game Solo Dev |
|
||||
|
||||
---
|
||||
|
||||
## Multi-Agent Collaboration
|
||||
|
||||
### Party Mode
|
||||
@@ -388,8 +372,6 @@ Game QA integrates at multiple points:
|
||||
- During Implementation: Create automated tests
|
||||
- Before Release: Performance and certification testing
|
||||
|
||||
---
|
||||
|
||||
## Project Context
|
||||
|
||||
All agents share the principle:
|
||||
@@ -398,10 +380,8 @@ All agents share the principle:
|
||||
|
||||
The `project-context.md` file (if present) serves as the authoritative source for project decisions and constraints.
|
||||
|
||||
---
|
||||
|
||||
## Next Steps
|
||||
|
||||
- **[Quick Start Guide](./quick-start.md)** - Get started with BMGD
|
||||
- **[Workflows Guide](./workflows-guide.md)** - Detailed workflow reference
|
||||
- **[Game Types Guide](./game-types-guide.md)** - Game type templates
|
||||
- **[Quick Start Guide](/docs/tutorials/getting-started/quick-start-bmgd.md)** - Get started with BMGD
|
||||
- **[Workflows Guide](/docs/reference/workflows/index.md)** - Detailed workflow reference
|
||||
- **[Game Types Guide](/docs/explanation/game-dev/game-types.md)** - Game type templates
|
||||
125
docs/explanation/game-dev/bmgd-vs-bmm.md
Normal file
125
docs/explanation/game-dev/bmgd-vs-bmm.md
Normal file
@@ -0,0 +1,125 @@
|
||||
---
|
||||
title: "BMGD vs BMM"
|
||||
description: Understanding the differences between BMGD and BMM
|
||||
---
|
||||
|
||||
BMGD (BMad Game Development) extends BMM (BMad Method) with game-specific capabilities. This page explains the key differences.
|
||||
|
||||
## Quick Comparison
|
||||
|
||||
| Aspect | BMM | BMGD |
|
||||
| -------------- | ------------------------------------- | ------------------------------------------------------------------------ |
|
||||
| **Focus** | General software | Game development |
|
||||
| **Agents** | PM, Architect, Dev, SM, TEA, Solo Dev | Game Designer, Game Dev, Game Architect, Game SM, Game QA, Game Solo Dev |
|
||||
| **Planning** | PRD, Tech Spec | Game Brief, GDD |
|
||||
| **Types** | N/A | 24 game type templates |
|
||||
| **Narrative** | N/A | Full narrative workflow |
|
||||
| **Testing** | Web-focused | Engine-specific (Unity, Unreal, Godot) |
|
||||
| **Production** | BMM workflows | BMM workflows with game overrides |
|
||||
|
||||
## Agent Differences
|
||||
|
||||
### BMM Agents
|
||||
- PM (Product Manager)
|
||||
- Architect
|
||||
- DEV (Developer)
|
||||
- SM (Scrum Master)
|
||||
- TEA (Test Architect)
|
||||
- Quick Flow Solo Dev
|
||||
|
||||
### BMGD Agents
|
||||
- Game Designer
|
||||
- Game Developer
|
||||
- Game Architect
|
||||
- Game Scrum Master
|
||||
- Game QA
|
||||
- Game Solo Dev
|
||||
|
||||
BMGD agents understand game-specific concepts like:
|
||||
- Game mechanics and balance
|
||||
- Player psychology
|
||||
- Engine-specific patterns
|
||||
- Playtesting and QA
|
||||
|
||||
## Planning Documents
|
||||
|
||||
### BMM Planning
|
||||
- **Product Brief** → **PRD** → **Architecture**
|
||||
- Focus: Software requirements, user stories, system design
|
||||
|
||||
### BMGD Planning
|
||||
- **Game Brief** → **GDD** → **Architecture**
|
||||
- Focus: Game vision, mechanics, narrative, player experience
|
||||
|
||||
The GDD (Game Design Document) includes:
|
||||
- Core gameplay loop
|
||||
- Mechanics and systems
|
||||
- Progression and balance
|
||||
- Art and audio direction
|
||||
- Genre-specific sections
|
||||
|
||||
## Game Type Templates
|
||||
|
||||
BMGD includes 24 game type templates that auto-configure GDD sections:
|
||||
|
||||
- Action, Adventure, Puzzle
|
||||
- RPG, Strategy, Simulation
|
||||
- Sports, Racing, Fighting
|
||||
- Horror, Platformer, Shooter
|
||||
- And more...
|
||||
|
||||
Each template provides:
|
||||
- Genre-specific GDD sections
|
||||
- Relevant mechanics patterns
|
||||
- Testing considerations
|
||||
- Common pitfalls to avoid
|
||||
|
||||
## Narrative Support
|
||||
|
||||
BMGD includes full narrative workflow for story-driven games:
|
||||
|
||||
- **Narrative Design** workflow
|
||||
- Story structure templates
|
||||
- Character development
|
||||
- World-building guidelines
|
||||
- Dialogue systems
|
||||
|
||||
BMM has no equivalent for narrative design.
|
||||
|
||||
## Testing Differences
|
||||
|
||||
### BMM Testing (TEA)
|
||||
- Web-focused (Playwright, Cypress)
|
||||
- API testing
|
||||
- E2E for web applications
|
||||
|
||||
### BMGD Testing (Game QA)
|
||||
- Engine-specific frameworks (Unity, Unreal, Godot)
|
||||
- Gameplay testing
|
||||
- Performance profiling
|
||||
- Playtest planning
|
||||
- Balance validation
|
||||
|
||||
## Production Workflow
|
||||
|
||||
BMGD production workflows **inherit from BMM** and add game-specific:
|
||||
- Checklists
|
||||
- Templates
|
||||
- Quality gates
|
||||
- Engine-specific considerations
|
||||
|
||||
This means you get all of BMM's implementation structure plus game-specific enhancements.
|
||||
|
||||
## When to Use Each
|
||||
|
||||
### Use BMM when:
|
||||
- Building web applications
|
||||
- Creating APIs and services
|
||||
- Developing mobile apps (non-game)
|
||||
- Any general software project
|
||||
|
||||
### Use BMGD when:
|
||||
- Building video games
|
||||
- Creating interactive experiences
|
||||
- Game prototyping
|
||||
- Game jams
|
||||
@@ -1,15 +1,13 @@
|
||||
# BMGD Game Types Guide
|
||||
---
|
||||
title: "BMGD Game Types Guide"
|
||||
---
|
||||
|
||||
Reference for selecting and using BMGD's 24 supported game type templates.
|
||||
|
||||
---
|
||||
|
||||
## Overview
|
||||
|
||||
When creating a GDD, BMGD offers game type templates that provide genre-specific sections. This ensures your design document covers mechanics and systems relevant to your game's genre.
|
||||
|
||||
---
|
||||
|
||||
## Supported Game Types
|
||||
|
||||
### Action & Combat
|
||||
@@ -27,8 +25,6 @@ Side-scrolling or 3D platforming with combat mechanics. Think Hollow Knight, Cel
|
||||
- Level design patterns
|
||||
- Boss design
|
||||
|
||||
---
|
||||
|
||||
#### Shooter
|
||||
|
||||
**Tags:** shooter, combat, aiming, fps, tps
|
||||
@@ -43,8 +39,6 @@ Projectile combat with aiming mechanics. Covers FPS, TPS, and arena shooters.
|
||||
- Level/arena design
|
||||
- Multiplayer considerations
|
||||
|
||||
---
|
||||
|
||||
#### Fighting
|
||||
|
||||
**Tags:** fighting, combat, competitive, combos, pvp
|
||||
@@ -59,8 +53,6 @@ Projectile combat with aiming mechanics. Covers FPS, TPS, and arena shooters.
|
||||
- Competitive balance
|
||||
- Netcode requirements
|
||||
|
||||
---
|
||||
|
||||
### Strategy & Tactics
|
||||
|
||||
#### Strategy
|
||||
@@ -77,8 +69,6 @@ Resource management with tactical decisions. RTS, 4X, and grand strategy.
|
||||
- Map/scenario design
|
||||
- Victory conditions
|
||||
|
||||
---
|
||||
|
||||
#### Turn-Based Tactics
|
||||
|
||||
**Tags:** tactics, turn-based, grid, positioning
|
||||
@@ -93,8 +83,6 @@ Grid-based movement with turn order. XCOM-likes and tactical RPGs.
|
||||
- Unit progression
|
||||
- Procedural mission generation
|
||||
|
||||
---
|
||||
|
||||
#### Tower Defense
|
||||
|
||||
**Tags:** tower-defense, waves, placement, strategy
|
||||
@@ -109,8 +97,6 @@ Wave-based defense with tower placement.
|
||||
- Map design patterns
|
||||
- Meta-progression
|
||||
|
||||
---
|
||||
|
||||
### RPG & Progression
|
||||
|
||||
#### RPG
|
||||
@@ -127,8 +113,6 @@ Character progression with stats, inventory, and quests.
|
||||
- Combat system (action/turn-based)
|
||||
- Skill trees and builds
|
||||
|
||||
---
|
||||
|
||||
#### Roguelike
|
||||
|
||||
**Tags:** roguelike, procedural, permadeath, runs
|
||||
@@ -143,8 +127,6 @@ Procedural generation with permadeath and run-based progression.
|
||||
- Item/ability synergies
|
||||
- Meta-progression systems
|
||||
|
||||
---
|
||||
|
||||
#### Metroidvania
|
||||
|
||||
**Tags:** metroidvania, exploration, abilities, interconnected
|
||||
@@ -159,8 +141,6 @@ Interconnected world with ability gating.
|
||||
- Secret and collectible placement
|
||||
- Power-up progression
|
||||
|
||||
---
|
||||
|
||||
### Narrative & Story
|
||||
|
||||
#### Adventure
|
||||
@@ -177,8 +157,6 @@ Story-driven exploration and narrative. Point-and-click and narrative adventures
|
||||
- Dialogue systems
|
||||
- Story branching
|
||||
|
||||
---
|
||||
|
||||
#### Visual Novel
|
||||
|
||||
**Tags:** visual-novel, narrative, choices, story
|
||||
@@ -193,8 +171,6 @@ Narrative choices with branching story.
|
||||
- UI/presentation
|
||||
- Save/load states
|
||||
|
||||
---
|
||||
|
||||
#### Text-Based
|
||||
|
||||
**Tags:** text, parser, interactive-fiction, mud
|
||||
@@ -209,8 +185,6 @@ Text input/output games. Parser games, choice-based IF, MUDs.
|
||||
- Text presentation
|
||||
- Save state management
|
||||
|
||||
---
|
||||
|
||||
### Simulation & Management
|
||||
|
||||
#### Simulation
|
||||
@@ -227,8 +201,6 @@ Realistic systems with management and building. Includes tycoons and sim games.
|
||||
- Building/construction
|
||||
- Failure states
|
||||
|
||||
---
|
||||
|
||||
#### Sandbox
|
||||
|
||||
**Tags:** sandbox, creative, building, freedom
|
||||
@@ -243,8 +215,6 @@ Creative freedom with building and minimal objectives.
|
||||
- Sharing/community features
|
||||
- Optional objectives
|
||||
|
||||
---
|
||||
|
||||
### Sports & Racing
|
||||
|
||||
#### Racing
|
||||
@@ -261,8 +231,6 @@ Vehicle control with tracks and lap times.
|
||||
- Progression/career mode
|
||||
- Multiplayer racing
|
||||
|
||||
---
|
||||
|
||||
#### Sports
|
||||
|
||||
**Tags:** sports, teams, realistic, physics
|
||||
@@ -277,8 +245,6 @@ Team-based or individual sports simulation.
|
||||
- Season/career modes
|
||||
- Multiplayer modes
|
||||
|
||||
---
|
||||
|
||||
### Multiplayer
|
||||
|
||||
#### MOBA
|
||||
@@ -295,8 +261,6 @@ Multiplayer team battles with hero selection.
|
||||
- Matchmaking
|
||||
- Economy (gold/items)
|
||||
|
||||
---
|
||||
|
||||
#### Party Game
|
||||
|
||||
**Tags:** party, multiplayer, minigames, casual
|
||||
@@ -311,8 +275,6 @@ Local multiplayer with minigames.
|
||||
- Scoring systems
|
||||
- Player count flexibility
|
||||
|
||||
---
|
||||
|
||||
### Horror & Survival
|
||||
|
||||
#### Survival
|
||||
@@ -329,8 +291,6 @@ Resource gathering with crafting and persistent threats.
|
||||
- Threat systems
|
||||
- Base building
|
||||
|
||||
---
|
||||
|
||||
#### Horror
|
||||
|
||||
**Tags:** horror, atmosphere, tension, fear
|
||||
@@ -345,8 +305,6 @@ Atmosphere and tension with limited resources.
|
||||
- Lighting and visibility
|
||||
- Enemy/threat design
|
||||
|
||||
---
|
||||
|
||||
### Casual & Progression
|
||||
|
||||
#### Puzzle
|
||||
@@ -363,8 +321,6 @@ Logic-based challenges and problem-solving.
|
||||
- Level structure
|
||||
- Scoring/rating
|
||||
|
||||
---
|
||||
|
||||
#### Idle/Incremental
|
||||
|
||||
**Tags:** idle, incremental, automation, progression
|
||||
@@ -379,8 +335,6 @@ Passive progression with upgrades and automation.
|
||||
- Number scaling
|
||||
- Offline progress
|
||||
|
||||
---
|
||||
|
||||
#### Card Game
|
||||
|
||||
**Tags:** card, deck-building, strategy, turns
|
||||
@@ -395,8 +349,6 @@ Deck building with card mechanics.
|
||||
- Rarity and collection
|
||||
- Competitive balance
|
||||
|
||||
---
|
||||
|
||||
### Rhythm
|
||||
|
||||
#### Rhythm
|
||||
@@ -413,8 +365,6 @@ Music synchronization with timing-based gameplay.
|
||||
- Music licensing
|
||||
- Input methods
|
||||
|
||||
---
|
||||
|
||||
## Hybrid Game Types
|
||||
|
||||
Many games combine multiple genres. BMGD supports hybrid selection:
|
||||
@@ -446,8 +396,6 @@ You: It's a roguelike with card game combat
|
||||
Agent: I'll include sections for both Roguelike and Card Game...
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Game Type Selection Tips
|
||||
|
||||
### 1. Start with Core Fantasy
|
||||
@@ -479,8 +427,6 @@ One type should be primary (most gameplay time). Others add flavor:
|
||||
- **Primary:** Platformer (core movement and exploration)
|
||||
- **Secondary:** Metroidvania (ability gating structure)
|
||||
|
||||
---
|
||||
|
||||
## GDD Section Mapping
|
||||
|
||||
When you select a game type, BMGD adds these GDD sections:
|
||||
@@ -494,10 +440,8 @@ When you select a game type, BMGD adds these GDD sections:
|
||||
| Multiplayer | Matchmaking, Netcode, Balance |
|
||||
| Simulation | Systems, Economy, AI |
|
||||
|
||||
---
|
||||
|
||||
## Next Steps
|
||||
|
||||
- **[Quick Start Guide](./quick-start.md)** - Get started with BMGD
|
||||
- **[Workflows Guide](./workflows-guide.md)** - GDD workflow details
|
||||
- **[Glossary](./glossary.md)** - Game development terminology
|
||||
- **[Quick Start Guide](/docs/tutorials/getting-started/quick-start-bmgd.md)** - Get started with BMGD
|
||||
- **[Workflows Guide](/docs/reference/workflows/bmgd-workflows.md)** - GDD workflow details
|
||||
- **[Glossary](/docs/reference/glossary/index.md)** - Game development terminology
|
||||
70
docs/explanation/game-dev/index.md
Normal file
70
docs/explanation/game-dev/index.md
Normal file
@@ -0,0 +1,70 @@
|
||||
---
|
||||
title: "BMGD - Game Development Module"
|
||||
description: AI-powered workflows for game design and development with BMGD
|
||||
---
|
||||
|
||||
Complete guides for the BMad Game Development Module (BMGD) — AI-powered workflows for game design and development that adapt to your project's needs.
|
||||
|
||||
## Getting Started
|
||||
|
||||
**New to BMGD?** Start here:
|
||||
|
||||
- **[Quick Start Guide](/docs/tutorials/getting-started/quick-start-bmgd.md)** - Get started building your first game
|
||||
- Installation and setup
|
||||
- Understanding the game development phases
|
||||
- Running your first workflows
|
||||
- Agent-based development flow
|
||||
|
||||
:::tip[Quick Path]
|
||||
Install BMGD module → Game Brief → GDD → Architecture → Build
|
||||
:::
|
||||
|
||||
## Core Documentation
|
||||
|
||||
- **[Game Types Guide](/docs/explanation/game-dev/game-types.md)** - Selecting and using game type templates (24 supported types)
|
||||
- **[BMGD vs BMM](/docs/explanation/game-dev/bmgd-vs-bmm.md)** - Understanding the differences
|
||||
|
||||
## Game Development Phases
|
||||
|
||||
BMGD follows four phases aligned with game development:
|
||||
|
||||
### Phase 1: Preproduction
|
||||
- **Brainstorm Game** - Ideation with game-specific techniques
|
||||
- **Game Brief** - Capture vision, market, and fundamentals
|
||||
|
||||
### Phase 2: Design
|
||||
- **GDD (Game Design Document)** - Comprehensive game design
|
||||
- **Narrative Design** - Story, characters, world (for story-driven games)
|
||||
|
||||
### Phase 3: Technical
|
||||
- **Game Architecture** - Engine, systems, patterns, structure
|
||||
|
||||
### Phase 4: Production
|
||||
- **Sprint Planning** - Epic and story management
|
||||
- **Story Development** - Implementation workflow
|
||||
- **Code Review** - Quality assurance
|
||||
- **Testing** - Automated tests, playtesting, performance
|
||||
- **Retrospective** - Continuous improvement
|
||||
|
||||
## Choose Your Path
|
||||
|
||||
### I need to...
|
||||
|
||||
**Start a new game project**
|
||||
→ Start with [Quick Start Guide](/docs/tutorials/getting-started/quick-start-bmgd.md)
|
||||
→ Run `brainstorm-game` for ideation
|
||||
→ Create a Game Brief with `create-brief`
|
||||
|
||||
**Design my game**
|
||||
→ Create a GDD with `create-gdd`
|
||||
→ If story-heavy, add Narrative Design with `create-narrative`
|
||||
|
||||
**Plan the technical architecture**
|
||||
→ Run `create-architecture` with the Game Architect
|
||||
|
||||
**Build my game**
|
||||
→ Use Phase 4 production workflows
|
||||
→ Follow the sprint-based development cycle
|
||||
|
||||
**Quickly test an idea**
|
||||
→ Use [Quick-Flow](/docs/how-to/workflows/bmgd-quick-flow.md) for rapid prototyping
|
||||
106
docs/explanation/philosophy/facilitation-over-generation.md
Normal file
106
docs/explanation/philosophy/facilitation-over-generation.md
Normal file
@@ -0,0 +1,106 @@
|
||||
---
|
||||
title: "Facilitation Over Generation"
|
||||
description: Understanding CIS's facilitation-first approach to creative work
|
||||
---
|
||||
|
||||
|
||||
The Creative Intelligence Suite (CIS) takes a fundamentally different approach from typical AI tools. Instead of generating solutions directly, CIS agents act as master facilitators who guide you to discover insights yourself.
|
||||
|
||||
## The Problem with Generation
|
||||
|
||||
Traditional AI approaches to creative work:
|
||||
|
||||
:::note[Generation Example]
|
||||
**User:** "Give me marketing ideas"
|
||||
|
||||
**AI:** "Here are 10 marketing ideas..."
|
||||
:::
|
||||
|
||||
This approach:
|
||||
- Produces generic, predictable outputs
|
||||
- Removes human ownership of ideas
|
||||
- Misses context and nuance
|
||||
- Limits creative exploration
|
||||
|
||||
## The Facilitation Approach
|
||||
|
||||
CIS agents use strategic questioning:
|
||||
|
||||
:::note[Facilitation Example]
|
||||
**User:** "I need marketing ideas"
|
||||
|
||||
**CIS:** "What makes your customers choose you over alternatives? What's the one thing they always mention?"
|
||||
|
||||
**User:** "They say our support is exceptional"
|
||||
|
||||
**CIS:** "Interesting! How might you make that exceptional support visible before they become customers?"
|
||||
:::
|
||||
|
||||
This approach:
|
||||
- Draws out insights already within you
|
||||
- Maintains human ownership of ideas
|
||||
- Captures context and nuance
|
||||
- Enables deeper creative exploration
|
||||
|
||||
## Key Principles
|
||||
|
||||
### 1. Questions Over Answers
|
||||
|
||||
CIS agents ask strategic questions rather than providing direct answers. This:
|
||||
- Activates your own creative thinking
|
||||
- Uncovers assumptions
|
||||
- Reveals blind spots
|
||||
- Builds on your domain knowledge
|
||||
|
||||
### 2. Energy-Aware Sessions
|
||||
|
||||
CIS monitors engagement and adapts:
|
||||
- Adjusts pace when energy flags
|
||||
- Suggests breaks when needed
|
||||
- Changes techniques to maintain momentum
|
||||
- Recognizes productive vs. unproductive struggle
|
||||
|
||||
### 3. Process Trust
|
||||
|
||||
CIS uses proven methodologies:
|
||||
- Design Thinking's 5 phases
|
||||
- Structured brainstorming techniques
|
||||
- Root cause analysis frameworks
|
||||
- Innovation strategy patterns
|
||||
|
||||
You're not just having a conversation—you're following time-tested creative processes.
|
||||
|
||||
### 4. Persona-Driven Engagement
|
||||
|
||||
Each CIS agent has a distinct personality:
|
||||
- **Carson** - Energetic, encouraging
|
||||
- **Maya** - Jazz-like, improvisational
|
||||
- **Dr. Quinn** - Analytical, methodical
|
||||
- **Victor** - Bold, strategic
|
||||
- **Sophia** - Narrative, imaginative
|
||||
|
||||
These personas create engaging experiences that maintain creative flow.
|
||||
|
||||
## When Generation is Appropriate
|
||||
|
||||
CIS does generate when appropriate:
|
||||
- Synthesizing session outputs
|
||||
- Documenting decisions
|
||||
- Creating structured artifacts
|
||||
- Providing technique examples
|
||||
|
||||
But the core creative work happens through facilitated discovery.
|
||||
|
||||
## Benefits
|
||||
|
||||
### For Individuals
|
||||
- Deeper insights than pure generation
|
||||
- Ownership of creative outputs
|
||||
- Skill development in creative thinking
|
||||
- More memorable and actionable ideas
|
||||
|
||||
### For Teams
|
||||
- Shared creative experience
|
||||
- Aligned understanding
|
||||
- Documented rationale
|
||||
- Stronger buy-in to outcomes
|
||||
112
docs/explanation/philosophy/testing-as-engineering.md
Normal file
112
docs/explanation/philosophy/testing-as-engineering.md
Normal file
@@ -0,0 +1,112 @@
|
||||
---
|
||||
title: "AI-Generated Testing: Why Most Approaches Fail"
|
||||
description: How Playwright-Utils, TEA workflows, and Playwright MCPs solve AI test quality problems
|
||||
---
|
||||
|
||||
|
||||
AI-generated tests frequently fail in production because they lack systematic quality standards. This document explains the problem and presents a solution combining three components: Playwright-Utils, TEA (Test Architect), and Playwright MCPs.
|
||||
|
||||
:::note[Source]
|
||||
This article is adapted from [The Testing Meta Most Teams Have Not Caught Up To Yet](https://dev.to/muratkeremozcan/the-testing-meta-most-teams-have-not-caught-up-to-yet-5765) by Murat K Ozcan.
|
||||
:::
|
||||
|
||||
## The Problem with AI-Generated Tests
|
||||
|
||||
When teams use AI to generate tests without structure, they often produce what can be called "slop factory" outputs:
|
||||
|
||||
| Issue | Description |
|
||||
|-------|-------------|
|
||||
| Redundant coverage | Multiple tests covering the same functionality |
|
||||
| Incorrect assertions | Tests that pass but don't actually verify behavior |
|
||||
| Flaky tests | Non-deterministic tests that randomly pass or fail |
|
||||
| Unreviewable diffs | Generated code too verbose or inconsistent to review |
|
||||
|
||||
The core problem is that prompt-driven testing paths lean into nondeterminism, which is the exact opposite of what testing exists to protect.
|
||||
|
||||
:::caution[The Paradox]
|
||||
AI excels at generating code quickly, but testing requires precision and consistency. Without guardrails, AI-generated tests amplify the chaos they're meant to prevent.
|
||||
:::
|
||||
|
||||
## The Solution: A Three-Part Stack
|
||||
|
||||
The solution combines three components that work together to enforce quality:
|
||||
|
||||
### Playwright-Utils
|
||||
|
||||
Bridges the gap between Cypress ergonomics and Playwright's capabilities by standardizing commonly reinvented primitives through utility functions.
|
||||
|
||||
| Utility | Purpose |
|
||||
|---------|---------|
|
||||
| api-request | API calls with schema validation |
|
||||
| auth-session | Authentication handling |
|
||||
| intercept-network-call | Network mocking and interception |
|
||||
| recurse | Retry logic and polling |
|
||||
| log | Structured logging |
|
||||
| network-recorder | Record and replay network traffic |
|
||||
| burn-in | Smart test selection for CI |
|
||||
| network-error-monitor | HTTP error detection |
|
||||
| file-utils | CSV/PDF handling |
|
||||
|
||||
These utilities eliminate the need to reinvent authentication, API calls, retries, and logging for every project.
|
||||
|
||||
### TEA (Test Architect Agent)
|
||||
|
||||
A quality operating model packaged as eight executable workflows spanning test design, CI/CD gates, and release readiness. TEA encodes test architecture expertise into repeatable processes.
|
||||
|
||||
| Workflow | Purpose |
|
||||
|----------|---------|
|
||||
| `*test-design` | Risk-based test planning per epic |
|
||||
| `*framework` | Scaffold production-ready test infrastructure |
|
||||
| `*ci` | CI pipeline with selective testing |
|
||||
| `*atdd` | Acceptance test-driven development |
|
||||
| `*automate` | Prioritized test automation |
|
||||
| `*test-review` | Test quality audits (0-100 score) |
|
||||
| `*nfr-assess` | Non-functional requirements assessment |
|
||||
| `*trace` | Coverage traceability and gate decisions |
|
||||
|
||||
:::tip[Key Insight]
|
||||
TEA doesn't just generate tests—it provides a complete quality operating model with workflows for planning, execution, and release gates.
|
||||
:::
|
||||
|
||||
### Playwright MCPs
|
||||
|
||||
Model Context Protocols enable real-time verification during test generation. Instead of inferring selectors and behavior from documentation, MCPs allow agents to:
|
||||
|
||||
- Run flows and confirm the DOM against the accessibility tree
|
||||
- Validate network responses in real-time
|
||||
- Discover actual functionality through interactive exploration
|
||||
- Verify generated tests against live applications
|
||||
|
||||
## How They Work Together
|
||||
|
||||
The three components form a quality pipeline:
|
||||
|
||||
| Stage | Component | Action |
|
||||
|-------|-----------|--------|
|
||||
| Standards | Playwright-Utils | Provides production-ready patterns and utilities |
|
||||
| Process | TEA Workflows | Enforces systematic test planning and review |
|
||||
| Verification | Playwright MCPs | Validates generated tests against live applications |
|
||||
|
||||
**Before (AI-only):** 20 tests with redundant coverage, incorrect assertions, and flaky behavior.
|
||||
|
||||
**After (Full Stack):** Risk-based selection, verified selectors, validated behavior, reviewable code.
|
||||
|
||||
## Why This Matters
|
||||
|
||||
Traditional AI testing approaches fail because they:
|
||||
|
||||
- **Lack quality standards** — No consistent patterns or utilities
|
||||
- **Skip planning** — Jump straight to test generation without risk assessment
|
||||
- **Can't verify** — Generate tests without validating against actual behavior
|
||||
- **Don't review** — No systematic audit of generated test quality
|
||||
|
||||
The three-part stack addresses each gap:
|
||||
|
||||
| Gap | Solution |
|
||||
|-----|----------|
|
||||
| No standards | Playwright-Utils provides production-ready patterns |
|
||||
| No planning | TEA `*test-design` workflow creates risk-based test plans |
|
||||
| No verification | Playwright MCPs validate against live applications |
|
||||
| No review | TEA `*test-review` audits quality with scoring |
|
||||
|
||||
This approach is sometimes called *context engineering*—loading domain-specific standards into AI context automatically rather than relying on prompts alone. TEA's `tea-index.csv` manifest loads relevant knowledge fragments so the AI doesn't relearn testing patterns each session.
|
||||
710
docs/explanation/tea/engagement-models.md
Normal file
710
docs/explanation/tea/engagement-models.md
Normal file
@@ -0,0 +1,710 @@
|
||||
---
|
||||
title: "TEA Engagement Models Explained"
|
||||
description: Understanding the five ways to use TEA - from standalone to full BMad Method integration
|
||||
---
|
||||
|
||||
# TEA Engagement Models Explained
|
||||
|
||||
TEA is optional and flexible. There are five valid ways to engage with TEA - choose intentionally based on your project needs and methodology.
|
||||
|
||||
## Overview
|
||||
|
||||
**TEA is not mandatory.** Pick the engagement model that fits your context:
|
||||
|
||||
1. **No TEA** - Skip all TEA workflows, use existing testing approach
|
||||
2. **TEA Solo** - Use TEA standalone without BMad Method
|
||||
3. **TEA Lite** - Beginner approach using just `*automate`
|
||||
4. **TEA Integrated (Greenfield)** - Full BMad Method integration from scratch
|
||||
5. **TEA Integrated (Brownfield)** - Full BMad Method integration with existing code
|
||||
|
||||
## The Problem
|
||||
|
||||
### One-Size-Fits-All Doesn't Work
|
||||
|
||||
**Traditional testing tools force one approach:**
|
||||
- Must use entire framework
|
||||
- All-or-nothing adoption
|
||||
- No flexibility for different project types
|
||||
- Teams abandon tool if it doesn't fit
|
||||
|
||||
**TEA recognizes:**
|
||||
- Different projects have different needs
|
||||
- Different teams have different maturity levels
|
||||
- Different contexts require different approaches
|
||||
- Flexibility increases adoption
|
||||
|
||||
## The Five Engagement Models
|
||||
|
||||
### Model 1: No TEA
|
||||
|
||||
**What:** Skip all TEA workflows, use your existing testing approach.
|
||||
|
||||
**When to Use:**
|
||||
- Team has established testing practices
|
||||
- Quality is already high
|
||||
- Testing tools already in place
|
||||
- TEA doesn't add value
|
||||
|
||||
**What You Miss:**
|
||||
- Risk-based test planning
|
||||
- Systematic quality review
|
||||
- Gate decisions with evidence
|
||||
- Knowledge base patterns
|
||||
|
||||
**What You Keep:**
|
||||
- Full control
|
||||
- Existing tools
|
||||
- Team expertise
|
||||
- No learning curve
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Your team:
|
||||
- 10-year veteran QA team
|
||||
- Established testing practices
|
||||
- High-quality test suite
|
||||
- No problems to solve
|
||||
|
||||
Decision: Skip TEA, keep what works
|
||||
```
|
||||
|
||||
**Verdict:** Valid choice if existing approach works.
|
||||
|
||||
---
|
||||
|
||||
### Model 2: TEA Solo
|
||||
|
||||
**What:** Use TEA workflows standalone without full BMad Method integration.
|
||||
|
||||
**When to Use:**
|
||||
- Non-BMad projects
|
||||
- Want TEA's quality operating model only
|
||||
- Don't need full planning workflow
|
||||
- Bring your own requirements
|
||||
|
||||
**Typical Sequence:**
|
||||
```
|
||||
1. *test-design (system or epic)
|
||||
2. *atdd or *automate
|
||||
3. *test-review (optional)
|
||||
4. *trace (coverage + gate decision)
|
||||
```
|
||||
|
||||
**You Bring:**
|
||||
- Requirements (user stories, acceptance criteria)
|
||||
- Development environment
|
||||
- Project context
|
||||
|
||||
**TEA Provides:**
|
||||
- Risk-based test planning (`*test-design`)
|
||||
- Test generation (`*atdd`, `*automate`)
|
||||
- Quality review (`*test-review`)
|
||||
- Coverage traceability (`*trace`)
|
||||
|
||||
**Optional:**
|
||||
- Framework setup (`*framework`) if needed
|
||||
- CI configuration (`*ci`) if needed
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Your project:
|
||||
- Using Scrum (not BMad Method)
|
||||
- Jira for story management
|
||||
- Need better test strategy
|
||||
|
||||
Workflow:
|
||||
1. Export stories from Jira
|
||||
2. Run *test-design on epic
|
||||
3. Run *atdd for each story
|
||||
4. Implement features
|
||||
5. Run *trace for coverage
|
||||
```
|
||||
|
||||
**Verdict:** Best for teams wanting TEA benefits without BMad Method commitment.
|
||||
|
||||
---
|
||||
|
||||
### Model 3: TEA Lite
|
||||
|
||||
**What:** Beginner approach using just `*automate` to test existing features.
|
||||
|
||||
**When to Use:**
|
||||
- Learning TEA fundamentals
|
||||
- Want quick results
|
||||
- Testing existing application
|
||||
- No time for full methodology
|
||||
|
||||
**Workflow:**
|
||||
```
|
||||
1. *framework (setup test infrastructure)
|
||||
2. *test-design (optional, risk assessment)
|
||||
3. *automate (generate tests for existing features)
|
||||
4. Run tests (they pass immediately)
|
||||
```
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Beginner developer:
|
||||
- Never used TEA before
|
||||
- Want to add tests to existing app
|
||||
- 30 minutes available
|
||||
|
||||
Steps:
|
||||
1. Run *framework
|
||||
2. Run *automate on TodoMVC demo
|
||||
3. Tests generated and passing
|
||||
4. Learn TEA basics
|
||||
```
|
||||
|
||||
**What You Get:**
|
||||
- Working test framework
|
||||
- Passing tests for existing features
|
||||
- Learning experience
|
||||
- Foundation to expand
|
||||
|
||||
**What You Miss:**
|
||||
- TDD workflow (ATDD)
|
||||
- Risk-based planning (test-design depth)
|
||||
- Quality gates (trace Phase 2)
|
||||
- Full TEA capabilities
|
||||
|
||||
**Verdict:** Perfect entry point for beginners.
|
||||
|
||||
---
|
||||
|
||||
### Model 4: TEA Integrated (Greenfield)
|
||||
|
||||
**What:** Full BMad Method integration with TEA workflows across all phases.
|
||||
|
||||
**When to Use:**
|
||||
- New projects starting from scratch
|
||||
- Using BMad Method or Enterprise track
|
||||
- Want complete quality operating model
|
||||
- Testing is critical to success
|
||||
|
||||
**Lifecycle:**
|
||||
|
||||
**Phase 2: Planning**
|
||||
- PM creates PRD with NFRs
|
||||
- (Optional) TEA runs `*nfr-assess` (Enterprise only)
|
||||
|
||||
**Phase 3: Solutioning**
|
||||
- Architect creates architecture
|
||||
- TEA runs `*test-design` (system-level) → testability review
|
||||
- TEA runs `*framework` → test infrastructure
|
||||
- TEA runs `*ci` → CI/CD pipeline
|
||||
- Architect runs `*implementation-readiness` (fed by test design)
|
||||
|
||||
**Phase 4: Implementation (Per Epic)**
|
||||
- SM runs `*sprint-planning`
|
||||
- TEA runs `*test-design` (epic-level) → risk assessment for THIS epic
|
||||
- SM creates stories
|
||||
- (Optional) TEA runs `*atdd` → failing tests before dev
|
||||
- DEV implements story
|
||||
- TEA runs `*automate` → expand coverage
|
||||
- (Optional) TEA runs `*test-review` → quality audit
|
||||
- TEA runs `*trace` Phase 1 → refresh coverage
|
||||
|
||||
**Release Gate:**
|
||||
- (Optional) TEA runs `*test-review` → final audit
|
||||
- (Optional) TEA runs `*nfr-assess` → validate NFRs
|
||||
- TEA runs `*trace` Phase 2 → gate decision (PASS/CONCERNS/FAIL/WAIVED)
|
||||
|
||||
**What You Get:**
|
||||
- Complete quality operating model
|
||||
- Systematic test planning
|
||||
- Risk-based prioritization
|
||||
- Evidence-based gate decisions
|
||||
- Consistent patterns across epics
|
||||
|
||||
**Example:**
|
||||
```
|
||||
New SaaS product:
|
||||
- 50 stories across 8 epics
|
||||
- Security critical
|
||||
- Need quality gates
|
||||
|
||||
Workflow:
|
||||
- Phase 2: Define NFRs in PRD
|
||||
- Phase 3: Architecture → test design → framework → CI
|
||||
- Phase 4: Per epic: test design → ATDD → dev → automate → review → trace
|
||||
- Gate: NFR assess → trace Phase 2 → decision
|
||||
```
|
||||
|
||||
**Verdict:** Most comprehensive TEA usage, best for structured teams.
|
||||
|
||||
---
|
||||
|
||||
### Model 5: TEA Integrated (Brownfield)
|
||||
|
||||
**What:** Full BMad Method integration with TEA for existing codebases.
|
||||
|
||||
**When to Use:**
|
||||
- Existing codebase with legacy tests
|
||||
- Want to improve test quality incrementally
|
||||
- Adding features to existing application
|
||||
- Need to establish coverage baseline
|
||||
|
||||
**Differences from Greenfield:**
|
||||
|
||||
**Phase 0: Documentation (if needed)**
|
||||
```
|
||||
- Run *document-project
|
||||
- Create baseline documentation
|
||||
```
|
||||
|
||||
**Phase 2: Planning**
|
||||
```
|
||||
- TEA runs *trace Phase 1 → establish coverage baseline
|
||||
- PM creates PRD (with existing system context)
|
||||
```
|
||||
|
||||
**Phase 3: Solutioning**
|
||||
```
|
||||
- Architect creates architecture (with brownfield constraints)
|
||||
- TEA runs *test-design (system-level) → testability review
|
||||
- TEA runs *framework (only if modernizing test infra)
|
||||
- TEA runs *ci (update existing CI or create new)
|
||||
```
|
||||
|
||||
**Phase 4: Implementation**
|
||||
```
|
||||
- TEA runs *test-design (epic-level) → focus on REGRESSION HOTSPOTS
|
||||
- Per story: ATDD → dev → automate
|
||||
- TEA runs *test-review → improve legacy test quality
|
||||
- TEA runs *trace Phase 1 → track coverage improvement
|
||||
```
|
||||
|
||||
**Brownfield-Specific:**
|
||||
- Baseline coverage BEFORE planning
|
||||
- Focus on regression hotspots (bug-prone areas)
|
||||
- Incremental quality improvement
|
||||
- Compare coverage to baseline (trending up?)
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Legacy e-commerce platform:
|
||||
- 200 existing tests (30% passing, 70% flaky)
|
||||
- Adding new checkout flow
|
||||
- Want to improve quality
|
||||
|
||||
Workflow:
|
||||
1. Phase 2: *trace baseline → 30% coverage
|
||||
2. Phase 3: *test-design → identify regression risks
|
||||
3. Phase 4: Fix top 20 flaky tests + add tests for new checkout
|
||||
4. Gate: *trace → 60% coverage (2x improvement)
|
||||
```
|
||||
|
||||
**Verdict:** Best for incrementally improving legacy systems.
|
||||
|
||||
---
|
||||
|
||||
## Decision Guide: Which Model?
|
||||
|
||||
### Quick Decision Tree
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
flowchart TD
|
||||
Start([Choose TEA Model]) --> BMad{Using<br/>BMad Method?}
|
||||
|
||||
BMad -->|No| NonBMad{Project Type?}
|
||||
NonBMad -->|Learning| Lite[TEA Lite<br/>Just *automate<br/>30 min tutorial]
|
||||
NonBMad -->|Serious Project| Solo[TEA Solo<br/>Standalone workflows<br/>Full capabilities]
|
||||
|
||||
BMad -->|Yes| WantTEA{Want TEA?}
|
||||
WantTEA -->|No| None[No TEA<br/>Use existing approach<br/>Valid choice]
|
||||
WantTEA -->|Yes| ProjectType{New or<br/>Existing?}
|
||||
|
||||
ProjectType -->|New Project| Green[TEA Integrated<br/>Greenfield<br/>Full lifecycle]
|
||||
ProjectType -->|Existing Code| Brown[TEA Integrated<br/>Brownfield<br/>Baseline + improve]
|
||||
|
||||
Green --> Compliance{Compliance<br/>Needs?}
|
||||
Compliance -->|Yes| Enterprise[Enterprise Track<br/>NFR + audit trails]
|
||||
Compliance -->|No| Method[BMad Method Track<br/>Standard quality]
|
||||
|
||||
style Lite fill:#bbdefb,stroke:#1565c0,stroke-width:2px
|
||||
style Solo fill:#c5cae9,stroke:#283593,stroke-width:2px
|
||||
style None fill:#e0e0e0,stroke:#616161,stroke-width:1px
|
||||
style Green fill:#c8e6c9,stroke:#2e7d32,stroke-width:2px
|
||||
style Brown fill:#fff9c4,stroke:#f57f17,stroke-width:2px
|
||||
style Enterprise fill:#f3e5f5,stroke:#6a1b9a,stroke-width:2px
|
||||
style Method fill:#e1f5fe,stroke:#01579b,stroke-width:2px
|
||||
```
|
||||
|
||||
**Decision Path Examples:**
|
||||
- Learning TEA → TEA Lite (blue)
|
||||
- Non-BMad project → TEA Solo (purple)
|
||||
- BMad + new project + compliance → Enterprise (purple)
|
||||
- BMad + existing code → Brownfield (yellow)
|
||||
- Don't want TEA → No TEA (gray)
|
||||
|
||||
### By Project Type
|
||||
|
||||
| Project Type | Recommended Model | Why |
|
||||
|--------------|------------------|-----|
|
||||
| **New SaaS product** | TEA Integrated (Greenfield) | Full quality operating model from day one |
|
||||
| **Existing app + new feature** | TEA Integrated (Brownfield) | Improve incrementally while adding features |
|
||||
| **Bug fix** | TEA Lite or No TEA | Quick flow, minimal overhead |
|
||||
| **Learning project** | TEA Lite | Learn basics with immediate results |
|
||||
| **Non-BMad enterprise** | TEA Solo | Quality model without full methodology |
|
||||
| **High-quality existing tests** | No TEA | Keep what works |
|
||||
|
||||
### By Team Maturity
|
||||
|
||||
| Team Maturity | Recommended Model | Why |
|
||||
|---------------|------------------|-----|
|
||||
| **Beginners** | TEA Lite → TEA Solo | Learn basics, then expand |
|
||||
| **Intermediate** | TEA Solo or Integrated | Depends on methodology |
|
||||
| **Advanced** | TEA Integrated or No TEA | Full model or existing expertise |
|
||||
|
||||
### By Compliance Needs
|
||||
|
||||
| Compliance | Recommended Model | Why |
|
||||
|------------|------------------|-----|
|
||||
| **None** | Any model | Choose based on project needs |
|
||||
| **Light** (internal audit) | TEA Solo or Integrated | Gate decisions helpful |
|
||||
| **Heavy** (SOC 2, HIPAA) | TEA Integrated (Enterprise) | NFR assessment mandatory |
|
||||
|
||||
## Switching Between Models
|
||||
|
||||
### Can Change Models Mid-Project
|
||||
|
||||
**Scenario:** Start with TEA Lite, expand to TEA Solo
|
||||
|
||||
```
|
||||
Week 1: TEA Lite
|
||||
- Run *framework
|
||||
- Run *automate
|
||||
- Learn basics
|
||||
|
||||
Week 2: Expand to TEA Solo
|
||||
- Add *test-design
|
||||
- Use *atdd for new features
|
||||
- Add *test-review
|
||||
|
||||
Week 3: Continue expanding
|
||||
- Add *trace for coverage
|
||||
- Setup *ci
|
||||
- Full TEA Solo workflow
|
||||
```
|
||||
|
||||
**Benefit:** Start small, expand as comfortable.
|
||||
|
||||
### Can Mix Models
|
||||
|
||||
**Scenario:** TEA Integrated for main features, No TEA for bug fixes
|
||||
|
||||
```
|
||||
Main features (epics):
|
||||
- Use full TEA workflow
|
||||
- Risk assessment, ATDD, quality gates
|
||||
|
||||
Bug fixes:
|
||||
- Skip TEA
|
||||
- Quick Flow + manual testing
|
||||
- Move fast
|
||||
|
||||
Result: TEA where it adds value, skip where it doesn't
|
||||
```
|
||||
|
||||
**Benefit:** Flexible, pragmatic, not dogmatic.
|
||||
|
||||
## Comparison Table
|
||||
|
||||
| Aspect | No TEA | TEA Lite | TEA Solo | Integrated (Green) | Integrated (Brown) |
|
||||
|--------|--------|----------|----------|-------------------|-------------------|
|
||||
| **BMad Required** | No | No | No | Yes | Yes |
|
||||
| **Learning Curve** | None | Low | Medium | High | High |
|
||||
| **Setup Time** | 0 | 30 min | 2 hours | 1 day | 2 days |
|
||||
| **Workflows Used** | 0 | 2-3 | 4-6 | 8 | 8 |
|
||||
| **Test Planning** | Manual | Optional | Yes | Systematic | + Regression focus |
|
||||
| **Quality Gates** | No | No | Optional | Yes | Yes + baseline |
|
||||
| **NFR Assessment** | No | No | No | Optional | Recommended |
|
||||
| **Coverage Tracking** | Manual | No | Optional | Yes | Yes + trending |
|
||||
| **Best For** | Experts | Beginners | Standalone | New projects | Legacy code |
|
||||
|
||||
## Real-World Examples
|
||||
|
||||
### Example 1: Startup (TEA Lite → TEA Integrated)
|
||||
|
||||
**Month 1:** TEA Lite
|
||||
```
|
||||
Team: 3 developers, no QA
|
||||
Testing: Manual only
|
||||
Decision: Start with TEA Lite
|
||||
|
||||
Result:
|
||||
- Run *framework (Playwright setup)
|
||||
- Run *automate (20 tests generated)
|
||||
- Learning TEA basics
|
||||
```
|
||||
|
||||
**Month 3:** TEA Solo
|
||||
```
|
||||
Team: Growing to 5 developers
|
||||
Testing: Automated tests exist
|
||||
Decision: Expand to TEA Solo
|
||||
|
||||
Result:
|
||||
- Add *test-design (risk assessment)
|
||||
- Add *atdd (TDD workflow)
|
||||
- Add *test-review (quality audits)
|
||||
```
|
||||
|
||||
**Month 6:** TEA Integrated
|
||||
```
|
||||
Team: 8 developers, 1 QA
|
||||
Testing: Critical to business
|
||||
Decision: Full BMad Method + TEA Integrated
|
||||
|
||||
Result:
|
||||
- Full lifecycle integration
|
||||
- Quality gates before releases
|
||||
- NFR assessment for enterprise customers
|
||||
```
|
||||
|
||||
### Example 2: Enterprise (TEA Integrated - Brownfield)
|
||||
|
||||
**Project:** Legacy banking application
|
||||
|
||||
**Challenge:**
|
||||
- 500 existing tests (50% flaky)
|
||||
- Adding new features
|
||||
- SOC 2 compliance required
|
||||
|
||||
**Model:** TEA Integrated (Brownfield)
|
||||
|
||||
**Phase 2:**
|
||||
```
|
||||
- *trace baseline → 45% coverage (lots of gaps)
|
||||
- Document current state
|
||||
```
|
||||
|
||||
**Phase 3:**
|
||||
```
|
||||
- *test-design (system) → identify regression hotspots
|
||||
- *framework → modernize test infrastructure
|
||||
- *ci → add selective testing
|
||||
```
|
||||
|
||||
**Phase 4:**
|
||||
```
|
||||
Per epic:
|
||||
- *test-design → focus on regression + new features
|
||||
- Fix top 10 flaky tests
|
||||
- *atdd for new features
|
||||
- *automate for coverage expansion
|
||||
- *test-review → track quality improvement
|
||||
- *trace → compare to baseline
|
||||
```
|
||||
|
||||
**Result after 6 months:**
|
||||
- Coverage: 45% → 85%
|
||||
- Quality score: 52 → 82
|
||||
- Flakiness: 50% → 2%
|
||||
- SOC 2 compliant (traceability + NFR evidence)
|
||||
|
||||
### Example 3: Consultancy (TEA Solo)
|
||||
|
||||
**Context:** Testing consultancy working with multiple clients
|
||||
|
||||
**Challenge:**
|
||||
- Different clients use different methodologies
|
||||
- Need consistent testing approach
|
||||
- Not always using BMad Method
|
||||
|
||||
**Model:** TEA Solo (bring to any client project)
|
||||
|
||||
**Workflow:**
|
||||
```
|
||||
Client project 1 (Scrum):
|
||||
- Import Jira stories
|
||||
- Run *test-design
|
||||
- Generate tests with *atdd/*automate
|
||||
- Deliver quality report with *test-review
|
||||
|
||||
Client project 2 (Kanban):
|
||||
- Import requirements from Notion
|
||||
- Same TEA workflow
|
||||
- Consistent quality across clients
|
||||
|
||||
Client project 3 (Ad-hoc):
|
||||
- Document requirements manually
|
||||
- Same TEA workflow
|
||||
- Same patterns, different context
|
||||
```
|
||||
|
||||
**Benefit:** Consistent testing approach regardless of client methodology.
|
||||
|
||||
## Choosing Your Model
|
||||
|
||||
### Start Here Questions
|
||||
|
||||
**Question 1:** Are you using BMad Method?
|
||||
- **No** → TEA Solo or TEA Lite or No TEA
|
||||
- **Yes** → TEA Integrated or No TEA
|
||||
|
||||
**Question 2:** Is this a new project?
|
||||
- **Yes** → TEA Integrated (Greenfield) or TEA Lite
|
||||
- **No** → TEA Integrated (Brownfield) or TEA Solo
|
||||
|
||||
**Question 3:** What's your testing maturity?
|
||||
- **Beginner** → TEA Lite
|
||||
- **Intermediate** → TEA Solo or Integrated
|
||||
- **Advanced** → TEA Integrated or No TEA (already expert)
|
||||
|
||||
**Question 4:** Do you need compliance/quality gates?
|
||||
- **Yes** → TEA Integrated (Enterprise)
|
||||
- **No** → Any model
|
||||
|
||||
**Question 5:** How much time can you invest?
|
||||
- **30 minutes** → TEA Lite
|
||||
- **Few hours** → TEA Solo
|
||||
- **Multiple days** → TEA Integrated
|
||||
|
||||
### Recommendation Matrix
|
||||
|
||||
| Your Context | Recommended Model | Alternative |
|
||||
|--------------|------------------|-------------|
|
||||
| BMad Method + new project | TEA Integrated (Greenfield) | TEA Lite (learning) |
|
||||
| BMad Method + existing code | TEA Integrated (Brownfield) | TEA Solo |
|
||||
| Non-BMad + need quality | TEA Solo | TEA Lite |
|
||||
| Just learning testing | TEA Lite | No TEA (learn basics first) |
|
||||
| Enterprise + compliance | TEA Integrated (Enterprise) | TEA Solo |
|
||||
| Established QA team | No TEA | TEA Solo (supplement) |
|
||||
|
||||
## Transitioning Between Models
|
||||
|
||||
### TEA Lite → TEA Solo
|
||||
|
||||
**When:** Outgrow beginner approach, need more workflows.
|
||||
|
||||
**Steps:**
|
||||
1. Continue using `*framework` and `*automate`
|
||||
2. Add `*test-design` for planning
|
||||
3. Add `*atdd` for TDD workflow
|
||||
4. Add `*test-review` for quality audits
|
||||
5. Add `*trace` for coverage tracking
|
||||
|
||||
**Timeline:** 2-4 weeks of gradual expansion
|
||||
|
||||
### TEA Solo → TEA Integrated
|
||||
|
||||
**When:** Adopt BMad Method, want full integration.
|
||||
|
||||
**Steps:**
|
||||
1. Install BMad Method (see installation guide)
|
||||
2. Run planning workflows (PRD, architecture)
|
||||
3. Integrate TEA into Phase 3 (system-level test design)
|
||||
4. Follow integrated lifecycle (per epic workflows)
|
||||
5. Add release gates (trace Phase 2)
|
||||
|
||||
**Timeline:** 1-2 sprints of transition
|
||||
|
||||
### TEA Integrated → TEA Solo
|
||||
|
||||
**When:** Moving away from BMad Method, keep TEA.
|
||||
|
||||
**Steps:**
|
||||
1. Export BMad artifacts (PRD, architecture, stories)
|
||||
2. Continue using TEA workflows standalone
|
||||
3. Skip BMad-specific integration
|
||||
4. Bring your own requirements to TEA
|
||||
|
||||
**Timeline:** Immediate (just skip BMad workflows)
|
||||
|
||||
## Common Patterns
|
||||
|
||||
### Pattern 1: TEA Lite for Learning, Then Choose
|
||||
|
||||
```
|
||||
Phase 1 (Week 1-2): TEA Lite
|
||||
- Learn with *automate on demo app
|
||||
- Understand TEA fundamentals
|
||||
- Low commitment
|
||||
|
||||
Phase 2 (Week 3-4): Evaluate
|
||||
- Try *test-design (planning)
|
||||
- Try *atdd (TDD)
|
||||
- See if value justifies investment
|
||||
|
||||
Phase 3 (Month 2+): Decide
|
||||
- Valuable → Expand to TEA Solo or Integrated
|
||||
- Not valuable → Stay with TEA Lite or No TEA
|
||||
```
|
||||
|
||||
### Pattern 2: TEA Solo for Quality, Skip Full Method
|
||||
|
||||
```
|
||||
Team decision:
|
||||
- Don't want full BMad Method (too heavyweight)
|
||||
- Want systematic testing (TEA benefits)
|
||||
|
||||
Approach: TEA Solo only
|
||||
- Use existing project management (Jira, Linear)
|
||||
- Use TEA for testing only
|
||||
- Get quality without methodology commitment
|
||||
```
|
||||
|
||||
### Pattern 3: Integrated for Critical, Lite for Non-Critical
|
||||
|
||||
```
|
||||
Critical features (payment, auth):
|
||||
- Full TEA Integrated workflow
|
||||
- Risk assessment, ATDD, quality gates
|
||||
- High confidence required
|
||||
|
||||
Non-critical features (UI tweaks):
|
||||
- TEA Lite or No TEA
|
||||
- Quick tests, minimal overhead
|
||||
- Move fast
|
||||
```
|
||||
|
||||
## Technical Implementation
|
||||
|
||||
Each model uses different TEA workflows. See:
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Model details
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - Workflow reference
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Setup options
|
||||
|
||||
## Related Concepts
|
||||
|
||||
**Core TEA Concepts:**
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Risk assessment in different models
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Quality across all models
|
||||
- [Knowledge Base System](/docs/explanation/tea/knowledge-base-system.md) - Consistent patterns across models
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Infrastructure in different models
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Reliability in all models
|
||||
|
||||
**Overview:**
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - 5 engagement models with cheat sheets
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - Design philosophy
|
||||
|
||||
## Practical Guides
|
||||
|
||||
**Getting Started:**
|
||||
- [TEA Lite Quickstart Tutorial](/docs/tutorials/getting-started/tea-lite-quickstart.md) - Model 3: TEA Lite
|
||||
|
||||
**Use-Case Guides:**
|
||||
- [Using TEA with Existing Tests](/docs/how-to/brownfield/use-tea-with-existing-tests.md) - Model 5: Brownfield
|
||||
- [Running TEA for Enterprise](/docs/how-to/enterprise/use-tea-for-enterprise.md) - Enterprise integration
|
||||
|
||||
**All Workflow Guides:**
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Used in TEA Solo and Integrated
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md)
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md)
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md)
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md)
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - All workflows explained
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Config per model
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - TEA Lite, TEA Solo, TEA Integrated terms
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
457
docs/explanation/tea/fixture-architecture.md
Normal file
457
docs/explanation/tea/fixture-architecture.md
Normal file
@@ -0,0 +1,457 @@
|
||||
---
|
||||
title: "Fixture Architecture Explained"
|
||||
description: Understanding TEA's pure function → fixture → composition pattern for reusable test utilities
|
||||
---
|
||||
|
||||
# Fixture Architecture Explained
|
||||
|
||||
Fixture architecture is TEA's pattern for building reusable, testable, and composable test utilities. The core principle: build pure functions first, wrap in framework fixtures second.
|
||||
|
||||
## Overview
|
||||
|
||||
**The Pattern:**
|
||||
1. Write utility as pure function (unit-testable)
|
||||
2. Wrap in framework fixture (Playwright, Cypress)
|
||||
3. Compose fixtures with mergeTests (combine capabilities)
|
||||
4. Package for reuse across projects
|
||||
|
||||
**Why this order?**
|
||||
- Pure functions are easier to test
|
||||
- Fixtures depend on framework (less portable)
|
||||
- Composition happens at fixture level
|
||||
- Reusability maximized
|
||||
|
||||
### Fixture Architecture Flow
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
flowchart TD
|
||||
Start([Testing Need]) --> Pure[Step 1: Pure Function<br/>helpers/api-request.ts]
|
||||
Pure -->|Unit testable<br/>Framework agnostic| Fixture[Step 2: Fixture Wrapper<br/>fixtures/api-request.ts]
|
||||
Fixture -->|Injects framework<br/>dependencies| Compose[Step 3: Composition<br/>fixtures/index.ts]
|
||||
Compose -->|mergeTests| Use[Step 4: Use in Tests<br/>tests/**.spec.ts]
|
||||
|
||||
Pure -.->|Can test in isolation| UnitTest[Unit Tests<br/>No framework needed]
|
||||
Fixture -.->|Reusable pattern| Other[Other Projects<br/>Package export]
|
||||
Compose -.->|Combine utilities| Multi[Multiple Fixtures<br/>One test]
|
||||
|
||||
style Pure fill:#e3f2fd,stroke:#1565c0,stroke-width:2px
|
||||
style Fixture fill:#fff3e0,stroke:#e65100,stroke-width:2px
|
||||
style Compose fill:#f3e5f5,stroke:#6a1b9a,stroke-width:2px
|
||||
style Use fill:#e8f5e9,stroke:#2e7d32,stroke-width:2px
|
||||
style UnitTest fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px
|
||||
style Other fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px
|
||||
style Multi fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px
|
||||
```
|
||||
|
||||
**Benefits at Each Step:**
|
||||
1. **Pure Function:** Testable, portable, reusable
|
||||
2. **Fixture:** Framework integration, clean API
|
||||
3. **Composition:** Combine capabilities, flexible
|
||||
4. **Usage:** Simple imports, type-safe
|
||||
|
||||
## The Problem
|
||||
|
||||
### Framework-First Approach (Common Anti-Pattern)
|
||||
|
||||
```typescript
|
||||
// ❌ Bad: Built as fixture from the start
|
||||
export const test = base.extend({
|
||||
apiRequest: async ({ request }, use) => {
|
||||
await use(async (options) => {
|
||||
const response = await request.fetch(options.url, {
|
||||
method: options.method,
|
||||
data: options.data
|
||||
});
|
||||
|
||||
if (!response.ok()) {
|
||||
throw new Error(`API request failed: ${response.status()}`);
|
||||
}
|
||||
|
||||
return response.json();
|
||||
});
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Cannot unit test (requires Playwright context)
|
||||
- Tied to framework (not reusable in other tools)
|
||||
- Hard to compose with other fixtures
|
||||
- Difficult to mock for testing the utility itself
|
||||
|
||||
### Copy-Paste Utilities
|
||||
|
||||
```typescript
|
||||
// test-1.spec.ts
|
||||
test('test 1', async ({ request }) => {
|
||||
const response = await request.post('/api/users', { data: {...} });
|
||||
const body = await response.json();
|
||||
if (!response.ok()) throw new Error('Failed');
|
||||
// ... repeated in every test
|
||||
});
|
||||
|
||||
// test-2.spec.ts
|
||||
test('test 2', async ({ request }) => {
|
||||
const response = await request.post('/api/users', { data: {...} });
|
||||
const body = await response.json();
|
||||
if (!response.ok()) throw new Error('Failed');
|
||||
// ... same code repeated
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Code duplication (violates DRY)
|
||||
- Inconsistent error handling
|
||||
- Hard to update (change 50 tests)
|
||||
- No shared behavior
|
||||
|
||||
## The Solution: Three-Step Pattern
|
||||
|
||||
### Step 1: Pure Function
|
||||
|
||||
```typescript
|
||||
// helpers/api-request.ts
|
||||
|
||||
/**
|
||||
* Make API request with automatic error handling
|
||||
* Pure function - no framework dependencies
|
||||
*/
|
||||
export async function apiRequest({
|
||||
request, // Passed in (dependency injection)
|
||||
method,
|
||||
url,
|
||||
data,
|
||||
headers = {}
|
||||
}: ApiRequestParams): Promise<ApiResponse> {
|
||||
const response = await request.fetch(url, {
|
||||
method,
|
||||
data,
|
||||
headers
|
||||
});
|
||||
|
||||
if (!response.ok()) {
|
||||
throw new Error(`API request failed: ${response.status()}`);
|
||||
}
|
||||
|
||||
return {
|
||||
status: response.status(),
|
||||
body: await response.json()
|
||||
};
|
||||
}
|
||||
|
||||
// ✅ Can unit test this function!
|
||||
describe('apiRequest', () => {
|
||||
it('should throw on non-OK response', async () => {
|
||||
const mockRequest = {
|
||||
fetch: vi.fn().mockResolvedValue({ ok: () => false, status: () => 500 })
|
||||
};
|
||||
|
||||
await expect(apiRequest({
|
||||
request: mockRequest,
|
||||
method: 'GET',
|
||||
url: '/api/test'
|
||||
})).rejects.toThrow('API request failed: 500');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Unit testable (mock dependencies)
|
||||
- Framework-agnostic (works with any HTTP client)
|
||||
- Easy to reason about (pure function)
|
||||
- Portable (can use in Node scripts, CLI tools)
|
||||
|
||||
### Step 2: Fixture Wrapper
|
||||
|
||||
```typescript
|
||||
// fixtures/api-request.ts
|
||||
import { test as base } from '@playwright/test';
|
||||
import { apiRequest as apiRequestFn } from '../helpers/api-request';
|
||||
|
||||
/**
|
||||
* Playwright fixture wrapping the pure function
|
||||
*/
|
||||
export const test = base.extend<{ apiRequest: typeof apiRequestFn }>({
|
||||
apiRequest: async ({ request }, use) => {
|
||||
// Inject framework dependency (request)
|
||||
await use((params) => apiRequestFn({ request, ...params }));
|
||||
}
|
||||
});
|
||||
|
||||
export { expect } from '@playwright/test';
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Fixture provides framework context (request)
|
||||
- Pure function handles logic
|
||||
- Clean separation of concerns
|
||||
- Can swap frameworks (Cypress, etc.) by changing wrapper only
|
||||
|
||||
### Step 3: Composition with mergeTests
|
||||
|
||||
```typescript
|
||||
// fixtures/index.ts
|
||||
import { mergeTests } from '@playwright/test';
|
||||
import { test as apiRequestTest } from './api-request';
|
||||
import { test as authSessionTest } from './auth-session';
|
||||
import { test as logTest } from './log';
|
||||
|
||||
/**
|
||||
* Compose all fixtures into one test
|
||||
*/
|
||||
export const test = mergeTests(
|
||||
apiRequestTest,
|
||||
authSessionTest,
|
||||
logTest
|
||||
);
|
||||
|
||||
export { expect } from '@playwright/test';
|
||||
```
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
// tests/profile.spec.ts
|
||||
import { test, expect } from '../support/fixtures';
|
||||
|
||||
test('should update profile', async ({ apiRequest, authToken, log }) => {
|
||||
log.info('Starting profile update test');
|
||||
|
||||
// Use API request fixture (matches pure function signature)
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
url: '/api/profile',
|
||||
data: { name: 'New Name' },
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
expect(status).toBe(200);
|
||||
expect(body.name).toBe('New Name');
|
||||
|
||||
log.info('Profile updated successfully');
|
||||
});
|
||||
```
|
||||
|
||||
**Note:** This example uses the vanilla pure function signature (`url`, `data`). Playwright Utils uses different parameter names (`path`, `body`). See [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) for the utilities API.
|
||||
|
||||
**Note:** `authToken` requires auth-session fixture setup with provider configuration. See [auth-session documentation](https://seontechnologies.github.io/playwright-utils/auth-session.html).
|
||||
|
||||
**Benefits:**
|
||||
- Use multiple fixtures in one test
|
||||
- No manual composition needed
|
||||
- Type-safe (TypeScript knows all fixture types)
|
||||
- Clean imports
|
||||
|
||||
## How It Works in TEA
|
||||
|
||||
### TEA Generates This Pattern
|
||||
|
||||
When you run `*framework` with `tea_use_playwright_utils: true`:
|
||||
|
||||
**TEA scaffolds:**
|
||||
```
|
||||
tests/
|
||||
├── support/
|
||||
│ ├── helpers/ # Pure functions
|
||||
│ │ ├── api-request.ts
|
||||
│ │ └── auth-session.ts
|
||||
│ └── fixtures/ # Framework wrappers
|
||||
│ ├── api-request.ts
|
||||
│ ├── auth-session.ts
|
||||
│ └── index.ts # Composition
|
||||
└── e2e/
|
||||
└── example.spec.ts # Uses composed fixtures
|
||||
```
|
||||
|
||||
### TEA Reviews Against This Pattern
|
||||
|
||||
When you run `*test-review`:
|
||||
|
||||
**TEA checks:**
|
||||
- Are utilities pure functions? ✓
|
||||
- Are fixtures minimal wrappers? ✓
|
||||
- Is composition used? ✓
|
||||
- Can utilities be unit tested? ✓
|
||||
|
||||
## Package Export Pattern
|
||||
|
||||
### Make Fixtures Reusable Across Projects
|
||||
|
||||
**Option 1: Build Your Own (Vanilla)**
|
||||
```json
|
||||
// package.json
|
||||
{
|
||||
"name": "@company/test-utils",
|
||||
"exports": {
|
||||
"./api-request": "./fixtures/api-request.ts",
|
||||
"./auth-session": "./fixtures/auth-session.ts",
|
||||
"./log": "./fixtures/log.ts"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test as apiTest } from '@company/test-utils/api-request';
|
||||
import { test as authTest } from '@company/test-utils/auth-session';
|
||||
import { mergeTests } from '@playwright/test';
|
||||
|
||||
export const test = mergeTests(apiTest, authTest);
|
||||
```
|
||||
|
||||
**Option 2: Use Playwright Utils (Recommended)**
|
||||
```bash
|
||||
npm install -D @seontechnologies/playwright-utils
|
||||
```
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test as base } from '@playwright/test';
|
||||
import { mergeTests } from '@playwright/test';
|
||||
import { test as apiRequestFixture } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { createAuthFixtures } from '@seontechnologies/playwright-utils/auth-session';
|
||||
|
||||
const authFixtureTest = base.extend(createAuthFixtures());
|
||||
export const test = mergeTests(apiRequestFixture, authFixtureTest);
|
||||
// Production-ready utilities, battle-tested!
|
||||
```
|
||||
|
||||
**Note:** Auth-session requires provider configuration. See [auth-session setup guide](https://seontechnologies.github.io/playwright-utils/auth-session.html).
|
||||
|
||||
**Why Playwright Utils:**
|
||||
- Already built, tested, and maintained
|
||||
- Consistent patterns across projects
|
||||
- 11 utilities available (API, auth, network, logging, files)
|
||||
- Community support and documentation
|
||||
- Regular updates and improvements
|
||||
|
||||
**When to Build Your Own:**
|
||||
- Company-specific patterns
|
||||
- Custom authentication systems
|
||||
- Unique requirements not covered by utilities
|
||||
|
||||
## Comparison: Good vs Bad Patterns
|
||||
|
||||
### Anti-Pattern: God Fixture
|
||||
|
||||
```typescript
|
||||
// ❌ Bad: Everything in one fixture
|
||||
export const test = base.extend({
|
||||
testUtils: async ({ page, request, context }, use) => {
|
||||
await use({
|
||||
// 50 different methods crammed into one fixture
|
||||
apiRequest: async (...) => { },
|
||||
login: async (...) => { },
|
||||
createUser: async (...) => { },
|
||||
deleteUser: async (...) => { },
|
||||
uploadFile: async (...) => { },
|
||||
// ... 45 more methods
|
||||
});
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Cannot test individual utilities
|
||||
- Cannot compose (all-or-nothing)
|
||||
- Cannot reuse specific utilities
|
||||
- Hard to maintain (1000+ line file)
|
||||
|
||||
### Good Pattern: Single-Concern Fixtures
|
||||
|
||||
```typescript
|
||||
// ✅ Good: One concern per fixture
|
||||
|
||||
// api-request.ts
|
||||
export const test = base.extend({ apiRequest });
|
||||
|
||||
// auth-session.ts
|
||||
export const test = base.extend({ authSession });
|
||||
|
||||
// log.ts
|
||||
export const test = base.extend({ log });
|
||||
|
||||
// Compose as needed
|
||||
import { mergeTests } from '@playwright/test';
|
||||
export const test = mergeTests(apiRequestTest, authSessionTest, logTest);
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Each fixture is unit-testable
|
||||
- Compose only what you need
|
||||
- Reuse individual fixtures
|
||||
- Easy to maintain (small files)
|
||||
|
||||
## Technical Implementation
|
||||
|
||||
For detailed fixture architecture patterns, see the knowledge base:
|
||||
- [Knowledge Base Index - Architecture & Fixtures](/docs/reference/tea/knowledge-base.md)
|
||||
- [Complete Knowledge Base Index](/docs/reference/tea/knowledge-base.md)
|
||||
|
||||
## When to Use This Pattern
|
||||
|
||||
### Always Use For:
|
||||
|
||||
**Reusable utilities:**
|
||||
- API request helpers
|
||||
- Authentication handlers
|
||||
- File operations
|
||||
- Network mocking
|
||||
|
||||
**Test infrastructure:**
|
||||
- Shared fixtures across teams
|
||||
- Packaged utilities (playwright-utils)
|
||||
- Company-wide test standards
|
||||
|
||||
### Consider Skipping For:
|
||||
|
||||
**One-off test setup:**
|
||||
```typescript
|
||||
// Simple one-time setup - inline is fine
|
||||
test.beforeEach(async ({ page }) => {
|
||||
await page.goto('/');
|
||||
await page.click('#accept-cookies');
|
||||
});
|
||||
```
|
||||
|
||||
**Test-specific helpers:**
|
||||
```typescript
|
||||
// Used in one test file only - keep local
|
||||
function createTestUser(name: string) {
|
||||
return { name, email: `${name}@test.com` };
|
||||
}
|
||||
```
|
||||
|
||||
## Related Concepts
|
||||
|
||||
**Core TEA Concepts:**
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Quality standards fixtures enforce
|
||||
- [Knowledge Base System](/docs/explanation/tea/knowledge-base-system.md) - Fixture patterns in knowledge base
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Network fixtures explained
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Fixture complexity matches risk
|
||||
|
||||
**Overview:**
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Fixture architecture in workflows
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - Why fixtures matter
|
||||
|
||||
## Practical Guides
|
||||
|
||||
**Setup Guides:**
|
||||
- [How to Set Up Test Framework](/docs/how-to/workflows/setup-test-framework.md) - TEA scaffolds fixtures
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - Production-ready fixtures
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Using fixtures in tests
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Fixture composition examples
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - *framework command
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Fixture architecture fragments
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - Fixture architecture term
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
554
docs/explanation/tea/knowledge-base-system.md
Normal file
554
docs/explanation/tea/knowledge-base-system.md
Normal file
@@ -0,0 +1,554 @@
|
||||
---
|
||||
title: "Knowledge Base System Explained"
|
||||
description: Understanding how TEA uses tea-index.csv for context engineering and consistent test quality
|
||||
---
|
||||
|
||||
# Knowledge Base System Explained
|
||||
|
||||
TEA's knowledge base system is how context engineering works - automatically loading domain-specific standards into AI context so tests are consistently high-quality regardless of prompt variation.
|
||||
|
||||
## Overview
|
||||
|
||||
**The Problem:** AI without context produces inconsistent results.
|
||||
|
||||
**Traditional approach:**
|
||||
```
|
||||
User: "Write tests for login"
|
||||
AI: [Generates tests with random quality]
|
||||
- Sometimes uses hard waits
|
||||
- Sometimes uses good patterns
|
||||
- Inconsistent across sessions
|
||||
- Quality depends on prompt
|
||||
```
|
||||
|
||||
**TEA with knowledge base:**
|
||||
```
|
||||
User: "Write tests for login"
|
||||
TEA: [Loads test-quality.md, network-first.md, auth-session.md]
|
||||
TEA: [Generates tests following established patterns]
|
||||
- Always uses network-first patterns
|
||||
- Always uses proper fixtures
|
||||
- Consistent across all sessions
|
||||
- Quality independent of prompt
|
||||
```
|
||||
|
||||
**Result:** Systematic quality, not random chance.
|
||||
|
||||
## The Problem
|
||||
|
||||
### Prompt-Driven Testing = Inconsistency
|
||||
|
||||
**Session 1:**
|
||||
```
|
||||
User: "Write tests for profile editing"
|
||||
|
||||
AI: [No context loaded]
|
||||
// Generates test with hard waits
|
||||
await page.waitForTimeout(3000);
|
||||
```
|
||||
|
||||
**Session 2:**
|
||||
```
|
||||
User: "Write comprehensive tests for profile editing with best practices"
|
||||
|
||||
AI: [Still no systematic context]
|
||||
// Generates test with some improvements, but still issues
|
||||
await page.waitForSelector('.success', { timeout: 10000 });
|
||||
```
|
||||
|
||||
**Session 3:**
|
||||
```
|
||||
User: "Write tests using network-first patterns and proper fixtures"
|
||||
|
||||
AI: [Better prompt, but still reinventing patterns]
|
||||
// Generates test with network-first, but inconsistent with other tests
|
||||
```
|
||||
|
||||
**Problem:** Quality depends on prompt engineering skill, no consistency.
|
||||
|
||||
### Knowledge Drift
|
||||
|
||||
Without a knowledge base:
|
||||
- Team A uses pattern X
|
||||
- Team B uses pattern Y
|
||||
- Both work, but inconsistent
|
||||
- No single source of truth
|
||||
- Patterns drift over time
|
||||
|
||||
## The Solution: tea-index.csv Manifest
|
||||
|
||||
### How It Works
|
||||
|
||||
**1. Manifest Defines Fragments**
|
||||
|
||||
`src/modules/bmm/testarch/tea-index.csv`:
|
||||
```csv
|
||||
id,name,description,tags,fragment_file
|
||||
test-quality,Test Quality,Execution limits and isolation rules,quality;standards,knowledge/test-quality.md
|
||||
network-first,Network-First Safeguards,Intercept-before-navigate workflow,network;stability,knowledge/network-first.md
|
||||
fixture-architecture,Fixture Architecture,Composable fixture patterns,fixtures;architecture,knowledge/fixture-architecture.md
|
||||
```
|
||||
|
||||
**2. Workflow Loads Relevant Fragments**
|
||||
|
||||
When user runs `*atdd`:
|
||||
```
|
||||
TEA reads tea-index.csv
|
||||
Identifies fragments needed for ATDD:
|
||||
- test-quality.md (quality standards)
|
||||
- network-first.md (avoid flakiness)
|
||||
- component-tdd.md (TDD patterns)
|
||||
- fixture-architecture.md (reusable fixtures)
|
||||
- data-factories.md (test data)
|
||||
|
||||
Loads only these 5 fragments (not all 33)
|
||||
Generates tests following these patterns
|
||||
```
|
||||
|
||||
**3. Consistent Output**
|
||||
|
||||
Every time `*atdd` runs:
|
||||
- Same fragments loaded
|
||||
- Same patterns applied
|
||||
- Same quality standards
|
||||
- Consistent test structure
|
||||
|
||||
**Result:** Tests look like they were written by the same expert, every time.
|
||||
|
||||
### Knowledge Base Loading Diagram
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
flowchart TD
|
||||
User([User: *atdd]) --> Workflow[TEA Workflow<br/>Triggered]
|
||||
Workflow --> Read[Read Manifest<br/>tea-index.csv]
|
||||
|
||||
Read --> Identify{Identify Relevant<br/>Fragments for ATDD}
|
||||
|
||||
Identify -->|Needed| L1[✓ test-quality.md]
|
||||
Identify -->|Needed| L2[✓ network-first.md]
|
||||
Identify -->|Needed| L3[✓ component-tdd.md]
|
||||
Identify -->|Needed| L4[✓ data-factories.md]
|
||||
Identify -->|Needed| L5[✓ fixture-architecture.md]
|
||||
|
||||
Identify -.->|Skip| S1[✗ contract-testing.md]
|
||||
Identify -.->|Skip| S2[✗ burn-in.md]
|
||||
Identify -.->|Skip| S3[+ 26 other fragments]
|
||||
|
||||
L1 --> Context[AI Context<br/>5 fragments loaded]
|
||||
L2 --> Context
|
||||
L3 --> Context
|
||||
L4 --> Context
|
||||
L5 --> Context
|
||||
|
||||
Context --> Gen[Generate Tests<br/>Following patterns]
|
||||
Gen --> Out([Consistent Output<br/>Same quality every time])
|
||||
|
||||
style User fill:#e3f2fd,stroke:#1565c0,stroke-width:2px
|
||||
style Read fill:#fff3e0,stroke:#e65100,stroke-width:2px
|
||||
style L1 fill:#c8e6c9,stroke:#2e7d32,stroke-width:2px
|
||||
style L2 fill:#c8e6c9,stroke:#2e7d32,stroke-width:2px
|
||||
style L3 fill:#c8e6c9,stroke:#2e7d32,stroke-width:2px
|
||||
style L4 fill:#c8e6c9,stroke:#2e7d32,stroke-width:2px
|
||||
style L5 fill:#c8e6c9,stroke:#2e7d32,stroke-width:2px
|
||||
style S1 fill:#e0e0e0,stroke:#616161,stroke-width:1px
|
||||
style S2 fill:#e0e0e0,stroke:#616161,stroke-width:1px
|
||||
style S3 fill:#e0e0e0,stroke:#616161,stroke-width:1px
|
||||
style Context fill:#f3e5f5,stroke:#6a1b9a,stroke-width:3px
|
||||
style Out fill:#4caf50,stroke:#1b5e20,stroke-width:3px,color:#fff
|
||||
```
|
||||
|
||||
## Fragment Structure
|
||||
|
||||
### Anatomy of a Fragment
|
||||
|
||||
Each fragment follows this structure:
|
||||
|
||||
```markdown
|
||||
# Fragment Name
|
||||
|
||||
## Principle
|
||||
[One sentence - what is this pattern?]
|
||||
|
||||
## Rationale
|
||||
[Why use this instead of alternatives?]
|
||||
Why this pattern exists
|
||||
Problems it solves
|
||||
Benefits it provides
|
||||
|
||||
## Pattern Examples
|
||||
|
||||
### Example 1: Basic Usage
|
||||
```code
|
||||
[Runnable code example]
|
||||
```
|
||||
[Explanation of example]
|
||||
|
||||
### Example 2: Advanced Pattern
|
||||
```code
|
||||
[More complex example]
|
||||
```
|
||||
[Explanation]
|
||||
|
||||
## Anti-Patterns
|
||||
|
||||
### Don't Do This
|
||||
```code
|
||||
[Bad code example]
|
||||
```
|
||||
[Why it's bad]
|
||||
[What breaks]
|
||||
|
||||
## Related Patterns
|
||||
- [Link to related fragment]
|
||||
```
|
||||
|
||||
<!-- markdownlint-disable MD024 -->
|
||||
### Example: test-quality.md Fragment
|
||||
|
||||
```markdown
|
||||
# Test Quality
|
||||
|
||||
## Principle
|
||||
Tests must be deterministic, isolated, explicit, focused, and fast.
|
||||
|
||||
## Rationale
|
||||
Tests that fail randomly, depend on each other, or take too long lose team trust.
|
||||
[... detailed explanation ...]
|
||||
|
||||
## Pattern Examples
|
||||
|
||||
### Example 1: Deterministic Test
|
||||
```typescript
|
||||
// ✅ Wait for actual response, not timeout
|
||||
const promise = page.waitForResponse(matcher);
|
||||
await page.click('button');
|
||||
await promise;
|
||||
```
|
||||
|
||||
### Example 2: Isolated Test
|
||||
```typescript
|
||||
// ✅ Self-cleaning test
|
||||
test('test', async ({ page }) => {
|
||||
const userId = await createTestUser();
|
||||
// ... test logic ...
|
||||
await deleteTestUser(userId); // Cleanup
|
||||
});
|
||||
```
|
||||
|
||||
## Anti-Patterns
|
||||
|
||||
### Hard Waits
|
||||
```typescript
|
||||
// ❌ Non-deterministic
|
||||
await page.waitForTimeout(3000);
|
||||
```
|
||||
[Why this causes flakiness]
|
||||
```
|
||||
|
||||
**Total:** 24.5 KB, 12 code examples
|
||||
<!-- markdownlint-enable MD024 -->
|
||||
|
||||
## How TEA Uses the Knowledge Base
|
||||
|
||||
### Workflow-Specific Loading
|
||||
|
||||
**Different workflows load different fragments:**
|
||||
|
||||
| Workflow | Fragments Loaded | Purpose |
|
||||
|----------|-----------------|---------|
|
||||
| `*framework` | fixture-architecture, playwright-config, fixtures-composition | Infrastructure patterns |
|
||||
| `*test-design` | test-quality, test-priorities-matrix, risk-governance | Planning standards |
|
||||
| `*atdd` | test-quality, component-tdd, network-first, data-factories | TDD patterns |
|
||||
| `*automate` | test-quality, test-levels-framework, selector-resilience | Comprehensive generation |
|
||||
| `*test-review` | All quality/resilience/debugging fragments | Full audit patterns |
|
||||
| `*ci` | ci-burn-in, burn-in, selective-testing | CI/CD optimization |
|
||||
|
||||
**Benefit:** Only load what's needed (focused context, no bloat).
|
||||
|
||||
### Dynamic Fragment Selection
|
||||
|
||||
TEA doesn't load all 33 fragments at once:
|
||||
|
||||
```
|
||||
User runs: *atdd for authentication feature
|
||||
|
||||
TEA analyzes context:
|
||||
- Feature type: Authentication
|
||||
- Relevant fragments:
|
||||
- test-quality.md (always loaded)
|
||||
- auth-session.md (auth patterns)
|
||||
- network-first.md (avoid flakiness)
|
||||
- email-auth.md (if email-based auth)
|
||||
- data-factories.md (test users)
|
||||
|
||||
Skips:
|
||||
- contract-testing.md (not relevant)
|
||||
- feature-flags.md (not relevant)
|
||||
- file-utils.md (not relevant)
|
||||
|
||||
Result: 5 relevant fragments loaded, 28 skipped
|
||||
```
|
||||
|
||||
**Benefit:** Focused context = better results, lower token usage.
|
||||
|
||||
## Context Engineering in Practice
|
||||
|
||||
### Example: Consistent Test Generation
|
||||
|
||||
**Without Knowledge Base (Vanilla Playwright, Random Quality):**
|
||||
```
|
||||
Session 1: User runs *atdd
|
||||
AI: [Guesses patterns from general knowledge]
|
||||
|
||||
Generated:
|
||||
test('api test', async ({ request }) => {
|
||||
const response = await request.get('/api/users');
|
||||
await page.waitForTimeout(2000); // Hard wait
|
||||
const users = await response.json();
|
||||
// Random quality
|
||||
});
|
||||
|
||||
Session 2: User runs *atdd (different day)
|
||||
AI: [Different random patterns]
|
||||
|
||||
Generated:
|
||||
test('api test', async ({ request }) => {
|
||||
const response = await request.get('/api/users');
|
||||
const users = await response.json();
|
||||
// Better but inconsistent
|
||||
});
|
||||
|
||||
Result: Inconsistent quality, random patterns
|
||||
```
|
||||
|
||||
**With Knowledge Base (TEA + Playwright Utils):**
|
||||
```
|
||||
Session 1: User runs *atdd
|
||||
TEA: [Loads test-quality.md, network-first.md, api-request.md from tea-index.csv]
|
||||
|
||||
Generated:
|
||||
import { test } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
|
||||
test('should fetch users', async ({ apiRequest }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/users'
|
||||
}).validateSchema(UsersSchema); // Chained validation
|
||||
|
||||
expect(status).toBe(200);
|
||||
expect(body).toBeInstanceOf(Array);
|
||||
});
|
||||
|
||||
Session 2: User runs *atdd (different day)
|
||||
TEA: [Loads same fragments from tea-index.csv]
|
||||
|
||||
Generated: Identical pattern, same quality
|
||||
|
||||
Result: Systematic quality, established patterns (ALWAYS uses apiRequest utility when playwright-utils enabled)
|
||||
```
|
||||
|
||||
**Key Difference:**
|
||||
- **Without KB:** Random patterns, inconsistent APIs
|
||||
- **With KB:** Always uses `apiRequest` utility, always validates schemas, always returns `{ status, body }`
|
||||
|
||||
### Example: Test Review Consistency
|
||||
|
||||
**Without Knowledge Base:**
|
||||
```
|
||||
*test-review session 1:
|
||||
"This test looks okay" [50 issues missed]
|
||||
|
||||
*test-review session 2:
|
||||
"This test has some issues" [Different issues flagged]
|
||||
|
||||
Result: Inconsistent feedback
|
||||
```
|
||||
|
||||
**With Knowledge Base:**
|
||||
```
|
||||
*test-review session 1:
|
||||
[Loads all quality fragments]
|
||||
Flags: 12 hard waits, 5 conditionals (based on test-quality.md)
|
||||
|
||||
*test-review session 2:
|
||||
[Loads same fragments]
|
||||
Flags: Same issues with same explanations
|
||||
|
||||
Result: Consistent, reliable feedback
|
||||
```
|
||||
|
||||
## Maintaining the Knowledge Base
|
||||
|
||||
### When to Add a Fragment
|
||||
|
||||
**Good reasons:**
|
||||
- Pattern is used across multiple workflows
|
||||
- Standard is non-obvious (needs documentation)
|
||||
- Team asks "how should we handle X?" repeatedly
|
||||
- New tool integration (e.g., new testing library)
|
||||
|
||||
**Bad reasons:**
|
||||
- One-off pattern (document in test file instead)
|
||||
- Obvious pattern (everyone knows this)
|
||||
- Experimental (not proven yet)
|
||||
|
||||
### Fragment Quality Standards
|
||||
|
||||
**Good fragment:**
|
||||
- Principle stated in one sentence
|
||||
- Rationale explains why clearly
|
||||
- 3+ pattern examples with code
|
||||
- Anti-patterns shown (what not to do)
|
||||
- Self-contained (minimal dependencies)
|
||||
|
||||
**Example size:** 10-30 KB optimal
|
||||
|
||||
### Updating Existing Fragments
|
||||
|
||||
**When to update:**
|
||||
- Pattern evolved (better approach discovered)
|
||||
- Tool updated (new Playwright API)
|
||||
- Team feedback (pattern unclear)
|
||||
- Bug in example code
|
||||
|
||||
**How to update:**
|
||||
1. Edit fragment markdown file
|
||||
2. Update examples
|
||||
3. Test with affected workflows
|
||||
4. Ensure no breaking changes
|
||||
|
||||
**No need to update tea-index.csv** unless description/tags change.
|
||||
|
||||
## Benefits of Knowledge Base System
|
||||
|
||||
### 1. Consistency
|
||||
|
||||
**Before:** Test quality varies by who wrote it
|
||||
**After:** All tests follow same patterns (TEA-generated or reviewed)
|
||||
|
||||
### 2. Onboarding
|
||||
|
||||
**Before:** New team member reads 20 documents, asks 50 questions
|
||||
**After:** New team member runs `*atdd`, sees patterns in generated code, learns by example
|
||||
|
||||
### 3. Quality Gates
|
||||
|
||||
**Before:** "Is this test good?" → subjective opinion
|
||||
**After:** "*test-review" → objective score against knowledge base
|
||||
|
||||
### 4. Pattern Evolution
|
||||
|
||||
**Before:** Update tests manually across 100 files
|
||||
**After:** Update fragment once, all new tests use new pattern
|
||||
|
||||
### 5. Cross-Project Reuse
|
||||
|
||||
**Before:** Reinvent patterns for each project
|
||||
**After:** Same fragments across all BMad projects (consistency at scale)
|
||||
|
||||
## Comparison: With vs Without Knowledge Base
|
||||
|
||||
### Scenario: Testing Async Background Job
|
||||
|
||||
**Without Knowledge Base:**
|
||||
|
||||
Developer 1:
|
||||
```typescript
|
||||
// Uses hard wait
|
||||
await page.click('button');
|
||||
await page.waitForTimeout(10000); // Hope job finishes
|
||||
```
|
||||
|
||||
Developer 2:
|
||||
```typescript
|
||||
// Uses polling
|
||||
await page.click('button');
|
||||
for (let i = 0; i < 10; i++) {
|
||||
const status = await page.locator('.status').textContent();
|
||||
if (status === 'complete') break;
|
||||
await page.waitForTimeout(1000);
|
||||
}
|
||||
```
|
||||
|
||||
Developer 3:
|
||||
```typescript
|
||||
// Uses waitForSelector
|
||||
await page.click('button');
|
||||
await page.waitForSelector('.success', { timeout: 30000 });
|
||||
```
|
||||
|
||||
**Result:** 3 different patterns, all suboptimal.
|
||||
|
||||
**With Knowledge Base (recurse.md fragment):**
|
||||
|
||||
All developers:
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('job completion', async ({ apiRequest, recurse }) => {
|
||||
// Start async job
|
||||
const { body: job } = await apiRequest({
|
||||
method: 'POST',
|
||||
path: '/api/jobs'
|
||||
});
|
||||
|
||||
// Poll until complete (correct API: command, predicate, options)
|
||||
const result = await recurse(
|
||||
() => apiRequest({ method: 'GET', path: `/api/jobs/${job.id}` }),
|
||||
(response) => response.body.status === 'completed', // response.body from apiRequest
|
||||
{
|
||||
timeout: 30000,
|
||||
interval: 2000,
|
||||
log: 'Waiting for job to complete'
|
||||
}
|
||||
);
|
||||
|
||||
expect(result.body.status).toBe('completed');
|
||||
});
|
||||
```
|
||||
|
||||
**Result:** Consistent pattern using correct playwright-utils API (command, predicate, options).
|
||||
|
||||
## Technical Implementation
|
||||
|
||||
For details on the knowledge base index, see:
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md)
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md)
|
||||
|
||||
## Related Concepts
|
||||
|
||||
**Core TEA Concepts:**
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Standards in knowledge base
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Risk patterns in knowledge base
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - Knowledge base across all models
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Fixture patterns in knowledge base
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Network patterns in knowledge base
|
||||
|
||||
**Overview:**
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Knowledge base in workflows
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - **Foundation: Context engineering philosophy** (why knowledge base solves AI test problems)
|
||||
|
||||
## Practical Guides
|
||||
|
||||
**All Workflow Guides Use Knowledge Base:**
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md)
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md)
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md)
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md)
|
||||
|
||||
**Integration:**
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - PW-Utils in knowledge base
|
||||
|
||||
## Reference
|
||||
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Complete fragment index
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - Which workflows load which fragments
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Config affects fragment loading
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - Context engineering, knowledge fragment terms
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
853
docs/explanation/tea/network-first-patterns.md
Normal file
853
docs/explanation/tea/network-first-patterns.md
Normal file
@@ -0,0 +1,853 @@
|
||||
---
|
||||
title: "Network-First Patterns Explained"
|
||||
description: Understanding how TEA eliminates test flakiness by waiting for actual network responses
|
||||
---
|
||||
|
||||
# Network-First Patterns Explained
|
||||
|
||||
Network-first patterns are TEA's solution to test flakiness. Instead of guessing how long to wait with fixed timeouts, wait for the actual network event that causes UI changes.
|
||||
|
||||
## Overview
|
||||
|
||||
**The Core Principle:**
|
||||
UI changes because APIs respond. Wait for the API response, not an arbitrary timeout.
|
||||
|
||||
**Traditional approach:**
|
||||
```typescript
|
||||
await page.click('button');
|
||||
await page.waitForTimeout(3000); // Hope 3 seconds is enough
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
```
|
||||
|
||||
**Network-first approach:**
|
||||
```typescript
|
||||
const responsePromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/submit') && resp.ok()
|
||||
);
|
||||
await page.click('button');
|
||||
await responsePromise; // Wait for actual response
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
```
|
||||
|
||||
**Result:** Deterministic tests that wait exactly as long as needed.
|
||||
|
||||
## The Problem
|
||||
|
||||
### Hard Waits Create Flakiness
|
||||
|
||||
```typescript
|
||||
// ❌ The flaky test pattern
|
||||
test('should submit form', async ({ page }) => {
|
||||
await page.fill('#name', 'Test User');
|
||||
await page.click('button[type="submit"]');
|
||||
|
||||
await page.waitForTimeout(2000); // Wait 2 seconds
|
||||
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Why this fails:**
|
||||
- **Fast network:** Wastes 1.5 seconds waiting
|
||||
- **Slow network:** Not enough time, test fails
|
||||
- **CI environment:** Slower than local, fails randomly
|
||||
- **Under load:** API takes 3 seconds, test fails
|
||||
|
||||
**Result:** "Works on my machine" syndrome, flaky CI.
|
||||
|
||||
### The Timeout Escalation Trap
|
||||
|
||||
```typescript
|
||||
// Developer sees flaky test
|
||||
await page.waitForTimeout(2000); // Failed in CI
|
||||
|
||||
// Increases timeout
|
||||
await page.waitForTimeout(5000); // Still fails sometimes
|
||||
|
||||
// Increases again
|
||||
await page.waitForTimeout(10000); // Now it passes... slowly
|
||||
|
||||
// Problem: Now EVERY test waits 10 seconds
|
||||
// Suite that took 5 minutes now takes 30 minutes
|
||||
```
|
||||
|
||||
**Result:** Slow, still-flaky tests.
|
||||
|
||||
### Race Conditions
|
||||
|
||||
```typescript
|
||||
// ❌ Navigate-then-wait race condition
|
||||
test('should load dashboard data', async ({ page }) => {
|
||||
await page.goto('/dashboard'); // Navigation starts
|
||||
|
||||
// Race condition! API might not have responded yet
|
||||
await expect(page.locator('.data-table')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**What happens:**
|
||||
1. `goto()` starts navigation
|
||||
2. Page loads HTML
|
||||
3. JavaScript requests `/api/dashboard`
|
||||
4. Test checks for `.data-table` BEFORE API responds
|
||||
5. Test fails intermittently
|
||||
|
||||
**Result:** "Sometimes it works, sometimes it doesn't."
|
||||
|
||||
## The Solution: Intercept-Before-Navigate
|
||||
|
||||
### Wait for Response Before Asserting
|
||||
|
||||
```typescript
|
||||
// ✅ Good: Network-first pattern
|
||||
test('should load dashboard data', async ({ page }) => {
|
||||
// Set up promise BEFORE navigation
|
||||
const dashboardPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/dashboard') && resp.ok()
|
||||
);
|
||||
|
||||
// Navigate
|
||||
await page.goto('/dashboard');
|
||||
|
||||
// Wait for API response
|
||||
const response = await dashboardPromise;
|
||||
const data = await response.json();
|
||||
|
||||
// Now assert UI
|
||||
await expect(page.locator('.data-table')).toBeVisible();
|
||||
await expect(page.locator('.data-table tr')).toHaveCount(data.items.length);
|
||||
});
|
||||
```
|
||||
|
||||
**Why this works:**
|
||||
- Wait set up BEFORE navigation (no race)
|
||||
- Wait for actual API response (deterministic)
|
||||
- No fixed timeout (fast when API is fast)
|
||||
- Validates API response (catch backend errors)
|
||||
|
||||
**With Playwright Utils (Even Cleaner):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('should load dashboard data', async ({ page, interceptNetworkCall }) => {
|
||||
// Set up interception BEFORE navigation
|
||||
const dashboardCall = interceptNetworkCall({
|
||||
method: 'GET',
|
||||
url: '**/api/dashboard'
|
||||
});
|
||||
|
||||
// Navigate
|
||||
await page.goto('/dashboard');
|
||||
|
||||
// Wait for API response (automatic JSON parsing)
|
||||
const { status, responseJson: data } = await dashboardCall;
|
||||
|
||||
// Validate API response
|
||||
expect(status).toBe(200);
|
||||
expect(data.items).toBeDefined();
|
||||
|
||||
// Assert UI matches API data
|
||||
await expect(page.locator('.data-table')).toBeVisible();
|
||||
await expect(page.locator('.data-table tr')).toHaveCount(data.items.length);
|
||||
});
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- Automatic JSON parsing (no `await response.json()`)
|
||||
- Returns `{ status, responseJson, requestJson }` structure
|
||||
- Cleaner API (no need to check `resp.ok()`)
|
||||
- Same intercept-before-navigate pattern
|
||||
|
||||
### Intercept-Before-Navigate Pattern
|
||||
|
||||
**Key insight:** Set up wait BEFORE triggering the action.
|
||||
|
||||
```typescript
|
||||
// ✅ Pattern: Intercept → Action → Await
|
||||
|
||||
// 1. Intercept (set up wait)
|
||||
const promise = page.waitForResponse(matcher);
|
||||
|
||||
// 2. Action (trigger request)
|
||||
await page.click('button');
|
||||
|
||||
// 3. Await (wait for actual response)
|
||||
await promise;
|
||||
```
|
||||
|
||||
**Why this order:**
|
||||
- `waitForResponse()` starts listening immediately
|
||||
- Then trigger the action that makes the request
|
||||
- Then wait for the promise to resolve
|
||||
- No race condition possible
|
||||
|
||||
#### Intercept-Before-Navigate Flow
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
sequenceDiagram
|
||||
participant Test
|
||||
participant Playwright
|
||||
participant Browser
|
||||
participant API
|
||||
|
||||
rect rgb(200, 230, 201)
|
||||
Note over Test,Playwright: ✅ CORRECT: Intercept First
|
||||
Test->>Playwright: 1. waitForResponse(matcher)
|
||||
Note over Playwright: Starts listening for response
|
||||
Test->>Browser: 2. click('button')
|
||||
Browser->>API: 3. POST /api/submit
|
||||
API-->>Browser: 4. 200 OK {success: true}
|
||||
Browser-->>Playwright: 5. Response captured
|
||||
Test->>Playwright: 6. await promise
|
||||
Playwright-->>Test: 7. Returns response
|
||||
Note over Test: No race condition!
|
||||
end
|
||||
|
||||
rect rgb(255, 205, 210)
|
||||
Note over Test,API: ❌ WRONG: Action First
|
||||
Test->>Browser: 1. click('button')
|
||||
Browser->>API: 2. POST /api/submit
|
||||
API-->>Browser: 3. 200 OK (already happened!)
|
||||
Test->>Playwright: 4. waitForResponse(matcher)
|
||||
Note over Test,Playwright: Too late - response already occurred
|
||||
Note over Test: Race condition! Test hangs or fails
|
||||
end
|
||||
```
|
||||
|
||||
**Correct Order (Green):**
|
||||
1. Set up listener (`waitForResponse`)
|
||||
2. Trigger action (`click`)
|
||||
3. Wait for response (`await promise`)
|
||||
|
||||
**Wrong Order (Red):**
|
||||
1. Trigger action first
|
||||
2. Set up listener too late
|
||||
3. Response already happened - missed!
|
||||
|
||||
## How It Works in TEA
|
||||
|
||||
### TEA Generates Network-First Tests
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// When you run *atdd or *automate, TEA generates:
|
||||
|
||||
test('should create user', async ({ page }) => {
|
||||
// TEA automatically includes network wait
|
||||
const createUserPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/users') &&
|
||||
resp.request().method() === 'POST' &&
|
||||
resp.ok()
|
||||
);
|
||||
|
||||
await page.fill('#name', 'Test User');
|
||||
await page.click('button[type="submit"]');
|
||||
|
||||
const response = await createUserPromise;
|
||||
const user = await response.json();
|
||||
|
||||
// Validate both API and UI
|
||||
expect(user.id).toBeDefined();
|
||||
await expect(page.locator('.success')).toContainText(user.name);
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils (if `tea_use_playwright_utils: true`):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('should create user', async ({ page, interceptNetworkCall }) => {
|
||||
// TEA uses interceptNetworkCall for cleaner interception
|
||||
const createUserCall = interceptNetworkCall({
|
||||
method: 'POST',
|
||||
url: '**/api/users'
|
||||
});
|
||||
|
||||
await page.getByLabel('Name').fill('Test User');
|
||||
await page.getByRole('button', { name: 'Submit' }).click();
|
||||
|
||||
// Wait for response (automatic JSON parsing)
|
||||
const { status, responseJson: user } = await createUserCall;
|
||||
|
||||
// Validate both API and UI
|
||||
expect(status).toBe(201);
|
||||
expect(user.id).toBeDefined();
|
||||
await expect(page.locator('.success')).toContainText(user.name);
|
||||
});
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- Automatic JSON parsing (`responseJson` ready to use)
|
||||
- No manual `await response.json()`
|
||||
- Returns `{ status, responseJson }` structure
|
||||
- Cleaner, more readable code
|
||||
|
||||
### TEA Reviews for Hard Waits
|
||||
|
||||
When you run `*test-review`:
|
||||
|
||||
```markdown
|
||||
## Critical Issue: Hard Wait Detected
|
||||
|
||||
**File:** tests/e2e/submit.spec.ts:45
|
||||
**Issue:** Using `page.waitForTimeout(3000)`
|
||||
**Severity:** Critical (causes flakiness)
|
||||
|
||||
**Current Code:**
|
||||
```typescript
|
||||
await page.click('button');
|
||||
await page.waitForTimeout(3000); // ❌
|
||||
```
|
||||
|
||||
**Fix:**
|
||||
```typescript
|
||||
const responsePromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/submit') && resp.ok()
|
||||
);
|
||||
await page.click('button');
|
||||
await responsePromise; // ✅
|
||||
```
|
||||
|
||||
**Why:** Hard waits are non-deterministic. Use network-first patterns.
|
||||
```
|
||||
|
||||
## Pattern Variations
|
||||
|
||||
### Basic Response Wait
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// Wait for any successful response
|
||||
const promise = page.waitForResponse(resp => resp.ok());
|
||||
await page.click('button');
|
||||
await promise;
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('basic wait', async ({ page, interceptNetworkCall }) => {
|
||||
const responseCall = interceptNetworkCall({ url: '**' }); // Match any
|
||||
await page.click('button');
|
||||
const { status } = await responseCall;
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### Specific URL Match
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// Wait for specific endpoint
|
||||
const promise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/users/123')
|
||||
);
|
||||
await page.goto('/user/123');
|
||||
await promise;
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
test('specific URL', async ({ page, interceptNetworkCall }) => {
|
||||
const userCall = interceptNetworkCall({ url: '**/api/users/123' });
|
||||
await page.goto('/user/123');
|
||||
const { status, responseJson } = await userCall;
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### Method + Status Match
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// Wait for POST that returns 201
|
||||
const promise = page.waitForResponse(
|
||||
resp =>
|
||||
resp.url().includes('/api/users') &&
|
||||
resp.request().method() === 'POST' &&
|
||||
resp.status() === 201
|
||||
);
|
||||
await page.click('button[type="submit"]');
|
||||
await promise;
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
test('method and status', async ({ page, interceptNetworkCall }) => {
|
||||
const createCall = interceptNetworkCall({
|
||||
method: 'POST',
|
||||
url: '**/api/users'
|
||||
});
|
||||
await page.click('button[type="submit"]');
|
||||
const { status, responseJson } = await createCall;
|
||||
expect(status).toBe(201); // Explicit status check
|
||||
});
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### Multiple Responses
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// Wait for multiple API calls
|
||||
const [usersResp, postsResp] = await Promise.all([
|
||||
page.waitForResponse(resp => resp.url().includes('/api/users')),
|
||||
page.waitForResponse(resp => resp.url().includes('/api/posts')),
|
||||
page.goto('/dashboard') // Triggers both requests
|
||||
]);
|
||||
|
||||
const users = await usersResp.json();
|
||||
const posts = await postsResp.json();
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
test('multiple responses', async ({ page, interceptNetworkCall }) => {
|
||||
const usersCall = interceptNetworkCall({ url: '**/api/users' });
|
||||
const postsCall = interceptNetworkCall({ url: '**/api/posts' });
|
||||
|
||||
await page.goto('/dashboard'); // Triggers both
|
||||
|
||||
const [{ responseJson: users }, { responseJson: posts }] = await Promise.all([
|
||||
usersCall,
|
||||
postsCall
|
||||
]);
|
||||
|
||||
expect(users).toBeInstanceOf(Array);
|
||||
expect(posts).toBeInstanceOf(Array);
|
||||
});
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### Validate Response Data
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// Verify API response before asserting UI
|
||||
const promise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/checkout') && resp.ok()
|
||||
);
|
||||
|
||||
await page.click('button:has-text("Complete Order")');
|
||||
|
||||
const response = await promise;
|
||||
const order = await response.json();
|
||||
|
||||
// Response validation
|
||||
expect(order.status).toBe('confirmed');
|
||||
expect(order.total).toBeGreaterThan(0);
|
||||
|
||||
// UI validation
|
||||
await expect(page.locator('.order-confirmation')).toContainText(order.id);
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
test('validate response data', async ({ page, interceptNetworkCall }) => {
|
||||
const checkoutCall = interceptNetworkCall({
|
||||
method: 'POST',
|
||||
url: '**/api/checkout'
|
||||
});
|
||||
|
||||
await page.click('button:has-text("Complete Order")');
|
||||
|
||||
const { status, responseJson: order } = await checkoutCall;
|
||||
|
||||
// Response validation (automatic JSON parsing)
|
||||
expect(status).toBe(200);
|
||||
expect(order.status).toBe('confirmed');
|
||||
expect(order.total).toBeGreaterThan(0);
|
||||
|
||||
// UI validation
|
||||
await expect(page.locator('.order-confirmation')).toContainText(order.id);
|
||||
});
|
||||
```
|
||||
|
||||
## Advanced Patterns
|
||||
|
||||
### HAR Recording for Offline Testing
|
||||
|
||||
**Vanilla Playwright (Manual HAR Handling):**
|
||||
|
||||
```typescript
|
||||
// First run: Record mode (saves HAR file)
|
||||
test('offline testing - RECORD', async ({ page, context }) => {
|
||||
// Record mode: Save network traffic to HAR
|
||||
await context.routeFromHAR('./hars/dashboard.har', {
|
||||
url: '**/api/**',
|
||||
update: true // Update HAR file
|
||||
});
|
||||
|
||||
await page.goto('/dashboard');
|
||||
// All network traffic saved to dashboard.har
|
||||
});
|
||||
|
||||
// Subsequent runs: Playback mode (uses saved HAR)
|
||||
test('offline testing - PLAYBACK', async ({ page, context }) => {
|
||||
// Playback mode: Use saved network traffic
|
||||
await context.routeFromHAR('./hars/dashboard.har', {
|
||||
url: '**/api/**',
|
||||
update: false // Use existing HAR, no network calls
|
||||
});
|
||||
|
||||
await page.goto('/dashboard');
|
||||
// Uses recorded responses, no backend needed
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils (Automatic HAR Management):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/network-recorder/fixtures';
|
||||
|
||||
// Record mode: Set environment variable
|
||||
process.env.PW_NET_MODE = 'record';
|
||||
|
||||
test('should work offline', async ({ page, context, networkRecorder }) => {
|
||||
await networkRecorder.setup(context); // Handles HAR automatically
|
||||
|
||||
await page.goto('/dashboard');
|
||||
await page.click('#add-item');
|
||||
// All network traffic recorded, CRUD operations detected
|
||||
});
|
||||
```
|
||||
|
||||
**Switch to playback:**
|
||||
```bash
|
||||
# Playback mode (offline)
|
||||
PW_NET_MODE=playback npx playwright test
|
||||
# Uses HAR file, no backend needed!
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- Automatic HAR file management (naming, paths)
|
||||
- CRUD operation detection (stateful mocking)
|
||||
- Environment variable control (easy switching)
|
||||
- Works for complex interactions (create, update, delete)
|
||||
- No manual route configuration
|
||||
|
||||
### Network Request Interception
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
test('should handle API error', async ({ page }) => {
|
||||
// Manual route setup
|
||||
await page.route('**/api/users', (route) => {
|
||||
route.fulfill({
|
||||
status: 500,
|
||||
body: JSON.stringify({ error: 'Internal server error' })
|
||||
});
|
||||
});
|
||||
|
||||
await page.goto('/users');
|
||||
|
||||
const response = await page.waitForResponse('**/api/users');
|
||||
const error = await response.json();
|
||||
|
||||
expect(error.error).toContain('Internal server');
|
||||
await expect(page.locator('.error-message')).toContainText('Server error');
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('should handle API error', async ({ page, interceptNetworkCall }) => {
|
||||
// Stub API to return error (set up BEFORE navigation)
|
||||
const usersCall = interceptNetworkCall({
|
||||
method: 'GET',
|
||||
url: '**/api/users',
|
||||
fulfillResponse: {
|
||||
status: 500,
|
||||
body: { error: 'Internal server error' }
|
||||
}
|
||||
});
|
||||
|
||||
await page.goto('/users');
|
||||
|
||||
// Wait for mocked response and access parsed data
|
||||
const { status, responseJson } = await usersCall;
|
||||
|
||||
expect(status).toBe(500);
|
||||
expect(responseJson.error).toContain('Internal server');
|
||||
await expect(page.locator('.error-message')).toContainText('Server error');
|
||||
});
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- Automatic JSON parsing (`responseJson` ready to use)
|
||||
- Returns promise with `{ status, responseJson, requestJson }`
|
||||
- No need to pass `page` (auto-injected by fixture)
|
||||
- Glob pattern matching (simpler than regex)
|
||||
- Single declarative call (setup + wait in one)
|
||||
|
||||
## Comparison: Traditional vs Network-First
|
||||
|
||||
### Loading Dashboard Data
|
||||
|
||||
**Traditional (Flaky):**
|
||||
```typescript
|
||||
test('dashboard loads data', async ({ page }) => {
|
||||
await page.goto('/dashboard');
|
||||
await page.waitForTimeout(2000); // ❌ Magic number
|
||||
await expect(page.locator('table tr')).toHaveCount(5);
|
||||
});
|
||||
```
|
||||
|
||||
**Failure modes:**
|
||||
- API takes 2.5s → test fails
|
||||
- API returns 3 items not 5 → hard to debug (which issue?)
|
||||
- CI slower than local → fails in CI only
|
||||
|
||||
**Network-First (Deterministic):**
|
||||
```typescript
|
||||
test('dashboard loads data', async ({ page }) => {
|
||||
const apiPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/dashboard') && resp.ok()
|
||||
);
|
||||
|
||||
await page.goto('/dashboard');
|
||||
|
||||
const response = await apiPromise;
|
||||
const { items } = await response.json();
|
||||
|
||||
// Validate API response
|
||||
expect(items).toHaveLength(5);
|
||||
|
||||
// Validate UI matches API
|
||||
await expect(page.locator('table tr')).toHaveCount(items.length);
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Waits exactly as long as needed (100ms or 5s, doesn't matter)
|
||||
- Validates API response (catch backend errors)
|
||||
- Validates UI matches API (catch frontend bugs)
|
||||
- Works in any environment (local, CI, staging)
|
||||
|
||||
**With Playwright Utils (Even Better):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('dashboard loads data', async ({ page, interceptNetworkCall }) => {
|
||||
const dashboardCall = interceptNetworkCall({
|
||||
method: 'GET',
|
||||
url: '**/api/dashboard'
|
||||
});
|
||||
|
||||
await page.goto('/dashboard');
|
||||
|
||||
const { status, responseJson: { items } } = await dashboardCall;
|
||||
|
||||
// Validate API response (automatic JSON parsing)
|
||||
expect(status).toBe(200);
|
||||
expect(items).toHaveLength(5);
|
||||
|
||||
// Validate UI matches API
|
||||
await expect(page.locator('table tr')).toHaveCount(items.length);
|
||||
});
|
||||
```
|
||||
|
||||
**Additional Benefits:**
|
||||
- No manual `await response.json()` (automatic parsing)
|
||||
- Cleaner destructuring of nested data
|
||||
- Consistent API across all network calls
|
||||
|
||||
---
|
||||
|
||||
### Form Submission
|
||||
|
||||
**Traditional (Flaky):**
|
||||
```typescript
|
||||
test('form submission', async ({ page }) => {
|
||||
await page.fill('#email', 'test@example.com');
|
||||
await page.click('button[type="submit"]');
|
||||
await page.waitForTimeout(3000); // ❌ Hope it's enough
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Network-First (Deterministic):**
|
||||
```typescript
|
||||
test('form submission', async ({ page }) => {
|
||||
const submitPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/submit') &&
|
||||
resp.request().method() === 'POST' &&
|
||||
resp.ok()
|
||||
);
|
||||
|
||||
await page.fill('#email', 'test@example.com');
|
||||
await page.click('button[type="submit"]');
|
||||
|
||||
const response = await submitPromise;
|
||||
const result = await response.json();
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('form submission', async ({ page, interceptNetworkCall }) => {
|
||||
const submitCall = interceptNetworkCall({
|
||||
method: 'POST',
|
||||
url: '**/api/submit'
|
||||
});
|
||||
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByRole('button', { name: 'Submit' }).click();
|
||||
|
||||
const { status, responseJson: result } = await submitCall;
|
||||
|
||||
// Automatic JSON parsing, no manual await
|
||||
expect(status).toBe(200);
|
||||
expect(result.success).toBe(true);
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Progression:**
|
||||
- Traditional: Hard waits (flaky)
|
||||
- Network-First (Vanilla): waitForResponse (deterministic)
|
||||
- Network-First (PW-Utils): interceptNetworkCall (deterministic + cleaner API)
|
||||
|
||||
---
|
||||
|
||||
## Common Misconceptions
|
||||
|
||||
### "I Already Use waitForSelector"
|
||||
|
||||
```typescript
|
||||
// This is still a hard wait in disguise
|
||||
await page.click('button');
|
||||
await page.waitForSelector('.success', { timeout: 5000 });
|
||||
```
|
||||
|
||||
**Problem:** Waiting for DOM, not for the API that caused DOM change.
|
||||
|
||||
**Better:**
|
||||
```typescript
|
||||
await page.waitForResponse(matcher); // Wait for root cause
|
||||
await page.waitForSelector('.success'); // Then validate UI
|
||||
```
|
||||
|
||||
### "My Tests Are Fast, Why Add Complexity?"
|
||||
|
||||
**Short-term:** Tests are fast locally
|
||||
|
||||
**Long-term problems:**
|
||||
- Different environments (CI slower)
|
||||
- Under load (API slower)
|
||||
- Network variability (random)
|
||||
- Scaling test suite (100 → 1000 tests)
|
||||
|
||||
**Network-first prevents these issues before they appear.**
|
||||
|
||||
### "Too Much Boilerplate"
|
||||
|
||||
**Problem:** `waitForResponse` is verbose, repeated in every test.
|
||||
|
||||
**Solution:** Use Playwright Utils `interceptNetworkCall` - built-in fixture that reduces boilerplate.
|
||||
|
||||
**Vanilla Playwright (Repetitive):**
|
||||
```typescript
|
||||
test('test 1', async ({ page }) => {
|
||||
const promise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/submit') && resp.ok()
|
||||
);
|
||||
await page.click('button');
|
||||
await promise;
|
||||
});
|
||||
|
||||
test('test 2', async ({ page }) => {
|
||||
const promise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/load') && resp.ok()
|
||||
);
|
||||
await page.click('button');
|
||||
await promise;
|
||||
});
|
||||
// Repeated pattern in every test
|
||||
```
|
||||
|
||||
**With Playwright Utils (Cleaner):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('test 1', async ({ page, interceptNetworkCall }) => {
|
||||
const submitCall = interceptNetworkCall({ url: '**/api/submit' });
|
||||
await page.click('button');
|
||||
const { status, responseJson } = await submitCall;
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
|
||||
test('test 2', async ({ page, interceptNetworkCall }) => {
|
||||
const loadCall = interceptNetworkCall({ url: '**/api/load' });
|
||||
await page.click('button');
|
||||
const { responseJson } = await loadCall;
|
||||
// Automatic JSON parsing, cleaner API
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Less boilerplate (fixture handles complexity)
|
||||
- Automatic JSON parsing
|
||||
- Glob pattern matching (`**/api/**`)
|
||||
- Consistent API across all tests
|
||||
|
||||
See [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md#intercept-network-call) for setup.
|
||||
|
||||
## Technical Implementation
|
||||
|
||||
For detailed network-first patterns, see the knowledge base:
|
||||
- [Knowledge Base Index - Network & Reliability](/docs/reference/tea/knowledge-base.md)
|
||||
- [Complete Knowledge Base Index](/docs/reference/tea/knowledge-base.md)
|
||||
|
||||
## Related Concepts
|
||||
|
||||
**Core TEA Concepts:**
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Determinism requires network-first
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - High-risk features need reliable tests
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Network utilities as fixtures
|
||||
- [Knowledge Base System](/docs/explanation/tea/knowledge-base-system.md) - Network patterns in knowledge base
|
||||
|
||||
**Overview:**
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Network-first in workflows
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - Why flakiness matters
|
||||
|
||||
## Practical Guides
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Review for hard waits
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Generate network-first tests
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Expand with network patterns
|
||||
|
||||
**Use-Case Guides:**
|
||||
- [Using TEA with Existing Tests](/docs/how-to/brownfield/use-tea-with-existing-tests.md) - Fix flaky legacy tests
|
||||
|
||||
**Customization:**
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - Network utilities (recorder, interceptor, error monitor)
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - All workflows use network-first
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Network-first fragment
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - Network-first pattern term
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
586
docs/explanation/tea/risk-based-testing.md
Normal file
586
docs/explanation/tea/risk-based-testing.md
Normal file
@@ -0,0 +1,586 @@
|
||||
---
|
||||
title: "Risk-Based Testing Explained"
|
||||
description: Understanding how TEA uses probability × impact scoring to prioritize testing effort
|
||||
---
|
||||
|
||||
# Risk-Based Testing Explained
|
||||
|
||||
Risk-based testing is TEA's core principle: testing depth scales with business impact. Instead of testing everything equally, focus effort where failures hurt most.
|
||||
|
||||
## Overview
|
||||
|
||||
Traditional testing approaches treat all features equally:
|
||||
- Every feature gets same test coverage
|
||||
- Same level of scrutiny regardless of impact
|
||||
- No systematic prioritization
|
||||
- Testing becomes checkbox exercise
|
||||
|
||||
**Risk-based testing asks:**
|
||||
- What's the probability this will fail?
|
||||
- What's the impact if it does fail?
|
||||
- How much testing is appropriate for this risk level?
|
||||
|
||||
**Result:** Testing effort matches business criticality.
|
||||
|
||||
## The Problem
|
||||
|
||||
### Equal Testing for Unequal Risk
|
||||
|
||||
```markdown
|
||||
Feature A: User login (critical path, millions of users)
|
||||
Feature B: Export to PDF (nice-to-have, rarely used)
|
||||
|
||||
Traditional approach:
|
||||
- Both get 10 tests
|
||||
- Both get same review scrutiny
|
||||
- Both take same development time
|
||||
|
||||
Problem: Wasting effort on low-impact features while under-testing critical paths.
|
||||
```
|
||||
|
||||
### No Objective Prioritization
|
||||
|
||||
```markdown
|
||||
PM: "We need more tests for checkout"
|
||||
QA: "How many tests?"
|
||||
PM: "I don't know... a lot?"
|
||||
QA: "How do we know when we have enough?"
|
||||
PM: "When it feels safe?"
|
||||
|
||||
Problem: Subjective decisions, no data, political debates.
|
||||
```
|
||||
|
||||
## The Solution: Probability × Impact Scoring
|
||||
|
||||
### Risk Score = Probability × Impact
|
||||
|
||||
**Probability** (How likely to fail?)
|
||||
- **1 (Low):** Stable, well-tested, simple logic
|
||||
- **2 (Medium):** Moderate complexity, some unknowns
|
||||
- **3 (High):** Complex, untested, many edge cases
|
||||
|
||||
**Impact** (How bad if it fails?)
|
||||
- **1 (Low):** Minor inconvenience, few users affected
|
||||
- **2 (Medium):** Degraded experience, workarounds exist
|
||||
- **3 (High):** Critical path broken, business impact
|
||||
|
||||
**Score Range:** 1-9
|
||||
|
||||
#### Risk Scoring Matrix
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
graph TD
|
||||
subgraph Matrix[" "]
|
||||
direction TB
|
||||
subgraph Impact3["Impact: HIGH (3)"]
|
||||
P1I3["Score: 3<br/>Low Risk"]
|
||||
P2I3["Score: 6<br/>HIGH RISK<br/>Mitigation Required"]
|
||||
P3I3["Score: 9<br/>CRITICAL<br/>Blocks Release"]
|
||||
end
|
||||
subgraph Impact2["Impact: MEDIUM (2)"]
|
||||
P1I2["Score: 2<br/>Low Risk"]
|
||||
P2I2["Score: 4<br/>Medium Risk"]
|
||||
P3I2["Score: 6<br/>HIGH RISK<br/>Mitigation Required"]
|
||||
end
|
||||
subgraph Impact1["Impact: LOW (1)"]
|
||||
P1I1["Score: 1<br/>Low Risk"]
|
||||
P2I1["Score: 2<br/>Low Risk"]
|
||||
P3I1["Score: 3<br/>Low Risk"]
|
||||
end
|
||||
end
|
||||
|
||||
Prob1["Probability: LOW (1)"] -.-> P1I1
|
||||
Prob1 -.-> P1I2
|
||||
Prob1 -.-> P1I3
|
||||
|
||||
Prob2["Probability: MEDIUM (2)"] -.-> P2I1
|
||||
Prob2 -.-> P2I2
|
||||
Prob2 -.-> P2I3
|
||||
|
||||
Prob3["Probability: HIGH (3)"] -.-> P3I1
|
||||
Prob3 -.-> P3I2
|
||||
Prob3 -.-> P3I3
|
||||
|
||||
style P3I3 fill:#f44336,stroke:#b71c1c,stroke-width:3px,color:#fff
|
||||
style P2I3 fill:#ff9800,stroke:#e65100,stroke-width:2px,color:#000
|
||||
style P3I2 fill:#ff9800,stroke:#e65100,stroke-width:2px,color:#000
|
||||
style P2I2 fill:#fff9c4,stroke:#f57f17,stroke-width:1px,color:#000
|
||||
style P1I1 fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
style P2I1 fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
style P3I1 fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
style P1I2 fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
style P1I3 fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
```
|
||||
|
||||
**Legend:**
|
||||
- 🔴 Red (Score 9): CRITICAL - Blocks release
|
||||
- 🟠 Orange (Score 6-8): HIGH RISK - Mitigation required
|
||||
- 🟡 Yellow (Score 4-5): MEDIUM - Mitigation recommended
|
||||
- 🟢 Green (Score 1-3): LOW - Optional mitigation
|
||||
|
||||
### Scoring Examples
|
||||
|
||||
**Score 9 (Critical):**
|
||||
```
|
||||
Feature: Payment processing
|
||||
Probability: 3 (complex third-party integration)
|
||||
Impact: 3 (broken payments = lost revenue)
|
||||
Score: 3 × 3 = 9
|
||||
|
||||
Action: Extensive testing required
|
||||
- E2E tests for all payment flows
|
||||
- API tests for all payment scenarios
|
||||
- Error handling for all failure modes
|
||||
- Security testing for payment data
|
||||
- Load testing for high traffic
|
||||
- Monitoring and alerts
|
||||
```
|
||||
|
||||
**Score 1 (Low):**
|
||||
```
|
||||
Feature: Change profile theme color
|
||||
Probability: 1 (simple UI toggle)
|
||||
Impact: 1 (cosmetic only)
|
||||
Score: 1 × 1 = 1
|
||||
|
||||
Action: Minimal testing
|
||||
- One E2E smoke test
|
||||
- Skip edge cases
|
||||
- No API tests needed
|
||||
```
|
||||
|
||||
**Score 6 (Medium-High):**
|
||||
```
|
||||
Feature: User profile editing
|
||||
Probability: 2 (moderate complexity)
|
||||
Impact: 3 (users can't update info)
|
||||
Score: 2 × 3 = 6
|
||||
|
||||
Action: Focused testing
|
||||
- E2E test for happy path
|
||||
- API tests for CRUD operations
|
||||
- Validation testing
|
||||
- Skip low-value edge cases
|
||||
```
|
||||
|
||||
## How It Works in TEA
|
||||
|
||||
### 1. Risk Categories
|
||||
|
||||
TEA assesses risk across 6 categories:
|
||||
|
||||
**TECH** - Technical debt, architecture fragility
|
||||
```
|
||||
Example: Migrating from REST to GraphQL
|
||||
Probability: 3 (major architectural change)
|
||||
Impact: 3 (affects all API consumers)
|
||||
Score: 9 - Extensive integration testing required
|
||||
```
|
||||
|
||||
**SEC** - Security vulnerabilities
|
||||
```
|
||||
Example: Adding OAuth integration
|
||||
Probability: 2 (third-party dependency)
|
||||
Impact: 3 (auth breach = data exposure)
|
||||
Score: 6 - Security testing mandatory
|
||||
```
|
||||
|
||||
**PERF** - Performance degradation
|
||||
```
|
||||
Example: Adding real-time notifications
|
||||
Probability: 2 (WebSocket complexity)
|
||||
Impact: 2 (slower experience)
|
||||
Score: 4 - Load testing recommended
|
||||
```
|
||||
|
||||
**DATA** - Data integrity, corruption
|
||||
```
|
||||
Example: Database migration
|
||||
Probability: 2 (schema changes)
|
||||
Impact: 3 (data loss unacceptable)
|
||||
Score: 6 - Data validation tests required
|
||||
```
|
||||
|
||||
**BUS** - Business logic errors
|
||||
```
|
||||
Example: Discount calculation
|
||||
Probability: 2 (business rules complex)
|
||||
Impact: 3 (wrong prices = revenue loss)
|
||||
Score: 6 - Business logic tests mandatory
|
||||
```
|
||||
|
||||
**OPS** - Operational issues
|
||||
```
|
||||
Example: Logging system update
|
||||
Probability: 1 (straightforward)
|
||||
Impact: 2 (debugging harder without logs)
|
||||
Score: 2 - Basic smoke test sufficient
|
||||
```
|
||||
|
||||
### 2. Test Priorities (P0-P3)
|
||||
|
||||
Risk scores inform test priorities (but aren't the only factor):
|
||||
|
||||
**P0 - Critical Path**
|
||||
- **Risk Scores:** Typically 6-9 (high risk)
|
||||
- **Other Factors:** Revenue impact, security-critical, regulatory compliance, frequent usage
|
||||
- **Coverage Target:** 100%
|
||||
- **Test Levels:** E2E + API
|
||||
- **Example:** Login, checkout, payment processing
|
||||
|
||||
**P1 - High Value**
|
||||
- **Risk Scores:** Typically 4-6 (medium-high risk)
|
||||
- **Other Factors:** Core user journeys, complex logic, integration points
|
||||
- **Coverage Target:** 90%
|
||||
- **Test Levels:** API + selective E2E
|
||||
- **Example:** Profile editing, search, filters
|
||||
|
||||
**P2 - Medium Value**
|
||||
- **Risk Scores:** Typically 2-4 (medium risk)
|
||||
- **Other Factors:** Secondary features, admin functionality, reporting
|
||||
- **Coverage Target:** 50%
|
||||
- **Test Levels:** API happy path only
|
||||
- **Example:** Export features, advanced settings
|
||||
|
||||
**P3 - Low Value**
|
||||
- **Risk Scores:** Typically 1-2 (low risk)
|
||||
- **Other Factors:** Rarely used, nice-to-have, cosmetic
|
||||
- **Coverage Target:** 20% (smoke test)
|
||||
- **Test Levels:** E2E smoke test only
|
||||
- **Example:** Theme customization, experimental features
|
||||
|
||||
**Note:** Priorities consider risk scores plus business context (usage frequency, user impact, etc.). See [Test Priorities Matrix](/docs/reference/tea/knowledge-base.md#test-priorities-matrix) for complete criteria.
|
||||
|
||||
### 3. Mitigation Plans
|
||||
|
||||
**Scores ≥6 require documented mitigation:**
|
||||
|
||||
```markdown
|
||||
## Risk Mitigation
|
||||
|
||||
**Risk:** Payment integration failure (Score: 9)
|
||||
|
||||
**Mitigation Plan:**
|
||||
- Create comprehensive test suite (20+ tests)
|
||||
- Add payment sandbox environment
|
||||
- Implement retry logic with idempotency
|
||||
- Add monitoring and alerts
|
||||
- Document rollback procedure
|
||||
|
||||
**Owner:** Backend team lead
|
||||
**Deadline:** Before production deployment
|
||||
**Status:** In progress
|
||||
```
|
||||
|
||||
**Gate Rules:**
|
||||
- **Score = 9** (Critical): Mandatory FAIL - blocks release without mitigation
|
||||
- **Score 6-8** (High): Requires mitigation plan, becomes CONCERNS if incomplete
|
||||
- **Score 4-5** (Medium): Mitigation recommended but not required
|
||||
- **Score 1-3** (Low): No mitigation needed
|
||||
|
||||
## Comparison: Traditional vs Risk-Based
|
||||
|
||||
### Traditional Approach
|
||||
|
||||
```typescript
|
||||
// Test everything equally
|
||||
describe('User profile', () => {
|
||||
test('should display name');
|
||||
test('should display email');
|
||||
test('should display phone');
|
||||
test('should display address');
|
||||
test('should display bio');
|
||||
test('should display avatar');
|
||||
test('should display join date');
|
||||
test('should display last login');
|
||||
test('should display theme preference');
|
||||
test('should display language preference');
|
||||
// 10 tests for profile display (all equal priority)
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Same effort for critical (name) vs trivial (theme)
|
||||
- No guidance on what matters
|
||||
- Wastes time on low-value tests
|
||||
|
||||
### Risk-Based Approach
|
||||
|
||||
```typescript
|
||||
// Test based on risk
|
||||
|
||||
describe('User profile - Critical (P0)', () => {
|
||||
test('should display name and email'); // Score: 9 (identity critical)
|
||||
test('should allow editing name and email');
|
||||
test('should validate email format');
|
||||
test('should prevent unauthorized edits');
|
||||
// 4 focused tests on high-risk areas
|
||||
});
|
||||
|
||||
describe('User profile - High Value (P1)', () => {
|
||||
test('should upload avatar'); // Score: 6 (users care about this)
|
||||
test('should update bio');
|
||||
// 2 tests for high-value features
|
||||
});
|
||||
|
||||
// P2: Theme preference - single smoke test
|
||||
// P3: Last login display - skip (read-only, low value)
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- 6 focused tests vs 10 unfocused tests
|
||||
- Effort matches business impact
|
||||
- Clear priorities guide development
|
||||
- No wasted effort on trivial features
|
||||
|
||||
## When to Use Risk-Based Testing
|
||||
|
||||
### Always Use For:
|
||||
|
||||
**Enterprise projects:**
|
||||
- High stakes (revenue, compliance, security)
|
||||
- Many features competing for test effort
|
||||
- Need objective prioritization
|
||||
|
||||
**Large codebases:**
|
||||
- Can't test everything exhaustively
|
||||
- Need to focus limited QA resources
|
||||
- Want data-driven decisions
|
||||
|
||||
**Regulated industries:**
|
||||
- Must justify testing decisions
|
||||
- Auditors want risk assessments
|
||||
- Compliance requires evidence
|
||||
|
||||
### Consider Skipping For:
|
||||
|
||||
**Tiny projects:**
|
||||
- 5 features total
|
||||
- Can test everything thoroughly
|
||||
- Risk scoring is overhead
|
||||
|
||||
**Prototypes:**
|
||||
- Throw-away code
|
||||
- Speed over quality
|
||||
- Learning experiments
|
||||
|
||||
## Real-World Example
|
||||
|
||||
### Scenario: E-Commerce Checkout Redesign
|
||||
|
||||
**Feature:** Redesigning checkout flow from 5 steps to 3 steps
|
||||
|
||||
**Risk Assessment:**
|
||||
|
||||
| Component | Probability | Impact | Score | Priority | Testing |
|
||||
|-----------|-------------|--------|-------|----------|---------|
|
||||
| **Payment processing** | 3 | 3 | 9 | P0 | 15 E2E + 20 API tests |
|
||||
| **Order validation** | 2 | 3 | 6 | P1 | 5 E2E + 10 API tests |
|
||||
| **Shipping calculation** | 2 | 2 | 4 | P1 | 3 E2E + 8 API tests |
|
||||
| **Promo code validation** | 2 | 2 | 4 | P1 | 2 E2E + 5 API tests |
|
||||
| **Gift message** | 1 | 1 | 1 | P3 | 1 E2E smoke test |
|
||||
|
||||
**Test Budget:** 40 hours
|
||||
|
||||
**Allocation:**
|
||||
- Payment (Score 9): 20 hours (50%)
|
||||
- Order validation (Score 6): 8 hours (20%)
|
||||
- Shipping (Score 4): 6 hours (15%)
|
||||
- Promo codes (Score 4): 4 hours (10%)
|
||||
- Gift message (Score 1): 2 hours (5%)
|
||||
|
||||
**Result:** 50% of effort on highest-risk feature (payment), proportional allocation for others.
|
||||
|
||||
### Without Risk-Based Testing:
|
||||
|
||||
**Equal allocation:** 8 hours per component = wasted effort on gift message, under-testing payment.
|
||||
|
||||
**Result:** Payment bugs slip through (critical), perfect testing of gift message (trivial).
|
||||
|
||||
## Mitigation Strategies by Risk Level
|
||||
|
||||
### Score 9: Mandatory Mitigation (Blocks Release)
|
||||
|
||||
```markdown
|
||||
**Gate Impact:** FAIL - Cannot deploy without mitigation
|
||||
|
||||
**Actions:**
|
||||
- Comprehensive test suite (E2E, API, security)
|
||||
- Multiple test environments (dev, staging, prod-mirror)
|
||||
- Load testing and performance validation
|
||||
- Security audit and penetration testing
|
||||
- Monitoring and alerting
|
||||
- Rollback plan documented
|
||||
- On-call rotation assigned
|
||||
|
||||
**Cannot deploy until score is mitigated below 9.**
|
||||
```
|
||||
|
||||
### Score 6-8: Required Mitigation (Gate: CONCERNS)
|
||||
|
||||
```markdown
|
||||
**Gate Impact:** CONCERNS - Can deploy with documented mitigation plan
|
||||
|
||||
**Actions:**
|
||||
- Targeted test suite (happy path + critical errors)
|
||||
- Test environment setup
|
||||
- Monitoring plan
|
||||
- Document mitigation and owners
|
||||
|
||||
**Can deploy with approved mitigation plan.**
|
||||
```
|
||||
|
||||
### Score 4-5: Recommended Mitigation
|
||||
|
||||
```markdown
|
||||
**Gate Impact:** Advisory - Does not affect gate decision
|
||||
|
||||
**Actions:**
|
||||
- Basic test coverage
|
||||
- Standard monitoring
|
||||
- Document known limitations
|
||||
|
||||
**Can deploy, mitigation recommended but not required.**
|
||||
```
|
||||
|
||||
### Score 1-3: Optional Mitigation
|
||||
|
||||
```markdown
|
||||
**Gate Impact:** None
|
||||
|
||||
**Actions:**
|
||||
- Smoke test if desired
|
||||
- Feature flag for easy disable (optional)
|
||||
|
||||
**Can deploy without mitigation.**
|
||||
```
|
||||
|
||||
## Technical Implementation
|
||||
|
||||
For detailed risk governance patterns, see the knowledge base:
|
||||
- [Knowledge Base Index - Risk & Gates](/docs/reference/tea/knowledge-base.md)
|
||||
- [TEA Command Reference - *test-design](/docs/reference/tea/commands.md#test-design)
|
||||
|
||||
### Risk Scoring Matrix
|
||||
|
||||
TEA uses this framework in `*test-design`:
|
||||
|
||||
```
|
||||
Impact
|
||||
1 2 3
|
||||
┌────┬────┬────┐
|
||||
1 │ 1 │ 2 │ 3 │ Low risk
|
||||
P 2 │ 2 │ 4 │ 6 │ Medium risk
|
||||
r 3 │ 3 │ 6 │ 9 │ High risk
|
||||
o └────┴────┴────┘
|
||||
b Low Med High
|
||||
```
|
||||
|
||||
### Gate Decision Rules
|
||||
|
||||
| Score | Mitigation Required | Gate Impact |
|
||||
|-------|-------------------|-------------|
|
||||
| **9** | Mandatory, blocks release | FAIL if no mitigation |
|
||||
| **6-8** | Required, documented plan | CONCERNS if incomplete |
|
||||
| **4-5** | Recommended | Advisory only |
|
||||
| **1-3** | Optional | No impact |
|
||||
|
||||
#### Gate Decision Flow
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
flowchart TD
|
||||
Start([Risk Assessment]) --> Score{Risk Score?}
|
||||
|
||||
Score -->|Score = 9| Critical[CRITICAL RISK<br/>Score: 9]
|
||||
Score -->|Score 6-8| High[HIGH RISK<br/>Score: 6-8]
|
||||
Score -->|Score 4-5| Medium[MEDIUM RISK<br/>Score: 4-5]
|
||||
Score -->|Score 1-3| Low[LOW RISK<br/>Score: 1-3]
|
||||
|
||||
Critical --> HasMit9{Mitigation<br/>Plan?}
|
||||
HasMit9 -->|Yes| Concerns9[CONCERNS ⚠️<br/>Can deploy with plan]
|
||||
HasMit9 -->|No| Fail[FAIL ❌<br/>Blocks release]
|
||||
|
||||
High --> HasMit6{Mitigation<br/>Plan?}
|
||||
HasMit6 -->|Yes| Pass6[PASS ✅<br/>or CONCERNS ⚠️]
|
||||
HasMit6 -->|No| Concerns6[CONCERNS ⚠️<br/>Document plan needed]
|
||||
|
||||
Medium --> Advisory[Advisory Only<br/>No gate impact]
|
||||
Low --> NoAction[No Action<br/>Proceed]
|
||||
|
||||
style Critical fill:#f44336,stroke:#b71c1c,stroke-width:3px,color:#fff
|
||||
style Fail fill:#d32f2f,stroke:#b71c1c,stroke-width:3px,color:#fff
|
||||
style High fill:#ff9800,stroke:#e65100,stroke-width:2px,color:#000
|
||||
style Concerns9 fill:#ffc107,stroke:#f57f17,stroke-width:2px,color:#000
|
||||
style Concerns6 fill:#ffc107,stroke:#f57f17,stroke-width:2px,color:#000
|
||||
style Pass6 fill:#4caf50,stroke:#1b5e20,stroke-width:2px,color:#fff
|
||||
style Medium fill:#fff9c4,stroke:#f57f17,stroke-width:1px,color:#000
|
||||
style Low fill:#c8e6c9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
style Advisory fill:#e8f5e9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
style NoAction fill:#e8f5e9,stroke:#2e7d32,stroke-width:1px,color:#000
|
||||
```
|
||||
|
||||
## Common Misconceptions
|
||||
|
||||
### "Risk-based = Less Testing"
|
||||
|
||||
**Wrong:** Risk-based testing often means MORE testing where it matters.
|
||||
|
||||
**Example:**
|
||||
- Traditional: 50 tests spread equally
|
||||
- Risk-based: 70 tests focused on P0/P1 (more total, better allocated)
|
||||
|
||||
### "Low Priority = Skip Testing"
|
||||
|
||||
**Wrong:** P3 still gets smoke tests.
|
||||
|
||||
**Correct:**
|
||||
- P3: Smoke test (feature works at all)
|
||||
- P2: Happy path (feature works correctly)
|
||||
- P1: Happy path + errors
|
||||
- P0: Comprehensive (all scenarios)
|
||||
|
||||
### "Risk Scores Are Permanent"
|
||||
|
||||
**Wrong:** Risk changes over time.
|
||||
|
||||
**Correct:**
|
||||
- Initial launch: Payment is Score 9 (untested integration)
|
||||
- After 6 months: Payment is Score 6 (proven in production)
|
||||
- Re-assess risk quarterly
|
||||
|
||||
## Related Concepts
|
||||
|
||||
**Core TEA Concepts:**
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Quality complements risk assessment
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - When risk-based testing matters most
|
||||
- [Knowledge Base System](/docs/explanation/tea/knowledge-base-system.md) - How risk patterns are loaded
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Building risk-appropriate test infrastructure
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Quality patterns for high-risk features
|
||||
|
||||
**Overview:**
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Risk assessment in TEA lifecycle
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - Design philosophy
|
||||
|
||||
## Practical Guides
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Apply risk scoring
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md) - Gate decisions based on risk
|
||||
- [How to Run NFR Assessment](/docs/how-to/workflows/run-nfr-assess.md) - NFR risk assessment
|
||||
|
||||
**Use-Case Guides:**
|
||||
- [Running TEA for Enterprise](/docs/how-to/enterprise/use-tea-for-enterprise.md) - Enterprise risk management
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - `*test-design`, `*nfr-assess`, `*trace`
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Risk governance fragments
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - Risk-based testing term
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
907
docs/explanation/tea/test-quality-standards.md
Normal file
907
docs/explanation/tea/test-quality-standards.md
Normal file
@@ -0,0 +1,907 @@
|
||||
---
|
||||
title: "Test Quality Standards Explained"
|
||||
description: Understanding TEA's Definition of Done for deterministic, isolated, and maintainable tests
|
||||
---
|
||||
|
||||
# Test Quality Standards Explained
|
||||
|
||||
Test quality standards define what makes a test "good" in TEA. These aren't suggestions - they're the Definition of Done that prevents tests from rotting in review.
|
||||
|
||||
## Overview
|
||||
|
||||
**TEA's Quality Principles:**
|
||||
- **Deterministic** - Same result every run
|
||||
- **Isolated** - No dependencies on other tests
|
||||
- **Explicit** - Assertions visible in test body
|
||||
- **Focused** - Single responsibility, appropriate size
|
||||
- **Fast** - Execute in reasonable time
|
||||
|
||||
**Why these matter:** Tests that violate these principles create maintenance burden, slow down development, and lose team trust.
|
||||
|
||||
## The Problem
|
||||
|
||||
### Tests That Rot in Review
|
||||
|
||||
```typescript
|
||||
// ❌ The anti-pattern: This test will rot
|
||||
test('user can do stuff', async ({ page }) => {
|
||||
await page.goto('/');
|
||||
await page.waitForTimeout(5000); // Non-deterministic
|
||||
|
||||
if (await page.locator('.banner').isVisible()) { // Conditional
|
||||
await page.click('.dismiss');
|
||||
}
|
||||
|
||||
try { // Try-catch for flow control
|
||||
await page.click('#load-more');
|
||||
} catch (e) {
|
||||
// Silently continue
|
||||
}
|
||||
|
||||
// ... 300 more lines of test logic
|
||||
// ... no clear assertions
|
||||
});
|
||||
```
|
||||
|
||||
**What's wrong:**
|
||||
- **Hard wait** - Flaky, wastes time
|
||||
- **Conditional** - Non-deterministic behavior
|
||||
- **Try-catch** - Hides failures
|
||||
- **Too large** - Hard to maintain
|
||||
- **Vague name** - Unclear purpose
|
||||
- **No explicit assertions** - What's being tested?
|
||||
|
||||
**Result:** PR review comments: "This test is flaky, please fix" → never merged → test deleted → coverage lost
|
||||
|
||||
### AI-Generated Tests Without Standards
|
||||
|
||||
AI-generated tests without quality guardrails:
|
||||
|
||||
```typescript
|
||||
// AI generates 50 tests like this:
|
||||
test('test1', async ({ page }) => {
|
||||
await page.goto('/');
|
||||
await page.waitForTimeout(3000);
|
||||
// ... flaky, vague, redundant
|
||||
});
|
||||
|
||||
test('test2', async ({ page }) => {
|
||||
await page.goto('/');
|
||||
await page.waitForTimeout(3000);
|
||||
// ... duplicates test1
|
||||
});
|
||||
|
||||
// ... 48 more similar tests
|
||||
```
|
||||
|
||||
**Result:** 50 tests, 80% redundant, 90% flaky, 0% trusted by team - low-quality outputs that create maintenance burden.
|
||||
|
||||
## The Solution: TEA's Quality Standards
|
||||
|
||||
### 1. Determinism (No Flakiness)
|
||||
|
||||
**Rule:** Test produces same result every run.
|
||||
|
||||
**Requirements:**
|
||||
- ❌ No hard waits (`waitForTimeout`)
|
||||
- ❌ No conditionals for flow control (`if/else`)
|
||||
- ❌ No try-catch for flow control
|
||||
- ✅ Use network-first patterns (wait for responses)
|
||||
- ✅ Use explicit waits (waitForSelector, waitForResponse)
|
||||
|
||||
**Bad Example:**
|
||||
```typescript
|
||||
test('flaky test', async ({ page }) => {
|
||||
await page.click('button');
|
||||
await page.waitForTimeout(2000); // ❌ Might be too short
|
||||
|
||||
if (await page.locator('.modal').isVisible()) { // ❌ Non-deterministic
|
||||
await page.click('.dismiss');
|
||||
}
|
||||
|
||||
try { // ❌ Silently handles errors
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
} catch (e) {
|
||||
// Test passes even if assertion fails!
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
**Good Example (Vanilla Playwright):**
|
||||
```typescript
|
||||
test('deterministic test', async ({ page }) => {
|
||||
const responsePromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/submit') && resp.ok()
|
||||
);
|
||||
|
||||
await page.click('button');
|
||||
await responsePromise; // ✅ Wait for actual response
|
||||
|
||||
// Modal should ALWAYS show (make it deterministic)
|
||||
await expect(page.locator('.modal')).toBeVisible();
|
||||
await page.click('.dismiss');
|
||||
|
||||
// Explicit assertion (fails if not visible)
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils (Even Cleaner):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('deterministic test', async ({ page, interceptNetworkCall }) => {
|
||||
const submitCall = interceptNetworkCall({
|
||||
method: 'POST',
|
||||
url: '**/api/submit'
|
||||
});
|
||||
|
||||
await page.click('button');
|
||||
|
||||
// Wait for actual response (automatic JSON parsing)
|
||||
const { status, responseJson } = await submitCall;
|
||||
expect(status).toBe(200);
|
||||
|
||||
// Modal should ALWAYS show (make it deterministic)
|
||||
await expect(page.locator('.modal')).toBeVisible();
|
||||
await page.click('.dismiss');
|
||||
|
||||
// Explicit assertion (fails if not visible)
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Why both work:**
|
||||
- Waits for actual event (network response)
|
||||
- No conditionals (behavior is deterministic)
|
||||
- Assertions fail loudly (no silent failures)
|
||||
- Same result every run (deterministic)
|
||||
|
||||
**Playwright Utils additional benefits:**
|
||||
- Automatic JSON parsing
|
||||
- `{ status, responseJson }` structure (can validate response data)
|
||||
- No manual `await response.json()`
|
||||
|
||||
### 2. Isolation (No Dependencies)
|
||||
|
||||
**Rule:** Test runs independently, no shared state.
|
||||
|
||||
**Requirements:**
|
||||
- ✅ Self-cleaning (cleanup after test)
|
||||
- ✅ No global state dependencies
|
||||
- ✅ Can run in parallel
|
||||
- ✅ Can run in any order
|
||||
- ✅ Use unique test data
|
||||
|
||||
**Bad Example:**
|
||||
```typescript
|
||||
// ❌ Tests depend on execution order
|
||||
let userId: string; // Shared global state
|
||||
|
||||
test('create user', async ({ apiRequest }) => {
|
||||
const { body } = await apiRequest({
|
||||
method: 'POST',
|
||||
path: '/api/users',
|
||||
body: { email: 'test@example.com' } (hard-coded)
|
||||
});
|
||||
userId = body.id; // Store in global
|
||||
});
|
||||
|
||||
test('update user', async ({ apiRequest }) => {
|
||||
// Depends on previous test setting userId
|
||||
await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: `/api/users/${userId}`,
|
||||
body: { name: 'Updated' }
|
||||
});
|
||||
// No cleanup - leaves user in database
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Tests must run in order (can't parallelize)
|
||||
- Second test fails if first skipped (`.only`)
|
||||
- Hard-coded data causes conflicts
|
||||
- No cleanup (database fills with test data)
|
||||
|
||||
**Good Example (Vanilla Playwright):**
|
||||
```typescript
|
||||
test('should update user profile', async ({ request }) => {
|
||||
// Create unique test data
|
||||
const testEmail = `test-${Date.now()}@example.com`;
|
||||
|
||||
// Setup: Create user
|
||||
const createResp = await request.post('/api/users', {
|
||||
data: { email: testEmail, name: 'Original' }
|
||||
});
|
||||
const user = await createResp.json();
|
||||
|
||||
// Test: Update user
|
||||
const updateResp = await request.patch(`/api/users/${user.id}`, {
|
||||
data: { name: 'Updated' }
|
||||
});
|
||||
const updated = await updateResp.json();
|
||||
|
||||
expect(updated.name).toBe('Updated');
|
||||
|
||||
// Cleanup: Delete user
|
||||
await request.delete(`/api/users/${user.id}`);
|
||||
});
|
||||
```
|
||||
|
||||
**Even Better (With Playwright Utils):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
import { faker } from '@faker-js/faker';
|
||||
|
||||
test('should update user profile', async ({ apiRequest }) => {
|
||||
// Dynamic unique test data
|
||||
const testEmail = faker.internet.email();
|
||||
|
||||
// Setup: Create user
|
||||
const { status: createStatus, body: user } = await apiRequest({
|
||||
method: 'POST',
|
||||
path: '/api/users',
|
||||
body: { email: testEmail, name: faker.person.fullName() }
|
||||
});
|
||||
|
||||
expect(createStatus).toBe(201);
|
||||
|
||||
// Test: Update user
|
||||
const { status, body: updated } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: `/api/users/${user.id}`,
|
||||
body: { name: 'Updated Name' }
|
||||
});
|
||||
|
||||
expect(status).toBe(200);
|
||||
expect(updated.name).toBe('Updated Name');
|
||||
|
||||
// Cleanup: Delete user
|
||||
await apiRequest({
|
||||
method: 'DELETE',
|
||||
path: `/api/users/${user.id}`
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- `{ status, body }` destructuring (cleaner than `response.status()` + `await response.json()`)
|
||||
- No manual `await response.json()`
|
||||
- Automatic retry for 5xx errors
|
||||
- Optional schema validation with `.validateSchema()`
|
||||
|
||||
**Why it works:**
|
||||
- No global state
|
||||
- Unique test data (no conflicts)
|
||||
- Self-cleaning (deletes user)
|
||||
- Can run in parallel
|
||||
- Can run in any order
|
||||
|
||||
### 3. Explicit Assertions (No Hidden Validation)
|
||||
|
||||
**Rule:** Assertions visible in test body, not abstracted.
|
||||
|
||||
**Requirements:**
|
||||
- ✅ Assertions in test code (not helper functions)
|
||||
- ✅ Specific assertions (not generic `toBeTruthy`)
|
||||
- ✅ Meaningful expectations (test actual behavior)
|
||||
|
||||
**Bad Example:**
|
||||
```typescript
|
||||
// ❌ Assertions hidden in helper
|
||||
async function verifyProfilePage(page: Page) {
|
||||
// Assertions buried in helper (not visible in test)
|
||||
await expect(page.locator('h1')).toBeVisible();
|
||||
await expect(page.locator('.email')).toContainText('@');
|
||||
await expect(page.locator('.name')).not.toBeEmpty();
|
||||
}
|
||||
|
||||
test('profile page', async ({ page }) => {
|
||||
await page.goto('/profile');
|
||||
await verifyProfilePage(page); // What's being verified?
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Can't see what's tested (need to read helper)
|
||||
- Hard to debug failures (which assertion failed?)
|
||||
- Reduces test readability
|
||||
- Hides important validation
|
||||
|
||||
**Good Example:**
|
||||
```typescript
|
||||
// ✅ Assertions explicit in test
|
||||
test('should display profile with correct data', async ({ page }) => {
|
||||
await page.goto('/profile');
|
||||
|
||||
// Explicit assertions - clear what's tested
|
||||
await expect(page.locator('h1')).toContainText('Test User');
|
||||
await expect(page.locator('.email')).toContainText('test@example.com');
|
||||
await expect(page.locator('.bio')).toContainText('Software Engineer');
|
||||
await expect(page.locator('img[alt="Avatar"]')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Why it works:**
|
||||
- See what's tested at a glance
|
||||
- Debug failures easily (know which assertion failed)
|
||||
- Test is self-documenting
|
||||
- No hidden behavior
|
||||
|
||||
**Exception:** Use helper for setup/cleanup, not assertions.
|
||||
|
||||
### 4. Focused Tests (Appropriate Size)
|
||||
|
||||
**Rule:** Test has single responsibility, reasonable size.
|
||||
|
||||
**Requirements:**
|
||||
- ✅ Test size < 300 lines
|
||||
- ✅ Single responsibility (test one thing well)
|
||||
- ✅ Clear describe/test names
|
||||
- ✅ Appropriate scope (not too granular, not too broad)
|
||||
|
||||
**Bad Example:**
|
||||
```typescript
|
||||
// ❌ 500-line test testing everything
|
||||
test('complete user flow', async ({ page }) => {
|
||||
// Registration (50 lines)
|
||||
await page.goto('/register');
|
||||
await page.fill('#email', 'test@example.com');
|
||||
// ... 48 more lines
|
||||
|
||||
// Profile setup (100 lines)
|
||||
await page.goto('/profile');
|
||||
// ... 98 more lines
|
||||
|
||||
// Settings configuration (150 lines)
|
||||
await page.goto('/settings');
|
||||
// ... 148 more lines
|
||||
|
||||
// Data export (200 lines)
|
||||
await page.goto('/export');
|
||||
// ... 198 more lines
|
||||
|
||||
// Total: 500 lines, testing 4 different features
|
||||
});
|
||||
```
|
||||
|
||||
**Problems:**
|
||||
- Failure in line 50 prevents testing lines 51-500
|
||||
- Hard to understand (what's being tested?)
|
||||
- Slow to execute (testing too much)
|
||||
- Hard to debug (which feature failed?)
|
||||
|
||||
**Good Example:**
|
||||
```typescript
|
||||
// ✅ Focused tests - one responsibility each
|
||||
|
||||
test('should register new user', async ({ page }) => {
|
||||
await page.goto('/register');
|
||||
await page.fill('#email', 'test@example.com');
|
||||
await page.fill('#password', 'password123');
|
||||
await page.click('button[type="submit"]');
|
||||
|
||||
await expect(page).toHaveURL('/welcome');
|
||||
await expect(page.locator('h1')).toContainText('Welcome');
|
||||
});
|
||||
|
||||
test('should configure user profile', async ({ page, authSession }) => {
|
||||
await authSession.login({ email: 'test@example.com', password: 'pass' });
|
||||
await page.goto('/profile');
|
||||
|
||||
await page.fill('#name', 'Test User');
|
||||
await page.fill('#bio', 'Software Engineer');
|
||||
await page.click('button:has-text("Save")');
|
||||
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
});
|
||||
|
||||
// ... separate tests for settings, export (each < 50 lines)
|
||||
```
|
||||
|
||||
**Why it works:**
|
||||
- Each test has one responsibility
|
||||
- Failure is easy to diagnose
|
||||
- Can run tests independently
|
||||
- Test names describe exactly what's tested
|
||||
|
||||
### 5. Fast Execution (Performance Budget)
|
||||
|
||||
**Rule:** Individual test executes in < 1.5 minutes.
|
||||
|
||||
**Requirements:**
|
||||
- ✅ Test execution < 90 seconds
|
||||
- ✅ Efficient selectors (getByRole > XPath)
|
||||
- ✅ Minimal redundant actions
|
||||
- ✅ Parallel execution enabled
|
||||
|
||||
**Bad Example:**
|
||||
```typescript
|
||||
// ❌ Slow test (3+ minutes)
|
||||
test('slow test', async ({ page }) => {
|
||||
await page.goto('/');
|
||||
await page.waitForTimeout(10000); // 10s wasted
|
||||
|
||||
// Navigate through 10 pages (2 minutes)
|
||||
for (let i = 1; i <= 10; i++) {
|
||||
await page.click(`a[href="/page-${i}"]`);
|
||||
await page.waitForTimeout(5000); // 5s per page = 50s wasted
|
||||
}
|
||||
|
||||
// Complex XPath selector (slow)
|
||||
await page.locator('//div[@class="container"]/section[3]/div[2]/p').click();
|
||||
|
||||
// More waiting
|
||||
await page.waitForTimeout(30000); // 30s wasted
|
||||
|
||||
await expect(page.locator('.result')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Total time:** 3+ minutes (95 seconds wasted on hard waits)
|
||||
|
||||
**Good Example (Vanilla Playwright):**
|
||||
```typescript
|
||||
// ✅ Fast test (< 10 seconds)
|
||||
test('fast test', async ({ page }) => {
|
||||
// Set up response wait
|
||||
const apiPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/result') && resp.ok()
|
||||
);
|
||||
|
||||
await page.goto('/');
|
||||
|
||||
// Direct navigation (skip intermediate pages)
|
||||
await page.goto('/page-10');
|
||||
|
||||
// Efficient selector
|
||||
await page.getByRole('button', { name: 'Submit' }).click();
|
||||
|
||||
// Wait for actual response (fast when API is fast)
|
||||
await apiPromise;
|
||||
|
||||
await expect(page.locator('.result')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('fast test', async ({ page, interceptNetworkCall }) => {
|
||||
// Set up interception
|
||||
const resultCall = interceptNetworkCall({
|
||||
method: 'GET',
|
||||
url: '**/api/result'
|
||||
});
|
||||
|
||||
await page.goto('/');
|
||||
|
||||
// Direct navigation (skip intermediate pages)
|
||||
await page.goto('/page-10');
|
||||
|
||||
// Efficient selector
|
||||
await page.getByRole('button', { name: 'Submit' }).click();
|
||||
|
||||
// Wait for actual response (automatic JSON parsing)
|
||||
const { status, responseJson } = await resultCall;
|
||||
|
||||
expect(status).toBe(200);
|
||||
await expect(page.locator('.result')).toBeVisible();
|
||||
|
||||
// Can also validate response data if needed
|
||||
// expect(responseJson.data).toBeDefined();
|
||||
});
|
||||
```
|
||||
|
||||
**Total time:** < 10 seconds (no wasted waits)
|
||||
|
||||
**Both examples achieve:**
|
||||
- No hard waits (wait for actual events)
|
||||
- Direct navigation (skip unnecessary steps)
|
||||
- Efficient selectors (getByRole)
|
||||
- Fast execution
|
||||
|
||||
**Playwright Utils bonus:**
|
||||
- Can validate API response data easily
|
||||
- Automatic JSON parsing
|
||||
- Cleaner API
|
||||
|
||||
## TEA's Quality Scoring
|
||||
|
||||
TEA reviews tests against these standards in `*test-review`:
|
||||
|
||||
### Scoring Categories (100 points total)
|
||||
|
||||
**Determinism (35 points):**
|
||||
- No hard waits: 10 points
|
||||
- No conditionals: 10 points
|
||||
- No try-catch flow: 10 points
|
||||
- Network-first patterns: 5 points
|
||||
|
||||
**Isolation (25 points):**
|
||||
- Self-cleaning: 15 points
|
||||
- No global state: 5 points
|
||||
- Parallel-safe: 5 points
|
||||
|
||||
**Assertions (20 points):**
|
||||
- Explicit in test body: 10 points
|
||||
- Specific and meaningful: 10 points
|
||||
|
||||
**Structure (10 points):**
|
||||
- Test size < 300 lines: 5 points
|
||||
- Clear naming: 5 points
|
||||
|
||||
**Performance (10 points):**
|
||||
- Execution time < 1.5 min: 10 points
|
||||
|
||||
#### Quality Scoring Breakdown
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'14px'}}}%%
|
||||
pie title Test Quality Score (100 points)
|
||||
"Determinism" : 35
|
||||
"Isolation" : 25
|
||||
"Assertions" : 20
|
||||
"Structure" : 10
|
||||
"Performance" : 10
|
||||
```
|
||||
|
||||
```mermaid
|
||||
%%{init: {'theme':'base', 'themeVariables': { 'fontSize':'13px'}}}%%
|
||||
flowchart LR
|
||||
subgraph Det[Determinism - 35 pts]
|
||||
D1[No hard waits<br/>10 pts]
|
||||
D2[No conditionals<br/>10 pts]
|
||||
D3[No try-catch flow<br/>10 pts]
|
||||
D4[Network-first<br/>5 pts]
|
||||
end
|
||||
|
||||
subgraph Iso[Isolation - 25 pts]
|
||||
I1[Self-cleaning<br/>15 pts]
|
||||
I2[No global state<br/>5 pts]
|
||||
I3[Parallel-safe<br/>5 pts]
|
||||
end
|
||||
|
||||
subgraph Assrt[Assertions - 20 pts]
|
||||
A1[Explicit in body<br/>10 pts]
|
||||
A2[Specific/meaningful<br/>10 pts]
|
||||
end
|
||||
|
||||
subgraph Struct[Structure - 10 pts]
|
||||
S1[Size < 300 lines<br/>5 pts]
|
||||
S2[Clear naming<br/>5 pts]
|
||||
end
|
||||
|
||||
subgraph Perf[Performance - 10 pts]
|
||||
P1[Time < 1.5 min<br/>10 pts]
|
||||
end
|
||||
|
||||
Det --> Total([Total: 100 points])
|
||||
Iso --> Total
|
||||
Assrt --> Total
|
||||
Struct --> Total
|
||||
Perf --> Total
|
||||
|
||||
style Det fill:#ffebee,stroke:#c62828,stroke-width:2px
|
||||
style Iso fill:#e3f2fd,stroke:#1565c0,stroke-width:2px
|
||||
style Assrt fill:#f3e5f5,stroke:#6a1b9a,stroke-width:2px
|
||||
style Struct fill:#fff9c4,stroke:#f57f17,stroke-width:2px
|
||||
style Perf fill:#e8f5e9,stroke:#2e7d32,stroke-width:2px
|
||||
style Total fill:#fff,stroke:#000,stroke-width:3px
|
||||
```
|
||||
|
||||
### Score Interpretation
|
||||
|
||||
| Score | Interpretation | Action |
|
||||
| ---------- | -------------- | -------------------------------------- |
|
||||
| **90-100** | Excellent | Production-ready, minimal changes |
|
||||
| **80-89** | Good | Minor improvements recommended |
|
||||
| **70-79** | Acceptable | Address recommendations before release |
|
||||
| **60-69** | Needs Work | Fix critical issues |
|
||||
| **< 60** | Critical | Significant refactoring needed |
|
||||
|
||||
## Comparison: Good vs Bad Tests
|
||||
|
||||
### Example: User Login
|
||||
|
||||
**Bad Test (Score: 45/100):**
|
||||
```typescript
|
||||
test('login test', async ({ page }) => { // Vague name
|
||||
await page.goto('/login');
|
||||
await page.waitForTimeout(3000); // -10 (hard wait)
|
||||
|
||||
await page.fill('[name="email"]', 'test@example.com');
|
||||
await page.fill('[name="password"]', 'password');
|
||||
|
||||
if (await page.locator('.remember-me').isVisible()) { // -10 (conditional)
|
||||
await page.click('.remember-me');
|
||||
}
|
||||
|
||||
await page.click('button');
|
||||
|
||||
try { // -10 (try-catch flow)
|
||||
await page.waitForURL('/dashboard', { timeout: 5000 });
|
||||
} catch (e) {
|
||||
// Ignore navigation failure
|
||||
}
|
||||
|
||||
// No assertions! -10
|
||||
// No cleanup! -10
|
||||
});
|
||||
```
|
||||
|
||||
**Issues:**
|
||||
- Determinism: 5/35 (hard wait, conditional, try-catch)
|
||||
- Isolation: 10/25 (no cleanup)
|
||||
- Assertions: 0/20 (no assertions!)
|
||||
- Structure: 15/10 (okay)
|
||||
- Performance: 5/10 (slow)
|
||||
- **Total: 45/100**
|
||||
|
||||
**Good Test (Score: 95/100):**
|
||||
```typescript
|
||||
test('should login with valid credentials and redirect to dashboard', async ({ page, authSession }) => {
|
||||
// Use fixture for deterministic auth
|
||||
const loginPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/auth/login') && resp.ok()
|
||||
);
|
||||
|
||||
await page.goto('/login');
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByLabel('Password').fill('password123');
|
||||
await page.getByRole('button', { name: 'Sign in' }).click();
|
||||
|
||||
// Wait for actual API response
|
||||
const response = await loginPromise;
|
||||
const { token } = await response.json();
|
||||
|
||||
// Explicit assertions
|
||||
expect(token).toBeDefined();
|
||||
await expect(page).toHaveURL('/dashboard');
|
||||
await expect(page.getByText('Welcome back')).toBeVisible();
|
||||
|
||||
// Cleanup handled by authSession fixture
|
||||
});
|
||||
```
|
||||
|
||||
**Quality:**
|
||||
- Determinism: 35/35 (network-first, no conditionals)
|
||||
- Isolation: 25/25 (fixture handles cleanup)
|
||||
- Assertions: 20/20 (explicit and specific)
|
||||
- Structure: 10/10 (clear name, focused)
|
||||
- Performance: 5/10 (< 1 min)
|
||||
- **Total: 95/100**
|
||||
|
||||
### Example: API Testing
|
||||
|
||||
**Bad Test (Score: 50/100):**
|
||||
```typescript
|
||||
test('api test', async ({ request }) => {
|
||||
const response = await request.post('/api/users', {
|
||||
data: { email: 'test@example.com' } // Hard-coded (conflicts)
|
||||
});
|
||||
|
||||
if (response.ok()) { // Conditional
|
||||
const user = await response.json();
|
||||
// Weak assertion
|
||||
expect(user).toBeTruthy();
|
||||
}
|
||||
|
||||
// No cleanup - user left in database
|
||||
});
|
||||
```
|
||||
|
||||
**Good Test (Score: 92/100):**
|
||||
```typescript
|
||||
test('should create user with valid data', async ({ apiRequest }) => {
|
||||
// Unique test data
|
||||
const testEmail = `test-${Date.now()}@example.com`;
|
||||
|
||||
// Create user
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'POST',
|
||||
path: '/api/users',
|
||||
body: { email: testEmail, name: 'Test User' }
|
||||
});
|
||||
|
||||
// Explicit assertions
|
||||
expect(status).toBe(201);
|
||||
expect(body.id).toBeDefined();
|
||||
expect(body.email).toBe(testEmail);
|
||||
expect(body.name).toBe('Test User');
|
||||
|
||||
// Cleanup
|
||||
await apiRequest({
|
||||
method: 'DELETE',
|
||||
path: `/api/users/${body.id}`
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## How TEA Enforces Standards
|
||||
|
||||
### During Test Generation (`*atdd`, `*automate`)
|
||||
|
||||
TEA generates tests following standards by default:
|
||||
|
||||
```typescript
|
||||
// TEA-generated test (automatically follows standards)
|
||||
test('should submit contact form', async ({ page }) => {
|
||||
// Network-first pattern (no hard waits)
|
||||
const submitPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/contact') && resp.ok()
|
||||
);
|
||||
|
||||
// Accessible selectors (resilient)
|
||||
await page.getByLabel('Name').fill('Test User');
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByLabel('Message').fill('Test message');
|
||||
await page.getByRole('button', { name: 'Send' }).click();
|
||||
|
||||
const response = await submitPromise;
|
||||
const result = await response.json();
|
||||
|
||||
// Explicit assertions
|
||||
expect(result.success).toBe(true);
|
||||
await expect(page.getByText('Message sent')).toBeVisible();
|
||||
|
||||
// Size: 15 lines (< 300 ✓)
|
||||
// Execution: ~2 seconds (< 90s ✓)
|
||||
});
|
||||
```
|
||||
|
||||
### During Test Review (*test-review)
|
||||
|
||||
TEA audits tests and flags violations:
|
||||
|
||||
```markdown
|
||||
## Critical Issues
|
||||
|
||||
### Hard Wait Detected (tests/login.spec.ts:23)
|
||||
**Issue:** `await page.waitForTimeout(3000)`
|
||||
**Score Impact:** -10 (Determinism)
|
||||
**Fix:** Use network-first pattern
|
||||
|
||||
### Conditional Flow Control (tests/profile.spec.ts:45)
|
||||
**Issue:** `if (await page.locator('.banner').isVisible())`
|
||||
**Score Impact:** -10 (Determinism)
|
||||
**Fix:** Make banner presence deterministic
|
||||
|
||||
## Recommendations
|
||||
|
||||
### Extract Fixture (tests/auth.spec.ts)
|
||||
**Issue:** Login code repeated 5 times
|
||||
**Score Impact:** -3 (Structure)
|
||||
**Fix:** Extract to authSession fixture
|
||||
```
|
||||
|
||||
## Definition of Done Checklist
|
||||
|
||||
When is a test "done"?
|
||||
|
||||
**Test Quality DoD:**
|
||||
- [ ] No hard waits (`waitForTimeout`)
|
||||
- [ ] No conditionals for flow control
|
||||
- [ ] No try-catch for flow control
|
||||
- [ ] Network-first patterns used
|
||||
- [ ] Assertions explicit in test body
|
||||
- [ ] Test size < 300 lines
|
||||
- [ ] Clear, descriptive test name
|
||||
- [ ] Self-cleaning (cleanup in afterEach or test)
|
||||
- [ ] Unique test data (no hard-coded values)
|
||||
- [ ] Execution time < 1.5 minutes
|
||||
- [ ] Can run in parallel
|
||||
- [ ] Can run in any order
|
||||
|
||||
**Code Review DoD:**
|
||||
- [ ] Test quality score > 80
|
||||
- [ ] No critical issues from `*test-review`
|
||||
- [ ] Follows project patterns (fixtures, selectors)
|
||||
- [ ] Test reviewed by team member
|
||||
|
||||
## Common Quality Issues
|
||||
|
||||
### Issue: "My test needs conditionals for optional elements"
|
||||
|
||||
**Wrong approach:**
|
||||
```typescript
|
||||
if (await page.locator('.banner').isVisible()) {
|
||||
await page.click('.dismiss');
|
||||
}
|
||||
```
|
||||
|
||||
**Right approach - Make it deterministic:**
|
||||
```typescript
|
||||
// Option 1: Always expect banner
|
||||
await expect(page.locator('.banner')).toBeVisible();
|
||||
await page.click('.dismiss');
|
||||
|
||||
// Option 2: Test both scenarios separately
|
||||
test('should show banner for new users', ...);
|
||||
test('should not show banner for returning users', ...);
|
||||
```
|
||||
|
||||
### Issue: "My test needs try-catch for error handling"
|
||||
|
||||
**Wrong approach:**
|
||||
```typescript
|
||||
try {
|
||||
await page.click('#optional-button');
|
||||
} catch (e) {
|
||||
// Silently continue
|
||||
}
|
||||
```
|
||||
|
||||
**Right approach - Make failures explicit:**
|
||||
```typescript
|
||||
// Option 1: Button should exist
|
||||
await page.click('#optional-button'); // Fails loudly if missing
|
||||
|
||||
// Option 2: Button might not exist (test both)
|
||||
test('should work with optional button', async ({ page }) => {
|
||||
const hasButton = await page.locator('#optional-button').count() > 0;
|
||||
if (hasButton) {
|
||||
await page.click('#optional-button');
|
||||
}
|
||||
// But now you're testing optional behavior explicitly
|
||||
});
|
||||
```
|
||||
|
||||
### Issue: "Hard waits are easier than network patterns"
|
||||
|
||||
**Short-term:** Hard waits seem simpler
|
||||
**Long-term:** Flaky tests waste more time than learning network patterns
|
||||
|
||||
**Investment:**
|
||||
- 30 minutes to learn network-first patterns
|
||||
- Prevents hundreds of hours debugging flaky tests
|
||||
- Tests run faster (no wasted waits)
|
||||
- Team trusts test suite
|
||||
|
||||
## Technical Implementation
|
||||
|
||||
For detailed test quality patterns, see:
|
||||
- [Test Quality Fragment](/docs/reference/tea/knowledge-base.md#test-quality)
|
||||
- [Test Levels Framework Fragment](/docs/reference/tea/knowledge-base.md#test-levels-framework)
|
||||
- [Complete Knowledge Base Index](/docs/reference/tea/knowledge-base.md)
|
||||
|
||||
## Related Concepts
|
||||
|
||||
**Core TEA Concepts:**
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Quality scales with risk
|
||||
- [Knowledge Base System](/docs/explanation/tea/knowledge-base-system.md) - How standards are enforced
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - Quality in different models
|
||||
|
||||
**Technical Patterns:**
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Determinism explained
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Isolation through fixtures
|
||||
|
||||
**Overview:**
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Quality standards in lifecycle
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - Why quality matters
|
||||
|
||||
## Practical Guides
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Audit against these standards
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Generate quality tests
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Expand with quality
|
||||
|
||||
**Use-Case Guides:**
|
||||
- [Using TEA with Existing Tests](/docs/how-to/brownfield/use-tea-with-existing-tests.md) - Improve legacy quality
|
||||
- [Running TEA for Enterprise](/docs/how-to/enterprise/use-tea-for-enterprise.md) - Enterprise quality thresholds
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - *test-review command
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Test quality fragment
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - TEA terminology
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
@@ -1,76 +0,0 @@
|
||||
# Installation
|
||||
|
||||
Get BMAD Method running in your project in under 2 minutes.
|
||||
|
||||
## Quick Install
|
||||
|
||||
```bash
|
||||
npx bmad-method@alpha install
|
||||
```
|
||||
|
||||
This interactive installer will:
|
||||
|
||||
1. Detect your IDE (Claude Code, Cursor, VS Code, etc.)
|
||||
2. Let you choose which modules to install
|
||||
3. Configure agents and workflows for your project
|
||||
|
||||
## Requirements
|
||||
|
||||
- **Node.js** 18+ (for the installer)
|
||||
- **Git** (recommended for version control)
|
||||
- An **AI-powered IDE** or access to Claude/ChatGPT/Gemini
|
||||
|
||||
## Module Options
|
||||
|
||||
During installation, you'll choose which modules to install:
|
||||
|
||||
| Module | Description | Best For |
|
||||
| -------- | ---------------- | ---------------------------------------- |
|
||||
| **BMM** | BMAD Method Core | Software development projects |
|
||||
| **BMGD** | Game Development | Game projects with specialized workflows |
|
||||
| **BMB** | Builder | Creating custom agents and workflows |
|
||||
|
||||
## Post-Installation
|
||||
|
||||
After installation, your project will have:
|
||||
|
||||
```
|
||||
your-project/
|
||||
├── _bmad/ # BMAD configuration and agents
|
||||
│ ├── bmm/ # Method module (if installed)
|
||||
│ ├── bmgd/ # Game dev module (if installed)
|
||||
│ └── config.yaml # Your project configuration
|
||||
├── .claude/ # IDE-specific setup (varies by IDE)
|
||||
└── ... your code
|
||||
```
|
||||
|
||||
## Next Steps
|
||||
|
||||
1. **Read the [Quick Start Guide](../modules/bmm/quick-start.md)** to build your first feature
|
||||
2. **Check your [IDE Guide](../ide-info/index.md)** for IDE-specific tips
|
||||
3. **Explore [Workflows](../modules/bmm/workflows-planning.md)** to understand the methodology
|
||||
|
||||
## Alternative: Web Bundles
|
||||
|
||||
Don't want to install? Use BMAD agents directly in:
|
||||
|
||||
- **Claude Projects** - Upload the web bundle
|
||||
- **ChatGPT** - Use custom GPT bundles
|
||||
- **Gemini** - Import agent prompts
|
||||
|
||||
See the [Web Bundles Guide](../web-bundles-gemini-gpt-guide.md) for details.
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
**"Command not found: npx"**
|
||||
: Install Node.js 18+ from [nodejs.org](https://nodejs.org)
|
||||
|
||||
**"Permission denied"**
|
||||
: Run with appropriate permissions or check your npm configuration
|
||||
|
||||
**IDE not detected**
|
||||
: The installer will prompt you to select your IDE manually
|
||||
|
||||
For more help, see [Troubleshooting](../modules/bmm/troubleshooting.md) or join our [Discord](https://discord.gg/bmad).
|
||||
74
docs/how-to/brownfield/add-feature-to-existing.md
Normal file
74
docs/how-to/brownfield/add-feature-to-existing.md
Normal file
@@ -0,0 +1,74 @@
|
||||
---
|
||||
title: "How to Add a Feature to an Existing Project"
|
||||
description: How to add new features to an existing brownfield project
|
||||
---
|
||||
|
||||
Use the `workflow-init` workflow to add new functionality to your brownfield codebase while respecting existing patterns and architecture.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Adding a new feature to an existing codebase
|
||||
- Major enhancements that need proper planning
|
||||
- Features that touch multiple parts of the system
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- Existing project documentation (run `document-project` first if needed)
|
||||
- Clear understanding of what you want to build
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Run workflow-init
|
||||
|
||||
```
|
||||
Run workflow-init
|
||||
```
|
||||
|
||||
The workflow should recognize you're in an existing project. If not, explicitly clarify that this is brownfield development.
|
||||
|
||||
### 2. Choose Your Approach
|
||||
|
||||
| Feature Scope | Recommended Approach |
|
||||
|---------------|---------------------|
|
||||
| Small (1-5 stories) | Quick Flow with tech-spec |
|
||||
| Medium (5-15 stories) | BMad Method with PRD |
|
||||
| Large (15+ stories) | Full BMad Method with architecture |
|
||||
|
||||
### 3. Create Planning Documents
|
||||
|
||||
**For Quick Flow:**
|
||||
- Load PM agent
|
||||
- Run tech-spec workflow
|
||||
- The agent will analyze your existing codebase and create a context-aware spec
|
||||
|
||||
**For BMad Method:**
|
||||
- Load PM agent
|
||||
- Run PRD workflow
|
||||
- Ensure the agent reads your existing documentation
|
||||
- Review that integration points are clearly identified
|
||||
|
||||
### 4. Consider Architecture Impact
|
||||
|
||||
If your feature affects system architecture:
|
||||
|
||||
- Load Architect agent
|
||||
- Run architecture workflow
|
||||
- Ensure alignment with existing patterns
|
||||
- Document any new ADRs (Architecture Decision Records)
|
||||
|
||||
### 5. Implement
|
||||
|
||||
Follow the standard Phase 4 implementation workflows:
|
||||
|
||||
1. `sprint-planning` - Organize your work
|
||||
2. `create-story` - Prepare each story
|
||||
3. `dev-story` - Implement with tests
|
||||
4. `code-review` - Quality assurance
|
||||
|
||||
## Tips
|
||||
|
||||
- Always ensure agents read your existing documentation
|
||||
- Pay attention to integration points with existing code
|
||||
- Follow existing conventions unless deliberately changing them
|
||||
- Document why you're adding new patterns (if any)
|
||||
66
docs/how-to/brownfield/document-existing-project.md
Normal file
66
docs/how-to/brownfield/document-existing-project.md
Normal file
@@ -0,0 +1,66 @@
|
||||
---
|
||||
title: "How to Document an Existing Project"
|
||||
description: How to document an existing brownfield codebase using BMad Method
|
||||
---
|
||||
|
||||
Use the `document-project` workflow to scan your entire codebase and generate comprehensive documentation about its current state.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Starting work on an undocumented legacy project
|
||||
- Documentation is outdated and needs refresh
|
||||
- AI agents need context about existing code patterns
|
||||
- Onboarding new team members
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed in your project
|
||||
- Access to the codebase you want to document
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the Analyst Agent
|
||||
|
||||
Start a fresh chat and load the Analyst agent.
|
||||
|
||||
### 2. Run the document-project Workflow
|
||||
|
||||
Tell the agent:
|
||||
|
||||
```
|
||||
Run the document-project workflow
|
||||
```
|
||||
|
||||
### 3. Let the Agent Scan Your Codebase
|
||||
|
||||
The workflow will:
|
||||
|
||||
- Scan your codebase structure
|
||||
- Identify architecture patterns
|
||||
- Document the technology stack
|
||||
- Create reference documentation
|
||||
- Generate a PRD-like document from existing code
|
||||
|
||||
### 4. Review the Generated Documentation
|
||||
|
||||
The output will be saved to `project-documentation-{date}.md` in your output folder.
|
||||
|
||||
Review the documentation for:
|
||||
|
||||
- Accuracy of detected patterns
|
||||
- Completeness of architecture description
|
||||
- Any missing business rules or intent
|
||||
|
||||
## What You Get
|
||||
|
||||
- **Project overview** - High-level description of what the project does
|
||||
- **Technology stack** - Detected frameworks, libraries, and tools
|
||||
- **Architecture patterns** - Code organization and design patterns found
|
||||
- **Business rules** - Logic extracted from the codebase
|
||||
- **Integration points** - External APIs and services
|
||||
|
||||
## Tips
|
||||
|
||||
- Run this before any major brownfield work
|
||||
- Keep the documentation updated as the project evolves
|
||||
- Use it as input for future PRD creation
|
||||
84
docs/how-to/brownfield/index.md
Normal file
84
docs/how-to/brownfield/index.md
Normal file
@@ -0,0 +1,84 @@
|
||||
---
|
||||
title: "Brownfield Development"
|
||||
description: How to use BMad Method on existing codebases
|
||||
---
|
||||
|
||||
Use BMad Method effectively when working on existing projects and legacy codebases.
|
||||
|
||||
## What is Brownfield Development?
|
||||
|
||||
**Brownfield** refers to working on existing projects with established codebases and patterns, as opposed to **greenfield** which means starting from scratch with a clean slate.
|
||||
|
||||
This guide covers the essential workflow for onboarding to brownfield projects with BMad Method.
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed (`npx bmad-method install`)
|
||||
- An existing codebase you want to work on
|
||||
- Access to an AI-powered IDE (Claude Code, Cursor, or Windsurf)
|
||||
:::
|
||||
|
||||
## Step 1: Clean Up Completed Planning Artifacts
|
||||
|
||||
If you have completed all PRD epics and stories through the BMad process, clean up those files. Archive them, delete them, or rely on version history if needed. Do not keep these files in:
|
||||
|
||||
- `docs/`
|
||||
- `_bmad-output/planning-artifacts/`
|
||||
- `_bmad-output/implementation-artifacts/`
|
||||
|
||||
## Step 2: Maintain Quality Project Documentation
|
||||
|
||||
Your `docs/` folder should contain succinct, well-organized documentation that accurately represents your project:
|
||||
|
||||
- Intent and business rationale
|
||||
- Business rules
|
||||
- Architecture
|
||||
- Any other relevant project information
|
||||
|
||||
For complex projects, consider using the `document-project` workflow. It offers runtime variants that will scan your entire project and document its actual current state.
|
||||
|
||||
## Step 3: Initialize for Brownfield Work
|
||||
|
||||
Run `workflow-init`. It should recognize you are in an existing project. If not, explicitly clarify that this is brownfield development for a new feature.
|
||||
|
||||
### Choosing Your Approach
|
||||
|
||||
You have two primary options depending on the scope of changes:
|
||||
|
||||
| Scope | Recommended Approach |
|
||||
| ------------------------------ | ----------------------------------------------------------------------------------------------------------------------------- |
|
||||
| **Small updates or additions** | Use `quick-flow-solo-dev` to create a tech-spec and implement the change. The full four-phase BMad method is likely overkill. |
|
||||
| **Major changes or additions** | Start with the BMad method, applying as much or as little rigor as needed. |
|
||||
|
||||
### During PRD Creation
|
||||
|
||||
When creating a brief or jumping directly into the PRD, ensure the agent:
|
||||
|
||||
- Finds and analyzes your existing project documentation
|
||||
- Reads the proper context about your current system
|
||||
|
||||
You can guide the agent explicitly, but the goal is to ensure the new feature integrates well with your existing system.
|
||||
|
||||
### UX Considerations
|
||||
|
||||
UX work is optional. The decision depends not on whether your project has a UX, but on:
|
||||
|
||||
- Whether you will be working on UX changes
|
||||
- Whether significant new UX designs or patterns are needed
|
||||
|
||||
If your changes amount to simple updates to existing screens you are happy with, a full UX process is unnecessary.
|
||||
|
||||
### Architecture Considerations
|
||||
|
||||
When doing architecture, ensure the architect:
|
||||
|
||||
- Uses the proper documented files
|
||||
- Scans the existing codebase
|
||||
|
||||
Pay close attention here to prevent reinventing the wheel or making decisions that misalign with your existing architecture.
|
||||
|
||||
## Next Steps
|
||||
|
||||
- **[Document Existing Project](/docs/how-to/brownfield/document-existing-project.md)** - How to document your brownfield codebase
|
||||
- **[Add Feature to Existing Project](/docs/how-to/brownfield/add-feature-to-existing.md)** - Adding new functionality
|
||||
- **[Quick Fix in Brownfield](/docs/how-to/brownfield/quick-fix-in-brownfield.md)** - Bug fixes and ad-hoc changes
|
||||
- **[Brownfield FAQ](/docs/explanation/faq/brownfield-faq.md)** - Common questions about brownfield development
|
||||
77
docs/how-to/brownfield/quick-fix-in-brownfield.md
Normal file
77
docs/how-to/brownfield/quick-fix-in-brownfield.md
Normal file
@@ -0,0 +1,77 @@
|
||||
---
|
||||
title: "How to Make Quick Fixes in Brownfield Projects"
|
||||
description: How to make quick fixes and ad-hoc changes in brownfield projects
|
||||
---
|
||||
|
||||
Use the **DEV agent** directly for bug fixes, refactorings, or small targeted changes that don't require the full BMad method or Quick Flow.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Bug fixes
|
||||
- Small refactorings
|
||||
- Targeted code improvements
|
||||
- Learning about your codebase
|
||||
- One-off changes that don't need planning
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load an Agent
|
||||
|
||||
For quick fixes, you can use:
|
||||
|
||||
- **DEV agent** - For implementation-focused work
|
||||
- **Quick Flow Solo Dev** - For slightly larger changes that still need a tech-spec
|
||||
|
||||
### 2. Describe the Change
|
||||
|
||||
Simply tell the agent what you need:
|
||||
|
||||
```
|
||||
Fix the login validation bug that allows empty passwords
|
||||
```
|
||||
|
||||
or
|
||||
|
||||
```
|
||||
Refactor the UserService to use async/await instead of callbacks
|
||||
```
|
||||
|
||||
### 3. Let the Agent Work
|
||||
|
||||
The agent will:
|
||||
|
||||
- Analyze the relevant code
|
||||
- Propose a solution
|
||||
- Implement the change
|
||||
- Run tests (if available)
|
||||
|
||||
### 4. Review and Commit
|
||||
|
||||
Review the changes made and commit when satisfied.
|
||||
|
||||
## Learning Your Codebase
|
||||
|
||||
This approach is also excellent for exploring unfamiliar code:
|
||||
|
||||
```
|
||||
Explain how the authentication system works in this codebase
|
||||
```
|
||||
|
||||
```
|
||||
Show me where error handling happens in the API layer
|
||||
```
|
||||
|
||||
LLMs are excellent at interpreting and analyzing code—whether it was AI-generated or not. Use the agent to:
|
||||
|
||||
- Learn about your project
|
||||
- Understand how things are built
|
||||
- Explore unfamiliar parts of the codebase
|
||||
|
||||
## When to Upgrade to Formal Planning
|
||||
|
||||
Consider using Quick Flow or full BMad Method when:
|
||||
|
||||
- The change affects multiple files or systems
|
||||
- You're unsure about the scope
|
||||
- The fix keeps growing in complexity
|
||||
- You need documentation for the change
|
||||
577
docs/how-to/brownfield/use-tea-with-existing-tests.md
Normal file
577
docs/how-to/brownfield/use-tea-with-existing-tests.md
Normal file
@@ -0,0 +1,577 @@
|
||||
---
|
||||
title: "Using TEA with Existing Tests (Brownfield)"
|
||||
description: Apply TEA workflows to legacy codebases with existing test suites
|
||||
---
|
||||
|
||||
# Using TEA with Existing Tests (Brownfield)
|
||||
|
||||
Use TEA on brownfield projects (existing codebases with legacy tests) to establish coverage baselines, identify gaps, and improve test quality without starting from scratch.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Existing codebase with some tests already written
|
||||
- Legacy test suite needs quality improvement
|
||||
- Adding features to existing application
|
||||
- Need to understand current test coverage
|
||||
- Want to prevent regression as you add features
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Existing codebase with tests (even if incomplete or low quality)
|
||||
- Tests run successfully (or at least can be executed)
|
||||
|
||||
**Note:** If your codebase is completely undocumented, run `*document-project` first to create baseline documentation.
|
||||
|
||||
## Brownfield Strategy
|
||||
|
||||
### Phase 1: Establish Baseline
|
||||
|
||||
Understand what you have before changing anything.
|
||||
|
||||
#### Step 1: Baseline Coverage with *trace
|
||||
|
||||
Run `*trace` Phase 1 to map existing tests to requirements:
|
||||
|
||||
```
|
||||
*trace
|
||||
```
|
||||
|
||||
**Select:** Phase 1 (Requirements Traceability)
|
||||
|
||||
**Provide:**
|
||||
- Existing requirements docs (PRD, user stories, feature specs)
|
||||
- Test location (`tests/` or wherever tests live)
|
||||
- Focus areas (specific features if large codebase)
|
||||
|
||||
**Output:** `traceability-matrix.md` showing:
|
||||
- Which requirements have tests
|
||||
- Which requirements lack coverage
|
||||
- Coverage classification (FULL/PARTIAL/NONE)
|
||||
- Gap prioritization
|
||||
|
||||
**Example Baseline:**
|
||||
```markdown
|
||||
# Baseline Coverage (Before Improvements)
|
||||
|
||||
**Total Requirements:** 50
|
||||
**Full Coverage:** 15 (30%)
|
||||
**Partial Coverage:** 20 (40%)
|
||||
**No Coverage:** 15 (30%)
|
||||
|
||||
**By Priority:**
|
||||
- P0: 50% coverage (5/10) ❌ Critical gap
|
||||
- P1: 40% coverage (8/20) ⚠️ Needs improvement
|
||||
- P2: 20% coverage (2/10) ✅ Acceptable
|
||||
```
|
||||
|
||||
This baseline becomes your improvement target.
|
||||
|
||||
#### Step 2: Quality Audit with *test-review
|
||||
|
||||
Run `*test-review` on existing tests:
|
||||
|
||||
```
|
||||
*test-review tests/
|
||||
```
|
||||
|
||||
**Output:** `test-review.md` with quality score and issues.
|
||||
|
||||
**Common Brownfield Issues:**
|
||||
- Hard waits everywhere (`page.waitForTimeout(5000)`)
|
||||
- Fragile CSS selectors (`.class > div:nth-child(3)`)
|
||||
- No test isolation (tests depend on execution order)
|
||||
- Try-catch for flow control
|
||||
- Tests don't clean up (leave test data in DB)
|
||||
|
||||
**Example Baseline Quality:**
|
||||
```markdown
|
||||
# Quality Score: 55/100
|
||||
|
||||
**Critical Issues:** 12
|
||||
- 8 hard waits
|
||||
- 4 conditional flow control
|
||||
|
||||
**Recommendations:** 25
|
||||
- Extract fixtures
|
||||
- Improve selectors
|
||||
- Add network assertions
|
||||
```
|
||||
|
||||
This shows where to focus improvement efforts.
|
||||
|
||||
### Phase 2: Prioritize Improvements
|
||||
|
||||
Don't try to fix everything at once.
|
||||
|
||||
#### Focus on Critical Path First
|
||||
|
||||
**Priority 1: P0 Requirements**
|
||||
```
|
||||
Goal: Get P0 coverage to 100%
|
||||
|
||||
Actions:
|
||||
1. Identify P0 requirements with no tests (from trace)
|
||||
2. Run *automate to generate tests for missing P0 scenarios
|
||||
3. Fix critical quality issues in P0 tests (from test-review)
|
||||
```
|
||||
|
||||
**Priority 2: Fix Flaky Tests**
|
||||
```
|
||||
Goal: Eliminate flakiness
|
||||
|
||||
Actions:
|
||||
1. Identify tests with hard waits (from test-review)
|
||||
2. Replace with network-first patterns
|
||||
3. Run burn-in loops to verify stability
|
||||
```
|
||||
|
||||
**Example Modernization:**
|
||||
|
||||
**Before (Flaky - Hard Waits):**
|
||||
```typescript
|
||||
test('checkout completes', async ({ page }) => {
|
||||
await page.click('button[name="checkout"]');
|
||||
await page.waitForTimeout(5000); // ❌ Flaky
|
||||
await expect(page.locator('.confirmation')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**After (Network-First - Vanilla):**
|
||||
```typescript
|
||||
test('checkout completes', async ({ page }) => {
|
||||
const checkoutPromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/checkout') && resp.ok()
|
||||
);
|
||||
await page.click('button[name="checkout"]');
|
||||
await checkoutPromise; // ✅ Deterministic
|
||||
await expect(page.locator('.confirmation')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**After (With Playwright Utils - Cleaner API):**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('checkout completes', async ({ page, interceptNetworkCall }) => {
|
||||
// Use interceptNetworkCall for cleaner network interception
|
||||
const checkoutCall = interceptNetworkCall({
|
||||
method: 'POST',
|
||||
url: '**/api/checkout'
|
||||
});
|
||||
|
||||
await page.click('button[name="checkout"]');
|
||||
|
||||
// Wait for response (automatic JSON parsing)
|
||||
const { status, responseJson: order } = await checkoutCall;
|
||||
|
||||
// Validate API response
|
||||
expect(status).toBe(200);
|
||||
expect(order.status).toBe('confirmed');
|
||||
|
||||
// Validate UI
|
||||
await expect(page.locator('.confirmation')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- `interceptNetworkCall` for cleaner network interception
|
||||
- Automatic JSON parsing (`responseJson` ready to use)
|
||||
- No manual `await response.json()`
|
||||
- Glob pattern matching (`**/api/checkout`)
|
||||
- Cleaner, more maintainable code
|
||||
|
||||
**For automatic error detection,** use `network-error-monitor` fixture separately. See [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md#network-error-monitor).
|
||||
|
||||
**Priority 3: P1 Requirements**
|
||||
```
|
||||
Goal: Get P1 coverage to 80%+
|
||||
|
||||
Actions:
|
||||
1. Generate tests for highest-risk P1 gaps
|
||||
2. Improve test quality incrementally
|
||||
```
|
||||
|
||||
#### Create Improvement Roadmap
|
||||
|
||||
```markdown
|
||||
# Test Improvement Roadmap
|
||||
|
||||
## Week 1: Critical Path (P0)
|
||||
- [ ] Add 5 missing P0 tests (Epic 1: Auth)
|
||||
- [ ] Fix 8 hard waits in auth tests
|
||||
- [ ] Verify P0 coverage = 100%
|
||||
|
||||
## Week 2: Flakiness
|
||||
- [ ] Replace all hard waits with network-first
|
||||
- [ ] Fix conditional flow control
|
||||
- [ ] Run burn-in loops (target: 0 failures in 10 runs)
|
||||
|
||||
## Week 3: High-Value Coverage (P1)
|
||||
- [ ] Add 10 missing P1 tests
|
||||
- [ ] Improve selector resilience
|
||||
- [ ] P1 coverage target: 80%
|
||||
|
||||
## Week 4: Quality Polish
|
||||
- [ ] Extract fixtures for common patterns
|
||||
- [ ] Add network assertions
|
||||
- [ ] Quality score target: 75+
|
||||
```
|
||||
|
||||
### Phase 3: Incremental Improvement
|
||||
|
||||
Apply TEA workflows to new work while improving legacy tests.
|
||||
|
||||
#### For New Features (Greenfield Within Brownfield)
|
||||
|
||||
**Use full TEA workflow:**
|
||||
```
|
||||
1. *test-design (epic-level) - Plan tests for new feature
|
||||
2. *atdd - Generate failing tests first (TDD)
|
||||
3. Implement feature
|
||||
4. *automate - Expand coverage
|
||||
5. *test-review - Ensure quality
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- New code has high-quality tests from day one
|
||||
- Gradually raises overall quality
|
||||
- Team learns good patterns
|
||||
|
||||
#### For Bug Fixes (Regression Prevention)
|
||||
|
||||
**Add regression tests:**
|
||||
```
|
||||
1. Reproduce bug with failing test
|
||||
2. Fix bug
|
||||
3. Verify test passes
|
||||
4. Run *test-review on regression test
|
||||
5. Add to regression test suite
|
||||
```
|
||||
|
||||
#### For Refactoring (Regression Safety)
|
||||
|
||||
**Before refactoring:**
|
||||
```
|
||||
1. Run *trace - Baseline coverage
|
||||
2. Note current coverage %
|
||||
3. Refactor code
|
||||
4. Run *trace - Verify coverage maintained
|
||||
5. No coverage should decrease
|
||||
```
|
||||
|
||||
### Phase 4: Continuous Improvement
|
||||
|
||||
Track improvement over time.
|
||||
|
||||
#### Quarterly Quality Audits
|
||||
|
||||
**Q1 Baseline:**
|
||||
```
|
||||
Coverage: 30%
|
||||
Quality Score: 55/100
|
||||
Flakiness: 15% fail rate
|
||||
```
|
||||
|
||||
**Q2 Target:**
|
||||
```
|
||||
Coverage: 50% (focus on P0)
|
||||
Quality Score: 65/100
|
||||
Flakiness: 5%
|
||||
```
|
||||
|
||||
**Q3 Target:**
|
||||
```
|
||||
Coverage: 70%
|
||||
Quality Score: 75/100
|
||||
Flakiness: 1%
|
||||
```
|
||||
|
||||
**Q4 Target:**
|
||||
```
|
||||
Coverage: 85%
|
||||
Quality Score: 85/100
|
||||
Flakiness: <0.5%
|
||||
```
|
||||
|
||||
## Brownfield-Specific Tips
|
||||
|
||||
### Don't Rewrite Everything
|
||||
|
||||
**Common mistake:**
|
||||
```
|
||||
"Our tests are bad, let's delete them all and start over!"
|
||||
```
|
||||
|
||||
**Better approach:**
|
||||
```
|
||||
"Our tests are bad, let's:
|
||||
1. Keep tests that work (even if not perfect)
|
||||
2. Fix critical quality issues incrementally
|
||||
3. Add tests for gaps
|
||||
4. Gradually improve over time"
|
||||
```
|
||||
|
||||
**Why:**
|
||||
- Rewriting is risky (might lose coverage)
|
||||
- Incremental improvement is safer
|
||||
- Team learns gradually
|
||||
- Business value delivered continuously
|
||||
|
||||
### Use Regression Hotspots
|
||||
|
||||
**Identify regression-prone areas:**
|
||||
```markdown
|
||||
## Regression Hotspots
|
||||
|
||||
**Based on:**
|
||||
- Bug reports (last 6 months)
|
||||
- Customer complaints
|
||||
- Code complexity (cyclomatic complexity >10)
|
||||
- Frequent changes (git log analysis)
|
||||
|
||||
**High-Risk Areas:**
|
||||
1. Authentication flow (12 bugs in 6 months)
|
||||
2. Checkout process (8 bugs)
|
||||
3. Payment integration (6 bugs)
|
||||
|
||||
**Test Priority:**
|
||||
- Add regression tests for these areas FIRST
|
||||
- Ensure P0 coverage before touching code
|
||||
```
|
||||
|
||||
### Quarantine Flaky Tests
|
||||
|
||||
Don't let flaky tests block improvement:
|
||||
|
||||
```typescript
|
||||
// Mark flaky tests with .skip temporarily
|
||||
test.skip('flaky test - needs fixing', async ({ page }) => {
|
||||
// TODO: Fix hard wait on line 45
|
||||
// TODO: Add network-first pattern
|
||||
});
|
||||
```
|
||||
|
||||
**Track quarantined tests:**
|
||||
```markdown
|
||||
# Quarantined Tests
|
||||
|
||||
| Test | Reason | Owner | Target Fix Date |
|
||||
| ------------------- | -------------------------- | -------- | --------------- |
|
||||
| checkout.spec.ts:45 | Hard wait causes flakiness | QA Team | 2026-01-20 |
|
||||
| profile.spec.ts:28 | Conditional flow control | Dev Team | 2026-01-25 |
|
||||
```
|
||||
|
||||
**Fix systematically:**
|
||||
- Don't accumulate quarantined tests
|
||||
- Set deadlines for fixes
|
||||
- Review quarantine list weekly
|
||||
|
||||
### Migrate One Directory at a Time
|
||||
|
||||
**Large test suite?** Improve incrementally:
|
||||
|
||||
**Week 1:** `tests/auth/`
|
||||
```
|
||||
1. Run *test-review on auth tests
|
||||
2. Fix critical issues
|
||||
3. Re-review
|
||||
4. Mark directory as "modernized"
|
||||
```
|
||||
|
||||
**Week 2:** `tests/api/`
|
||||
```
|
||||
Same process
|
||||
```
|
||||
|
||||
**Week 3:** `tests/e2e/`
|
||||
```
|
||||
Same process
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Focused improvement
|
||||
- Visible progress
|
||||
- Team learns patterns
|
||||
- Lower risk
|
||||
|
||||
### Document Migration Status
|
||||
|
||||
**Track which tests are modernized:**
|
||||
|
||||
```markdown
|
||||
# Test Suite Status
|
||||
|
||||
| Directory | Tests | Quality Score | Status | Notes |
|
||||
| ------------------ | ----- | ------------- | ------------- | -------------- |
|
||||
| tests/auth/ | 15 | 85/100 | ✅ Modernized | Week 1 cleanup |
|
||||
| tests/api/ | 32 | 78/100 | ⚠️ In Progress | Week 2 |
|
||||
| tests/e2e/ | 28 | 62/100 | ❌ Legacy | Week 3 planned |
|
||||
| tests/integration/ | 12 | 45/100 | ❌ Legacy | Week 4 planned |
|
||||
|
||||
**Legend:**
|
||||
- ✅ Modernized: Quality >80, no critical issues
|
||||
- ⚠️ In Progress: Active improvement
|
||||
- ❌ Legacy: Not yet touched
|
||||
```
|
||||
|
||||
## Common Brownfield Challenges
|
||||
|
||||
### "We Don't Know What Tests Cover"
|
||||
|
||||
**Problem:** No documentation, unclear what tests do.
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
1. Run *trace - TEA analyzes tests and maps to requirements
|
||||
2. Review traceability matrix
|
||||
3. Document findings
|
||||
4. Use as baseline for improvement
|
||||
```
|
||||
|
||||
TEA reverse-engineers test coverage even without documentation.
|
||||
|
||||
### "Tests Are Too Brittle to Touch"
|
||||
|
||||
**Problem:** Afraid to modify tests (might break them).
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
1. Run tests, capture current behavior (baseline)
|
||||
2. Make small improvement (fix one hard wait)
|
||||
3. Run tests again
|
||||
4. If still pass, continue
|
||||
5. If fail, investigate why
|
||||
|
||||
Incremental changes = lower risk
|
||||
```
|
||||
|
||||
### "No One Knows How to Run Tests"
|
||||
|
||||
**Problem:** Test documentation is outdated or missing.
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
1. Document manually or ask TEA to help analyze test structure
|
||||
2. Create tests/README.md with:
|
||||
- How to install dependencies
|
||||
- How to run tests (npx playwright test, npm test, etc.)
|
||||
- What each test directory contains
|
||||
- Common issues and troubleshooting
|
||||
3. Commit documentation for team
|
||||
```
|
||||
|
||||
**Note:** `*framework` is for new test setup, not existing tests. For brownfield, document what you have.
|
||||
|
||||
### "Tests Take Hours to Run"
|
||||
|
||||
**Problem:** Full test suite takes 4+ hours.
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
1. Configure parallel execution (shard tests across workers)
|
||||
2. Add selective testing (run only affected tests on PR)
|
||||
3. Run full suite nightly only
|
||||
4. Optimize slow tests (remove hard waits, improve selectors)
|
||||
|
||||
Before: 4 hours sequential
|
||||
After: 15 minutes with sharding + selective testing
|
||||
```
|
||||
|
||||
**How `*ci` helps:**
|
||||
- Scaffolds CI configuration with parallel sharding examples
|
||||
- Provides selective testing script templates
|
||||
- Documents burn-in and optimization strategies
|
||||
- But YOU configure workers, test selection, and optimization
|
||||
|
||||
**With Playwright Utils burn-in:**
|
||||
- Smart selective testing based on git diff
|
||||
- Volume control (run percentage of affected tests)
|
||||
- See [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md#burn-in)
|
||||
|
||||
### "We Have Tests But They Always Fail"
|
||||
|
||||
**Problem:** Tests are so flaky they're ignored.
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
1. Run *test-review to identify flakiness patterns
|
||||
2. Fix top 5 flaky tests (biggest impact)
|
||||
3. Quarantine remaining flaky tests
|
||||
4. Re-enable as you fix them
|
||||
|
||||
Don't let perfect be the enemy of good
|
||||
```
|
||||
|
||||
## Brownfield TEA Workflow
|
||||
|
||||
### Recommended Sequence
|
||||
|
||||
**1. Documentation (if needed):**
|
||||
```
|
||||
*document-project
|
||||
```
|
||||
|
||||
**2. Baseline (Phase 2):**
|
||||
```
|
||||
*trace Phase 1 - Establish coverage baseline
|
||||
*test-review - Establish quality baseline
|
||||
```
|
||||
|
||||
**3. Planning (Phase 2-3):**
|
||||
```
|
||||
*prd - Document requirements (if missing)
|
||||
*architecture - Document architecture (if missing)
|
||||
*test-design (system-level) - Testability review
|
||||
```
|
||||
|
||||
**4. Infrastructure (Phase 3):**
|
||||
```
|
||||
*framework - Modernize test framework (if needed)
|
||||
*ci - Setup or improve CI/CD
|
||||
```
|
||||
|
||||
**5. Per Epic (Phase 4):**
|
||||
```
|
||||
*test-design (epic-level) - Focus on regression hotspots
|
||||
*automate - Add missing tests
|
||||
*test-review - Ensure quality
|
||||
*trace Phase 1 - Refresh coverage
|
||||
```
|
||||
|
||||
**6. Release Gate:**
|
||||
```
|
||||
*nfr-assess - Validate NFRs (if enterprise)
|
||||
*trace Phase 2 - Gate decision
|
||||
```
|
||||
|
||||
## Related Guides
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md) - Baseline coverage analysis
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Quality audit
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Fill coverage gaps
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Risk assessment
|
||||
|
||||
**Customization:**
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - Modernize tests with utilities
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - Brownfield model explained
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - What makes tests good
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Fix flakiness
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Prioritize improvements
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - All 8 workflows
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Config options
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Testing patterns
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - TEA terminology
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
@@ -1,6 +1,15 @@
|
||||
# Agent Customization Guide
|
||||
---
|
||||
title: "Agent Customization Guide"
|
||||
---
|
||||
|
||||
Customize BMad agents without modifying core files. All customizations persist through updates.
|
||||
Use `.customize.yaml` files to customize BMad agents without modifying core files. All customizations persist through updates.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Change agent names or personas
|
||||
- Add project-specific memories or context
|
||||
- Add custom menu items and workflows
|
||||
- Define critical actions for consistent behavior
|
||||
|
||||
## Quick Start
|
||||
|
||||
@@ -26,10 +35,8 @@ After editing, IT IS CRITICAL to rebuild the agent to apply changes:
|
||||
|
||||
```bash
|
||||
npx bmad-method@alpha install # and then select option to compile all agents
|
||||
# OR for individual agent only
|
||||
npx bmad-method@alpha build <agent-name>
|
||||
|
||||
# Examples:
|
||||
npx bmad-method@alpha build bmm-dev
|
||||
npx bmad-method@alpha build core-bmad-master
|
||||
npx bmad-method@alpha build bmm-pm
|
||||
@@ -81,7 +88,7 @@ Add your own workflows to the agent's menu:
|
||||
```yaml
|
||||
menu:
|
||||
- trigger: my-workflow
|
||||
workflow: '{project-root}/custom/my-workflow.yaml'
|
||||
workflow: '{project-root}/my-custom/workflows/my-workflow.yaml'
|
||||
description: My custom workflow
|
||||
- trigger: deploy
|
||||
action: '#deploy-prompt'
|
||||
@@ -119,7 +126,6 @@ prompts:
|
||||
**Example 1: Customize Developer Agent for TDD**
|
||||
|
||||
```yaml
|
||||
# _bmad/_config/agents/bmm-dev.customize.yaml
|
||||
agent:
|
||||
metadata:
|
||||
name: 'TDD Developer'
|
||||
@@ -135,7 +141,6 @@ critical_actions:
|
||||
**Example 2: Add Custom Deployment Workflow**
|
||||
|
||||
```yaml
|
||||
# _bmad/_config/agents/bmm-dev.customize.yaml
|
||||
menu:
|
||||
- trigger: deploy-staging
|
||||
workflow: '{project-root}/_bmad/deploy-staging.yaml'
|
||||
@@ -148,7 +153,6 @@ menu:
|
||||
**Example 3: Multilingual Product Manager**
|
||||
|
||||
```yaml
|
||||
# _bmad/_config/agents/bmm-pm.customize.yaml
|
||||
persona:
|
||||
role: 'Bilingual Product Manager'
|
||||
identity: 'Expert in US and LATAM markets'
|
||||
@@ -203,6 +207,6 @@ memories:
|
||||
|
||||
## Next Steps
|
||||
|
||||
- **[BMM Agents Guide](./modules/bmm/agents-guide)** - Learn about the BMad Method agents
|
||||
- **[BMB Create Agent Workflow](./modules/bmb/agents/index)** - Build completely custom agents
|
||||
- **[BMM Complete Documentation](./modules/bmm/index)** - Full BMad Method reference
|
||||
- **[Learn about Agents](/docs/explanation/core-concepts/what-are-agents.md)** - Understand Simple vs Expert agents
|
||||
- **[Agent Creation Guide](/docs/tutorials/advanced/create-custom-agent.md)** - Build completely custom agents
|
||||
- **[BMM Complete Documentation](/docs/explanation/bmm/index.md)** - Full BMad Method reference
|
||||
424
docs/how-to/customization/enable-tea-mcp-enhancements.md
Normal file
424
docs/how-to/customization/enable-tea-mcp-enhancements.md
Normal file
@@ -0,0 +1,424 @@
|
||||
---
|
||||
title: "Enable TEA MCP Enhancements"
|
||||
description: Configure Playwright MCP servers for live browser verification during TEA workflows
|
||||
---
|
||||
|
||||
# Enable TEA MCP Enhancements
|
||||
|
||||
Configure Model Context Protocol (MCP) servers to enable live browser verification, exploratory mode, and recording mode in TEA workflows.
|
||||
|
||||
## What are MCP Enhancements?
|
||||
|
||||
MCP (Model Context Protocol) servers enable AI agents to interact with live browsers during test generation. This allows TEA to:
|
||||
|
||||
- **Explore UIs interactively** - Discover actual functionality through browser automation
|
||||
- **Verify selectors** - Generate accurate locators from real DOM
|
||||
- **Validate behavior** - Confirm test scenarios against live applications
|
||||
- **Debug visually** - Use trace viewer and screenshots during generation
|
||||
|
||||
## When to Use This
|
||||
|
||||
**For UI Testing:**
|
||||
- Want exploratory mode in `*test-design` (browser-based UI discovery)
|
||||
- Want recording mode in `*atdd` or `*automate` (verify selectors with live browser)
|
||||
- Want healing mode in `*automate` (fix tests with visual debugging)
|
||||
- Need accurate selectors from actual DOM
|
||||
- Debugging complex UI interactions
|
||||
|
||||
**For API Testing:**
|
||||
- Want healing mode in `*automate` (analyze failures with trace data)
|
||||
- Need to debug test failures (network responses, request/response data, timing)
|
||||
- Want to inspect trace files (network traffic, errors, race conditions)
|
||||
|
||||
**For Both:**
|
||||
- Visual debugging (trace viewer shows network + UI)
|
||||
- Test failure analysis (MCP can run tests and extract errors)
|
||||
- Understanding complex test failures (network + DOM together)
|
||||
|
||||
**Don't use if:**
|
||||
- You don't have MCP servers configured
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- IDE with MCP support (Cursor, VS Code with Claude extension)
|
||||
- Node.js v18 or later
|
||||
- Playwright installed
|
||||
|
||||
## Available MCP Servers
|
||||
|
||||
**Two Playwright MCP servers** (actively maintained, continuously updated):
|
||||
|
||||
### 1. Playwright MCP - Browser Automation
|
||||
|
||||
**Command:** `npx @playwright/mcp@latest`
|
||||
|
||||
**Capabilities:**
|
||||
- Navigate to URLs
|
||||
- Click elements
|
||||
- Fill forms
|
||||
- Take screenshots
|
||||
- Extract DOM information
|
||||
|
||||
**Best for:** Exploratory mode, recording mode
|
||||
|
||||
### 2. Playwright Test MCP - Test Runner
|
||||
|
||||
**Command:** `npx playwright run-test-mcp-server`
|
||||
|
||||
**Capabilities:**
|
||||
- Run test files
|
||||
- Analyze failures
|
||||
- Extract error messages
|
||||
- Show trace files
|
||||
|
||||
**Best for:** Healing mode, debugging
|
||||
|
||||
### Recommended: Configure Both
|
||||
|
||||
Both servers work together to provide full TEA MCP capabilities.
|
||||
|
||||
## Setup
|
||||
|
||||
### 1. Configure MCP Servers
|
||||
|
||||
Add to your IDE's MCP configuration:
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"playwright": {
|
||||
"command": "npx",
|
||||
"args": ["@playwright/mcp@latest"]
|
||||
},
|
||||
"playwright-test": {
|
||||
"command": "npx",
|
||||
"args": ["playwright", "run-test-mcp-server"]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
See [TEA Overview](/docs/explanation/features/tea-overview.md#playwright-mcp-enhancements) for IDE-specific config locations.
|
||||
|
||||
### 2. Enable in BMAD
|
||||
|
||||
Answer "Yes" when prompted during installation, or set in config:
|
||||
|
||||
```yaml
|
||||
# _bmad/bmm/config.yaml
|
||||
tea_use_mcp_enhancements: true
|
||||
```
|
||||
|
||||
### 3. Verify MCPs Running
|
||||
|
||||
Ensure your MCP servers are running in your IDE.
|
||||
|
||||
## How MCP Enhances TEA Workflows
|
||||
|
||||
### *test-design: Exploratory Mode
|
||||
|
||||
**Without MCP:**
|
||||
- TEA infers UI functionality from documentation
|
||||
- Relies on your description of features
|
||||
- May miss actual UI behavior
|
||||
|
||||
**With MCP:**
|
||||
TEA can open live browser to:
|
||||
```
|
||||
"Let me explore the profile page to understand the UI"
|
||||
|
||||
[TEA navigates to /profile]
|
||||
[Takes screenshot]
|
||||
[Extracts accessible elements]
|
||||
|
||||
"I see the profile has:
|
||||
- Name field (editable)
|
||||
- Email field (editable)
|
||||
- Avatar upload button
|
||||
- Save button
|
||||
- Cancel button
|
||||
|
||||
I'll design tests for these interactions."
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Accurate test design based on actual UI
|
||||
- Discovers functionality you might not describe
|
||||
- Validates test scenarios are possible
|
||||
|
||||
### *atdd: Recording Mode
|
||||
|
||||
**Without MCP:**
|
||||
- TEA generates selectors from best practices
|
||||
- TEA infers API patterns from documentation
|
||||
|
||||
**With MCP (Recording Mode):**
|
||||
|
||||
**For UI Tests:**
|
||||
```
|
||||
[TEA navigates to /login with live browser]
|
||||
[Inspects actual form fields]
|
||||
|
||||
"I see:
|
||||
- Email input has label 'Email Address' (not 'Email')
|
||||
- Password input has label 'Your Password'
|
||||
- Submit button has text 'Sign In' (not 'Login')
|
||||
|
||||
I'll use these exact selectors."
|
||||
```
|
||||
|
||||
**For API Tests:**
|
||||
```
|
||||
[TEA analyzes trace files from test runs]
|
||||
[Inspects network requests/responses]
|
||||
|
||||
"I see the API returns:
|
||||
- POST /api/login → 200 with { token, userId }
|
||||
- Response time: 150ms
|
||||
- Required headers: Content-Type, Authorization
|
||||
|
||||
I'll validate these in tests."
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- UI: Accurate selectors from real DOM
|
||||
- API: Validated request/response patterns from trace
|
||||
- Both: Tests work on first run
|
||||
|
||||
### *automate: Healing + Recording Modes
|
||||
|
||||
**Without MCP:**
|
||||
- TEA analyzes test code only
|
||||
- Suggests fixes based on static analysis
|
||||
- Generates tests from documentation/code
|
||||
|
||||
**With MCP:**
|
||||
|
||||
**Healing Mode (UI + API):**
|
||||
```
|
||||
[TEA opens trace file]
|
||||
[Analyzes screenshots + network tab]
|
||||
|
||||
UI failures: "Button selector changed from 'Save' to 'Save Changes'"
|
||||
API failures: "Response structure changed, expected {id} got {userId}"
|
||||
|
||||
[TEA makes fixes]
|
||||
[Verifies with trace analysis]
|
||||
```
|
||||
|
||||
**Recording Mode (UI + API):**
|
||||
```
|
||||
UI: [Inspects actual DOM, generates verified selectors]
|
||||
API: [Analyzes network traffic, validates request/response patterns]
|
||||
|
||||
[Generates tests with verified patterns]
|
||||
[Tests work on first run]
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Visual debugging + trace analysis (not just UI)
|
||||
- Verified selectors (UI) + network patterns (API)
|
||||
- Tests verified against actual application behavior
|
||||
|
||||
## Usage Examples
|
||||
|
||||
### Exploratory Mode Example
|
||||
|
||||
```
|
||||
User: Run *test-design
|
||||
|
||||
TEA: What feature are you designing tests for?
|
||||
|
||||
User: The checkout flow at https://shop.example.com
|
||||
|
||||
TEA: [With MCP] Let me explore the checkout flow
|
||||
[Opens browser]
|
||||
[Navigates to checkout]
|
||||
[Takes screenshots]
|
||||
[Identifies all interactive elements]
|
||||
|
||||
TEA: I've explored the checkout flow. I found:
|
||||
- 4-step wizard (Cart → Shipping → Payment → Confirm)
|
||||
- Shipping options (Standard, Express, Overnight)
|
||||
- Payment methods (Credit Card, PayPal)
|
||||
- Promo code field
|
||||
- Order summary with real-time totals
|
||||
|
||||
Test design will cover all these scenarios.
|
||||
```
|
||||
|
||||
### Recording Mode Example
|
||||
|
||||
```
|
||||
User: Run *atdd for user registration feature
|
||||
|
||||
TEA: [With MCP] Let me verify the registration form selectors
|
||||
[Opens browser at /register]
|
||||
[Inspects form]
|
||||
|
||||
TEA: I see the actual form has:
|
||||
- "Full Name" label (not "Name")
|
||||
- "Email Address" label (not "Email")
|
||||
- "Create Password" label (not "Password")
|
||||
- "I agree to Terms" checkbox
|
||||
- "Create Account" button (not "Register")
|
||||
|
||||
Generating tests with verified selectors...
|
||||
```
|
||||
|
||||
### Healing Mode Example
|
||||
|
||||
```
|
||||
User: Run *automate with healing mode
|
||||
|
||||
TEA: [With MCP] Running tests and checking for failures
|
||||
[Runs test suite]
|
||||
[Test fails: selector not found]
|
||||
|
||||
TEA: Test failed on selector '.submit-btn'
|
||||
[Opens trace viewer]
|
||||
[Sees button class changed to '.submit-button']
|
||||
|
||||
Fixing selector and verifying...
|
||||
[Updates test]
|
||||
[Re-runs with MCP]
|
||||
[Test passes]
|
||||
|
||||
Updated test with corrected selector.
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### MCP Servers Not Running
|
||||
|
||||
**Problem:** TEA says MCP enhancements aren't available.
|
||||
|
||||
**Causes:**
|
||||
1. MCP servers not configured in IDE
|
||||
2. Config syntax error in JSON
|
||||
3. IDE not restarted after config
|
||||
|
||||
**Solution:**
|
||||
```bash
|
||||
# Verify MCP config file exists
|
||||
ls ~/.cursor/config.json
|
||||
|
||||
# Validate JSON syntax
|
||||
cat ~/.cursor/config.json | python -m json.tool
|
||||
|
||||
# Restart IDE
|
||||
# Cmd+Q (quit) then reopen
|
||||
```
|
||||
|
||||
### Browser Doesn't Open
|
||||
|
||||
**Problem:** MCP enabled but browser never opens.
|
||||
|
||||
**Causes:**
|
||||
1. Playwright browsers not installed
|
||||
2. Headless mode enabled
|
||||
3. MCP server crashed
|
||||
|
||||
**Solution:**
|
||||
```bash
|
||||
# Install browsers
|
||||
npx playwright install
|
||||
|
||||
# Check MCP server logs (in IDE)
|
||||
# Look for error messages
|
||||
|
||||
# Try manual MCP server
|
||||
npx @playwright/mcp@latest
|
||||
# Should start without errors
|
||||
```
|
||||
|
||||
### TEA Doesn't Use MCP
|
||||
|
||||
**Problem:** `tea_use_mcp_enhancements: true` but TEA doesn't use browser.
|
||||
|
||||
**Causes:**
|
||||
1. Config not saved
|
||||
2. Workflow run before config update
|
||||
3. MCP servers not running
|
||||
|
||||
**Solution:**
|
||||
```bash
|
||||
# Verify config
|
||||
grep tea_use_mcp_enhancements _bmad/bmm/config.yaml
|
||||
# Should show: tea_use_mcp_enhancements: true
|
||||
|
||||
# Restart IDE (reload MCP servers)
|
||||
|
||||
# Start fresh chat (TEA loads config at start)
|
||||
```
|
||||
|
||||
### Selector Verification Fails
|
||||
|
||||
**Problem:** MCP can't find elements TEA is looking for.
|
||||
|
||||
**Causes:**
|
||||
1. Page not fully loaded
|
||||
2. Element behind modal/overlay
|
||||
3. Element requires authentication
|
||||
|
||||
**Solution:**
|
||||
TEA will handle this automatically:
|
||||
- Wait for page load
|
||||
- Dismiss modals if present
|
||||
- Handle auth if needed
|
||||
|
||||
If persistent, provide TEA more context:
|
||||
```
|
||||
"The element is behind a modal - dismiss the modal first"
|
||||
"The page requires login - use credentials X"
|
||||
```
|
||||
|
||||
### MCP Slows Down Workflows
|
||||
|
||||
**Problem:** Workflows take much longer with MCP enabled.
|
||||
|
||||
**Cause:** Browser automation adds overhead.
|
||||
|
||||
**Solution:**
|
||||
Use MCP selectively:
|
||||
- **Enable for:** Complex UIs, new projects, debugging
|
||||
- **Disable for:** Simple features, well-known patterns, API-only testing
|
||||
|
||||
Toggle quickly:
|
||||
```yaml
|
||||
# For this feature (complex UI)
|
||||
tea_use_mcp_enhancements: true
|
||||
|
||||
# For next feature (simple API)
|
||||
tea_use_mcp_enhancements: false
|
||||
```
|
||||
|
||||
## Related Guides
|
||||
|
||||
**Getting Started:**
|
||||
- [TEA Lite Quickstart Tutorial](/docs/tutorials/getting-started/tea-lite-quickstart.md) - Learn TEA basics first
|
||||
|
||||
**Workflow Guides (MCP-Enhanced):**
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Exploratory mode with browser
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Recording mode for accurate selectors
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Healing mode for debugging
|
||||
|
||||
**Other Customization:**
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - Production-ready utilities
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - MCP enhancements in lifecycle
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - When to use MCP enhancements
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - tea_use_mcp_enhancements option
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - MCP-enhanced workflows
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - MCP Enhancements term
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
23
docs/how-to/customization/index.md
Normal file
23
docs/how-to/customization/index.md
Normal file
@@ -0,0 +1,23 @@
|
||||
---
|
||||
title: "BMad Customization"
|
||||
---
|
||||
|
||||
Personalize agents and workflows to match your needs.
|
||||
|
||||
## Guides
|
||||
|
||||
| Guide | Description |
|
||||
|-------|-------------|
|
||||
| **[Agent Customization](/docs/how-to/customization/customize-agents.md)** | Modify agent behavior without editing core files |
|
||||
|
||||
## Overview
|
||||
|
||||
BMad provides two main customization approaches:
|
||||
|
||||
### Agent Customization
|
||||
|
||||
Modify any agent's persona, name, capabilities, or menu items using `.customize.yaml` files in `_bmad/_config/agents/`. Your customizations persist through updates.
|
||||
|
||||
### Workflow Customization
|
||||
|
||||
Replace or extend workflow steps to create tailored processes. (Coming soon)
|
||||
813
docs/how-to/customization/integrate-playwright-utils.md
Normal file
813
docs/how-to/customization/integrate-playwright-utils.md
Normal file
@@ -0,0 +1,813 @@
|
||||
---
|
||||
title: "Integrate Playwright Utils with TEA"
|
||||
description: Add production-ready fixtures and utilities to your TEA-generated tests
|
||||
---
|
||||
|
||||
# Integrate Playwright Utils with TEA
|
||||
|
||||
Integrate `@seontechnologies/playwright-utils` with TEA to get production-ready fixtures, utilities, and patterns in your test suite.
|
||||
|
||||
## What is Playwright Utils?
|
||||
|
||||
A production-ready utility library that provides:
|
||||
- Typed API request helper
|
||||
- Authentication session management
|
||||
- Network recording and replay (HAR)
|
||||
- Network request interception
|
||||
- Async polling (recurse)
|
||||
- Structured logging
|
||||
- File validation (CSV, PDF, XLSX, ZIP)
|
||||
- Burn-in testing utilities
|
||||
- Network error monitoring
|
||||
|
||||
**Repository:** [https://github.com/seontechnologies/playwright-utils](https://github.com/seontechnologies/playwright-utils)
|
||||
|
||||
**npm Package:** `@seontechnologies/playwright-utils`
|
||||
|
||||
## When to Use This
|
||||
|
||||
- You want production-ready fixtures (not DIY)
|
||||
- Your team benefits from standardized patterns
|
||||
- You need utilities like API testing, auth handling, network mocking
|
||||
- You want TEA to generate tests using these utilities
|
||||
- You're building reusable test infrastructure
|
||||
|
||||
**Don't use if:**
|
||||
- You're just learning testing (keep it simple first)
|
||||
- You have your own fixture library
|
||||
- You don't need the utilities
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Test framework setup complete (Playwright)
|
||||
- Node.js v18 or later
|
||||
|
||||
**Note:** Playwright Utils is for Playwright only (not Cypress).
|
||||
|
||||
## Installation
|
||||
|
||||
### Step 1: Install Package
|
||||
|
||||
```bash
|
||||
npm install -D @seontechnologies/playwright-utils
|
||||
```
|
||||
|
||||
### Step 2: Enable in TEA Config
|
||||
|
||||
Edit `_bmad/bmm/config.yaml`:
|
||||
|
||||
```yaml
|
||||
tea_use_playwright_utils: true
|
||||
```
|
||||
|
||||
**Note:** If you enabled this during BMad installation, it's already set.
|
||||
|
||||
### Step 3: Verify Installation
|
||||
|
||||
```bash
|
||||
# Check package installed
|
||||
npm list @seontechnologies/playwright-utils
|
||||
|
||||
# Check TEA config
|
||||
grep tea_use_playwright_utils _bmad/bmm/config.yaml
|
||||
```
|
||||
|
||||
Should show:
|
||||
```
|
||||
@seontechnologies/playwright-utils@2.x.x
|
||||
tea_use_playwright_utils: true
|
||||
```
|
||||
|
||||
## What Changes When Enabled
|
||||
|
||||
### *framework Workflow
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
// Basic Playwright fixtures only
|
||||
import { test, expect } from '@playwright/test';
|
||||
|
||||
test('api test', async ({ request }) => {
|
||||
const response = await request.get('/api/users');
|
||||
const users = await response.json();
|
||||
expect(response.status()).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils (Combined Fixtures):**
|
||||
```typescript
|
||||
// All utilities available via single import
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('api test', async ({ apiRequest, authToken, log }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/users',
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
log.info('Fetched users', body);
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils (Selective Merge):**
|
||||
```typescript
|
||||
import { mergeTests } from '@playwright/test';
|
||||
import { test as apiRequestFixture } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { test as logFixture } from '@seontechnologies/playwright-utils/log/fixtures';
|
||||
|
||||
export const test = mergeTests(apiRequestFixture, logFixture);
|
||||
export { expect } from '@playwright/test';
|
||||
|
||||
test('api test', async ({ apiRequest, log }) => {
|
||||
log.info('Fetching users');
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/users'
|
||||
});
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
### `*atdd` and `*automate` Workflows
|
||||
|
||||
**Without Playwright Utils:**
|
||||
```typescript
|
||||
// Manual API calls
|
||||
test('should fetch profile', async ({ page, request }) => {
|
||||
const response = await request.get('/api/profile');
|
||||
const profile = await response.json();
|
||||
// Manual parsing and validation
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
|
||||
test('should fetch profile', async ({ apiRequest }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/profile' // 'path' not 'url'
|
||||
}).validateSchema(ProfileSchema); // Chained validation
|
||||
|
||||
expect(status).toBe(200);
|
||||
// body is type-safe: { id: string, name: string, email: string }
|
||||
});
|
||||
```
|
||||
|
||||
### *test-review Workflow
|
||||
|
||||
**Without Playwright Utils:**
|
||||
Reviews against generic Playwright patterns
|
||||
|
||||
**With Playwright Utils:**
|
||||
Reviews against playwright-utils best practices:
|
||||
- Fixture composition patterns
|
||||
- Utility usage (apiRequest, authSession, etc.)
|
||||
- Network-first patterns
|
||||
- Structured logging
|
||||
|
||||
### *ci Workflow
|
||||
|
||||
**Without Playwright Utils:**
|
||||
- Parallel sharding
|
||||
- Burn-in loops (basic shell scripts)
|
||||
- CI triggers (PR, push, schedule)
|
||||
- Artifact collection
|
||||
|
||||
**With Playwright Utils:**
|
||||
Enhanced with smart testing:
|
||||
- Burn-in utility (git diff-based, volume control)
|
||||
- Selective testing (skip config/docs/types changes)
|
||||
- Test prioritization by file changes
|
||||
|
||||
## Available Utilities
|
||||
|
||||
### api-request
|
||||
|
||||
Typed HTTP client with schema validation.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/api-request.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Vanilla Playwright | api-request Utility |
|
||||
|-------------------|---------------------|
|
||||
| Manual `await response.json()` | Automatic JSON parsing |
|
||||
| `response.status()` + separate body parsing | Returns `{ status, body }` structure |
|
||||
| No built-in retry | Automatic retry for 5xx errors |
|
||||
| No schema validation | Single-line `.validateSchema()` |
|
||||
| Verbose status checking | Clean destructuring |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
import { z } from 'zod';
|
||||
|
||||
const UserSchema = z.object({
|
||||
id: z.string(),
|
||||
name: z.string(),
|
||||
email: z.string().email()
|
||||
});
|
||||
|
||||
test('should create user', async ({ apiRequest }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'POST',
|
||||
path: '/api/users', // Note: 'path' not 'url'
|
||||
body: { name: 'Test User', email: 'test@example.com' } // Note: 'body' not 'data'
|
||||
}).validateSchema(UserSchema); // Chained method (can await separately if needed)
|
||||
|
||||
expect(status).toBe(201);
|
||||
expect(body.id).toBeDefined();
|
||||
expect(body.email).toBe('test@example.com');
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Returns `{ status, body }` structure
|
||||
- Schema validation with `.validateSchema()` chained method
|
||||
- Automatic retry for 5xx errors
|
||||
- Type-safe response body
|
||||
|
||||
### auth-session
|
||||
|
||||
Authentication session management with token persistence.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/auth-session.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Vanilla Playwright Auth | auth-session |
|
||||
|------------------------|--------------|
|
||||
| Re-authenticate every test run (slow) | Authenticate once, persist to disk |
|
||||
| Single user per setup | Multi-user support (roles, accounts) |
|
||||
| No token expiration handling | Automatic token renewal |
|
||||
| Manual session management | Provider pattern (flexible auth) |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/auth-session/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('should access protected route', async ({ page, authToken }) => {
|
||||
// authToken automatically fetched and persisted
|
||||
// No manual login needed - handled by fixture
|
||||
|
||||
await page.goto('/dashboard');
|
||||
await expect(page).toHaveURL('/dashboard');
|
||||
|
||||
// Token is reused across tests (persisted to disk)
|
||||
});
|
||||
```
|
||||
|
||||
**Configuration required** (see auth-session docs for provider setup):
|
||||
```typescript
|
||||
// global-setup.ts
|
||||
import { authStorageInit, setAuthProvider, authGlobalInit } from '@seontechnologies/playwright-utils/auth-session';
|
||||
|
||||
async function globalSetup() {
|
||||
authStorageInit();
|
||||
setAuthProvider(myCustomProvider); // Define your auth mechanism
|
||||
await authGlobalInit(); // Fetch token once
|
||||
}
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Token fetched once, reused across all tests
|
||||
- Persisted to disk (faster subsequent runs)
|
||||
- Multi-user support via `authOptions.userIdentifier`
|
||||
- Automatic token renewal if expired
|
||||
|
||||
### network-recorder
|
||||
|
||||
Record and replay network traffic (HAR) for offline testing.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/network-recorder.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Vanilla Playwright HAR | network-recorder |
|
||||
|------------------------|------------------|
|
||||
| Manual `routeFromHAR()` configuration | Automatic HAR management with `PW_NET_MODE` |
|
||||
| Separate record/playback test files | Same test, switch env var |
|
||||
| No CRUD detection | Stateful mocking (POST/PUT/DELETE work) |
|
||||
| Manual HAR file paths | Auto-organized by test name |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/network-recorder/fixtures';
|
||||
|
||||
// Record mode: Set environment variable
|
||||
process.env.PW_NET_MODE = 'record';
|
||||
|
||||
test('should work with recorded traffic', async ({ page, context, networkRecorder }) => {
|
||||
// Setup recorder (records or replays based on PW_NET_MODE)
|
||||
await networkRecorder.setup(context);
|
||||
|
||||
// Your normal test code
|
||||
await page.goto('/dashboard');
|
||||
await page.click('#add-item');
|
||||
|
||||
// First run (record): Saves traffic to HAR file
|
||||
// Subsequent runs (playback): Uses HAR file, no backend needed
|
||||
});
|
||||
```
|
||||
|
||||
**Switch modes:**
|
||||
```bash
|
||||
# Record traffic
|
||||
PW_NET_MODE=record npx playwright test
|
||||
|
||||
# Playback traffic (offline)
|
||||
PW_NET_MODE=playback npx playwright test
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Offline testing (no backend needed)
|
||||
- Deterministic responses (same every time)
|
||||
- Faster execution (no network latency)
|
||||
- Stateful mocking (CRUD operations work)
|
||||
|
||||
### intercept-network-call
|
||||
|
||||
Spy or stub network requests with automatic JSON parsing.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/intercept-network-call.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Vanilla Playwright | interceptNetworkCall |
|
||||
|-------------------|----------------------|
|
||||
| Route setup + response waiting (separate steps) | Single declarative call |
|
||||
| Manual `await response.json()` | Automatic JSON parsing (`responseJson`) |
|
||||
| Complex filter predicates | Simple glob patterns (`**/api/**`) |
|
||||
| Verbose syntax | Concise, readable API |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('should handle API errors', async ({ page, interceptNetworkCall }) => {
|
||||
// Stub API to return error (set up BEFORE navigation)
|
||||
const profileCall = interceptNetworkCall({
|
||||
method: 'GET',
|
||||
url: '**/api/profile',
|
||||
fulfillResponse: {
|
||||
status: 500,
|
||||
body: { error: 'Server error' }
|
||||
}
|
||||
});
|
||||
|
||||
await page.goto('/profile');
|
||||
|
||||
// Wait for the intercepted response
|
||||
const { status, responseJson } = await profileCall;
|
||||
|
||||
expect(status).toBe(500);
|
||||
expect(responseJson.error).toBe('Server error');
|
||||
await expect(page.getByText('Server error occurred')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Automatic JSON parsing (`responseJson` ready to use)
|
||||
- Spy mode (observe real traffic) or stub mode (mock responses)
|
||||
- Glob pattern URL matching
|
||||
- Returns promise with `{ status, responseJson, requestJson }`
|
||||
|
||||
### recurse
|
||||
|
||||
Async polling for eventual consistency (Cypress-style).
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/recurse.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Manual Polling | recurse Utility |
|
||||
|----------------|-----------------|
|
||||
| `while` loops with `waitForTimeout` | Smart polling with exponential backoff |
|
||||
| Hard-coded retry logic | Configurable timeout/interval |
|
||||
| No logging visibility | Optional logging with custom messages |
|
||||
| Verbose, error-prone | Clean, readable API |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
|
||||
test('should wait for async job completion', async ({ apiRequest, recurse }) => {
|
||||
// Start async job
|
||||
const { body: job } = await apiRequest({
|
||||
method: 'POST',
|
||||
path: '/api/jobs'
|
||||
});
|
||||
|
||||
// Poll until complete (smart waiting)
|
||||
const completed = await recurse(
|
||||
() => apiRequest({ method: 'GET', path: `/api/jobs/${job.id}` }),
|
||||
(result) => result.body.status === 'completed',
|
||||
{
|
||||
timeout: 30000,
|
||||
interval: 2000,
|
||||
log: 'Waiting for job to complete'
|
||||
}
|
||||
});
|
||||
|
||||
expect(completed.body.status).toBe('completed');
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Smart polling with configurable interval
|
||||
- Handles async jobs, background tasks
|
||||
- Optional logging for debugging
|
||||
- Better than hard waits or manual polling loops
|
||||
|
||||
### log
|
||||
|
||||
Structured logging that integrates with Playwright reports.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/log.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Console.log / print | log Utility |
|
||||
|--------------------|-------------|
|
||||
| Not in test reports | Integrated with Playwright reports |
|
||||
| No step visualization | `.step()` shows in Playwright UI |
|
||||
| Manual object formatting | Logs objects seamlessly |
|
||||
| No structured output | JSON artifacts for debugging |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { log } from '@seontechnologies/playwright-utils';
|
||||
import { test, expect } from '@playwright/test';
|
||||
|
||||
test('should login', async ({ page }) => {
|
||||
await log.info('Starting login test');
|
||||
|
||||
await page.goto('/login');
|
||||
await log.step('Navigated to login page'); // Shows in Playwright UI
|
||||
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await log.debug('Filled email field');
|
||||
|
||||
await log.success('Login completed');
|
||||
// Logs appear in test output and Playwright reports
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Direct import (no fixture needed for basic usage)
|
||||
- Structured logs in test reports
|
||||
- `.step()` shows in Playwright UI
|
||||
- Logs objects seamlessly (no special handling needed)
|
||||
- Trace test execution
|
||||
|
||||
### file-utils
|
||||
|
||||
Read and validate CSV, PDF, XLSX, ZIP files.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/file-utils.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Vanilla Playwright | file-utils |
|
||||
|-------------------|------------|
|
||||
| ~80 lines per CSV flow | ~10 lines end-to-end |
|
||||
| Manual download event handling | `handleDownload()` encapsulates all |
|
||||
| External parsing libraries | Auto-parsing (CSV, XLSX, PDF, ZIP) |
|
||||
| No validation helpers | Built-in validation (headers, row count) |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { handleDownload, readCSV } from '@seontechnologies/playwright-utils/file-utils';
|
||||
import { expect } from '@playwright/test';
|
||||
import path from 'node:path';
|
||||
|
||||
const DOWNLOAD_DIR = path.join(__dirname, '../downloads');
|
||||
|
||||
test('should export valid CSV', async ({ page }) => {
|
||||
// Handle download and get file path
|
||||
const downloadPath = await handleDownload({
|
||||
page,
|
||||
downloadDir: DOWNLOAD_DIR,
|
||||
trigger: () => page.click('button:has-text("Export")')
|
||||
});
|
||||
|
||||
// Read and parse CSV
|
||||
const csvResult = await readCSV({ filePath: downloadPath });
|
||||
const { data, headers } = csvResult.content;
|
||||
|
||||
// Validate structure
|
||||
expect(headers).toEqual(['Name', 'Email', 'Status']);
|
||||
expect(data.length).toBeGreaterThan(0);
|
||||
expect(data[0]).toMatchObject({
|
||||
Name: expect.any(String),
|
||||
Email: expect.any(String),
|
||||
Status: expect.any(String)
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Handles downloads automatically
|
||||
- Auto-parses CSV, XLSX, PDF, ZIP
|
||||
- Type-safe access to parsed data
|
||||
- Returns structured `{ headers, data }`
|
||||
|
||||
### burn-in
|
||||
|
||||
Smart test selection with git diff analysis for CI optimization.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/burn-in.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Playwright `--only-changed` | burn-in Utility |
|
||||
|-----------------------------|-----------------|
|
||||
| Config changes trigger all tests | Smart filtering (skip configs, types, docs) |
|
||||
| All or nothing | Volume control (run percentage) |
|
||||
| No customization | Custom dependency analysis |
|
||||
| Slow CI on minor changes | Fast CI with intelligent selection |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
// scripts/burn-in-changed.ts
|
||||
import { runBurnIn } from '@seontechnologies/playwright-utils/burn-in';
|
||||
|
||||
async function main() {
|
||||
await runBurnIn({
|
||||
configPath: 'playwright.burn-in.config.ts',
|
||||
baseBranch: 'main'
|
||||
});
|
||||
}
|
||||
|
||||
main().catch(console.error);
|
||||
```
|
||||
|
||||
**Config:**
|
||||
```typescript
|
||||
// playwright.burn-in.config.ts
|
||||
import type { BurnInConfig } from '@seontechnologies/playwright-utils/burn-in';
|
||||
|
||||
const config: BurnInConfig = {
|
||||
skipBurnInPatterns: [
|
||||
'**/config/**',
|
||||
'**/*.md',
|
||||
'**/*types*'
|
||||
],
|
||||
burnInTestPercentage: 0.3,
|
||||
burnIn: {
|
||||
repeatEach: 3,
|
||||
retries: 1
|
||||
}
|
||||
};
|
||||
|
||||
export default config;
|
||||
```
|
||||
|
||||
**Package script:**
|
||||
```json
|
||||
{
|
||||
"scripts": {
|
||||
"test:burn-in": "tsx scripts/burn-in-changed.ts"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- **Ensure flake-free tests upfront** - Never deal with test flake again
|
||||
- Smart filtering (skip config, types, docs changes)
|
||||
- Volume control (run percentage of affected tests)
|
||||
- Git diff-based test selection
|
||||
- Faster CI feedback
|
||||
|
||||
### network-error-monitor
|
||||
|
||||
Automatically detect HTTP 4xx/5xx errors during tests.
|
||||
|
||||
**Official Docs:** <https://seontechnologies.github.io/playwright-utils/network-error-monitor.html>
|
||||
|
||||
**Why Use This?**
|
||||
|
||||
| Vanilla Playwright | network-error-monitor |
|
||||
|-------------------|----------------------|
|
||||
| UI passes, backend 500 ignored | Auto-fails on any 4xx/5xx |
|
||||
| Manual error checking | Zero boilerplate (auto-enabled) |
|
||||
| Silent failures slip through | Acts like Sentry for tests |
|
||||
| No domino effect prevention | Limits cascading failures |
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/network-error-monitor/fixtures';
|
||||
|
||||
// That's it! Network monitoring is automatically enabled
|
||||
test('should not have API errors', async ({ page }) => {
|
||||
await page.goto('/dashboard');
|
||||
await page.click('button');
|
||||
|
||||
// Test fails automatically if any HTTP 4xx/5xx errors occur
|
||||
// Error message shows: "Network errors detected: 2 request(s) failed"
|
||||
// GET 500 https://api.example.com/users
|
||||
// POST 503 https://api.example.com/metrics
|
||||
});
|
||||
```
|
||||
|
||||
**Opt-out for validation tests:**
|
||||
```typescript
|
||||
// When testing error scenarios, opt-out with annotation
|
||||
test('should show error message on 404',
|
||||
{ annotation: [{ type: 'skipNetworkMonitoring' }] }, // Array format
|
||||
async ({ page }) => {
|
||||
await page.goto('/invalid-page'); // Will 404
|
||||
await expect(page.getByText('Page not found')).toBeVisible();
|
||||
// Test won't fail on 404 because of annotation
|
||||
}
|
||||
);
|
||||
|
||||
// Or opt-out entire describe block
|
||||
test.describe('error handling',
|
||||
{ annotation: [{ type: 'skipNetworkMonitoring' }] },
|
||||
() => {
|
||||
test('handles 404', async ({ page }) => {
|
||||
// Monitoring disabled for all tests in block
|
||||
});
|
||||
}
|
||||
);
|
||||
```
|
||||
|
||||
**Benefits:**
|
||||
- Auto-enabled (zero setup)
|
||||
- Catches silent backend failures (500, 503, 504)
|
||||
- **Prevents domino effect** (limits cascading failures from one bad endpoint)
|
||||
- Opt-out with annotations for validation tests
|
||||
- Structured error reporting (JSON artifacts)
|
||||
|
||||
## Fixture Composition
|
||||
|
||||
**Option 1: Use Package's Combined Fixtures (Simplest)**
|
||||
|
||||
```typescript
|
||||
// Import all utilities at once
|
||||
import { test } from '@seontechnologies/playwright-utils/fixtures';
|
||||
import { log } from '@seontechnologies/playwright-utils';
|
||||
import { expect } from '@playwright/test';
|
||||
|
||||
test('api test', async ({ apiRequest, interceptNetworkCall }) => {
|
||||
await log.info('Fetching users');
|
||||
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/users'
|
||||
});
|
||||
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
**Option 2: Create Custom Merged Fixtures (Selective)**
|
||||
|
||||
**File 1: support/merged-fixtures.ts**
|
||||
```typescript
|
||||
import { test as base, mergeTests } from '@playwright/test';
|
||||
import { test as apiRequest } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { test as interceptNetworkCall } from '@seontechnologies/playwright-utils/intercept-network-call/fixtures';
|
||||
import { test as networkErrorMonitor } from '@seontechnologies/playwright-utils/network-error-monitor/fixtures';
|
||||
import { log } from '@seontechnologies/playwright-utils';
|
||||
|
||||
// Merge only what you need
|
||||
export const test = mergeTests(
|
||||
base,
|
||||
apiRequest,
|
||||
interceptNetworkCall,
|
||||
networkErrorMonitor
|
||||
);
|
||||
|
||||
export const expect = base.expect;
|
||||
export { log };
|
||||
```
|
||||
|
||||
**File 2: tests/api/users.spec.ts**
|
||||
```typescript
|
||||
import { test, expect, log } from '../support/merged-fixtures';
|
||||
|
||||
test('api test', async ({ apiRequest, interceptNetworkCall }) => {
|
||||
await log.info('Fetching users');
|
||||
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/users'
|
||||
});
|
||||
|
||||
expect(status).toBe(200);
|
||||
});
|
||||
```
|
||||
|
||||
**Contrast:**
|
||||
- Option 1: All utilities available, zero setup
|
||||
- Option 2: Pick utilities you need, one central file
|
||||
|
||||
**See working examples:** <https://github.com/seontechnologies/playwright-utils/tree/main/playwright/support>
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Import Errors
|
||||
|
||||
**Problem:** Cannot find module '@seontechnologies/playwright-utils/api-request'
|
||||
|
||||
**Solution:**
|
||||
```bash
|
||||
# Verify package installed
|
||||
npm list @seontechnologies/playwright-utils
|
||||
|
||||
# Check package.json has correct version
|
||||
"@seontechnologies/playwright-utils": "^2.0.0"
|
||||
|
||||
# Reinstall if needed
|
||||
npm install -D @seontechnologies/playwright-utils
|
||||
```
|
||||
|
||||
### TEA Not Using Utilities
|
||||
|
||||
**Problem:** TEA generates tests without playwright-utils.
|
||||
|
||||
**Causes:**
|
||||
1. Config not set: `tea_use_playwright_utils: false`
|
||||
2. Workflow run before config change
|
||||
3. Package not installed
|
||||
|
||||
**Solution:**
|
||||
```bash
|
||||
# Check config
|
||||
grep tea_use_playwright_utils _bmad/bmm/config.yaml
|
||||
|
||||
# Should show: tea_use_playwright_utils: true
|
||||
|
||||
# Start fresh chat (TEA loads config at start)
|
||||
```
|
||||
|
||||
### Type Errors with apiRequest
|
||||
|
||||
**Problem:** TypeScript errors on apiRequest response.
|
||||
|
||||
**Cause:** No schema validation.
|
||||
|
||||
**Solution:**
|
||||
```typescript
|
||||
// Add Zod schema for type safety
|
||||
import { z } from 'zod';
|
||||
|
||||
const ProfileSchema = z.object({
|
||||
id: z.string(),
|
||||
name: z.string(),
|
||||
email: z.string().email()
|
||||
});
|
||||
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/profile' // 'path' not 'url'
|
||||
}).validateSchema(ProfileSchema); // Chained method
|
||||
|
||||
expect(status).toBe(200);
|
||||
// body is typed as { id: string, name: string, email: string }
|
||||
```
|
||||
|
||||
## Migration Guide
|
||||
|
||||
## Related Guides
|
||||
|
||||
**Getting Started:**
|
||||
- [TEA Lite Quickstart Tutorial](/docs/tutorials/getting-started/tea-lite-quickstart.md) - Learn TEA basics
|
||||
- [How to Set Up Test Framework](/docs/how-to/workflows/setup-test-framework.md) - Initial framework setup
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Generate tests with utilities
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Expand coverage with utilities
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Review against PW-Utils patterns
|
||||
|
||||
**Other Customization:**
|
||||
- [Enable MCP Enhancements](/docs/how-to/customization/enable-tea-mcp-enhancements.md) - Live browser verification
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - **Why Playwright Utils matters** (part of TEA's three-part solution)
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Pure function → fixture pattern
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Network utilities explained
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Patterns PW-Utils enforces
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - tea_use_playwright_utils option
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Playwright Utils fragments
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - Playwright Utils term
|
||||
- [Official PW-Utils Docs](https://seontechnologies.github.io/playwright-utils/) - Complete API reference
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
101
docs/how-to/customization/shard-large-documents.md
Normal file
101
docs/how-to/customization/shard-large-documents.md
Normal file
@@ -0,0 +1,101 @@
|
||||
---
|
||||
title: "Document Sharding Guide"
|
||||
---
|
||||
|
||||
Use the `shard-doc` tool to split large markdown files into smaller, organized files for better context management.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Very large complex PRDs
|
||||
- Architecture documents with multiple system layers
|
||||
- Epic files with 4+ epics (especially for Phase 4)
|
||||
- UX design specs covering multiple subsystems
|
||||
|
||||
## What is Document Sharding?
|
||||
|
||||
Document sharding splits large markdown files into smaller, organized files based on level 2 headings (`## Heading`). This enables:
|
||||
|
||||
- **Selective Loading** - Workflows load only the sections they need
|
||||
- **Reduced Token Usage** - Massive efficiency gains for large projects
|
||||
- **Better Organization** - Logical section-based file structure
|
||||
- **Maintained Context** - Index file preserves document structure
|
||||
|
||||
### Architecture
|
||||
|
||||
```
|
||||
Before Sharding:
|
||||
docs/
|
||||
└── PRD.md (large 50k token file)
|
||||
|
||||
After Sharding:
|
||||
docs/
|
||||
└── prd/
|
||||
├── index.md # Table of contents with descriptions
|
||||
├── overview.md # Section 1
|
||||
├── user-requirements.md # Section 2
|
||||
├── technical-requirements.md # Section 3
|
||||
└── ... # Additional sections
|
||||
```
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Run the Shard-Doc Tool
|
||||
|
||||
```bash
|
||||
/bmad:core:tools:shard-doc
|
||||
```
|
||||
|
||||
### 2. Follow the Interactive Process
|
||||
|
||||
```
|
||||
Agent: Which document would you like to shard?
|
||||
User: docs/PRD.md
|
||||
|
||||
Agent: Default destination: docs/prd/
|
||||
Accept default? [y/n]
|
||||
User: y
|
||||
|
||||
Agent: Sharding PRD.md...
|
||||
✓ Created 12 section files
|
||||
✓ Generated index.md
|
||||
✓ Complete!
|
||||
```
|
||||
|
||||
## What You Get
|
||||
|
||||
**index.md structure:**
|
||||
|
||||
```markdown
|
||||
|
||||
## Sections
|
||||
|
||||
1. [Overview](./overview.md) - Project vision and objectives
|
||||
2. [User Requirements](./user-requirements.md) - Feature specifications
|
||||
3. [Epic 1: Authentication](./epic-1-authentication.md) - User auth system
|
||||
4. [Epic 2: Dashboard](./epic-2-dashboard.md) - Main dashboard UI
|
||||
...
|
||||
```
|
||||
|
||||
**Individual section files:**
|
||||
|
||||
- Named from heading text (kebab-case)
|
||||
- Contains complete section content
|
||||
- Preserves all markdown formatting
|
||||
- Can be read independently
|
||||
|
||||
## How Workflow Discovery Works
|
||||
|
||||
BMad workflows use a **dual discovery system**:
|
||||
|
||||
1. **Try whole document first** - Look for `document-name.md`
|
||||
2. **Check for sharded version** - Look for `document-name/index.md`
|
||||
3. **Priority rule** - Whole document takes precedence if both exist - remove the whole document if you want the sharded to be used instead
|
||||
|
||||
## Workflow Support
|
||||
|
||||
All BMM workflows support both formats:
|
||||
|
||||
- Whole documents
|
||||
- Sharded documents
|
||||
- Automatic detection
|
||||
- Transparent to user
|
||||
526
docs/how-to/enterprise/use-tea-for-enterprise.md
Normal file
526
docs/how-to/enterprise/use-tea-for-enterprise.md
Normal file
@@ -0,0 +1,526 @@
|
||||
---
|
||||
title: "Running TEA for Enterprise Projects"
|
||||
description: Use TEA with compliance, security, and regulatory requirements in enterprise environments
|
||||
---
|
||||
|
||||
# Running TEA for Enterprise Projects
|
||||
|
||||
Use TEA on enterprise projects with compliance, security, audit, and regulatory requirements. This guide covers NFR assessment, audit trails, and evidence collection.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Enterprise track projects (not Quick Flow or simple BMad Method)
|
||||
- Compliance requirements (SOC 2, HIPAA, GDPR, etc.)
|
||||
- Security-critical applications (finance, healthcare, government)
|
||||
- Audit trail requirements
|
||||
- Strict NFR thresholds (performance, security, reliability)
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed (Enterprise track selected)
|
||||
- TEA agent available
|
||||
- Compliance requirements documented
|
||||
- Stakeholders identified (who approves gates)
|
||||
|
||||
## Enterprise-Specific TEA Workflows
|
||||
|
||||
### NFR Assessment (*nfr-assess)
|
||||
|
||||
**Purpose:** Validate non-functional requirements with evidence.
|
||||
|
||||
**When:** Phase 2 (early) and Release Gate
|
||||
|
||||
**Why Enterprise Needs This:**
|
||||
- Compliance mandates specific thresholds
|
||||
- Audit trails required for certification
|
||||
- Security requirements are non-negotiable
|
||||
- Performance SLAs are contractual
|
||||
|
||||
**Example:**
|
||||
```
|
||||
*nfr-assess
|
||||
|
||||
Categories: Security, Performance, Reliability, Maintainability
|
||||
|
||||
Security thresholds:
|
||||
- Zero critical vulnerabilities (required by SOC 2)
|
||||
- All endpoints require authentication
|
||||
- Data encrypted at rest (FIPS 140-2)
|
||||
- Audit logging on all data access
|
||||
|
||||
Evidence:
|
||||
- Security scan: reports/nessus-scan.pdf
|
||||
- Penetration test: reports/pentest-2026-01.pdf
|
||||
- Compliance audit: reports/soc2-evidence.zip
|
||||
```
|
||||
|
||||
**Output:** NFR assessment with PASS/CONCERNS/FAIL for each category.
|
||||
|
||||
### Trace with Audit Evidence (*trace)
|
||||
|
||||
**Purpose:** Requirements traceability with audit trail.
|
||||
|
||||
**When:** Phase 2 (baseline), Phase 4 (refresh), Release Gate
|
||||
|
||||
**Why Enterprise Needs This:**
|
||||
- Auditors require requirements-to-test mapping
|
||||
- Compliance certifications need traceability
|
||||
- Regulatory bodies want evidence
|
||||
|
||||
**Example:**
|
||||
```
|
||||
*trace Phase 1
|
||||
|
||||
Requirements: PRD.md (with compliance requirements)
|
||||
Test location: tests/
|
||||
|
||||
Output: traceability-matrix.md with:
|
||||
- Requirement-to-test mapping
|
||||
- Compliance requirement coverage
|
||||
- Gap prioritization
|
||||
- Recommendations
|
||||
```
|
||||
|
||||
**For Release Gate:**
|
||||
```
|
||||
*trace Phase 2
|
||||
|
||||
Generate gate-decision-{gate_type}-{story_id}.md with:
|
||||
- Evidence references
|
||||
- Approver signatures
|
||||
- Compliance checklist
|
||||
- Decision rationale
|
||||
```
|
||||
|
||||
### Test Design with Compliance Focus (*test-design)
|
||||
|
||||
**Purpose:** Risk assessment with compliance and security focus.
|
||||
|
||||
**When:** Phase 3 (system-level), Phase 4 (epic-level)
|
||||
|
||||
**Why Enterprise Needs This:**
|
||||
- Security architecture alignment required
|
||||
- Compliance requirements must be testable
|
||||
- Performance requirements are contractual
|
||||
|
||||
**Example:**
|
||||
```
|
||||
*test-design
|
||||
|
||||
Mode: System-level
|
||||
|
||||
Focus areas:
|
||||
- Security architecture (authentication, authorization, encryption)
|
||||
- Performance requirements (SLA: P99 <200ms)
|
||||
- Compliance (HIPAA PHI handling, audit logging)
|
||||
|
||||
Output: test-design-system.md with:
|
||||
- Security testing strategy
|
||||
- Compliance requirement → test mapping
|
||||
- Performance testing plan
|
||||
- Audit logging validation
|
||||
```
|
||||
|
||||
## Enterprise TEA Lifecycle
|
||||
|
||||
### Phase 1: Discovery (Optional but Recommended)
|
||||
|
||||
**Research compliance requirements:**
|
||||
```
|
||||
Analyst: *research
|
||||
|
||||
Topics:
|
||||
- Industry compliance (SOC 2, HIPAA, GDPR)
|
||||
- Security standards (OWASP Top 10)
|
||||
- Performance benchmarks (industry P99)
|
||||
```
|
||||
|
||||
### Phase 2: Planning (Required)
|
||||
|
||||
**1. Define NFRs early:**
|
||||
```
|
||||
PM: *prd
|
||||
|
||||
Include in PRD:
|
||||
- Security requirements (authentication, encryption)
|
||||
- Performance SLAs (response time, throughput)
|
||||
- Reliability targets (uptime, RTO, RPO)
|
||||
- Compliance mandates (data retention, audit logs)
|
||||
```
|
||||
|
||||
**2. Assess NFRs:**
|
||||
```
|
||||
TEA: *nfr-assess
|
||||
|
||||
Categories: All (Security, Performance, Reliability, Maintainability)
|
||||
|
||||
Output: nfr-assessment.md
|
||||
- NFR requirements documented
|
||||
- Acceptance criteria defined
|
||||
- Test strategy planned
|
||||
```
|
||||
|
||||
**3. Baseline (brownfield only):**
|
||||
```
|
||||
TEA: *trace Phase 1
|
||||
|
||||
Establish baseline coverage before new work
|
||||
```
|
||||
|
||||
### Phase 3: Solutioning (Required)
|
||||
|
||||
**1. Architecture with testability review:**
|
||||
```
|
||||
Architect: *architecture
|
||||
|
||||
TEA: *test-design (system-level)
|
||||
|
||||
Focus:
|
||||
- Security architecture testability
|
||||
- Performance testing strategy
|
||||
- Compliance requirement mapping
|
||||
```
|
||||
|
||||
**2. Test infrastructure:**
|
||||
```
|
||||
TEA: *framework
|
||||
|
||||
Requirements:
|
||||
- Separate test environments (dev, staging, prod-mirror)
|
||||
- Secure test data handling (PHI, PII)
|
||||
- Audit logging in tests
|
||||
```
|
||||
|
||||
**3. CI/CD with compliance:**
|
||||
```
|
||||
TEA: *ci
|
||||
|
||||
Requirements:
|
||||
- Secrets management (Vault, AWS Secrets Manager)
|
||||
- Test isolation (no cross-contamination)
|
||||
- Artifact retention (compliance audit trail)
|
||||
- Access controls (who can run production tests)
|
||||
```
|
||||
|
||||
### Phase 4: Implementation (Required)
|
||||
|
||||
**Per epic:**
|
||||
```
|
||||
1. TEA: *test-design (epic-level)
|
||||
Focus: Compliance, security, performance for THIS epic
|
||||
|
||||
2. TEA: *atdd (optional)
|
||||
Generate tests including security/compliance scenarios
|
||||
|
||||
3. DEV: Implement story
|
||||
|
||||
4. TEA: *automate
|
||||
Expand coverage including compliance edge cases
|
||||
|
||||
5. TEA: *test-review
|
||||
Audit quality (score >80 per epic, rises to >85 at release)
|
||||
|
||||
6. TEA: *trace Phase 1
|
||||
Refresh coverage, verify compliance requirements tested
|
||||
```
|
||||
|
||||
### Release Gate (Required)
|
||||
|
||||
**1. Final NFR assessment:**
|
||||
```
|
||||
TEA: *nfr-assess
|
||||
|
||||
All categories (if not done earlier)
|
||||
Latest evidence (performance tests, security scans)
|
||||
```
|
||||
|
||||
**2. Final quality audit:**
|
||||
```
|
||||
TEA: *test-review tests/
|
||||
|
||||
Full suite review
|
||||
Quality target: >85 for enterprise
|
||||
```
|
||||
|
||||
**3. Gate decision:**
|
||||
```
|
||||
TEA: *trace Phase 2
|
||||
|
||||
Evidence required:
|
||||
- traceability-matrix.md (from Phase 1)
|
||||
- test-review.md (from quality audit)
|
||||
- nfr-assessment.md (from NFR assessment)
|
||||
- Test execution results (must have test results available)
|
||||
|
||||
Decision: PASS/CONCERNS/FAIL/WAIVED
|
||||
|
||||
Archive all artifacts for compliance audit
|
||||
```
|
||||
|
||||
**Note:** Phase 2 requires test execution results. If results aren't available, Phase 2 will be skipped.
|
||||
|
||||
**4. Archive for audit:**
|
||||
```
|
||||
Archive:
|
||||
- All test results
|
||||
- Coverage reports
|
||||
- NFR assessments
|
||||
- Gate decisions
|
||||
- Approver signatures
|
||||
|
||||
Retention: Per compliance requirements (7 years for HIPAA)
|
||||
```
|
||||
|
||||
## Enterprise-Specific Requirements
|
||||
|
||||
### Evidence Collection
|
||||
|
||||
**Required artifacts:**
|
||||
- Requirements traceability matrix
|
||||
- Test execution results (with timestamps)
|
||||
- NFR assessment reports
|
||||
- Security scan results
|
||||
- Performance test results
|
||||
- Gate decision records
|
||||
- Approver signatures
|
||||
|
||||
**Storage:**
|
||||
```
|
||||
compliance/
|
||||
├── 2026-Q1/
|
||||
│ ├── release-1.2.0/
|
||||
│ │ ├── traceability-matrix.md
|
||||
│ │ ├── test-review.md
|
||||
│ │ ├── nfr-assessment.md
|
||||
│ │ ├── gate-decision-release-v1.2.0.md
|
||||
│ │ ├── test-results/
|
||||
│ │ ├── security-scans/
|
||||
│ │ └── approvals.pdf
|
||||
```
|
||||
|
||||
**Retention:** 7 years (HIPAA), 3 years (SOC 2), per your compliance needs
|
||||
|
||||
### Approver Workflows
|
||||
|
||||
**Multi-level approval required:**
|
||||
|
||||
```markdown
|
||||
## Gate Approvals Required
|
||||
|
||||
### Technical Approval
|
||||
- [ ] QA Lead - Test coverage adequate
|
||||
- [ ] Tech Lead - Technical quality acceptable
|
||||
- [ ] Security Lead - Security requirements met
|
||||
|
||||
### Business Approval
|
||||
- [ ] Product Manager - Business requirements met
|
||||
- [ ] Compliance Officer - Regulatory requirements met
|
||||
|
||||
### Executive Approval (for major releases)
|
||||
- [ ] VP Engineering - Overall quality acceptable
|
||||
- [ ] CTO - Architecture approved for production
|
||||
```
|
||||
|
||||
### Compliance Checklists
|
||||
|
||||
**SOC 2 Example:**
|
||||
```markdown
|
||||
## SOC 2 Compliance Checklist
|
||||
|
||||
### Access Controls
|
||||
- [ ] All API endpoints require authentication
|
||||
- [ ] Authorization tested for all protected resources
|
||||
- [ ] Session management secure (token expiration tested)
|
||||
|
||||
### Audit Logging
|
||||
- [ ] All data access logged
|
||||
- [ ] Logs immutable (append-only)
|
||||
- [ ] Log retention policy enforced
|
||||
|
||||
### Data Protection
|
||||
- [ ] Data encrypted at rest (tested)
|
||||
- [ ] Data encrypted in transit (HTTPS enforced)
|
||||
- [ ] PII handling compliant (masking tested)
|
||||
|
||||
### Testing Evidence
|
||||
- [ ] Test coverage >80% (verified)
|
||||
- [ ] Security tests passing (100%)
|
||||
- [ ] Traceability matrix complete
|
||||
```
|
||||
|
||||
**HIPAA Example:**
|
||||
```markdown
|
||||
## HIPAA Compliance Checklist
|
||||
|
||||
### PHI Protection
|
||||
- [ ] PHI encrypted at rest (AES-256)
|
||||
- [ ] PHI encrypted in transit (TLS 1.3)
|
||||
- [ ] PHI access logged (audit trail)
|
||||
|
||||
### Access Controls
|
||||
- [ ] Role-based access control (RBAC tested)
|
||||
- [ ] Minimum necessary access (tested)
|
||||
- [ ] Authentication strong (MFA tested)
|
||||
|
||||
### Breach Notification
|
||||
- [ ] Breach detection tested
|
||||
- [ ] Notification workflow tested
|
||||
- [ ] Incident response plan tested
|
||||
```
|
||||
|
||||
## Enterprise Tips
|
||||
|
||||
### Start with Security
|
||||
|
||||
**Priority 1:** Security requirements
|
||||
```
|
||||
1. Document all security requirements
|
||||
2. Generate security tests with *atdd
|
||||
3. Run security test suite
|
||||
4. Pass security audit BEFORE moving forward
|
||||
```
|
||||
|
||||
**Why:** Security failures block everything in enterprise.
|
||||
|
||||
**Example: RBAC Testing**
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
test('should enforce role-based access', async ({ request }) => {
|
||||
// Login as regular user
|
||||
const userResp = await request.post('/api/auth/login', {
|
||||
data: { email: 'user@example.com', password: 'pass' }
|
||||
});
|
||||
const { token: userToken } = await userResp.json();
|
||||
|
||||
// Try to access admin endpoint
|
||||
const adminResp = await request.get('/api/admin/users', {
|
||||
headers: { Authorization: `Bearer ${userToken}` }
|
||||
});
|
||||
|
||||
expect(adminResp.status()).toBe(403); // Forbidden
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils (Cleaner, Reusable):**
|
||||
```typescript
|
||||
import { test as base, expect } from '@playwright/test';
|
||||
import { test as apiRequestFixture } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { createAuthFixtures } from '@seontechnologies/playwright-utils/auth-session';
|
||||
import { mergeTests } from '@playwright/test';
|
||||
|
||||
const authFixtureTest = base.extend(createAuthFixtures());
|
||||
export const testWithAuth = mergeTests(apiRequestFixture, authFixtureTest);
|
||||
|
||||
testWithAuth('should enforce role-based access', async ({ apiRequest, authToken }) => {
|
||||
// Auth token from fixture (configured for 'user' role)
|
||||
const { status } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/admin/users', // Admin endpoint
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
expect(status).toBe(403); // Regular user denied
|
||||
});
|
||||
|
||||
testWithAuth('admin can access admin endpoint', async ({ apiRequest, authToken, authOptions }) => {
|
||||
// Override to admin role
|
||||
authOptions.userIdentifier = 'admin';
|
||||
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/admin/users',
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
expect(status).toBe(200); // Admin allowed
|
||||
expect(body).toBeInstanceOf(Array);
|
||||
});
|
||||
```
|
||||
|
||||
**Note:** Auth-session requires provider setup in global-setup.ts. See [auth-session configuration](https://seontechnologies.github.io/playwright-utils/auth-session.html).
|
||||
|
||||
**Playwright Utils Benefits for Compliance:**
|
||||
- Multi-user auth testing (regular, admin, etc.)
|
||||
- Token persistence (faster test execution)
|
||||
- Consistent auth patterns (audit trail)
|
||||
- Automatic cleanup
|
||||
|
||||
### Set Higher Quality Thresholds
|
||||
|
||||
**Enterprise quality targets:**
|
||||
- Test coverage: >85% (vs 80% for non-enterprise)
|
||||
- Quality score: >85 (vs 75 for non-enterprise)
|
||||
- P0 coverage: 100% (non-negotiable)
|
||||
- P1 coverage: >95% (vs 90% for non-enterprise)
|
||||
|
||||
**Rationale:** Enterprise systems affect more users, higher stakes.
|
||||
|
||||
### Document Everything
|
||||
|
||||
**Auditors need:**
|
||||
- Why decisions were made (rationale)
|
||||
- Who approved (signatures)
|
||||
- When (timestamps)
|
||||
- What evidence (test results, scan reports)
|
||||
|
||||
**Use TEA's structured outputs:**
|
||||
- Reports have timestamps
|
||||
- Decisions have rationale
|
||||
- Evidence is referenced
|
||||
- Audit trail is automatic
|
||||
|
||||
### Budget for Compliance Testing
|
||||
|
||||
**Enterprise testing costs more:**
|
||||
- Penetration testing: $10k-50k
|
||||
- Security audits: $5k-20k
|
||||
- Performance testing tools: $500-5k/month
|
||||
- Compliance consulting: $200-500/hour
|
||||
|
||||
**Plan accordingly:**
|
||||
- Budget in project cost
|
||||
- Schedule early (3+ months for SOC 2)
|
||||
- Don't skip (non-negotiable for compliance)
|
||||
|
||||
### Use External Validators
|
||||
|
||||
**Don't self-certify:**
|
||||
- Penetration testing: Hire external firm
|
||||
- Security audits: Independent auditor
|
||||
- Compliance: Certification body
|
||||
- Performance: Load testing service
|
||||
|
||||
**TEA's role:** Prepare for external validation, don't replace it.
|
||||
|
||||
## Related Guides
|
||||
|
||||
**Workflow Guides:**
|
||||
- [How to Run NFR Assessment](/docs/how-to/workflows/run-nfr-assess.md) - Deep dive on NFRs
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md) - Gate decisions with evidence
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Quality audits
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Compliance-focused planning
|
||||
|
||||
**Use-Case Guides:**
|
||||
- [Using TEA with Existing Tests](/docs/how-to/brownfield/use-tea-with-existing-tests.md) - Brownfield patterns
|
||||
|
||||
**Customization:**
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - Production-ready utilities
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Engagement Models](/docs/explanation/tea/engagement-models.md) - Enterprise model explained
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Probability × impact scoring
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Enterprise quality thresholds
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Complete TEA lifecycle
|
||||
|
||||
## Reference
|
||||
|
||||
- [TEA Command Reference](/docs/reference/tea/commands.md) - All 8 workflows
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Enterprise config options
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Testing patterns
|
||||
- [Glossary](/docs/reference/glossary/index.md#test-architect-tea-concepts) - TEA terminology
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
102
docs/how-to/get-answers-about-bmad.md
Normal file
102
docs/how-to/get-answers-about-bmad.md
Normal file
@@ -0,0 +1,102 @@
|
||||
---
|
||||
title: "How to Get Answers About BMad"
|
||||
description: Use an LLM to quickly answer your own BMad questions
|
||||
---
|
||||
|
||||
Use your AI tool to get answers about BMad by pointing it at the source files.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- You have a question about how BMad works
|
||||
- You want to understand a specific agent or workflow
|
||||
- You need quick answers without waiting for Discord
|
||||
|
||||
:::note[Prerequisites]
|
||||
An AI tool (Claude Code, Cursor, ChatGPT, Claude.ai, etc.) and either BMad installed in your project or access to the GitHub repo.
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Choose Your Source
|
||||
|
||||
| Source | Best For | Examples |
|
||||
|--------|----------|----------|
|
||||
| **`_bmad` folder** | How BMad works—agents, workflows, prompts | "What does the PM agent do?" |
|
||||
| **Full GitHub repo** | History, installer, architecture | "What changed in v6?" |
|
||||
| **`llms-full.txt`** | Quick overview from docs | "Explain BMad's four phases" |
|
||||
|
||||
The `_bmad` folder is created when you install BMad. If you don't have it yet, clone the repo instead.
|
||||
|
||||
### 2. Point Your AI at the Source
|
||||
|
||||
**If your AI can read files (Claude Code, Cursor, etc.):**
|
||||
|
||||
- **BMad installed:** Point at the `_bmad` folder and ask directly
|
||||
- **Want deeper context:** Clone the [full repo](https://github.com/bmad-code-org/BMAD-METHOD)
|
||||
|
||||
**If you use ChatGPT or Claude.ai:**
|
||||
|
||||
Fetch `llms-full.txt` into your session:
|
||||
|
||||
```
|
||||
https://bmad-code-org.github.io/BMAD-METHOD/llms-full.txt
|
||||
```
|
||||
|
||||
See the [Downloads page](/docs/downloads.md) for other downloadable resources.
|
||||
|
||||
### 3. Ask Your Question
|
||||
|
||||
:::note[Example]
|
||||
**Q:** "Tell me the fastest way to build something with BMad"
|
||||
|
||||
**A:** Use Quick Flow: Run `quick-spec` to write a technical specification, then `quick-dev` to implement it—skipping the full planning phases.
|
||||
:::
|
||||
|
||||
## What You Get
|
||||
|
||||
Direct answers about BMad—how agents work, what workflows do, why things are structured the way they are—without waiting for someone else to respond.
|
||||
|
||||
## Tips
|
||||
|
||||
- **Verify surprising answers** — LLMs occasionally get things wrong. Check the source file or ask on Discord.
|
||||
- **Be specific** — "What does step 3 of the PRD workflow do?" beats "How does PRD work?"
|
||||
|
||||
## Still Stuck?
|
||||
|
||||
Tried the LLM approach and still need help? You now have a much better question to ask.
|
||||
|
||||
| Channel | Use For |
|
||||
|---------|---------|
|
||||
| `#bmad-method-help` | Quick questions (real-time chat) |
|
||||
| `help-requests` forum | Detailed questions (searchable, persistent) |
|
||||
| `#suggestions-feedback` | Ideas and feature requests |
|
||||
| `#report-bugs-and-issues` | Bug reports |
|
||||
|
||||
**Discord:** [discord.gg/gk8jAdXWmj](https://discord.gg/gk8jAdXWmj)
|
||||
|
||||
**GitHub Issues:** [github.com/bmad-code-org/BMAD-METHOD/issues](https://github.com/bmad-code-org/BMAD-METHOD/issues) (for clear bugs)
|
||||
|
||||
*You!*
|
||||
*Stuck*
|
||||
*in the queue—*
|
||||
*waiting*
|
||||
*for who?*
|
||||
|
||||
*The source*
|
||||
*is there,*
|
||||
*plain to see!*
|
||||
|
||||
*Point*
|
||||
*your machine.*
|
||||
*Set it free.*
|
||||
|
||||
*It reads.*
|
||||
*It speaks.*
|
||||
*Ask away—*
|
||||
|
||||
*Why wait*
|
||||
*for tomorrow*
|
||||
*when you have*
|
||||
*today?*
|
||||
|
||||
*—Claude*
|
||||
12
docs/how-to/installation/index.md
Normal file
12
docs/how-to/installation/index.md
Normal file
@@ -0,0 +1,12 @@
|
||||
---
|
||||
title: "Installation Guides"
|
||||
description: How to install and upgrade BMad Method
|
||||
---
|
||||
|
||||
How-to guides for installing and configuring the BMad Method.
|
||||
|
||||
| Guide | Description |
|
||||
|-------|-------------|
|
||||
| [Install BMad](/docs/how-to/installation/install-bmad.md) | Step-by-step installation instructions |
|
||||
| [Install Custom Modules](/docs/how-to/installation/install-custom-modules.md) | Add custom agents, workflows, and modules |
|
||||
| [Upgrade to v6](/docs/how-to/installation/upgrade-to-v6.md) | Migrate from BMad v4 to v6 |
|
||||
112
docs/how-to/installation/install-bmad.md
Normal file
112
docs/how-to/installation/install-bmad.md
Normal file
@@ -0,0 +1,112 @@
|
||||
---
|
||||
title: "How to Install BMad"
|
||||
description: Step-by-step guide to installing BMad in your project
|
||||
---
|
||||
|
||||
Use the `npx bmad-method install` command to set up BMad in your project with your choice of modules and AI tools.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Starting a new project with BMad
|
||||
- Adding BMad to an existing codebase
|
||||
- Setting up BMad on a new machine
|
||||
|
||||
:::note[Prerequisites]
|
||||
- **Node.js** 20+ (required for the installer)
|
||||
- **Git** (recommended)
|
||||
- **AI-powered IDE** (Claude Code, Cursor, Windsurf, or similar)
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Run the Installer
|
||||
|
||||
```bash
|
||||
npx bmad-method install
|
||||
```
|
||||
|
||||
### 2. Choose Installation Location
|
||||
|
||||
The installer will ask where to install BMad files:
|
||||
|
||||
- Current directory (recommended for new projects)
|
||||
- Subdirectory
|
||||
- Custom path
|
||||
|
||||
### 3. Select Your AI Tools
|
||||
|
||||
Choose which AI tools you'll be using:
|
||||
|
||||
- Claude Code
|
||||
- Cursor
|
||||
- Windsurf
|
||||
- Other
|
||||
|
||||
The installer configures BMad for your selected tools.
|
||||
|
||||
### 4. Choose Modules
|
||||
|
||||
Select which modules to install:
|
||||
|
||||
| Module | Purpose |
|
||||
|--------|---------|
|
||||
| **BMM** | Core methodology for software development |
|
||||
| **BMGD** | Game development workflows |
|
||||
| **CIS** | Creative intelligence and facilitation |
|
||||
| **BMB** | Building custom agents and workflows |
|
||||
|
||||
### 5. Add Custom Content (Optional)
|
||||
|
||||
If you have custom agents, workflows, or modules, point to their location and the installer will integrate them.
|
||||
|
||||
### 6. Configure Settings
|
||||
|
||||
For each module, either accept recommended defaults (faster) or customize settings (more control).
|
||||
|
||||
## What You Get
|
||||
|
||||
```
|
||||
your-project/
|
||||
├── _bmad/
|
||||
│ ├── bmm/ # Method module
|
||||
│ │ ├── agents/ # Agent files
|
||||
│ │ ├── workflows/ # Workflow files
|
||||
│ │ └── config.yaml # Module config
|
||||
│ ├── core/ # Core utilities
|
||||
│ └── ...
|
||||
├── _bmad-output/ # Generated artifacts
|
||||
└── .claude/ # IDE configuration
|
||||
```
|
||||
|
||||
## Verify Installation
|
||||
|
||||
1. Check the `_bmad/` directory exists
|
||||
2. Load an agent in your AI tool
|
||||
3. Run `*menu` to see available commands
|
||||
|
||||
## Configuration
|
||||
|
||||
Edit `_bmad/[module]/config.yaml` to customize:
|
||||
|
||||
```yaml
|
||||
output_folder: ./_bmad-output
|
||||
user_name: Your Name
|
||||
communication_language: english
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
**"Command not found: npx"** — Install Node.js 20+:
|
||||
```bash
|
||||
brew install node
|
||||
```
|
||||
|
||||
**"Permission denied"** — Check npm permissions:
|
||||
```bash
|
||||
npm config set prefix ~/.npm-global
|
||||
```
|
||||
|
||||
**Installer hangs** — Try running with verbose output:
|
||||
```bash
|
||||
npx bmad-method install --verbose
|
||||
```
|
||||
118
docs/how-to/installation/install-custom-modules.md
Normal file
118
docs/how-to/installation/install-custom-modules.md
Normal file
@@ -0,0 +1,118 @@
|
||||
---
|
||||
title: "How to Install Custom Modules"
|
||||
description: Add custom agents, workflows, and modules to BMad
|
||||
---
|
||||
|
||||
Use the BMad installer to add custom agents, workflows, and modules that extend BMad's functionality.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Adding third-party BMad modules to your project
|
||||
- Installing your own custom agents or workflows
|
||||
- Sharing custom content across projects or teams
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad installed in your project
|
||||
- Custom content with a valid `module.yaml` file
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Prepare Your Custom Content
|
||||
|
||||
Your custom content needs a `module.yaml` file. Choose the appropriate structure:
|
||||
|
||||
**For a cohesive module** (agents and workflows that work together):
|
||||
|
||||
```
|
||||
module-code/
|
||||
module.yaml
|
||||
agents/
|
||||
workflows/
|
||||
tools/
|
||||
templates/
|
||||
```
|
||||
|
||||
**For standalone items** (unrelated agents/workflows):
|
||||
|
||||
```
|
||||
module-name/
|
||||
module.yaml # Contains unitary: true
|
||||
agents/
|
||||
larry/larry.agent.md
|
||||
curly/curly.agent.md
|
||||
workflows/
|
||||
```
|
||||
|
||||
Add `unitary: true` in your `module.yaml` to indicate items don't depend on each other.
|
||||
|
||||
### 2. Run the Installer
|
||||
|
||||
**New project:**
|
||||
|
||||
```bash
|
||||
npx bmad-method install
|
||||
```
|
||||
|
||||
When prompted "Would you like to install a local custom module?", select 'y' and provide the path to your module folder.
|
||||
|
||||
**Existing project:**
|
||||
|
||||
```bash
|
||||
npx bmad-method install
|
||||
```
|
||||
|
||||
1. Select `Modify BMad Installation`
|
||||
2. Choose the option to add, modify, or update custom modules
|
||||
3. Provide the path to your module folder
|
||||
|
||||
### 3. Verify Installation
|
||||
|
||||
Check that your custom content appears in the `_bmad/` directory and is accessible from your AI tool.
|
||||
|
||||
## What You Get
|
||||
|
||||
- Custom agents available in your AI tool
|
||||
- Custom workflows accessible via `*workflow-name`
|
||||
- Content integrated with BMad's update system
|
||||
|
||||
## Content Types
|
||||
|
||||
BMad supports several categories of custom content:
|
||||
|
||||
| Type | Description |
|
||||
|------|-------------|
|
||||
| **Stand Alone Modules** | Complete modules with their own agents and workflows |
|
||||
| **Add On Modules** | Extensions that add to existing modules |
|
||||
| **Global Modules** | Content available across all modules |
|
||||
| **Custom Agents** | Individual agent definitions |
|
||||
| **Custom Workflows** | Individual workflow definitions |
|
||||
|
||||
For detailed information about content types, see [Custom Content Types](/docs/explanation/bmad-builder/custom-content-types.md).
|
||||
|
||||
## Updating Custom Content
|
||||
|
||||
When BMad Core or module updates are available, the quick update process:
|
||||
|
||||
1. Applies updates to core modules
|
||||
2. Recompiles all agents with your customizations
|
||||
3. Retains your custom content from cache
|
||||
4. Preserves your configurations
|
||||
|
||||
You don't need to keep source module files locally—just point to the updated location during updates.
|
||||
|
||||
## Tips
|
||||
|
||||
- **Use unique module codes** — Don't use `bmm` or other existing module codes
|
||||
- **Avoid naming conflicts** — Each module needs a distinct code
|
||||
- **Document dependencies** — Note any modules your custom content requires
|
||||
- **Test in isolation** — Verify custom modules work before sharing
|
||||
- **Version your content** — Track updates with version numbers
|
||||
|
||||
:::caution[Naming Conflicts]
|
||||
Don't create custom modules with codes like `bmm` (already used by BMad Method). Each custom module needs a unique code.
|
||||
:::
|
||||
|
||||
## Example Modules
|
||||
|
||||
Find example custom modules in the `samples/sample-custom-modules/` folder of the [BMad repository](https://github.com/bmad-code-org/BMAD-METHOD). Download either sample folder to try them out.
|
||||
131
docs/how-to/installation/upgrade-to-v6.md
Normal file
131
docs/how-to/installation/upgrade-to-v6.md
Normal file
@@ -0,0 +1,131 @@
|
||||
---
|
||||
title: "How to Upgrade to v6"
|
||||
description: Migrate from BMad v4 to v6
|
||||
---
|
||||
|
||||
Use the BMad installer to upgrade from v4 to v6, which includes automatic detection of legacy installations and migration assistance.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- You have BMad v4 installed (`.bmad-method` folder)
|
||||
- You want to migrate to the new v6 architecture
|
||||
- You have existing planning artifacts to preserve
|
||||
|
||||
:::note[Prerequisites]
|
||||
- Node.js 20+
|
||||
- Existing BMad v4 installation
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Run the Installer
|
||||
|
||||
```bash
|
||||
npx bmad-method install
|
||||
```
|
||||
|
||||
The installer automatically detects:
|
||||
|
||||
- **Legacy v4 folder**: `.bmad-method`
|
||||
- **IDE command artifacts**: Legacy bmad folders in `.claude/commands/`, `.cursor/commands/`, etc.
|
||||
|
||||
### 2. Handle Legacy Installation
|
||||
|
||||
When v4 is detected, you can:
|
||||
|
||||
- Allow the installer to back up and remove `.bmad-method`
|
||||
- Exit and handle cleanup manually
|
||||
- Keep both (not recommended for same project)
|
||||
|
||||
### 3. Clean Up IDE Commands
|
||||
|
||||
Manually remove legacy v4 IDE commands:
|
||||
|
||||
- `.claude/commands/BMad/agents`
|
||||
- `.claude/commands/BMad/tasks`
|
||||
|
||||
New v6 commands will be at `.claude/commands/bmad/<module>/agents|workflows`.
|
||||
|
||||
:::tip[Accidentally Deleted Commands?]
|
||||
If you delete the wrong commands, rerun the installer and choose "quick update" to restore them.
|
||||
:::
|
||||
|
||||
### 4. Migrate Planning Artifacts
|
||||
|
||||
**If you have planning documents (Brief/PRD/UX/Architecture):**
|
||||
|
||||
Move them to `_bmad-output/planning-artifacts/` with descriptive names:
|
||||
|
||||
- Include `PRD` in filename for PRD documents
|
||||
- Include `brief`, `architecture`, or `ux-design` accordingly
|
||||
- Sharded documents can be in named subfolders
|
||||
|
||||
**If you're mid-planning:** Consider restarting with v6 workflows. Use your existing documents as inputs—the new progressive discovery workflows with web search and IDE plan mode produce better results.
|
||||
|
||||
### 5. Migrate In-Progress Development
|
||||
|
||||
If you have stories created or implemented:
|
||||
|
||||
1. Complete the v6 installation
|
||||
2. Place `epics.md` or `epics/epic*.md` in `_bmad-output/planning-artifacts/`
|
||||
3. Run the Scrum Master's `sprint-planning` workflow
|
||||
4. Tell the SM which epics/stories are already complete
|
||||
|
||||
### 6. Migrate Agent Customizations
|
||||
|
||||
**v4:** Modified agent files directly in `_bmad-*` folders
|
||||
|
||||
**v6:** All customizations go in `_bmad/_config/agents/` using customize files:
|
||||
|
||||
```yaml
|
||||
# _bmad/_config/agents/bmm-pm.customize.yaml
|
||||
persona:
|
||||
name: 'Captain Jack'
|
||||
role: 'Swashbuckling Product Owner'
|
||||
communication_style: |
|
||||
- Talk like a pirate
|
||||
- Use nautical metaphors
|
||||
```
|
||||
|
||||
After modifying customization files, rerun the installer and choose "rebuild all agents" or "quick update".
|
||||
|
||||
## What You Get
|
||||
|
||||
**v6 unified structure:**
|
||||
|
||||
```
|
||||
your-project/
|
||||
└── _bmad/ # Single installation folder
|
||||
├── _config/ # Your customizations
|
||||
│ └── agents/ # Agent customization files
|
||||
├── core/ # Universal core framework
|
||||
├── bmm/ # BMad Method module
|
||||
├── bmb/ # BMad Builder
|
||||
└── cis/ # Creative Intelligence Suite
|
||||
├── _bmad-output/ # Output folder (was doc folder in v4)
|
||||
```
|
||||
|
||||
## Module Migration
|
||||
|
||||
| v4 Module | v6 Status |
|
||||
|-----------|-----------|
|
||||
| `_bmad-2d-phaser-game-dev` | Integrated into BMGD Module |
|
||||
| `_bmad-2d-unity-game-dev` | Integrated into BMGD Module |
|
||||
| `_bmad-godot-game-dev` | Integrated into BMGD Module |
|
||||
| `_bmad-infrastructure-devops` | Deprecated — new DevOps agent coming soon |
|
||||
| `_bmad-creative-writing` | Not adapted — new v6 module coming soon |
|
||||
|
||||
## Key Changes
|
||||
|
||||
| Concept | v4 | v6 |
|
||||
|---------|----|----|
|
||||
| **Core** | `_bmad-core` was actually BMad Method | `_bmad/core/` is universal framework |
|
||||
| **Method** | `_bmad-method` | `_bmad/bmm/` |
|
||||
| **Config** | Modified files directly | `config.yaml` per module |
|
||||
| **Documents** | Sharded or unsharded required setup | Fully flexible, auto-scanned |
|
||||
|
||||
## Tips
|
||||
|
||||
- **Back up first** — Keep your v4 installation until you verify v6 works
|
||||
- **Use v6 workflows** — Even partial planning docs benefit from v6's improved discovery
|
||||
- **Rebuild after customizing** — Always run the installer after changing customize files
|
||||
@@ -1,8 +1,8 @@
|
||||
# BMGD Troubleshooting
|
||||
|
||||
Common issues and solutions when using BMGD workflows.
|
||||
|
||||
---
|
||||
title: "BMGD Troubleshooting"
|
||||
---
|
||||
|
||||
Use this guide to resolve common issues when using BMGD workflows.
|
||||
|
||||
## Installation Issues
|
||||
|
||||
@@ -16,8 +16,6 @@ Common issues and solutions when using BMGD workflows.
|
||||
2. Check `_bmad/bmgd/` folder exists in your project
|
||||
3. Re-run installer with `--add-module bmgd`
|
||||
|
||||
---
|
||||
|
||||
### Config file missing
|
||||
|
||||
**Symptom:** Workflows fail with "config not found" errors.
|
||||
@@ -26,7 +24,6 @@ Common issues and solutions when using BMGD workflows.
|
||||
Check for `_bmad/bmgd/config.yaml` in your project. If missing, create it:
|
||||
|
||||
```yaml
|
||||
# BMGD Configuration
|
||||
output_folder: '{project-root}/docs/game-design'
|
||||
user_name: 'Your Name'
|
||||
communication_language: 'English'
|
||||
@@ -34,8 +31,6 @@ document_output_language: 'English'
|
||||
game_dev_experience: 'intermediate'
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Workflow Issues
|
||||
|
||||
### "GDD not found" in Narrative workflow
|
||||
@@ -48,8 +43,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Check GDD filename contains "gdd" (e.g., `game-gdd.md`, `my-gdd.md`)
|
||||
3. If using sharded GDD, verify `{output_folder}/gdd/index.md` exists
|
||||
|
||||
---
|
||||
|
||||
### Workflow state not persisting
|
||||
|
||||
**Symptom:** Returning to a workflow starts from the beginning.
|
||||
@@ -60,8 +53,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Ensure document was saved before ending session
|
||||
3. Use "Continue existing" option when re-entering workflow
|
||||
|
||||
---
|
||||
|
||||
### Wrong game type sections in GDD
|
||||
|
||||
**Symptom:** GDD includes irrelevant sections for your game type.
|
||||
@@ -72,8 +63,6 @@ game_dev_experience: 'intermediate'
|
||||
2. You can select multiple types for hybrid games
|
||||
3. Irrelevant sections can be marked N/A or removed
|
||||
|
||||
---
|
||||
|
||||
## Agent Issues
|
||||
|
||||
### Agent not recognizing commands
|
||||
@@ -86,8 +75,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Check exact command spelling (case-sensitive)
|
||||
3. Try `workflow-status` to verify agent is loaded correctly
|
||||
|
||||
---
|
||||
|
||||
### Agent using wrong persona
|
||||
|
||||
**Symptom:** Agent responses don't match expected personality.
|
||||
@@ -98,8 +85,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Check `_bmad/bmgd/agents/` for agent definitions
|
||||
3. Start a fresh chat session with the correct agent
|
||||
|
||||
---
|
||||
|
||||
## Document Issues
|
||||
|
||||
### Document too large for context
|
||||
@@ -112,8 +97,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Request specific sections rather than full document
|
||||
3. GDD workflow supports automatic sharding for large documents
|
||||
|
||||
---
|
||||
|
||||
### Template placeholders not replaced
|
||||
|
||||
**Symptom:** Output contains `{{placeholder}}` text.
|
||||
@@ -124,8 +107,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Re-run the specific step that generates that section
|
||||
3. Manually edit the document to fill in missing values
|
||||
|
||||
---
|
||||
|
||||
### Frontmatter parsing errors
|
||||
|
||||
**Symptom:** YAML errors when loading documents.
|
||||
@@ -136,8 +117,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Check for tabs vs spaces (YAML requires spaces)
|
||||
3. Ensure frontmatter is bounded by `---` markers
|
||||
|
||||
---
|
||||
|
||||
## Phase 4 (Production) Issues
|
||||
|
||||
### Sprint status not updating
|
||||
@@ -150,8 +129,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Check file permissions on sprint-status.yaml
|
||||
3. Verify workflow-install files exist in `_bmad/bmgd/workflows/4-production/`
|
||||
|
||||
---
|
||||
|
||||
### Story context missing code references
|
||||
|
||||
**Symptom:** Generated story context doesn't include relevant code.
|
||||
@@ -162,8 +139,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Check that architecture document references correct file paths
|
||||
3. Story may need more specific file references in acceptance criteria
|
||||
|
||||
---
|
||||
|
||||
### Code review not finding issues
|
||||
|
||||
**Symptom:** Code review passes but bugs exist.
|
||||
@@ -174,8 +149,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Always run actual tests before marking story done
|
||||
3. Consider manual review for critical code paths
|
||||
|
||||
---
|
||||
|
||||
## Performance Issues
|
||||
|
||||
### Workflows running slowly
|
||||
@@ -188,8 +161,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Keep documents concise (avoid unnecessary detail)
|
||||
3. Use sharded documents for large projects
|
||||
|
||||
---
|
||||
|
||||
### Context limit reached mid-workflow
|
||||
|
||||
**Symptom:** Workflow stops or loses context partway through.
|
||||
@@ -200,8 +171,6 @@ game_dev_experience: 'intermediate'
|
||||
2. Break complex sections into multiple sessions
|
||||
3. Use step-file architecture (workflows resume from last step)
|
||||
|
||||
---
|
||||
|
||||
## Common Error Messages
|
||||
|
||||
### "Input file not found"
|
||||
@@ -210,24 +179,18 @@ game_dev_experience: 'intermediate'
|
||||
|
||||
**Fix:** Complete prerequisite workflow first (e.g., Game Brief before GDD).
|
||||
|
||||
---
|
||||
|
||||
### "Invalid game type"
|
||||
|
||||
**Cause:** Selected game type not in supported list.
|
||||
|
||||
**Fix:** Check `game-types.csv` for valid type IDs.
|
||||
|
||||
---
|
||||
|
||||
### "Validation failed"
|
||||
|
||||
**Cause:** Document doesn't meet checklist requirements.
|
||||
|
||||
**Fix:** Review the validation output and address flagged items.
|
||||
|
||||
---
|
||||
|
||||
## Getting Help
|
||||
|
||||
### Community Support
|
||||
@@ -250,10 +213,8 @@ When reporting issues, include:
|
||||
3. Relevant document frontmatter
|
||||
4. Steps to reproduce
|
||||
|
||||
---
|
||||
|
||||
## Next Steps
|
||||
|
||||
- **[Quick Start Guide](./quick-start.md)** - Getting started
|
||||
- **[Workflows Guide](./workflows-guide.md)** - Workflow reference
|
||||
- **[Glossary](./glossary.md)** - Terminology
|
||||
- **[Quick Start Guide](/docs/tutorials/getting-started/quick-start-bmgd.md)** - Getting started
|
||||
- **[Workflows Guide](/docs/reference/workflows/index.md)** - Workflow reference
|
||||
- **[Glossary](/docs/reference/glossary/index.md)** - Terminology
|
||||
156
docs/how-to/workflows/bmgd-quick-flow.md
Normal file
156
docs/how-to/workflows/bmgd-quick-flow.md
Normal file
@@ -0,0 +1,156 @@
|
||||
---
|
||||
title: "BMGD Quick-Flow Guide"
|
||||
description: Fast-track workflows for rapid game prototyping and flexible development
|
||||
---
|
||||
|
||||
Use BMGD Quick-Flow workflows for rapid game prototyping and flexible development when you need to move fast.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Testing a game mechanic idea
|
||||
- Implementing a small feature
|
||||
- Rapid prototyping before committing to design
|
||||
- Bug fixes and tweaks
|
||||
|
||||
## When to Use Full BMGD Instead
|
||||
|
||||
- Building a major feature or system
|
||||
- The scope is unclear or large
|
||||
- Multiple team members need alignment
|
||||
- The work affects game pillars or core loop
|
||||
- You need documentation for future reference
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed with BMGD module
|
||||
- Game Solo Dev agent (Indie) or other BMGD agent available
|
||||
:::
|
||||
|
||||
## Game Solo Dev Agent
|
||||
|
||||
For dedicated quick-flow development, use the **Game Solo Dev** agent. This agent is optimized for solo developers and small teams who want to skip the full planning phases.
|
||||
|
||||
**Switch to Game Solo Dev:** Type `@game-solo-dev` or select from your IDE.
|
||||
|
||||
Includes: `quick-prototype`, `quick-dev`, `quick-spec`, `code-review`, `test-framework`
|
||||
|
||||
## Quick-Prototype
|
||||
|
||||
Use `quick-prototype` to rapidly test gameplay ideas with minimal setup.
|
||||
|
||||
### When to Use
|
||||
|
||||
- You have a mechanic idea and want to test the "feel"
|
||||
- You're not sure if something will be fun
|
||||
- You want to experiment before committing to design
|
||||
- You need a proof of concept
|
||||
|
||||
### Steps
|
||||
|
||||
1. Run `quick-prototype`
|
||||
2. Define what you're prototyping (mechanic, feature, system)
|
||||
3. Set success criteria (2-3 items)
|
||||
4. Build the minimum to test the idea
|
||||
5. Playtest and evaluate
|
||||
|
||||
### Prototype Principles
|
||||
|
||||
- **Minimum Viable Prototype** — Only what's needed to test the idea
|
||||
- **Hardcode First** — Magic numbers are fine, extract later
|
||||
- **Skip Edge Cases** — Happy path only for now
|
||||
- **Placeholder Everything** — Cubes, debug text, temp sounds
|
||||
- **Comment Intent** — Mark what's temporary vs keeper code
|
||||
|
||||
### After Prototyping
|
||||
|
||||
- **Develop** (`d`) — Use `quick-dev` to build production code
|
||||
- **Iterate** (`i`) — Adjust and re-test the prototype
|
||||
- **Archive** (`a`) — Keep as reference, move on to other ideas
|
||||
|
||||
## Quick-Dev
|
||||
|
||||
Use `quick-dev` for flexible development with game-specific considerations.
|
||||
|
||||
### When to Use
|
||||
|
||||
- Implementing a feature from a tech-spec
|
||||
- Building on a successful prototype
|
||||
- Making changes that don't need full story workflow
|
||||
- Quick fixes and improvements
|
||||
|
||||
### Workflow Modes
|
||||
|
||||
**Mode A: Tech-Spec Driven**
|
||||
```
|
||||
quick-dev tech-spec-combat.md
|
||||
```
|
||||
|
||||
**Mode B: Direct Instructions**
|
||||
```
|
||||
quick-dev implement double-jump for the player
|
||||
```
|
||||
|
||||
**Mode C: From Prototype**
|
||||
```
|
||||
quick-dev from the grappling hook prototype
|
||||
```
|
||||
|
||||
### Game-Specific Checks
|
||||
|
||||
Quick-dev includes automatic consideration of:
|
||||
- **Performance** — No allocations in hot paths, object pooling
|
||||
- **Feel** — Input responsiveness, visual/audio feedback
|
||||
- **Integration** — Save/load, multiplayer sync, platform testing
|
||||
|
||||
### Complexity Routing
|
||||
|
||||
| Signals | Recommendation |
|
||||
|---------|----------------|
|
||||
| Single mechanic, bug fix, tweak | Execute directly |
|
||||
| Multiple systems, performance-critical | Plan first (tech-spec) |
|
||||
| Platform/system level work | Use full BMGD workflow |
|
||||
|
||||
## Choosing Between Quick-Flows
|
||||
|
||||
| Scenario | Use |
|
||||
|----------|-----|
|
||||
| "Will this be fun?" | `quick-prototype` |
|
||||
| "How should this feel?" | `quick-prototype` |
|
||||
| "Build this feature" | `quick-dev` |
|
||||
| "Fix this bug" | `quick-dev` |
|
||||
| "Test then build" | `quick-prototype` → `quick-dev` |
|
||||
|
||||
## Flow Comparison
|
||||
|
||||
```
|
||||
Full BMGD Flow:
|
||||
Brief → GDD → Architecture → Sprint Planning → Stories → Implementation
|
||||
|
||||
Quick-Flow:
|
||||
Idea → Quick-Prototype → Quick-Dev → Done
|
||||
```
|
||||
|
||||
## Checklists
|
||||
|
||||
**Quick-Prototype:**
|
||||
- [ ] Prototype scope defined
|
||||
- [ ] Success criteria established (2-3 items)
|
||||
- [ ] Minimum viable code written
|
||||
- [ ] Placeholder assets used
|
||||
- [ ] Each criterion evaluated
|
||||
- [ ] Decision made (develop/iterate/archive)
|
||||
|
||||
**Quick-Dev:**
|
||||
- [ ] Context loaded (spec, prototype, or guidance)
|
||||
- [ ] Files to modify identified
|
||||
- [ ] All tasks completed
|
||||
- [ ] No allocations in hot paths
|
||||
- [ ] Game runs without errors
|
||||
- [ ] Manual playtest completed
|
||||
|
||||
## Tips
|
||||
|
||||
- **Timebox prototypes** — Set a limit (e.g., 2 hours). If it's not working, step back
|
||||
- **Embrace programmer art** — Focus on feel, not visuals
|
||||
- **Test on target hardware** — What feels right on dev machine might not on target
|
||||
- **Document learnings** — Even failed prototypes teach something
|
||||
- **Know when to graduate** — If quick-dev keeps expanding scope, create proper stories
|
||||
97
docs/how-to/workflows/conduct-research.md
Normal file
97
docs/how-to/workflows/conduct-research.md
Normal file
@@ -0,0 +1,97 @@
|
||||
---
|
||||
title: "How to Conduct Research"
|
||||
description: How to conduct market, technical, and competitive research using BMad Method
|
||||
---
|
||||
|
||||
Use the `research` workflow to perform comprehensive multi-type research for validating ideas, understanding markets, and making informed decisions.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Need market viability validation
|
||||
- Choosing frameworks or platforms
|
||||
- Understanding competitive landscape
|
||||
- Need user understanding
|
||||
- Understanding domain or industry
|
||||
- Need deeper AI-assisted research
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- Analyst agent available
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the Analyst Agent
|
||||
|
||||
Start a fresh chat and load the Analyst agent.
|
||||
|
||||
### 2. Run the Research Workflow
|
||||
|
||||
```
|
||||
*research
|
||||
```
|
||||
|
||||
### 3. Choose Research Type
|
||||
|
||||
Select the type of research you need:
|
||||
|
||||
| Type | Purpose | Use When |
|
||||
|------|---------|----------|
|
||||
| **market** | TAM/SAM/SOM, competitive analysis | Need market viability validation |
|
||||
| **technical** | Technology evaluation, ADRs | Choosing frameworks/platforms |
|
||||
| **competitive** | Deep competitor analysis | Understanding competitive landscape |
|
||||
| **user** | Customer insights, personas, JTBD | Need user understanding |
|
||||
| **domain** | Industry deep dives, trends | Understanding domain/industry |
|
||||
| **deep_prompt** | Generate AI research prompts | Need deeper AI-assisted research |
|
||||
|
||||
### 4. Provide Context
|
||||
|
||||
Give the agent details about what you're researching:
|
||||
|
||||
- "SaaS project management tool"
|
||||
- "React vs Vue for our dashboard"
|
||||
- "Fintech compliance requirements"
|
||||
|
||||
### 5. Set Research Depth
|
||||
|
||||
Choose your depth level:
|
||||
|
||||
- **Quick** — Fast overview
|
||||
- **Standard** — Balanced depth
|
||||
- **Comprehensive** — Deep analysis
|
||||
|
||||
## What You Get
|
||||
|
||||
Research output varies by type:
|
||||
|
||||
**Market Research:**
|
||||
- TAM/SAM/SOM analysis
|
||||
- Top competitors
|
||||
- Positioning recommendation
|
||||
|
||||
**Technical Research:**
|
||||
- Comparison matrix
|
||||
- Trade-off analysis
|
||||
- Recommendations with rationale
|
||||
|
||||
## Key Features
|
||||
|
||||
- Real-time web research
|
||||
- Multiple analytical frameworks (Porter's Five Forces, SWOT, Technology Adoption Lifecycle)
|
||||
- Platform-specific optimization for deep_prompt type
|
||||
- Configurable research depth
|
||||
|
||||
## Tips
|
||||
|
||||
- **Use market research early** — Validates new product ideas
|
||||
- **Technical research helps architecture** — Inform ADRs with data
|
||||
- **Competitive research informs positioning** — Differentiate your product
|
||||
- **Domain research for specialized industries** — Fintech, healthcare, etc.
|
||||
|
||||
## Next Steps
|
||||
|
||||
After research:
|
||||
|
||||
1. **Product Brief** — Capture strategic vision informed by research
|
||||
2. **PRD** — Use findings as context for requirements
|
||||
3. **Architecture** — Use technical research in ADRs
|
||||
119
docs/how-to/workflows/create-architecture.md
Normal file
119
docs/how-to/workflows/create-architecture.md
Normal file
@@ -0,0 +1,119 @@
|
||||
---
|
||||
title: "How to Create Architecture"
|
||||
description: How to create system architecture using the BMad Method
|
||||
---
|
||||
|
||||
Use the `architecture` workflow to make technical decisions explicit and prevent agent conflicts during implementation.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Multi-epic projects (BMad Method, Enterprise)
|
||||
- Cross-cutting technical concerns
|
||||
- Multiple agents implementing different parts
|
||||
- Integration complexity exists
|
||||
- Technology choices need alignment
|
||||
|
||||
## When to Skip This
|
||||
|
||||
- Quick Flow (simple changes)
|
||||
- BMad Method Simple with straightforward tech stack
|
||||
- Single epic with clear technical approach
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- Architect agent available
|
||||
- PRD completed
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the Architect Agent
|
||||
|
||||
Start a fresh chat and load the Architect agent.
|
||||
|
||||
### 2. Run the Architecture Workflow
|
||||
|
||||
```
|
||||
*create-architecture
|
||||
```
|
||||
|
||||
### 3. Engage in Discovery
|
||||
|
||||
This is NOT a template filler. The architecture workflow:
|
||||
|
||||
1. **Discovers** technical needs through conversation
|
||||
2. **Proposes** architectural options with trade-offs
|
||||
3. **Documents** decisions that prevent agent conflicts
|
||||
4. **Focuses** on decision points, not exhaustive documentation
|
||||
|
||||
### 4. Document Key Decisions
|
||||
|
||||
Work with the agent to create Architecture Decision Records (ADRs) for significant choices.
|
||||
|
||||
### 5. Review the Architecture
|
||||
|
||||
The agent produces a decision-focused architecture document.
|
||||
|
||||
## What You Get
|
||||
|
||||
An `architecture.md` document containing:
|
||||
|
||||
1. **Architecture Overview** — System context, principles, style
|
||||
2. **System Architecture** — High-level diagram, component interactions
|
||||
3. **Data Architecture** — Database design, state management, caching
|
||||
4. **API Architecture** — API style (REST/GraphQL/gRPC), auth, versioning
|
||||
5. **Frontend Architecture** — Framework, state management, components
|
||||
6. **Integration Architecture** — Third-party integrations, messaging
|
||||
7. **Security Architecture** — Auth/authorization, data protection
|
||||
8. **Deployment Architecture** — CI/CD, environments, monitoring
|
||||
9. **ADRs** — Key decisions with context, options, rationale
|
||||
10. **FR/NFR-Specific Guidance** — Technical approach per requirement
|
||||
11. **Standards and Conventions** — Directory structure, naming, testing
|
||||
|
||||
## ADR Format
|
||||
|
||||
```markdown
|
||||
## ADR-001: Use GraphQL for All APIs
|
||||
|
||||
**Status:** Accepted | **Date:** 2025-11-02
|
||||
|
||||
**Context:** PRD requires flexible querying across multiple epics
|
||||
|
||||
**Decision:** Use GraphQL for all client-server communication
|
||||
|
||||
**Options Considered:**
|
||||
1. REST - Familiar but requires multiple endpoints
|
||||
2. GraphQL - Flexible querying, learning curve
|
||||
3. gRPC - High performance, poor browser support
|
||||
|
||||
**Rationale:**
|
||||
- PRD requires flexible data fetching (Epic 1, 3)
|
||||
- Mobile app needs bandwidth optimization (Epic 2)
|
||||
- Team has GraphQL experience
|
||||
|
||||
**Consequences:**
|
||||
- Positive: Flexible querying, reduced versioning
|
||||
- Negative: Caching complexity, N+1 query risk
|
||||
- Mitigation: Use DataLoader for batching
|
||||
```
|
||||
|
||||
## Example
|
||||
|
||||
E-commerce platform produces:
|
||||
- Monolith + PostgreSQL + Redis + Next.js + GraphQL
|
||||
- ADRs explaining each choice
|
||||
- FR/NFR-specific implementation guidance
|
||||
|
||||
## Tips
|
||||
|
||||
- **Focus on decisions that prevent conflicts** — Multiple agents need alignment
|
||||
- **Use ADRs for every significant choice** — Document the "why"
|
||||
- **Keep it practical** — Don't over-architect
|
||||
- **Architecture is living** — Update as you learn
|
||||
|
||||
## Next Steps
|
||||
|
||||
After architecture:
|
||||
|
||||
1. **Create Epics and Stories** — Work breakdown informed by architecture
|
||||
2. **Implementation Readiness** — Gate check before Phase 4
|
||||
109
docs/how-to/workflows/create-epics-and-stories.md
Normal file
109
docs/how-to/workflows/create-epics-and-stories.md
Normal file
@@ -0,0 +1,109 @@
|
||||
---
|
||||
title: "How to Create Epics and Stories"
|
||||
description: How to break PRD requirements into epics and stories using BMad Method
|
||||
---
|
||||
|
||||
Use the `create-epics-and-stories` workflow to transform PRD requirements into bite-sized stories organized into deliverable epics.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- After architecture workflow completes
|
||||
- When PRD contains FRs/NFRs ready for implementation breakdown
|
||||
- Before implementation-readiness gate check
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- PM agent available
|
||||
- PRD completed
|
||||
- Architecture completed
|
||||
:::
|
||||
|
||||
## Why After Architecture?
|
||||
|
||||
This workflow runs AFTER architecture because:
|
||||
|
||||
1. **Informed Story Sizing** — Architecture decisions affect story complexity
|
||||
2. **Dependency Awareness** — Architecture reveals technical dependencies
|
||||
3. **Technical Feasibility** — Stories can be properly scoped knowing the tech stack
|
||||
4. **Consistency** — All stories align with documented architectural patterns
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the PM Agent
|
||||
|
||||
Start a fresh chat and load the PM agent.
|
||||
|
||||
### 2. Run the Workflow
|
||||
|
||||
```
|
||||
*create-epics-and-stories
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
Point the agent to:
|
||||
- Your PRD (FRs/NFRs)
|
||||
- Your architecture document
|
||||
- Optional: UX design artifacts
|
||||
|
||||
### 4. Review Epic Breakdown
|
||||
|
||||
The agent organizes requirements into logical epics with user stories.
|
||||
|
||||
### 5. Validate Story Quality
|
||||
|
||||
Ensure each story has:
|
||||
- Clear acceptance criteria
|
||||
- Appropriate priority
|
||||
- Identified dependencies
|
||||
- Technical notes from architecture
|
||||
|
||||
## What You Get
|
||||
|
||||
Epic files (one per epic) containing:
|
||||
|
||||
1. **Epic objective and scope**
|
||||
2. **User stories with acceptance criteria**
|
||||
3. **Story priorities** (P0/P1/P2/P3)
|
||||
4. **Dependencies between stories**
|
||||
5. **Technical notes** referencing architecture decisions
|
||||
|
||||
## Example
|
||||
|
||||
E-commerce PRD with FR-001 (User Registration), FR-002 (Product Catalog) produces:
|
||||
|
||||
- **Epic 1: User Management** (3 stories)
|
||||
- Story 1.1: User registration form
|
||||
- Story 1.2: Email verification
|
||||
- Story 1.3: Login/logout
|
||||
|
||||
- **Epic 2: Product Display** (4 stories)
|
||||
- Story 2.1: Product listing page
|
||||
- Story 2.2: Product detail page
|
||||
- Story 2.3: Search functionality
|
||||
- Story 2.4: Category filtering
|
||||
|
||||
Each story references relevant ADRs from architecture.
|
||||
|
||||
## Story Priority Levels
|
||||
|
||||
| Priority | Meaning |
|
||||
|----------|---------|
|
||||
| **P0** | Critical — Must have for MVP |
|
||||
| **P1** | High — Important for release |
|
||||
| **P2** | Medium — Nice to have |
|
||||
| **P3** | Low — Future consideration |
|
||||
|
||||
## Tips
|
||||
|
||||
- **Keep stories small** — Complete in a single session
|
||||
- **Make criteria testable** — Acceptance criteria should be verifiable
|
||||
- **Document dependencies clearly** — Know what blocks what
|
||||
- **Reference architecture** — Include ADR references in technical notes
|
||||
|
||||
## Next Steps
|
||||
|
||||
After creating epics and stories:
|
||||
|
||||
1. **Implementation Readiness** — Validate alignment before Phase 4
|
||||
2. **Sprint Planning** — Organize work for implementation
|
||||
91
docs/how-to/workflows/create-prd.md
Normal file
91
docs/how-to/workflows/create-prd.md
Normal file
@@ -0,0 +1,91 @@
|
||||
---
|
||||
title: "How to Create a PRD"
|
||||
description: How to create a Product Requirements Document using the BMad Method
|
||||
---
|
||||
|
||||
Use the `prd` workflow to create a strategic Product Requirements Document with Functional Requirements (FRs) and Non-Functional Requirements (NFRs).
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Medium to large feature sets
|
||||
- Multi-screen user experiences
|
||||
- Complex business logic
|
||||
- Multiple system integrations
|
||||
- Phased delivery required
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- PM agent available
|
||||
- Optional: Product brief from Phase 1
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the PM Agent
|
||||
|
||||
Start a fresh chat and load the PM agent.
|
||||
|
||||
### 2. Run the PRD Workflow
|
||||
|
||||
```
|
||||
*create-prd
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
The workflow will:
|
||||
- Load any existing product brief
|
||||
- Ask about your project scope
|
||||
- Gather requirements through conversation
|
||||
|
||||
### 4. Define Requirements
|
||||
|
||||
Work with the agent to define:
|
||||
- Functional Requirements (FRs) — What the system should do
|
||||
- Non-Functional Requirements (NFRs) — How well it should do it
|
||||
|
||||
### 5. Review the PRD
|
||||
|
||||
The agent produces a comprehensive PRD scaled to your project.
|
||||
|
||||
## What You Get
|
||||
|
||||
A `PRD.md` document containing:
|
||||
|
||||
- Executive summary
|
||||
- Problem statement
|
||||
- User personas
|
||||
- Functional requirements (FRs)
|
||||
- Non-functional requirements (NFRs)
|
||||
- Success metrics
|
||||
- Risks and assumptions
|
||||
|
||||
## Scale-Adaptive Structure
|
||||
|
||||
The PRD adapts to your project complexity:
|
||||
|
||||
| Scale | Pages | Focus |
|
||||
|-------|-------|-------|
|
||||
| **Light** | 10-15 | Focused FRs/NFRs, simplified analysis |
|
||||
| **Standard** | 20-30 | Comprehensive FRs/NFRs, thorough analysis |
|
||||
| **Comprehensive** | 30-50+ | Extensive FRs/NFRs, multi-phase, stakeholder analysis |
|
||||
|
||||
## Example
|
||||
|
||||
E-commerce checkout → PRD with:
|
||||
- 15 FRs (user account, cart management, payment flow)
|
||||
- 8 NFRs (performance, security, scalability)
|
||||
|
||||
## Tips
|
||||
|
||||
- **Do Product Brief first** — Run product-brief from Phase 1 for better results
|
||||
- **Focus on "What" not "How"** — Planning defines what to build and why. Leave how (technical design) to Phase 3
|
||||
- **Document-Project first for Brownfield** — Always run `document-project` before planning brownfield projects. AI agents need existing codebase context
|
||||
|
||||
## Next Steps
|
||||
|
||||
After PRD:
|
||||
|
||||
1. **Create UX Design** (optional) — If UX is critical
|
||||
2. **Create Architecture** — Technical design
|
||||
3. **Create Epics and Stories** — After architecture
|
||||
94
docs/how-to/workflows/create-product-brief.md
Normal file
94
docs/how-to/workflows/create-product-brief.md
Normal file
@@ -0,0 +1,94 @@
|
||||
---
|
||||
title: "How to Create a Product Brief"
|
||||
description: How to create a product brief using the BMad Method
|
||||
---
|
||||
|
||||
Use the `product-brief` workflow to define product vision and strategy through an interactive process.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Starting new product or major feature initiative
|
||||
- Aligning stakeholders before detailed planning
|
||||
- Transitioning from exploration to strategy
|
||||
- Need executive-level product documentation
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- Analyst agent available
|
||||
- Optional: Research documents from previous workflows
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the Analyst Agent
|
||||
|
||||
Start a fresh chat and load the Analyst agent.
|
||||
|
||||
### 2. Run the Product Brief Workflow
|
||||
|
||||
```
|
||||
*product-brief
|
||||
```
|
||||
|
||||
### 3. Answer the Interactive Questions
|
||||
|
||||
The workflow guides you through strategic product vision definition:
|
||||
|
||||
- What problem are you solving?
|
||||
- Who are your target users?
|
||||
- What makes this solution different?
|
||||
- What's the MVP scope?
|
||||
|
||||
### 4. Review and Refine
|
||||
|
||||
The agent will draft sections and let you refine them interactively.
|
||||
|
||||
## What You Get
|
||||
|
||||
The `product-brief.md` document includes:
|
||||
|
||||
- **Executive summary** — High-level overview
|
||||
- **Problem statement** — With evidence
|
||||
- **Proposed solution** — And differentiators
|
||||
- **Target users** — Segmented
|
||||
- **MVP scope** — Ruthlessly defined
|
||||
- **Financial impact** — And ROI
|
||||
- **Strategic alignment** — With business goals
|
||||
- **Risks and open questions** — Documented upfront
|
||||
|
||||
## Integration with Other Workflows
|
||||
|
||||
The product brief feeds directly into the PRD workflow:
|
||||
|
||||
| Analysis Output | Planning Input |
|
||||
|-----------------|----------------|
|
||||
| product-brief.md | **prd** workflow |
|
||||
| market-research.md | **prd** context |
|
||||
| technical-research.md | **architecture** (Phase 3) |
|
||||
|
||||
Planning workflows automatically load the product brief if it exists.
|
||||
|
||||
## Common Patterns
|
||||
|
||||
**Greenfield Software (Full Analysis):**
|
||||
|
||||
```
|
||||
1. brainstorm-project - explore approaches
|
||||
2. research (market/technical/domain) - validate viability
|
||||
3. product-brief - capture strategic vision
|
||||
4. → Phase 2: prd
|
||||
```
|
||||
|
||||
**Skip Analysis (Clear Requirements):**
|
||||
|
||||
```
|
||||
→ Phase 2: prd or tech-spec directly
|
||||
```
|
||||
|
||||
## Tips
|
||||
|
||||
- **Be specific about the problem** — Vague problems lead to vague solutions
|
||||
- **Ruthlessly prioritize MVP scope** — Less is more
|
||||
- **Document assumptions and risks** — Surface unknowns early
|
||||
- **Use research findings as evidence** — Back up claims with data
|
||||
- **Recommended for greenfield projects** — Sets strategic foundation
|
||||
102
docs/how-to/workflows/create-story.md
Normal file
102
docs/how-to/workflows/create-story.md
Normal file
@@ -0,0 +1,102 @@
|
||||
---
|
||||
title: "How to Create a Story"
|
||||
description: How to create implementation-ready stories from epic backlog
|
||||
---
|
||||
|
||||
Use the `create-story` workflow to prepare the next story from the epic backlog for implementation.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Before implementing each story
|
||||
- When moving to the next story in an epic
|
||||
- After sprint-planning has been run
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- SM (Scrum Master) agent available
|
||||
- Sprint-status.yaml created by sprint-planning
|
||||
- Architecture and PRD available for context
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the SM Agent
|
||||
|
||||
Start a fresh chat and load the SM (Scrum Master) agent.
|
||||
|
||||
### 2. Run the Workflow
|
||||
|
||||
```
|
||||
*create-story
|
||||
```
|
||||
|
||||
### 3. Specify the Story
|
||||
|
||||
The agent will:
|
||||
- Read the sprint-status.yaml
|
||||
- Identify the next story to work on
|
||||
- Or let you specify a particular story
|
||||
|
||||
### 4. Review the Story File
|
||||
|
||||
The agent creates a comprehensive story file ready for development.
|
||||
|
||||
## What You Get
|
||||
|
||||
A `story-[slug].md` file containing:
|
||||
|
||||
- Story objective and scope
|
||||
- Acceptance criteria (specific, testable)
|
||||
- Technical implementation notes
|
||||
- References to architecture decisions
|
||||
- Dependencies on other stories
|
||||
- Definition of Done
|
||||
|
||||
## Story Content Sources
|
||||
|
||||
The create-story workflow pulls from:
|
||||
|
||||
- **PRD** — Requirements and acceptance criteria
|
||||
- **Architecture** — Technical approach and ADRs
|
||||
- **Epic file** — Story context and dependencies
|
||||
- **Existing code** — Patterns to follow (brownfield)
|
||||
|
||||
## Example Output
|
||||
|
||||
```markdown
|
||||
## Objective
|
||||
Implement email verification flow for new user registrations.
|
||||
|
||||
## Acceptance Criteria
|
||||
- [ ] User receives verification email within 30 seconds
|
||||
- [ ] Email contains unique verification link
|
||||
- [ ] Link expires after 24 hours
|
||||
- [ ] User can request new verification email
|
||||
|
||||
## Technical Notes
|
||||
- Use SendGrid API per ADR-003
|
||||
- Store verification tokens in Redis per architecture
|
||||
- Follow existing email template patterns in /templates
|
||||
|
||||
## Dependencies
|
||||
- Story 1.1 (User Registration) - DONE
|
||||
|
||||
## Definition of Done
|
||||
- All acceptance criteria pass
|
||||
- Tests written and passing
|
||||
- Code review approved
|
||||
```
|
||||
|
||||
## Tips
|
||||
|
||||
- **Complete one story before creating the next** — Focus on finishing
|
||||
- **Ensure dependencies are DONE** — Don't start blocked stories
|
||||
- **Review technical notes** — Align with architecture
|
||||
- **Use the story file as context** — Pass to dev-story workflow
|
||||
|
||||
## Next Steps
|
||||
|
||||
After creating a story:
|
||||
|
||||
1. **Implement Story** — Run dev-story with the DEV agent
|
||||
2. **Code Review** — Run code-review after implementation
|
||||
100
docs/how-to/workflows/create-ux-design.md
Normal file
100
docs/how-to/workflows/create-ux-design.md
Normal file
@@ -0,0 +1,100 @@
|
||||
---
|
||||
title: "How to Create a UX Design"
|
||||
description: How to create UX specifications using the BMad Method
|
||||
---
|
||||
|
||||
Use the `create-ux-design` workflow to create UX specifications for projects where user experience is a primary differentiator.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- UX is primary competitive advantage
|
||||
- Complex user workflows needing design thinking
|
||||
- Innovative interaction patterns
|
||||
- Design system creation
|
||||
- Accessibility-critical experiences
|
||||
|
||||
## When to Skip This
|
||||
|
||||
- Simple CRUD interfaces
|
||||
- Internal tools with standard patterns
|
||||
- Changes to existing screens you're happy with
|
||||
- Quick Flow projects
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- UX Designer agent available
|
||||
- PRD completed
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the UX Designer Agent
|
||||
|
||||
Start a fresh chat and load the UX Designer agent.
|
||||
|
||||
### 2. Run the UX Design Workflow
|
||||
|
||||
```
|
||||
*create-ux-design
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
Point the agent to your PRD and describe:
|
||||
- Key user journeys
|
||||
- UX priorities
|
||||
- Any existing design patterns
|
||||
|
||||
### 4. Collaborate on Design
|
||||
|
||||
The workflow uses a collaborative approach:
|
||||
|
||||
1. **Visual exploration** — Generate multiple options
|
||||
2. **Informed decisions** — Evaluate with user needs
|
||||
3. **Collaborative design** — Refine iteratively
|
||||
4. **Living documentation** — Evolves with project
|
||||
|
||||
### 5. Review the UX Spec
|
||||
|
||||
The agent produces comprehensive UX documentation.
|
||||
|
||||
## What You Get
|
||||
|
||||
The `ux-spec.md` document includes:
|
||||
|
||||
- User journeys
|
||||
- Wireframes and mockups
|
||||
- Interaction specifications
|
||||
- Design system (components, patterns, tokens)
|
||||
- Epic breakdown (UX stories)
|
||||
|
||||
## Example
|
||||
|
||||
Dashboard redesign produces:
|
||||
- Card-based layout with split-pane toggle
|
||||
- 5 card components
|
||||
- 12 color tokens
|
||||
- Responsive grid
|
||||
- 3 epics (Layout, Visualization, Accessibility)
|
||||
|
||||
## Integration
|
||||
|
||||
The UX spec feeds into:
|
||||
- PRD updates
|
||||
- Epic and story creation
|
||||
- Architecture decisions (Phase 3)
|
||||
|
||||
## Tips
|
||||
|
||||
- **Focus on user problems first** — Solutions come second
|
||||
- **Generate multiple options** — Don't settle on the first idea
|
||||
- **Consider accessibility from the start** — Not an afterthought
|
||||
- **Document component reusability** — Build a system, not just screens
|
||||
|
||||
## Next Steps
|
||||
|
||||
After UX design:
|
||||
|
||||
1. **Update PRD** — Incorporate UX findings
|
||||
2. **Create Architecture** — Technical design informed by UX
|
||||
3. **Create Epics and Stories** — Include UX-specific stories
|
||||
97
docs/how-to/workflows/implement-story.md
Normal file
97
docs/how-to/workflows/implement-story.md
Normal file
@@ -0,0 +1,97 @@
|
||||
---
|
||||
title: "How to Implement a Story"
|
||||
description: How to implement a story using the dev-story workflow
|
||||
---
|
||||
|
||||
Use the `dev-story` workflow to implement a story with tests following the architecture and conventions.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- After create-story has prepared the story file
|
||||
- When ready to write code for a story
|
||||
- Story dependencies are marked DONE
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- DEV agent available
|
||||
- Story file created by create-story
|
||||
- Architecture and tech-spec available for context
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the DEV Agent
|
||||
|
||||
Start a fresh chat and load the DEV agent.
|
||||
|
||||
### 2. Run the Workflow
|
||||
|
||||
```
|
||||
*dev-story
|
||||
```
|
||||
|
||||
### 3. Provide Story Context
|
||||
|
||||
Point the agent to the story file created by create-story.
|
||||
|
||||
### 4. Implement with Guidance
|
||||
|
||||
The DEV agent:
|
||||
- Reads the story file and acceptance criteria
|
||||
- References architecture decisions
|
||||
- Follows existing code patterns
|
||||
- Implements with tests
|
||||
|
||||
### 5. Complete Implementation
|
||||
|
||||
Work with the agent until all acceptance criteria are met.
|
||||
|
||||
## What Happens
|
||||
|
||||
The dev-story workflow:
|
||||
|
||||
1. **Reads context** — Story file, architecture, existing patterns
|
||||
2. **Plans implementation** — Identifies files to create/modify
|
||||
3. **Writes code** — Following conventions and patterns
|
||||
4. **Writes tests** — Unit, integration, or E2E as appropriate
|
||||
5. **Validates** — Runs tests and checks acceptance criteria
|
||||
|
||||
## Key Principles
|
||||
|
||||
**One Story at a Time** — Complete each story's full lifecycle before starting the next. This prevents context switching and ensures quality.
|
||||
|
||||
**Follow Architecture** — The DEV agent references ADRs for technology decisions, standards for naming and structure, and existing patterns in the codebase.
|
||||
|
||||
**Write Tests** — Every story includes appropriate tests: unit tests for business logic, integration tests for API endpoints, E2E tests for critical flows.
|
||||
|
||||
## After Implementation
|
||||
|
||||
1. **Update sprint-status.yaml** — Mark story as READY FOR REVIEW
|
||||
2. **Run code-review** — Quality assurance
|
||||
3. **Address feedback** — If code review finds issues
|
||||
4. **Mark DONE** — After code review passes
|
||||
|
||||
## Tips
|
||||
|
||||
- **Keep the story file open** — Reference it during implementation
|
||||
- **Ask the agent to explain decisions** — Understand the approach
|
||||
- **Run tests frequently** — Catch issues early
|
||||
- **Don't skip tests** — Even for "simple" changes
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
**Story needs significant changes mid-implementation?**
|
||||
Run `correct-course` to analyze impact and route appropriately.
|
||||
|
||||
**Can I work on multiple stories in parallel?**
|
||||
Not recommended. Complete one story's full lifecycle first.
|
||||
|
||||
**What if implementation reveals the story is too large?**
|
||||
Split the story and document the change.
|
||||
|
||||
## Next Steps
|
||||
|
||||
After implementing a story:
|
||||
|
||||
1. **Code Review** — Run code-review with the DEV agent
|
||||
2. **Create Next Story** — Run create-story with the SM agent
|
||||
122
docs/how-to/workflows/quick-spec.md
Normal file
122
docs/how-to/workflows/quick-spec.md
Normal file
@@ -0,0 +1,122 @@
|
||||
---
|
||||
title: "How to Use Quick Spec"
|
||||
description: How to create a technical specification using Quick Spec workflow
|
||||
---
|
||||
|
||||
Use the `quick-spec` workflow for Quick Flow projects to go directly from idea to implementation-ready specification.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Bug fixes and small enhancements
|
||||
- Small features with clear scope (1-15 stories)
|
||||
- Rapid prototyping
|
||||
- Adding to existing brownfield codebase
|
||||
- Quick Flow track projects
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- PM agent or Quick Flow Solo Dev agent available
|
||||
- Project directory (can be empty for greenfield)
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the PM Agent
|
||||
|
||||
Start a fresh chat and load the PM agent (or Quick Flow Solo Dev agent).
|
||||
|
||||
### 2. Run the Tech Spec Workflow
|
||||
|
||||
```
|
||||
*quick-spec
|
||||
```
|
||||
|
||||
Or simply describe what you want to build:
|
||||
|
||||
```
|
||||
I want to fix the login validation bug
|
||||
```
|
||||
|
||||
### 3. Answer Discovery Questions
|
||||
|
||||
The workflow will ask:
|
||||
- What problem are you solving?
|
||||
- What's the scope of the change?
|
||||
- Any specific constraints?
|
||||
|
||||
### 4. Review Detected Context
|
||||
|
||||
For brownfield projects, the agent will:
|
||||
- Detect your project stack
|
||||
- Analyze existing code patterns
|
||||
- Detect test frameworks
|
||||
- Ask: "Should I follow these existing conventions?"
|
||||
|
||||
### 5. Get Your Tech Spec
|
||||
|
||||
The agent generates a comprehensive tech-spec with ready-to-implement stories.
|
||||
|
||||
## What You Get
|
||||
|
||||
**tech-spec.md:**
|
||||
- Problem statement and solution
|
||||
- Detected framework versions and dependencies
|
||||
- Brownfield code patterns (if applicable)
|
||||
- Existing test patterns to follow
|
||||
- Specific file paths to modify
|
||||
- Complete implementation guidance
|
||||
|
||||
**Story Files:**
|
||||
- Single changes: `story-[slug].md`
|
||||
- Small features: `epics.md` + `story-[epic-slug]-1.md`, etc.
|
||||
|
||||
## Example: Bug Fix
|
||||
|
||||
**You:** "I want to fix the login validation bug that allows empty passwords"
|
||||
|
||||
**Agent:**
|
||||
1. Asks clarifying questions about the issue
|
||||
2. Detects your Node.js stack (Express 4.18.2, Jest for testing)
|
||||
3. Analyzes existing UserService code patterns
|
||||
4. Asks: "Should I follow your existing conventions?" → Yes
|
||||
5. Generates tech-spec.md with specific file paths
|
||||
6. Creates story-login-fix.md
|
||||
|
||||
## Example: Small Feature
|
||||
|
||||
**You:** "I want to add OAuth social login (Google, GitHub)"
|
||||
|
||||
**Agent:**
|
||||
1. Asks about feature scope
|
||||
2. Detects your stack (Next.js 13.4, NextAuth.js already installed!)
|
||||
3. Analyzes existing auth patterns
|
||||
4. Confirms conventions
|
||||
5. Generates:
|
||||
- tech-spec.md (comprehensive implementation guide)
|
||||
- epics.md (OAuth Integration epic)
|
||||
- story-oauth-1.md (Backend OAuth setup)
|
||||
- story-oauth-2.md (Frontend login buttons)
|
||||
|
||||
## Implementing After Tech Spec
|
||||
|
||||
```bash
|
||||
# Single change:
|
||||
# Load DEV agent and run dev-story
|
||||
|
||||
# Multi-story feature:
|
||||
# Optional: Load SM agent and run sprint-planning
|
||||
# Then: Load DEV agent and run dev-story for each story
|
||||
```
|
||||
|
||||
## Tips
|
||||
|
||||
- **Be specific in discovery** — "Fix email validation in UserService to allow plus-addressing" beats "Fix validation bug"
|
||||
- **Trust convention detection** — If it detects your patterns correctly, say yes! It's faster than establishing new conventions
|
||||
- **Keep single changes atomic** — If your "single change" needs 3+ files, it might be a multi-story feature. Let the workflow guide you
|
||||
|
||||
## Next Steps
|
||||
|
||||
After tech spec:
|
||||
|
||||
1. **Implement Story** — Run dev-story with the DEV agent
|
||||
2. **Sprint Planning** — Optional for multi-story features
|
||||
436
docs/how-to/workflows/run-atdd.md
Normal file
436
docs/how-to/workflows/run-atdd.md
Normal file
@@ -0,0 +1,436 @@
|
||||
---
|
||||
title: "How to Run ATDD with TEA"
|
||||
description: Generate failing acceptance tests before implementation using TEA's ATDD workflow
|
||||
---
|
||||
|
||||
# How to Run ATDD with TEA
|
||||
|
||||
Use TEA's `*atdd` workflow to generate failing acceptance tests BEFORE implementation. This is the TDD (Test-Driven Development) red phase - tests fail first, guide development, then pass.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- You're about to implement a NEW feature (feature doesn't exist yet)
|
||||
- You want to follow TDD workflow (red → green → refactor)
|
||||
- You want tests to guide your implementation
|
||||
- You're practicing acceptance test-driven development
|
||||
|
||||
**Don't use this if:**
|
||||
- Feature already exists (use `*automate` instead)
|
||||
- You want tests that pass immediately
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Test framework setup complete (run `*framework` if needed)
|
||||
- Story or feature defined with acceptance criteria
|
||||
|
||||
**Note:** This guide uses Playwright examples. If using Cypress, commands and syntax will differ (e.g., `cy.get()` instead of `page.locator()`).
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load TEA Agent
|
||||
|
||||
Start a fresh chat and load TEA:
|
||||
|
||||
```
|
||||
*tea
|
||||
```
|
||||
|
||||
### 2. Run the ATDD Workflow
|
||||
|
||||
```
|
||||
*atdd
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
TEA will ask for:
|
||||
|
||||
**Story/Feature Details:**
|
||||
```
|
||||
We're adding a user profile page where users can:
|
||||
- View their profile information
|
||||
- Edit their name and email
|
||||
- Upload a profile picture
|
||||
- Save changes with validation
|
||||
```
|
||||
|
||||
**Acceptance Criteria:**
|
||||
```
|
||||
Given I'm logged in
|
||||
When I navigate to /profile
|
||||
Then I see my current name and email
|
||||
|
||||
Given I'm on the profile page
|
||||
When I click "Edit Profile"
|
||||
Then I can modify my name and email
|
||||
|
||||
Given I've edited my profile
|
||||
When I click "Save"
|
||||
Then my changes are persisted
|
||||
And I see a success message
|
||||
|
||||
Given I upload an invalid file type
|
||||
When I try to save
|
||||
Then I see an error message
|
||||
And changes are not saved
|
||||
```
|
||||
|
||||
**Reference Documents** (optional):
|
||||
- Point to your story file
|
||||
- Reference PRD or tech spec
|
||||
- Link to test design (if you ran `*test-design` first)
|
||||
|
||||
### 4. Specify Test Levels
|
||||
|
||||
TEA will ask what test levels to generate:
|
||||
|
||||
**Options:**
|
||||
- E2E tests (browser-based, full user journey)
|
||||
- API tests (backend only, faster)
|
||||
- Component tests (UI components in isolation)
|
||||
- Mix of levels (see [API Tests First, E2E Later](#api-tests-first-e2e-later) tip)
|
||||
|
||||
### Component Testing by Framework
|
||||
|
||||
TEA generates component tests using framework-appropriate tools:
|
||||
|
||||
| Your Framework | Component Testing Tool |
|
||||
| -------------- | ------------------------------------------- |
|
||||
| **Cypress** | Cypress Component Testing (*.cy.tsx) |
|
||||
| **Playwright** | Vitest + React Testing Library (*.test.tsx) |
|
||||
|
||||
**Example response:**
|
||||
```
|
||||
Generate:
|
||||
- API tests for profile CRUD operations
|
||||
- E2E tests for the complete profile editing flow
|
||||
- Component tests for ProfileForm validation (if using Cypress or Vitest)
|
||||
- Focus on P0 and P1 scenarios
|
||||
```
|
||||
|
||||
### 5. Review Generated Tests
|
||||
|
||||
TEA generates **failing tests** in appropriate directories:
|
||||
|
||||
#### API Tests (`tests/api/profile.spec.ts`):
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
import { test, expect } from '@playwright/test';
|
||||
|
||||
test.describe('Profile API', () => {
|
||||
test('should fetch user profile', async ({ request }) => {
|
||||
const response = await request.get('/api/profile');
|
||||
|
||||
expect(response.status()).toBe(200);
|
||||
const profile = await response.json();
|
||||
expect(profile).toHaveProperty('name');
|
||||
expect(profile).toHaveProperty('email');
|
||||
expect(profile).toHaveProperty('avatarUrl');
|
||||
});
|
||||
|
||||
test('should update user profile', async ({ request }) => {
|
||||
const response = await request.patch('/api/profile', {
|
||||
data: {
|
||||
name: 'Updated Name',
|
||||
email: 'updated@example.com'
|
||||
}
|
||||
});
|
||||
|
||||
expect(response.status()).toBe(200);
|
||||
const updated = await response.json();
|
||||
expect(updated.name).toBe('Updated Name');
|
||||
expect(updated.email).toBe('updated@example.com');
|
||||
});
|
||||
|
||||
test('should validate email format', async ({ request }) => {
|
||||
const response = await request.patch('/api/profile', {
|
||||
data: {
|
||||
email: 'invalid-email'
|
||||
}
|
||||
});
|
||||
|
||||
expect(response.status()).toBe(400);
|
||||
const error = await response.json();
|
||||
expect(error.message).toContain('Invalid email format');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { expect } from '@playwright/test';
|
||||
import { z } from 'zod';
|
||||
|
||||
const ProfileSchema = z.object({
|
||||
name: z.string(),
|
||||
email: z.string().email(),
|
||||
avatarUrl: z.string().url()
|
||||
});
|
||||
|
||||
test.describe('Profile API', () => {
|
||||
test('should fetch user profile', async ({ apiRequest }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/profile'
|
||||
}).validateSchema(ProfileSchema); // Chained validation
|
||||
|
||||
expect(status).toBe(200);
|
||||
// Schema already validated, type-safe access
|
||||
expect(body.name).toBeDefined();
|
||||
expect(body.email).toContain('@');
|
||||
});
|
||||
|
||||
test('should update user profile', async ({ apiRequest }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: '/api/profile',
|
||||
body: {
|
||||
name: 'Updated Name',
|
||||
email: 'updated@example.com'
|
||||
}
|
||||
}).validateSchema(ProfileSchema); // Chained validation
|
||||
|
||||
expect(status).toBe(200);
|
||||
expect(body.name).toBe('Updated Name');
|
||||
expect(body.email).toBe('updated@example.com');
|
||||
});
|
||||
|
||||
test('should validate email format', async ({ apiRequest }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: '/api/profile',
|
||||
body: { email: 'invalid-email' }
|
||||
});
|
||||
|
||||
expect(status).toBe(400);
|
||||
expect(body.message).toContain('Invalid email format');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**Key Benefits:**
|
||||
- Returns `{ status, body }` (cleaner than `response.status()` + `await response.json()`)
|
||||
- Automatic schema validation with Zod
|
||||
- Type-safe response bodies
|
||||
- Automatic retry for 5xx errors
|
||||
- Less boilerplate
|
||||
|
||||
#### E2E Tests (`tests/e2e/profile.spec.ts`):
|
||||
|
||||
```typescript
|
||||
import { test, expect } from '@playwright/test';
|
||||
|
||||
test('should edit and save profile', async ({ page }) => {
|
||||
// Login first
|
||||
await page.goto('/login');
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByLabel('Password').fill('password123');
|
||||
await page.getByRole('button', { name: 'Sign in' }).click();
|
||||
|
||||
// Navigate to profile
|
||||
await page.goto('/profile');
|
||||
|
||||
// Edit profile
|
||||
await page.getByRole('button', { name: 'Edit Profile' }).click();
|
||||
await page.getByLabel('Name').fill('Updated Name');
|
||||
await page.getByRole('button', { name: 'Save' }).click();
|
||||
|
||||
// Verify success
|
||||
await expect(page.getByText('Profile updated')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
TEA generates additional E2E tests for display, validation errors, etc. based on acceptance criteria.
|
||||
|
||||
#### Implementation Checklist
|
||||
|
||||
TEA also provides an implementation checklist:
|
||||
|
||||
```markdown
|
||||
## Implementation Checklist
|
||||
|
||||
### Backend
|
||||
- [ ] Create `GET /api/profile` endpoint
|
||||
- [ ] Create `PATCH /api/profile` endpoint
|
||||
- [ ] Add email validation middleware
|
||||
- [ ] Add profile picture upload handling
|
||||
- [ ] Write API unit tests
|
||||
|
||||
### Frontend
|
||||
- [ ] Create ProfilePage component
|
||||
- [ ] Implement profile form with validation
|
||||
- [ ] Add file upload for avatar
|
||||
- [ ] Handle API errors gracefully
|
||||
- [ ] Add loading states
|
||||
|
||||
### Tests
|
||||
- [x] API tests generated (failing)
|
||||
- [x] E2E tests generated (failing)
|
||||
- [ ] Run tests after implementation (should pass)
|
||||
```
|
||||
|
||||
### 6. Verify Tests Fail
|
||||
|
||||
This is the TDD red phase - tests MUST fail before implementation.
|
||||
|
||||
**For Playwright:**
|
||||
```bash
|
||||
npx playwright test
|
||||
```
|
||||
|
||||
**For Cypress:**
|
||||
```bash
|
||||
npx cypress run
|
||||
```
|
||||
|
||||
Expected output:
|
||||
```
|
||||
Running 6 tests using 1 worker
|
||||
|
||||
✗ tests/api/profile.spec.ts:3:3 › should fetch user profile
|
||||
Error: expect(received).toBe(expected)
|
||||
Expected: 200
|
||||
Received: 404
|
||||
|
||||
✗ tests/e2e/profile.spec.ts:10:3 › should display current profile information
|
||||
Error: page.goto: net::ERR_ABORTED
|
||||
```
|
||||
|
||||
**All tests should fail!** This confirms:
|
||||
- Feature doesn't exist yet
|
||||
- Tests will guide implementation
|
||||
- You have clear success criteria
|
||||
|
||||
### 7. Implement the Feature
|
||||
|
||||
Now implement the feature following the test guidance:
|
||||
|
||||
1. Start with API tests (backend first)
|
||||
2. Make API tests pass
|
||||
3. Move to E2E tests (frontend)
|
||||
4. Make E2E tests pass
|
||||
5. Refactor with confidence (tests protect you)
|
||||
|
||||
### 8. Verify Tests Pass
|
||||
|
||||
After implementation, run your test suite.
|
||||
|
||||
**For Playwright:**
|
||||
```bash
|
||||
npx playwright test
|
||||
```
|
||||
|
||||
**For Cypress:**
|
||||
```bash
|
||||
npx cypress run
|
||||
```
|
||||
|
||||
Expected output:
|
||||
```
|
||||
Running 6 tests using 1 worker
|
||||
|
||||
✓ tests/api/profile.spec.ts:3:3 › should fetch user profile (850ms)
|
||||
✓ tests/api/profile.spec.ts:15:3 › should update user profile (1.2s)
|
||||
✓ tests/api/profile.spec.ts:30:3 › should validate email format (650ms)
|
||||
✓ tests/e2e/profile.spec.ts:10:3 › should display current profile (2.1s)
|
||||
✓ tests/e2e/profile.spec.ts:18:3 › should edit and save profile (3.2s)
|
||||
✓ tests/e2e/profile.spec.ts:35:3 › should show validation error (1.8s)
|
||||
|
||||
6 passed (9.8s)
|
||||
```
|
||||
|
||||
**Green!** You've completed the TDD cycle: red → green → refactor.
|
||||
|
||||
## What You Get
|
||||
|
||||
### Failing Tests
|
||||
- API tests for backend endpoints
|
||||
- E2E tests for user workflows
|
||||
- Component tests (if requested)
|
||||
- All tests fail initially (red phase)
|
||||
|
||||
### Implementation Guidance
|
||||
- Clear checklist of what to build
|
||||
- Acceptance criteria translated to assertions
|
||||
- Edge cases and error scenarios identified
|
||||
|
||||
### TDD Workflow Support
|
||||
- Tests guide implementation
|
||||
- Confidence to refactor
|
||||
- Living documentation of features
|
||||
|
||||
## Tips
|
||||
|
||||
### Start with Test Design
|
||||
|
||||
Run `*test-design` before `*atdd` for better results:
|
||||
|
||||
```
|
||||
*test-design # Risk assessment and priorities
|
||||
*atdd # Generate tests based on design
|
||||
```
|
||||
|
||||
### MCP Enhancements (Optional)
|
||||
|
||||
If you have MCP servers configured (`tea_use_mcp_enhancements: true`), TEA can use them during `*atdd`.
|
||||
|
||||
**Note:** ATDD is for features that don't exist yet, so recording mode (verify selectors with live UI) only applies if you have skeleton/mockup UI already implemented. For typical ATDD (no UI yet), TEA infers selectors from best practices.
|
||||
|
||||
See [Enable MCP Enhancements](/docs/how-to/customization/enable-tea-mcp-enhancements.md) for setup.
|
||||
|
||||
### Focus on P0/P1 Scenarios
|
||||
|
||||
Don't generate tests for everything at once:
|
||||
|
||||
```
|
||||
Generate tests for:
|
||||
- P0: Critical path (happy path)
|
||||
- P1: High value (validation, errors)
|
||||
|
||||
Skip P2/P3 for now - add later with *automate
|
||||
```
|
||||
|
||||
### API Tests First, E2E Later
|
||||
|
||||
Recommended order:
|
||||
1. Generate API tests with `*atdd`
|
||||
2. Implement backend (make API tests pass)
|
||||
3. Generate E2E tests with `*atdd` (or `*automate`)
|
||||
4. Implement frontend (make E2E tests pass)
|
||||
|
||||
This "outside-in" approach is faster and more reliable.
|
||||
|
||||
### Keep Tests Deterministic
|
||||
|
||||
TEA generates deterministic tests by default:
|
||||
- No hard waits (`waitForTimeout`)
|
||||
- Network-first patterns (wait for responses)
|
||||
- Explicit assertions (no conditionals)
|
||||
|
||||
Don't modify these patterns - they prevent flakiness!
|
||||
|
||||
## Related Guides
|
||||
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Plan before generating
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Tests for existing features
|
||||
- [How to Set Up Test Framework](/docs/how-to/workflows/setup-test-framework.md) - Initial setup
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - **Why TEA generates quality tests** (foundational)
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Why P0 vs P3 matters
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - What makes tests good
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Avoiding flakiness
|
||||
|
||||
## Reference
|
||||
|
||||
- [Command: *atdd](/docs/reference/tea/commands.md#atdd) - Full command reference
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - MCP and Playwright Utils options
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
653
docs/how-to/workflows/run-automate.md
Normal file
653
docs/how-to/workflows/run-automate.md
Normal file
@@ -0,0 +1,653 @@
|
||||
---
|
||||
title: "How to Run Automate with TEA"
|
||||
description: Expand test automation coverage after implementation using TEA's automate workflow
|
||||
---
|
||||
|
||||
# How to Run Automate with TEA
|
||||
|
||||
Use TEA's `*automate` workflow to generate comprehensive tests for existing features. Unlike `*atdd`, these tests pass immediately because the feature already exists.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Feature already exists and works
|
||||
- Want to add test coverage to existing code
|
||||
- Need tests that pass immediately
|
||||
- Expanding existing test suite
|
||||
- Adding tests to legacy code
|
||||
|
||||
**Don't use this if:**
|
||||
- Feature doesn't exist yet (use `*atdd` instead)
|
||||
- Want failing tests to guide development (use `*atdd` for TDD)
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Test framework setup complete (run `*framework` if needed)
|
||||
- Feature implemented and working
|
||||
|
||||
**Note:** This guide uses Playwright examples. If using Cypress, commands and syntax will differ.
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load TEA Agent
|
||||
|
||||
Start a fresh chat and load TEA:
|
||||
|
||||
```
|
||||
*tea
|
||||
```
|
||||
|
||||
### 2. Run the Automate Workflow
|
||||
|
||||
```
|
||||
*automate
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
TEA will ask for context about what you're testing.
|
||||
|
||||
#### Option A: BMad-Integrated Mode (Recommended)
|
||||
|
||||
If you have BMad artifacts (stories, test designs, PRDs):
|
||||
|
||||
**What are you testing?**
|
||||
```
|
||||
I'm testing the user profile feature we just implemented.
|
||||
Story: story-profile-management.md
|
||||
Test Design: test-design-epic-1.md
|
||||
```
|
||||
|
||||
**Reference documents:**
|
||||
- Story file with acceptance criteria
|
||||
- Test design document (if available)
|
||||
- PRD sections relevant to this feature
|
||||
- Tech spec (if available)
|
||||
|
||||
**Existing tests:**
|
||||
```
|
||||
We have basic tests in tests/e2e/profile-view.spec.ts
|
||||
Avoid duplicating that coverage
|
||||
```
|
||||
|
||||
TEA will analyze your artifacts and generate comprehensive tests that:
|
||||
- Cover acceptance criteria from the story
|
||||
- Follow priorities from test design (P0 → P1 → P2)
|
||||
- Avoid duplicating existing tests
|
||||
- Include edge cases and error scenarios
|
||||
|
||||
#### Option B: Standalone Mode
|
||||
|
||||
If you're using TEA Solo or don't have BMad artifacts:
|
||||
|
||||
**What are you testing?**
|
||||
```
|
||||
TodoMVC React application at https://todomvc.com/examples/react/
|
||||
Features: Create todos, mark as complete, filter by status, delete todos
|
||||
```
|
||||
|
||||
**Specific scenarios to cover:**
|
||||
```
|
||||
- Creating todos (happy path)
|
||||
- Marking todos as complete/incomplete
|
||||
- Filtering (All, Active, Completed)
|
||||
- Deleting todos
|
||||
- Edge cases (empty input, long text)
|
||||
```
|
||||
|
||||
TEA will analyze the application and generate tests based on your description.
|
||||
|
||||
### 4. Specify Test Levels
|
||||
|
||||
TEA will ask which test levels to generate:
|
||||
|
||||
**Options:**
|
||||
- **E2E tests** - Full browser-based user workflows
|
||||
- **API tests** - Backend endpoint testing (faster, more reliable)
|
||||
- **Component tests** - UI component testing in isolation (framework-dependent)
|
||||
- **Mix** - Combination of levels (recommended)
|
||||
|
||||
**Example response:**
|
||||
```
|
||||
Generate:
|
||||
- API tests for all CRUD operations
|
||||
- E2E tests for critical user workflows (P0)
|
||||
- Focus on P0 and P1 scenarios
|
||||
- Skip P3 (low priority edge cases)
|
||||
```
|
||||
|
||||
### 5. Review Generated Tests
|
||||
|
||||
TEA generates a comprehensive test suite with multiple test levels.
|
||||
|
||||
#### API Tests (`tests/api/profile.spec.ts`):
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
import { test, expect } from '@playwright/test';
|
||||
|
||||
test.describe('Profile API', () => {
|
||||
let authToken: string;
|
||||
|
||||
test.beforeAll(async ({ request }) => {
|
||||
// Manual auth token fetch
|
||||
const response = await request.post('/api/auth/login', {
|
||||
data: { email: 'test@example.com', password: 'password123' }
|
||||
});
|
||||
const { token } = await response.json();
|
||||
authToken = token;
|
||||
});
|
||||
|
||||
test('should fetch user profile', async ({ request }) => {
|
||||
const response = await request.get('/api/profile', {
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
expect(response.ok()).toBeTruthy();
|
||||
const profile = await response.json();
|
||||
expect(profile).toMatchObject({
|
||||
id: expect.any(String),
|
||||
name: expect.any(String),
|
||||
email: expect.any(String)
|
||||
});
|
||||
});
|
||||
|
||||
test('should update profile successfully', async ({ request }) => {
|
||||
const response = await request.patch('/api/profile', {
|
||||
headers: { Authorization: `Bearer ${authToken}` },
|
||||
data: {
|
||||
name: 'Updated Name',
|
||||
bio: 'Test bio'
|
||||
}
|
||||
});
|
||||
|
||||
expect(response.ok()).toBeTruthy();
|
||||
const updated = await response.json();
|
||||
expect(updated.name).toBe('Updated Name');
|
||||
expect(updated.bio).toBe('Test bio');
|
||||
});
|
||||
|
||||
test('should validate email format', async ({ request }) => {
|
||||
const response = await request.patch('/api/profile', {
|
||||
headers: { Authorization: `Bearer ${authToken}` },
|
||||
data: { email: 'invalid-email' }
|
||||
});
|
||||
|
||||
expect(response.status()).toBe(400);
|
||||
const error = await response.json();
|
||||
expect(error.message).toContain('Invalid email');
|
||||
});
|
||||
|
||||
test('should require authentication', async ({ request }) => {
|
||||
const response = await request.get('/api/profile');
|
||||
expect(response.status()).toBe(401);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test as base, expect } from '@playwright/test';
|
||||
import { test as apiRequestFixture } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { createAuthFixtures } from '@seontechnologies/playwright-utils/auth-session';
|
||||
import { mergeTests } from '@playwright/test';
|
||||
import { z } from 'zod';
|
||||
|
||||
const ProfileSchema = z.object({
|
||||
id: z.string(),
|
||||
name: z.string(),
|
||||
email: z.string().email()
|
||||
});
|
||||
|
||||
// Merge API and auth fixtures
|
||||
const authFixtureTest = base.extend(createAuthFixtures());
|
||||
export const testWithAuth = mergeTests(apiRequestFixture, authFixtureTest);
|
||||
|
||||
testWithAuth.describe('Profile API', () => {
|
||||
testWithAuth('should fetch user profile', async ({ apiRequest, authToken }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'GET',
|
||||
path: '/api/profile',
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
}).validateSchema(ProfileSchema); // Chained validation
|
||||
|
||||
expect(status).toBe(200);
|
||||
// Schema already validated, type-safe access
|
||||
expect(body.name).toBeDefined();
|
||||
});
|
||||
|
||||
testWithAuth('should update profile successfully', async ({ apiRequest, authToken }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: '/api/profile',
|
||||
body: { name: 'Updated Name', bio: 'Test bio' },
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
}).validateSchema(ProfileSchema); // Chained validation
|
||||
|
||||
expect(status).toBe(200);
|
||||
expect(body.name).toBe('Updated Name');
|
||||
});
|
||||
|
||||
testWithAuth('should validate email format', async ({ apiRequest, authToken }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: '/api/profile',
|
||||
body: { email: 'invalid-email' },
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
expect(status).toBe(400);
|
||||
expect(body.message).toContain('Invalid email');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**Key Differences:**
|
||||
- `authToken` fixture (persisted, reused across tests)
|
||||
- `apiRequest` returns `{ status, body }` (cleaner)
|
||||
- Schema validation with Zod (type-safe)
|
||||
- Automatic retry for 5xx errors
|
||||
- Less boilerplate (no manual `await response.json()` everywhere)
|
||||
|
||||
#### E2E Tests (`tests/e2e/profile.spec.ts`):
|
||||
|
||||
```typescript
|
||||
import { test, expect } from '@playwright/test';
|
||||
|
||||
test('should edit profile', async ({ page }) => {
|
||||
// Login
|
||||
await page.goto('/login');
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByLabel('Password').fill('password123');
|
||||
await page.getByRole('button', { name: 'Sign in' }).click();
|
||||
|
||||
// Edit profile
|
||||
await page.goto('/profile');
|
||||
await page.getByRole('button', { name: 'Edit Profile' }).click();
|
||||
await page.getByLabel('Name').fill('New Name');
|
||||
await page.getByRole('button', { name: 'Save' }).click();
|
||||
|
||||
// Verify success
|
||||
await expect(page.getByText('Profile updated')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
TEA generates additional tests for validation, edge cases, etc. based on priorities.
|
||||
|
||||
#### Fixtures (`tests/support/fixtures/profile.ts`):
|
||||
|
||||
**Vanilla Playwright:**
|
||||
```typescript
|
||||
import { test as base, Page } from '@playwright/test';
|
||||
|
||||
type ProfileFixtures = {
|
||||
authenticatedPage: Page;
|
||||
testProfile: {
|
||||
name: string;
|
||||
email: string;
|
||||
bio: string;
|
||||
};
|
||||
};
|
||||
|
||||
export const test = base.extend<ProfileFixtures>({
|
||||
authenticatedPage: async ({ page }, use) => {
|
||||
// Manual login flow
|
||||
await page.goto('/login');
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByLabel('Password').fill('password123');
|
||||
await page.getByRole('button', { name: 'Sign in' }).click();
|
||||
await page.waitForURL(/\/dashboard/);
|
||||
|
||||
await use(page);
|
||||
},
|
||||
|
||||
testProfile: async ({ request }, use) => {
|
||||
// Static test data
|
||||
const profile = {
|
||||
name: 'Test User',
|
||||
email: 'test@example.com',
|
||||
bio: 'Test bio'
|
||||
};
|
||||
|
||||
await use(profile);
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
import { test as base } from '@playwright/test';
|
||||
import { createAuthFixtures } from '@seontechnologies/playwright-utils/auth-session';
|
||||
import { mergeTests } from '@playwright/test';
|
||||
import { faker } from '@faker-js/faker';
|
||||
|
||||
type ProfileFixtures = {
|
||||
testProfile: {
|
||||
name: string;
|
||||
email: string;
|
||||
bio: string;
|
||||
};
|
||||
};
|
||||
|
||||
// Merge auth fixtures with custom fixtures
|
||||
const authTest = base.extend(createAuthFixtures());
|
||||
const profileTest = base.extend<ProfileFixtures>({
|
||||
testProfile: async ({}, use) => {
|
||||
// Dynamic test data with faker
|
||||
const profile = {
|
||||
name: faker.person.fullName(),
|
||||
email: faker.internet.email(),
|
||||
bio: faker.person.bio()
|
||||
};
|
||||
|
||||
await use(profile);
|
||||
}
|
||||
});
|
||||
|
||||
export const test = mergeTests(authTest, profileTest);
|
||||
export { expect } from '@playwright/test';
|
||||
```
|
||||
|
||||
**Usage:**
|
||||
```typescript
|
||||
import { test, expect } from '../support/fixtures/profile';
|
||||
|
||||
test('should update profile', async ({ page, authToken, testProfile }) => {
|
||||
// authToken from auth-session (automatic, persisted)
|
||||
// testProfile from custom fixture (dynamic data)
|
||||
|
||||
await page.goto('/profile');
|
||||
// Test with dynamic, unique data
|
||||
});
|
||||
```
|
||||
|
||||
**Key Benefits:**
|
||||
- `authToken` fixture (persisted token, no manual login)
|
||||
- Dynamic test data with faker (no conflicts)
|
||||
- Fixture composition with mergeTests
|
||||
- Reusable across test files
|
||||
|
||||
### 6. Review Additional Artifacts
|
||||
|
||||
TEA also generates:
|
||||
|
||||
#### Updated README (`tests/README.md`):
|
||||
|
||||
```markdown
|
||||
# Test Suite
|
||||
|
||||
## Running Tests
|
||||
|
||||
### All Tests
|
||||
npm test
|
||||
|
||||
### Specific Levels
|
||||
npm run test:api # API tests only
|
||||
npm run test:e2e # E2E tests only
|
||||
npm run test:smoke # Smoke tests (@smoke tag)
|
||||
|
||||
### Single File
|
||||
npx playwright test tests/api/profile.spec.ts
|
||||
|
||||
## Test Structure
|
||||
|
||||
tests/
|
||||
├── api/ # API tests (fast, reliable)
|
||||
├── e2e/ # E2E tests (full workflows)
|
||||
├── fixtures/ # Shared test utilities
|
||||
└── README.md
|
||||
|
||||
## Writing Tests
|
||||
|
||||
Follow the patterns in existing tests:
|
||||
- Use fixtures for authentication
|
||||
- Network-first patterns (no hard waits)
|
||||
- Explicit assertions
|
||||
- Self-cleaning tests
|
||||
```
|
||||
|
||||
#### Definition of Done Summary:
|
||||
|
||||
```markdown
|
||||
## Test Quality Checklist
|
||||
|
||||
✅ All tests pass on first run
|
||||
✅ No hard waits (waitForTimeout)
|
||||
✅ No conditionals for flow control
|
||||
✅ Assertions are explicit
|
||||
✅ Tests clean up after themselves
|
||||
✅ Tests can run in parallel
|
||||
✅ Execution time < 1.5 minutes per test
|
||||
✅ Test files < 300 lines
|
||||
```
|
||||
|
||||
### 7. Run the Tests
|
||||
|
||||
All tests should pass immediately since the feature exists:
|
||||
|
||||
**For Playwright:**
|
||||
```bash
|
||||
npx playwright test
|
||||
```
|
||||
|
||||
**For Cypress:**
|
||||
```bash
|
||||
npx cypress run
|
||||
```
|
||||
|
||||
Expected output:
|
||||
```
|
||||
Running 15 tests using 4 workers
|
||||
|
||||
✓ tests/api/profile.spec.ts (4 tests) - 2.1s
|
||||
✓ tests/e2e/profile-workflow.spec.ts (2 tests) - 5.3s
|
||||
|
||||
15 passed (7.4s)
|
||||
```
|
||||
|
||||
**All green!** Tests pass because feature already exists.
|
||||
|
||||
### 8. Review Test Coverage
|
||||
|
||||
Check which scenarios are covered:
|
||||
|
||||
```bash
|
||||
# View test report
|
||||
npx playwright show-report
|
||||
|
||||
# Check coverage (if configured)
|
||||
npm run test:coverage
|
||||
```
|
||||
|
||||
Compare against:
|
||||
- Acceptance criteria from story
|
||||
- Test priorities from test design
|
||||
- Edge cases and error scenarios
|
||||
|
||||
## What You Get
|
||||
|
||||
### Comprehensive Test Suite
|
||||
- **API tests** - Fast, reliable backend testing
|
||||
- **E2E tests** - Critical user workflows
|
||||
- **Component tests** - UI component testing (if requested)
|
||||
- **Fixtures** - Shared utilities and setup
|
||||
|
||||
### Component Testing by Framework
|
||||
|
||||
TEA supports component testing using framework-appropriate tools:
|
||||
|
||||
| Your Framework | Component Testing Tool | Tests Location |
|
||||
| -------------- | ------------------------------ | ----------------------------------------- |
|
||||
| **Cypress** | Cypress Component Testing | `tests/component/` |
|
||||
| **Playwright** | Vitest + React Testing Library | `tests/component/` or `src/**/*.test.tsx` |
|
||||
|
||||
**Note:** Component tests use separate tooling from E2E tests:
|
||||
- Cypress users: TEA generates Cypress Component Tests
|
||||
- Playwright users: TEA generates Vitest + React Testing Library tests
|
||||
|
||||
### Quality Features
|
||||
- **Network-first patterns** - Wait for actual responses, not timeouts
|
||||
- **Deterministic tests** - No flakiness, no conditionals
|
||||
- **Self-cleaning** - Tests don't leave test data behind
|
||||
- **Parallel-safe** - Can run all tests concurrently
|
||||
|
||||
### Documentation
|
||||
- **Updated README** - How to run tests
|
||||
- **Test structure explanation** - Where tests live
|
||||
- **Definition of Done** - Quality standards
|
||||
|
||||
## Tips
|
||||
|
||||
### Start with Test Design
|
||||
|
||||
Run `*test-design` before `*automate` for better results:
|
||||
|
||||
```
|
||||
*test-design # Risk assessment, priorities
|
||||
*automate # Generate tests based on priorities
|
||||
```
|
||||
|
||||
TEA will focus on P0/P1 scenarios and skip low-value tests.
|
||||
|
||||
### Prioritize Test Levels
|
||||
|
||||
Not everything needs E2E tests:
|
||||
|
||||
**Good strategy:**
|
||||
```
|
||||
- P0 scenarios: API + E2E tests
|
||||
- P1 scenarios: API tests only
|
||||
- P2 scenarios: API tests (happy path)
|
||||
- P3 scenarios: Skip or add later
|
||||
```
|
||||
|
||||
**Why?**
|
||||
- API tests are 10x faster than E2E
|
||||
- API tests are more reliable (no browser flakiness)
|
||||
- E2E tests reserved for critical user journeys
|
||||
|
||||
### Avoid Duplicate Coverage
|
||||
|
||||
Tell TEA about existing tests:
|
||||
|
||||
```
|
||||
We already have tests in:
|
||||
- tests/e2e/profile-view.spec.ts (viewing profile)
|
||||
- tests/api/auth.spec.ts (authentication)
|
||||
|
||||
Don't duplicate that coverage
|
||||
```
|
||||
|
||||
TEA will analyze existing tests and only generate new scenarios.
|
||||
|
||||
### MCP Enhancements (Optional)
|
||||
|
||||
If you have MCP servers configured (`tea_use_mcp_enhancements: true`), TEA can use them during `*automate` for:
|
||||
|
||||
- **Healing mode:** Fix broken selectors, update assertions, enhance with trace analysis
|
||||
- **Recording mode:** Verify selectors with live browser, capture network requests
|
||||
|
||||
No prompts - TEA uses MCPs automatically when available. See [Enable MCP Enhancements](/docs/how-to/customization/enable-tea-mcp-enhancements.md) for setup.
|
||||
|
||||
### Generate Tests Incrementally
|
||||
|
||||
Don't generate all tests at once:
|
||||
|
||||
**Iteration 1:**
|
||||
```
|
||||
Generate P0 tests only (critical path)
|
||||
Run: *automate
|
||||
```
|
||||
|
||||
**Iteration 2:**
|
||||
```
|
||||
Generate P1 tests (high value scenarios)
|
||||
Run: *automate
|
||||
Tell TEA to avoid P0 coverage
|
||||
```
|
||||
|
||||
**Iteration 3:**
|
||||
```
|
||||
Generate P2 tests (if time permits)
|
||||
Run: *automate
|
||||
```
|
||||
|
||||
This iterative approach:
|
||||
- Provides fast feedback
|
||||
- Allows validation before proceeding
|
||||
- Keeps test generation focused
|
||||
|
||||
## Common Issues
|
||||
|
||||
### Tests Pass But Coverage Is Incomplete
|
||||
|
||||
**Problem:** Tests pass but don't cover all scenarios.
|
||||
|
||||
**Cause:** TEA wasn't given complete context.
|
||||
|
||||
**Solution:** Provide more details:
|
||||
```
|
||||
Generate tests for:
|
||||
- All acceptance criteria in story-profile.md
|
||||
- Error scenarios (validation, authorization)
|
||||
- Edge cases (empty fields, long inputs)
|
||||
```
|
||||
|
||||
### Too Many Tests Generated
|
||||
|
||||
**Problem:** TEA generated 50 tests for a simple feature.
|
||||
|
||||
**Cause:** Didn't specify priorities or scope.
|
||||
|
||||
**Solution:** Be specific:
|
||||
```
|
||||
Generate ONLY:
|
||||
- P0 and P1 scenarios
|
||||
- API tests for all scenarios
|
||||
- E2E tests only for critical workflows
|
||||
- Skip P2/P3 for now
|
||||
```
|
||||
|
||||
### Tests Duplicate Existing Coverage
|
||||
|
||||
**Problem:** New tests cover the same scenarios as existing tests.
|
||||
|
||||
**Cause:** Didn't tell TEA about existing tests.
|
||||
|
||||
**Solution:** Specify existing coverage:
|
||||
```
|
||||
We already have these tests:
|
||||
- tests/api/profile.spec.ts (GET /api/profile)
|
||||
- tests/e2e/profile-view.spec.ts (viewing profile)
|
||||
|
||||
Generate tests for scenarios NOT covered by those files
|
||||
```
|
||||
|
||||
### MCP Enhancements for Better Selectors
|
||||
|
||||
If you have MCP servers configured, TEA verifies selectors against live browser. Otherwise, TEA generates accessible selectors (`getByRole`, `getByLabel`) by default.
|
||||
|
||||
Setup: Answer "Yes" to MCPs in BMad installer + configure MCP servers in your IDE. See [Enable MCP Enhancements](/docs/how-to/customization/enable-tea-mcp-enhancements.md).
|
||||
|
||||
## Related Guides
|
||||
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Plan before generating
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Failing tests before implementation
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Audit generated quality
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Testing as Engineering](/docs/explanation/philosophy/testing-as-engineering.md) - **Why TEA generates quality tests** (foundational)
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Why prioritize P0 over P3
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - What makes tests good
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Reusable test patterns
|
||||
|
||||
## Reference
|
||||
|
||||
- [Command: *automate](/docs/reference/tea/commands.md#automate) - Full command reference
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - MCP and Playwright Utils options
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
73
docs/how-to/workflows/run-brainstorming-session.md
Normal file
73
docs/how-to/workflows/run-brainstorming-session.md
Normal file
@@ -0,0 +1,73 @@
|
||||
---
|
||||
title: "How to Run a Brainstorming Session"
|
||||
description: How to run a brainstorming session using the BMad Method
|
||||
---
|
||||
|
||||
Use the `brainstorm-project` workflow to explore solution approaches through parallel ideation tracks.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Very vague or seed kernel of an idea that needs exploration
|
||||
- Consider alternatives or enhancements to an idea
|
||||
- See your idea from different angles and viewpoints
|
||||
- No idea what you want to build, but want to find some inspiration
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- Analyst agent available
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the Analyst Agent
|
||||
|
||||
Start a fresh chat and load the Analyst agent.
|
||||
|
||||
### 2. Run the Brainstorm Workflow
|
||||
|
||||
```
|
||||
*brainstorm-project
|
||||
```
|
||||
|
||||
### 3. Describe Your Idea
|
||||
|
||||
Tell the agent about your project idea, even if it's vague:
|
||||
|
||||
- "I want to build something that helps developers manage their context"
|
||||
- "I have a game idea about resource management"
|
||||
- "I need a tool for my team but I'm not sure what exactly"
|
||||
|
||||
### 4. Explore the Tracks
|
||||
|
||||
The workflow generates solution approaches through parallel ideation tracks:
|
||||
|
||||
- **Architecture track** — Technical approaches and patterns
|
||||
- **UX track** — User experience possibilities
|
||||
- **Integration track** — How it connects with other systems
|
||||
- **Value track** — Business value and differentiation
|
||||
|
||||
### 5. Evaluate Options
|
||||
|
||||
Review the generated options with rationale for each approach.
|
||||
|
||||
## What You Get
|
||||
|
||||
- Multiple solution approaches with trade-offs
|
||||
- Different architectural options
|
||||
- UX and integration considerations
|
||||
- Clear rationale for each direction
|
||||
|
||||
## Tips
|
||||
|
||||
- **Don't worry about having a fully formed idea** — Vague is fine
|
||||
- **Let the agent guide exploration** — Follow the prompts
|
||||
- **Consider multiple tracks** — Don't settle on the first option
|
||||
- **Use outputs as input for product-brief** — Build on brainstorming results
|
||||
|
||||
## Next Steps
|
||||
|
||||
After brainstorming:
|
||||
|
||||
1. **Research** — Validate ideas with market/technical research
|
||||
2. **Product Brief** — Capture strategic vision
|
||||
3. **PRD** — Move to formal planning
|
||||
89
docs/how-to/workflows/run-code-review.md
Normal file
89
docs/how-to/workflows/run-code-review.md
Normal file
@@ -0,0 +1,89 @@
|
||||
---
|
||||
title: "How to Run Code Review"
|
||||
description: How to run code review for quality assurance
|
||||
---
|
||||
|
||||
Use the `code-review` workflow to perform a thorough quality review of implemented code.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- After dev-story completes implementation
|
||||
- Before marking a story as DONE
|
||||
- Every story goes through code review — no exceptions
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- DEV agent available
|
||||
- Story implementation complete
|
||||
- Tests written and passing
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the DEV Agent
|
||||
|
||||
Start a fresh chat (or continue from dev-story) and load the DEV agent.
|
||||
|
||||
### 2. Run the Workflow
|
||||
|
||||
```
|
||||
*code-review
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
Point the agent to:
|
||||
- The story file
|
||||
- Files changed during implementation
|
||||
- Test files
|
||||
|
||||
### 4. Review Findings
|
||||
|
||||
The agent performs a senior developer code review and reports findings.
|
||||
|
||||
### 5. Address Issues
|
||||
|
||||
If issues are found:
|
||||
1. Fix issues using dev-story
|
||||
2. Re-run tests
|
||||
3. Run code-review again
|
||||
|
||||
## What Gets Reviewed
|
||||
|
||||
| Category | Checks |
|
||||
|----------|--------|
|
||||
| **Code Quality** | Clean code, appropriate abstractions, no code smells, proper error handling |
|
||||
| **Architecture Alignment** | Follows ADRs, consistent with patterns, proper separation of concerns |
|
||||
| **Testing** | Adequate coverage, meaningful tests, edge cases, follows project patterns |
|
||||
| **Security** | No hardcoded secrets, input validation, proper auth, no common vulnerabilities |
|
||||
| **Performance** | No obvious issues, appropriate data structures, efficient queries |
|
||||
|
||||
## Review Outcomes
|
||||
|
||||
**Approved** — Code meets quality standards, tests pass. Mark story as DONE in sprint-status.yaml.
|
||||
|
||||
**Changes Requested** — Issues identified that need fixing. Fix issues in dev-story, then re-run code-review.
|
||||
|
||||
## Quality Gates
|
||||
|
||||
Every story goes through code-review before being marked done. This ensures:
|
||||
|
||||
- Consistent code quality
|
||||
- Architecture adherence
|
||||
- Test coverage
|
||||
- Security review
|
||||
|
||||
## Tips
|
||||
|
||||
- **Don't skip for "simple" changes** — Simple changes can have subtle bugs
|
||||
- **Address all findings** — Not just critical ones
|
||||
- **Use findings as learning opportunities** — Improve over time
|
||||
- **Re-run review after fixes** — Verify issues are resolved
|
||||
|
||||
## Next Steps
|
||||
|
||||
After code review:
|
||||
|
||||
1. **If approved** — Update sprint-status.yaml to mark story DONE
|
||||
2. **If changes requested** — Fix issues and re-run review
|
||||
3. **Move to next story** — Run create-story for the next item
|
||||
125
docs/how-to/workflows/run-implementation-readiness.md
Normal file
125
docs/how-to/workflows/run-implementation-readiness.md
Normal file
@@ -0,0 +1,125 @@
|
||||
---
|
||||
title: "How to Run Implementation Readiness"
|
||||
description: How to validate planning and solutioning before implementation
|
||||
---
|
||||
|
||||
Use the `implementation-readiness` workflow to validate that planning and solutioning are complete and aligned before Phase 4 implementation.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- **Always** before Phase 4 for BMad Method and Enterprise projects
|
||||
- After create-epics-and-stories workflow completes
|
||||
- Before sprint-planning workflow
|
||||
- When stakeholders request readiness check
|
||||
|
||||
## When to Skip This
|
||||
|
||||
- Quick Flow (no solutioning phase)
|
||||
- BMad Method Simple (no gate check required)
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- Architect agent available
|
||||
- PRD, Architecture, and Epics completed
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the Architect Agent
|
||||
|
||||
Start a fresh chat and load the Architect agent.
|
||||
|
||||
### 2. Run the Workflow
|
||||
|
||||
```
|
||||
*implementation-readiness
|
||||
```
|
||||
|
||||
### 3. Let the Agent Validate
|
||||
|
||||
The workflow systematically checks:
|
||||
- PRD completeness
|
||||
- Architecture completeness
|
||||
- Epic/Story completeness
|
||||
- Alignment between all documents
|
||||
|
||||
### 4. Review the Gate Decision
|
||||
|
||||
The agent produces a gate decision with rationale.
|
||||
|
||||
## Gate Decision Outcomes
|
||||
|
||||
| Decision | Meaning | Action |
|
||||
|----------|---------|--------|
|
||||
| **PASS** | All critical criteria met, minor gaps acceptable | Proceed to Phase 4 |
|
||||
| **CONCERNS** | Some criteria not met but not blockers | Proceed with caution, address gaps in parallel |
|
||||
| **FAIL** | Critical gaps or contradictions | BLOCK Phase 4, resolve issues first |
|
||||
|
||||
## What Gets Checked
|
||||
|
||||
**PRD/GDD Completeness:**
|
||||
- Problem statement clear and evidence-based
|
||||
- Success metrics defined
|
||||
- User personas identified
|
||||
- FRs and NFRs complete
|
||||
- Risks and assumptions documented
|
||||
|
||||
**Architecture Completeness:**
|
||||
- System, data, API architecture defined
|
||||
- Key ADRs documented
|
||||
- Security architecture addressed
|
||||
- FR/NFR-specific guidance provided
|
||||
- Standards and conventions defined
|
||||
|
||||
**Epic/Story Completeness:**
|
||||
- All PRD features mapped to stories
|
||||
- Stories have acceptance criteria
|
||||
- Stories prioritized (P0/P1/P2/P3)
|
||||
- Dependencies identified
|
||||
|
||||
**Alignment Checks:**
|
||||
- Architecture addresses all PRD FRs/NFRs
|
||||
- Epics align with architecture decisions
|
||||
- No contradictions between epics
|
||||
- Integration points clear
|
||||
|
||||
## What You Get
|
||||
|
||||
An `implementation-readiness.md` document containing:
|
||||
|
||||
1. **Executive Summary** (PASS/CONCERNS/FAIL)
|
||||
2. **Completeness Assessment** (scores for PRD, Architecture, Epics)
|
||||
3. **Alignment Assessment** (PRD↔Architecture, Architecture↔Epics)
|
||||
4. **Quality Assessment** (story quality, dependencies, risks)
|
||||
5. **Gaps and Recommendations** (critical/minor gaps, remediation)
|
||||
6. **Gate Decision** with rationale
|
||||
7. **Next Steps**
|
||||
|
||||
## Example
|
||||
|
||||
E-commerce platform → CONCERNS
|
||||
|
||||
**Gaps identified:**
|
||||
- Missing security architecture section
|
||||
- Undefined payment gateway
|
||||
|
||||
**Recommendation:**
|
||||
- Complete security section
|
||||
- Add payment gateway ADR
|
||||
|
||||
**Action:** Proceed with caution, address before payment epic.
|
||||
|
||||
## Tips
|
||||
|
||||
- **Run before every Phase 4 start** — It's a valuable checkpoint
|
||||
- **Take FAIL decisions seriously** — Fix issues first
|
||||
- **Use CONCERNS as a checklist** — Track parallel work
|
||||
- **Document why you proceed despite concerns** — Transparency matters
|
||||
|
||||
## Next Steps
|
||||
|
||||
After implementation readiness:
|
||||
|
||||
1. **If PASS** — Run sprint-planning to start Phase 4
|
||||
2. **If CONCERNS** — Proceed with documented gaps to address
|
||||
3. **If FAIL** — Return to relevant workflow to fix issues
|
||||
679
docs/how-to/workflows/run-nfr-assess.md
Normal file
679
docs/how-to/workflows/run-nfr-assess.md
Normal file
@@ -0,0 +1,679 @@
|
||||
---
|
||||
title: "How to Run NFR Assessment with TEA"
|
||||
description: Validate non-functional requirements for security, performance, reliability, and maintainability using TEA
|
||||
---
|
||||
|
||||
# How to Run NFR Assessment with TEA
|
||||
|
||||
Use TEA's `*nfr-assess` workflow to validate non-functional requirements (NFRs) with evidence-based assessment across security, performance, reliability, and maintainability.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Enterprise projects with compliance requirements
|
||||
- Projects with strict NFR thresholds
|
||||
- Before production release
|
||||
- When NFRs are critical to project success
|
||||
- Security or performance is mission-critical
|
||||
|
||||
**Best for:**
|
||||
- Enterprise track projects
|
||||
- Compliance-heavy industries (finance, healthcare, government)
|
||||
- High-traffic applications
|
||||
- Security-critical systems
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- NFRs defined in PRD or requirements doc
|
||||
- Evidence preferred but not required (test results, security scans, performance metrics)
|
||||
|
||||
**Note:** You can run NFR assessment without complete evidence. TEA will mark categories as CONCERNS where evidence is missing and document what's needed.
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Run the NFR Assessment Workflow
|
||||
|
||||
Start a fresh chat and run:
|
||||
|
||||
```
|
||||
*nfr-assess
|
||||
```
|
||||
|
||||
This loads TEA and starts the NFR assessment workflow.
|
||||
|
||||
### 2. Specify NFR Categories
|
||||
|
||||
TEA will ask which NFR categories to assess.
|
||||
|
||||
**Available Categories:**
|
||||
|
||||
| Category | Focus Areas |
|
||||
|----------|-------------|
|
||||
| **Security** | Authentication, authorization, encryption, vulnerabilities, security headers, input validation |
|
||||
| **Performance** | Response time, throughput, resource usage, database queries, frontend load time |
|
||||
| **Reliability** | Error handling, recovery mechanisms, availability, failover, data backup |
|
||||
| **Maintainability** | Code quality, test coverage, technical debt, documentation, dependency health |
|
||||
|
||||
**Example Response:**
|
||||
```
|
||||
Assess:
|
||||
- Security (critical for user data)
|
||||
- Performance (API must be fast)
|
||||
- Reliability (99.9% uptime requirement)
|
||||
|
||||
Skip maintainability for now
|
||||
```
|
||||
|
||||
### 3. Provide NFR Thresholds
|
||||
|
||||
TEA will ask for specific thresholds for each category.
|
||||
|
||||
**Critical Principle: Never guess thresholds.**
|
||||
|
||||
If you don't know the exact requirement, tell TEA to mark as CONCERNS and request clarification from stakeholders.
|
||||
|
||||
#### Security Thresholds
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Requirements:
|
||||
- All endpoints require authentication: YES
|
||||
- Data encrypted at rest: YES (PostgreSQL TDE)
|
||||
- Zero critical vulnerabilities: YES (npm audit)
|
||||
- Input validation on all endpoints: YES (Zod schemas)
|
||||
- Security headers configured: YES (helmet.js)
|
||||
```
|
||||
|
||||
#### Performance Thresholds
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Requirements:
|
||||
- API response time P99: < 200ms
|
||||
- API response time P95: < 150ms
|
||||
- Throughput: > 1000 requests/second
|
||||
- Frontend initial load: < 2 seconds
|
||||
- Database query time P99: < 50ms
|
||||
```
|
||||
|
||||
#### Reliability Thresholds
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Requirements:
|
||||
- Error handling: All endpoints return structured errors
|
||||
- Availability: 99.9% uptime
|
||||
- Recovery time: < 5 minutes (RTO)
|
||||
- Data backup: Daily automated backups
|
||||
- Failover: Automatic with < 30s downtime
|
||||
```
|
||||
|
||||
#### Maintainability Thresholds
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Requirements:
|
||||
- Test coverage: > 80%
|
||||
- Code quality: SonarQube grade A
|
||||
- Documentation: All APIs documented
|
||||
- Dependency age: < 6 months outdated
|
||||
- Technical debt: < 10% of codebase
|
||||
```
|
||||
|
||||
### 4. Provide Evidence
|
||||
|
||||
TEA will ask where to find evidence for each requirement.
|
||||
|
||||
**Evidence Sources:**
|
||||
|
||||
| Category | Evidence Type | Location |
|
||||
|----------|---------------|----------|
|
||||
| Security | Security scan reports | `/reports/security-scan.pdf` |
|
||||
| Security | Vulnerability scan | `npm audit`, `snyk test` results |
|
||||
| Security | Auth test results | Test reports showing auth coverage |
|
||||
| Performance | Load test results | `/reports/k6-load-test.json` |
|
||||
| Performance | APM data | Datadog, New Relic dashboards |
|
||||
| Performance | Lighthouse scores | `/reports/lighthouse.json` |
|
||||
| Reliability | Error rate metrics | Production monitoring dashboards |
|
||||
| Reliability | Uptime data | StatusPage, PagerDuty logs |
|
||||
| Maintainability | Coverage reports | `/reports/coverage/index.html` |
|
||||
| Maintainability | Code quality | SonarQube dashboard |
|
||||
|
||||
**Example Response:**
|
||||
```
|
||||
Evidence:
|
||||
- Security: npm audit results (clean), auth tests 15/15 passing
|
||||
- Performance: k6 load test at /reports/k6-results.json
|
||||
- Reliability: Error rate 0.01% in staging (logs in Datadog)
|
||||
|
||||
Don't have:
|
||||
- Uptime data (new system, no baseline)
|
||||
- Mark as CONCERNS and request monitoring setup
|
||||
```
|
||||
|
||||
### 5. Review NFR Assessment Report
|
||||
|
||||
TEA generates a comprehensive assessment report.
|
||||
|
||||
#### Assessment Report (`nfr-assessment.md`):
|
||||
|
||||
```markdown
|
||||
# Non-Functional Requirements Assessment
|
||||
|
||||
**Date:** 2026-01-13
|
||||
**Epic:** User Profile Management
|
||||
**Release:** v1.2.0
|
||||
**Overall Decision:** CONCERNS ⚠️
|
||||
|
||||
## Executive Summary
|
||||
|
||||
| Category | Status | Critical Issues |
|
||||
|----------|--------|-----------------|
|
||||
| Security | PASS ✅ | 0 |
|
||||
| Performance | CONCERNS ⚠️ | 2 |
|
||||
| Reliability | PASS ✅ | 0 |
|
||||
| Maintainability | PASS ✅ | 0 |
|
||||
|
||||
**Decision Rationale:**
|
||||
Performance metrics below target (P99 latency, throughput). Mitigation plan in place. Security and reliability meet all requirements.
|
||||
|
||||
---
|
||||
|
||||
## Security Assessment
|
||||
|
||||
**Status:** PASS ✅
|
||||
|
||||
### Requirements Met
|
||||
|
||||
| Requirement | Target | Actual | Status |
|
||||
|-------------|--------|--------|--------|
|
||||
| Authentication required | All endpoints | 100% enforced | ✅ |
|
||||
| Data encryption at rest | PostgreSQL TDE | Enabled | ✅ |
|
||||
| Critical vulnerabilities | 0 | 0 | ✅ |
|
||||
| Input validation | All endpoints | Zod schemas on 100% | ✅ |
|
||||
| Security headers | Configured | helmet.js enabled | ✅ |
|
||||
|
||||
### Evidence
|
||||
|
||||
**Security Scan:**
|
||||
```bash
|
||||
$ npm audit
|
||||
found 0 vulnerabilities
|
||||
```
|
||||
|
||||
**Authentication Tests:**
|
||||
- 15/15 auth tests passing
|
||||
- Tested unauthorized access (401 responses)
|
||||
- Token validation working
|
||||
|
||||
**Penetration Testing:**
|
||||
- Report: `/reports/pentest-2026-01.pdf`
|
||||
- Findings: 0 critical, 2 low (addressed)
|
||||
|
||||
**Conclusion:** All security requirements met. No blockers.
|
||||
|
||||
---
|
||||
|
||||
## Performance Assessment
|
||||
|
||||
**Status:** CONCERNS ⚠️
|
||||
|
||||
### Requirements Status
|
||||
|
||||
| Metric | Target | Actual | Status |
|
||||
|--------|--------|--------|--------|
|
||||
| API response P99 | < 200ms | 350ms | ❌ Exceeds |
|
||||
| API response P95 | < 150ms | 180ms | ⚠️ Exceeds |
|
||||
| Throughput | > 1000 rps | 850 rps | ⚠️ Below |
|
||||
| Frontend load | < 2s | 1.8s | ✅ Met |
|
||||
| DB query P99 | < 50ms | 85ms | ❌ Exceeds |
|
||||
|
||||
### Issues Identified
|
||||
|
||||
#### Issue 1: P99 Latency Exceeds Target
|
||||
|
||||
**Measured:** 350ms P99 (target: <200ms)
|
||||
**Root Cause:** Database queries not optimized
|
||||
- Missing indexes on profile queries
|
||||
- N+1 query problem in profile endpoint
|
||||
|
||||
**Impact:** User experience degraded for 1% of requests
|
||||
|
||||
**Mitigation Plan:**
|
||||
- Add composite index on `(user_id, profile_id)` - backend team, 2 days
|
||||
- Refactor profile endpoint to use joins instead of multiple queries - backend team, 3 days
|
||||
- Re-run load tests after optimization - QA team, 1 day
|
||||
|
||||
**Owner:** Backend team lead
|
||||
**Deadline:** Before release (January 20, 2026)
|
||||
|
||||
#### Issue 2: Throughput Below Target
|
||||
|
||||
**Measured:** 850 rps (target: >1000 rps)
|
||||
**Root Cause:** Connection pool size too small
|
||||
- PostgreSQL max_connections = 100 (too low)
|
||||
- No connection pooling in application
|
||||
|
||||
**Impact:** System cannot handle expected traffic
|
||||
|
||||
**Mitigation Plan:**
|
||||
- Increase PostgreSQL max_connections to 500 - DevOps, 1 day
|
||||
- Implement connection pooling with pg-pool - backend team, 2 days
|
||||
- Re-run load tests - QA team, 1 day
|
||||
|
||||
**Owner:** DevOps + Backend team
|
||||
**Deadline:** Before release (January 20, 2026)
|
||||
|
||||
### Evidence
|
||||
|
||||
**Load Testing:**
|
||||
```
|
||||
Tool: k6
|
||||
Duration: 10 minutes
|
||||
Virtual Users: 500 concurrent
|
||||
Report: /reports/k6-load-test.json
|
||||
```
|
||||
|
||||
**Results:**
|
||||
```
|
||||
scenarios: (100.00%) 1 scenario, 500 max VUs, 10m30s max duration
|
||||
✓ http_req_duration..............: avg=250ms min=45ms med=180ms max=2.1s p(90)=280ms p(95)=350ms
|
||||
http_reqs......................: 85000 (850/s)
|
||||
http_req_failed................: 0.1%
|
||||
```
|
||||
|
||||
**APM Data:**
|
||||
- Tool: Datadog
|
||||
- Dashboard: <https://app.datadoghq.com/dashboard/abc123>
|
||||
|
||||
**Conclusion:** Performance issues identified with mitigation plan. Re-assess after optimization.
|
||||
|
||||
---
|
||||
|
||||
## Reliability Assessment
|
||||
|
||||
**Status:** PASS ✅
|
||||
|
||||
### Requirements Met
|
||||
|
||||
| Requirement | Target | Actual | Status |
|
||||
|-------------|--------|--------|--------|
|
||||
| Error handling | Structured errors | 100% endpoints | ✅ |
|
||||
| Availability | 99.9% uptime | 99.95% (staging) | ✅ |
|
||||
| Recovery time | < 5 min (RTO) | 3 min (tested) | ✅ |
|
||||
| Data backup | Daily | Automated daily | ✅ |
|
||||
| Failover | < 30s downtime | 15s (tested) | ✅ |
|
||||
|
||||
### Evidence
|
||||
|
||||
**Error Handling Tests:**
|
||||
- All endpoints return structured JSON errors
|
||||
- Error codes standardized (400, 401, 403, 404, 500)
|
||||
- Error messages user-friendly (no stack traces)
|
||||
|
||||
**Chaos Engineering:**
|
||||
- Tested database failover: 15s downtime ✅
|
||||
- Tested service crash recovery: 3 min ✅
|
||||
- Tested network partition: Graceful degradation ✅
|
||||
|
||||
**Monitoring:**
|
||||
- Staging uptime (30 days): 99.95%
|
||||
- Error rate: 0.01% (target: <0.1%)
|
||||
- P50 availability: 100%
|
||||
|
||||
**Conclusion:** All reliability requirements exceeded. No issues.
|
||||
|
||||
---
|
||||
|
||||
## Maintainability Assessment
|
||||
|
||||
**Status:** PASS ✅
|
||||
|
||||
### Requirements Met
|
||||
|
||||
| Requirement | Target | Actual | Status |
|
||||
|-------------|--------|--------|--------|
|
||||
| Test coverage | > 80% | 85% | ✅ |
|
||||
| Code quality | Grade A | Grade A | ✅ |
|
||||
| Documentation | All APIs | 100% documented | ✅ |
|
||||
| Outdated dependencies | < 6 months | 3 months avg | ✅ |
|
||||
| Technical debt | < 10% | 7% | ✅ |
|
||||
|
||||
### Evidence
|
||||
|
||||
**Test Coverage:**
|
||||
```
|
||||
Statements : 85.2% ( 1205/1414 )
|
||||
Branches : 82.1% ( 412/502 )
|
||||
Functions : 88.5% ( 201/227 )
|
||||
Lines : 85.2% ( 1205/1414 )
|
||||
```
|
||||
|
||||
**Code Quality:**
|
||||
- SonarQube: Grade A
|
||||
- Maintainability rating: A
|
||||
- Technical debt ratio: 7%
|
||||
- Code smells: 12 (all minor)
|
||||
|
||||
**Documentation:**
|
||||
- API docs: 100% coverage (OpenAPI spec)
|
||||
- README: Complete and up-to-date
|
||||
- Architecture docs: ADRs for all major decisions
|
||||
|
||||
**Conclusion:** All maintainability requirements met. Codebase is healthy.
|
||||
|
||||
---
|
||||
|
||||
## Overall Gate Decision
|
||||
|
||||
### Decision: CONCERNS ⚠️
|
||||
|
||||
**Rationale:**
|
||||
- **Blockers:** None
|
||||
- **Concerns:** Performance metrics below target (P99 latency, throughput)
|
||||
- **Mitigation:** Plan in place with clear owners and deadlines (5 days total)
|
||||
- **Passing:** Security, reliability, maintainability all green
|
||||
|
||||
### Actions Required Before Release
|
||||
|
||||
1. **Optimize database queries** (backend team, 3 days)
|
||||
- Add indexes
|
||||
- Fix N+1 queries
|
||||
- Implement connection pooling
|
||||
|
||||
2. **Re-run performance tests** (QA team, 1 day)
|
||||
- Validate P99 < 200ms
|
||||
- Validate throughput > 1000 rps
|
||||
|
||||
3. **Update this assessment** (TEA, 1 hour)
|
||||
- Re-run `*nfr-assess` with new results
|
||||
- Confirm PASS status
|
||||
|
||||
### Waiver Option (If Business Approves)
|
||||
|
||||
If business decides to deploy with current performance:
|
||||
|
||||
**Waiver Justification:**
|
||||
```markdown
|
||||
## Performance Waiver
|
||||
|
||||
**Waived By:** VP Engineering, Product Manager
|
||||
**Date:** 2026-01-15
|
||||
**Reason:** Business priority to launch by Q1
|
||||
**Conditions:**
|
||||
- Set monitoring alerts for P99 > 300ms
|
||||
- Plan optimization for v1.3 (February release)
|
||||
- Document known performance limitations in release notes
|
||||
|
||||
**Accepted Risk:**
|
||||
- 1% of users experience slower response (350ms vs 200ms)
|
||||
- System can handle current traffic (850 rps sufficient for launch)
|
||||
- Optimization planned for next release
|
||||
```
|
||||
|
||||
### Approvals
|
||||
|
||||
- [ ] Product Manager - Review business impact
|
||||
- [ ] Tech Lead - Review mitigation plan
|
||||
- [ ] QA Lead - Validate test evidence
|
||||
- [ ] DevOps - Confirm infrastructure ready
|
||||
|
||||
---
|
||||
|
||||
## Monitoring Plan Post-Release
|
||||
|
||||
**Performance Alerts:**
|
||||
- P99 latency > 400ms (critical)
|
||||
- Throughput < 700 rps (warning)
|
||||
- Error rate > 1% (critical)
|
||||
|
||||
**Review Cadence:**
|
||||
- Daily: Check performance dashboards
|
||||
- Weekly: Review alert trends
|
||||
- Monthly: Re-assess NFRs
|
||||
```
|
||||
|
||||
## What You Get
|
||||
|
||||
### NFR Assessment Report
|
||||
- Category-by-category analysis (Security, Performance, Reliability, Maintainability)
|
||||
- Requirements status (target vs actual)
|
||||
- Evidence for each requirement
|
||||
- Issues identified with root cause analysis
|
||||
|
||||
### Gate Decision
|
||||
- **PASS** ✅ - All NFRs met, ready to release
|
||||
- **CONCERNS** ⚠️ - Some NFRs not met, mitigation plan exists
|
||||
- **FAIL** ❌ - Critical NFRs not met, blocks release
|
||||
- **WAIVED** ⏭️ - Business-approved waiver with documented risk
|
||||
|
||||
### Mitigation Plans
|
||||
- Specific actions to address concerns
|
||||
- Owners and deadlines
|
||||
- Re-assessment criteria
|
||||
|
||||
### Monitoring Plan
|
||||
- Post-release monitoring strategy
|
||||
- Alert thresholds
|
||||
- Review cadence
|
||||
|
||||
## Tips
|
||||
|
||||
### Run NFR Assessment Early
|
||||
|
||||
**Phase 2 (Enterprise):**
|
||||
Run `*nfr-assess` during planning to:
|
||||
- Identify NFR requirements early
|
||||
- Plan for performance testing
|
||||
- Budget for security audits
|
||||
- Set up monitoring infrastructure
|
||||
|
||||
**Phase 4 or Gate:**
|
||||
Re-run before release to validate all requirements met.
|
||||
|
||||
### Never Guess Thresholds
|
||||
|
||||
If you don't know the NFR target:
|
||||
|
||||
**Don't:**
|
||||
```
|
||||
API response time should probably be under 500ms
|
||||
```
|
||||
|
||||
**Do:**
|
||||
```
|
||||
Mark as CONCERNS - Request threshold from stakeholders
|
||||
"What is the acceptable API response time?"
|
||||
```
|
||||
|
||||
### Collect Evidence Beforehand
|
||||
|
||||
Before running `*nfr-assess`, gather:
|
||||
|
||||
**Security:**
|
||||
```bash
|
||||
npm audit # Vulnerability scan
|
||||
snyk test # Alternative security scan
|
||||
npm run test:security # Security test suite
|
||||
```
|
||||
|
||||
**Performance:**
|
||||
```bash
|
||||
npm run test:load # k6 or artillery load tests
|
||||
npm run test:lighthouse # Frontend performance
|
||||
npm run test:db-performance # Database query analysis
|
||||
```
|
||||
|
||||
**Reliability:**
|
||||
- Production error rate (last 30 days)
|
||||
- Uptime data (StatusPage, PagerDuty)
|
||||
- Incident response times
|
||||
|
||||
**Maintainability:**
|
||||
```bash
|
||||
npm run test:coverage # Test coverage report
|
||||
npm run lint # Code quality check
|
||||
npm outdated # Dependency freshness
|
||||
```
|
||||
|
||||
### Use Real Data, Not Assumptions
|
||||
|
||||
**Don't:**
|
||||
```
|
||||
System is probably fast enough
|
||||
Security seems fine
|
||||
```
|
||||
|
||||
**Do:**
|
||||
```
|
||||
Load test results show P99 = 350ms
|
||||
npm audit shows 0 vulnerabilities
|
||||
Test coverage report shows 85%
|
||||
```
|
||||
|
||||
Evidence-based decisions prevent surprises in production.
|
||||
|
||||
### Document Waivers Thoroughly
|
||||
|
||||
If business approves waiver:
|
||||
|
||||
**Required:**
|
||||
- Who approved (name, role, date)
|
||||
- Why (business justification)
|
||||
- Conditions (monitoring, future plans)
|
||||
- Accepted risk (quantified impact)
|
||||
|
||||
**Example:**
|
||||
```markdown
|
||||
Waived by: CTO, VP Product (2026-01-15)
|
||||
Reason: Q1 launch critical for investor demo
|
||||
Conditions: Optimize in v1.3, monitor closely
|
||||
Risk: 1% of users experience 350ms latency (acceptable for launch)
|
||||
```
|
||||
|
||||
### Re-Assess After Fixes
|
||||
|
||||
After implementing mitigations:
|
||||
|
||||
```
|
||||
1. Fix performance issues
|
||||
2. Run load tests again
|
||||
3. Run *nfr-assess with new evidence
|
||||
4. Verify PASS status
|
||||
```
|
||||
|
||||
Don't deploy with CONCERNS without mitigation or waiver.
|
||||
|
||||
### Integrate with Release Checklist
|
||||
|
||||
```markdown
|
||||
## Release Checklist
|
||||
|
||||
### Pre-Release
|
||||
- [ ] All tests passing
|
||||
- [ ] Test coverage > 80%
|
||||
- [ ] Run *nfr-assess
|
||||
- [ ] NFR status: PASS or WAIVED
|
||||
|
||||
### Performance
|
||||
- [ ] Load tests completed
|
||||
- [ ] P99 latency meets threshold
|
||||
- [ ] Throughput meets threshold
|
||||
|
||||
### Security
|
||||
- [ ] Security scan clean
|
||||
- [ ] Auth tests passing
|
||||
- [ ] Penetration test complete
|
||||
|
||||
### Post-Release
|
||||
- [ ] Monitoring alerts configured
|
||||
- [ ] Dashboards updated
|
||||
- [ ] Incident response plan ready
|
||||
```
|
||||
|
||||
## Common Issues
|
||||
|
||||
### No Evidence Available
|
||||
|
||||
**Problem:** Don't have performance data, security scans, etc.
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
Mark as CONCERNS for categories without evidence
|
||||
Document what evidence is needed
|
||||
Set up tests/scans before re-assessment
|
||||
```
|
||||
|
||||
**Don't block on missing evidence** - document what's needed and proceed.
|
||||
|
||||
### Thresholds Too Strict
|
||||
|
||||
**Problem:** Can't meet unrealistic thresholds.
|
||||
|
||||
**Symptoms:**
|
||||
- P99 < 50ms (impossible for complex queries)
|
||||
- 100% test coverage (impractical)
|
||||
- Zero technical debt (unrealistic)
|
||||
|
||||
**Solution:**
|
||||
```
|
||||
Negotiate thresholds with stakeholders:
|
||||
- "P99 < 50ms is unrealistic for our DB queries"
|
||||
- "Propose P99 < 200ms based on industry standards"
|
||||
- "Show evidence from load tests"
|
||||
```
|
||||
|
||||
Use data to negotiate realistic requirements.
|
||||
|
||||
### Assessment Takes Too Long
|
||||
|
||||
**Problem:** Gathering evidence for all categories is time-consuming.
|
||||
|
||||
**Solution:** Focus on critical categories first:
|
||||
|
||||
**For most projects:**
|
||||
```
|
||||
Priority 1: Security (always critical)
|
||||
Priority 2: Performance (if high-traffic)
|
||||
Priority 3: Reliability (if uptime critical)
|
||||
Priority 4: Maintainability (nice to have)
|
||||
```
|
||||
|
||||
Assess categories incrementally, not all at once.
|
||||
|
||||
### CONCERNS vs FAIL - When to Block?
|
||||
|
||||
**CONCERNS** ⚠️:
|
||||
- Issues exist but not critical
|
||||
- Mitigation plan in place
|
||||
- Business accepts risk (with waiver)
|
||||
- Can deploy with monitoring
|
||||
|
||||
**FAIL** ❌:
|
||||
- Critical security vulnerability (CVE critical)
|
||||
- System unusable (error rate >10%)
|
||||
- Data loss risk (no backups)
|
||||
- Zero mitigation possible
|
||||
|
||||
**Rule of thumb:** If you can mitigate or monitor, use CONCERNS. Reserve FAIL for absolute blockers.
|
||||
|
||||
## Related Guides
|
||||
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md) - Gate decision complements NFR
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Quality complements NFR
|
||||
- [Run TEA for Enterprise](/docs/how-to/enterprise/use-tea-for-enterprise.md) - Enterprise workflow
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Risk assessment principles
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - NFR in release gates
|
||||
|
||||
## Reference
|
||||
|
||||
- [Command: *nfr-assess](/docs/reference/tea/commands.md#nfr-assess) - Full command reference
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Enterprise config options
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
94
docs/how-to/workflows/run-sprint-planning.md
Normal file
94
docs/how-to/workflows/run-sprint-planning.md
Normal file
@@ -0,0 +1,94 @@
|
||||
---
|
||||
title: "How to Run Sprint Planning"
|
||||
description: How to initialize sprint tracking for implementation
|
||||
---
|
||||
|
||||
Use the `sprint-planning` workflow to initialize the sprint tracking file and organize work for implementation.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Once at the start of Phase 4 (Implementation)
|
||||
- After implementation-readiness gate passes
|
||||
- When starting a new sprint cycle
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- SM (Scrum Master) agent available
|
||||
- Epic files created from `create-epics-and-stories`
|
||||
- Implementation-readiness passed (for BMad Method/Enterprise)
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the SM Agent
|
||||
|
||||
Start a fresh chat and load the SM (Scrum Master) agent.
|
||||
|
||||
### 2. Run the Workflow
|
||||
|
||||
```
|
||||
*sprint-planning
|
||||
```
|
||||
|
||||
### 3. Provide Context
|
||||
|
||||
Point the agent to your epic files created during Phase 3.
|
||||
|
||||
### 4. Review Sprint Organization
|
||||
|
||||
The agent organizes stories into the sprint tracking file.
|
||||
|
||||
## What You Get
|
||||
|
||||
A `sprint-status.yaml` file containing:
|
||||
|
||||
- All epics with their stories
|
||||
- Story status tracking (TODO, IN PROGRESS, READY FOR REVIEW, DONE)
|
||||
- Dependencies between stories
|
||||
- Priority ordering
|
||||
|
||||
## Story Lifecycle States
|
||||
|
||||
| State | Description |
|
||||
|-------|-------------|
|
||||
| **TODO** | Story identified but not started |
|
||||
| **IN PROGRESS** | Story being implemented |
|
||||
| **READY FOR REVIEW** | Implementation complete, awaiting code review |
|
||||
| **DONE** | Accepted and complete |
|
||||
|
||||
## Typical Sprint Flow
|
||||
|
||||
**Sprint 0 (Planning Phase):**
|
||||
- Complete Phases 1-3
|
||||
- PRD/GDD + Architecture complete
|
||||
- Epics+Stories created via create-epics-and-stories
|
||||
|
||||
**Sprint 1+ (Implementation Phase):**
|
||||
|
||||
Start of Phase 4:
|
||||
1. SM runs `sprint-planning` (once)
|
||||
|
||||
Per Story (repeat until epic complete):
|
||||
1. SM runs `create-story`
|
||||
2. DEV runs `dev-story`
|
||||
3. DEV runs `code-review`
|
||||
4. Update sprint-status.yaml
|
||||
|
||||
After Epic Complete:
|
||||
- SM runs `retrospective`
|
||||
- Move to next epic
|
||||
|
||||
## Tips
|
||||
|
||||
- **Run sprint-planning only once** — At Phase 4 start
|
||||
- **Use sprint-status during Phase 4** — Check current state anytime
|
||||
- **Keep sprint-status.yaml as single source of truth** — All status updates go here
|
||||
- **Update story status after each stage** — Keep it current
|
||||
|
||||
## Next Steps
|
||||
|
||||
After sprint planning:
|
||||
|
||||
1. **Create Story** — Prepare the first story for implementation
|
||||
2. **Implement Story** — Run dev-story with the DEV agent
|
||||
3. **Code Review** — Quality assurance after implementation
|
||||
98
docs/how-to/workflows/run-test-design.md
Normal file
98
docs/how-to/workflows/run-test-design.md
Normal file
@@ -0,0 +1,98 @@
|
||||
---
|
||||
title: "How to Run Test Design"
|
||||
description: How to create comprehensive test plans using TEA's test-design workflow
|
||||
---
|
||||
|
||||
Use TEA's `*test-design` workflow to create comprehensive test plans with risk assessment and coverage strategies.
|
||||
|
||||
## When to Use This
|
||||
|
||||
**System-level (Phase 3):**
|
||||
- After architecture is complete
|
||||
- Before implementation-readiness gate
|
||||
- To validate architecture testability
|
||||
|
||||
**Epic-level (Phase 4):**
|
||||
- At the start of each epic
|
||||
- Before implementing stories in the epic
|
||||
- To identify epic-specific testing needs
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- For system-level: Architecture document complete
|
||||
- For epic-level: Epic defined with stories
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load the TEA Agent
|
||||
|
||||
Start a fresh chat and load the TEA (Test Architect) agent.
|
||||
|
||||
### 2. Run the Test Design Workflow
|
||||
|
||||
```
|
||||
*test-design
|
||||
```
|
||||
|
||||
### 3. Specify the Mode
|
||||
|
||||
TEA will ask if you want:
|
||||
|
||||
- **System-level** — For architecture testability review (Phase 3)
|
||||
- **Epic-level** — For epic-specific test planning (Phase 4)
|
||||
|
||||
### 4. Provide Context
|
||||
|
||||
For system-level:
|
||||
- Point to your architecture document
|
||||
- Reference any ADRs (Architecture Decision Records)
|
||||
|
||||
For epic-level:
|
||||
- Specify which epic you're planning
|
||||
- Reference the epic file with stories
|
||||
|
||||
### 5. Review the Output
|
||||
|
||||
TEA generates a comprehensive test design document.
|
||||
|
||||
## What You Get
|
||||
|
||||
**System-Level Output (`test-design-system.md`):**
|
||||
- Testability review of architecture
|
||||
- ADR → test mapping
|
||||
- Architecturally Significant Requirements (ASRs)
|
||||
- Environment needs
|
||||
- Test infrastructure recommendations
|
||||
|
||||
**Epic-Level Output (`test-design-epic-N.md`):**
|
||||
- Risk assessment for the epic
|
||||
- Test priorities
|
||||
- Coverage plan
|
||||
- Regression hotspots (for brownfield)
|
||||
- Integration risks
|
||||
- Mitigation strategies
|
||||
|
||||
## Test Design for Different Tracks
|
||||
|
||||
| Track | Phase 3 Focus | Phase 4 Focus |
|
||||
|-------|---------------|---------------|
|
||||
| **Greenfield** | System-level testability review | Per-epic risk assessment and test plan |
|
||||
| **Brownfield** | System-level + existing test baseline | Regression hotspots, integration risks |
|
||||
| **Enterprise** | Compliance-aware testability | Security/performance/compliance focus |
|
||||
|
||||
## Tips
|
||||
|
||||
- **Run system-level right after architecture** — Early testability review
|
||||
- **Run epic-level at the start of each epic** — Targeted test planning
|
||||
- **Update if ADRs change** — Keep test design aligned
|
||||
- **Use output to guide other workflows** — Feeds into `*atdd` and `*automate`
|
||||
|
||||
## Next Steps
|
||||
|
||||
After test design:
|
||||
|
||||
1. **Setup Test Framework** — If not already configured
|
||||
2. **Implementation Readiness** — System-level feeds into gate check
|
||||
3. **Story Implementation** — Epic-level guides testing during dev
|
||||
605
docs/how-to/workflows/run-test-review.md
Normal file
605
docs/how-to/workflows/run-test-review.md
Normal file
@@ -0,0 +1,605 @@
|
||||
---
|
||||
title: "How to Run Test Review with TEA"
|
||||
description: Audit test quality using TEA's comprehensive knowledge base and get 0-100 scoring
|
||||
---
|
||||
|
||||
# How to Run Test Review with TEA
|
||||
|
||||
Use TEA's `*test-review` workflow to audit test quality with objective scoring and actionable feedback. TEA reviews tests against its knowledge base of best practices.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Want to validate test quality objectively
|
||||
- Need quality metrics for release gates
|
||||
- Preparing for production deployment
|
||||
- Reviewing team-written tests
|
||||
- Auditing AI-generated tests
|
||||
- Onboarding new team members (show good patterns)
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Tests written (to review)
|
||||
- Test framework configured
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load TEA Agent
|
||||
|
||||
Start a fresh chat and load TEA:
|
||||
|
||||
```
|
||||
*tea
|
||||
```
|
||||
|
||||
### 2. Run the Test Review Workflow
|
||||
|
||||
```
|
||||
*test-review
|
||||
```
|
||||
|
||||
### 3. Specify Review Scope
|
||||
|
||||
TEA will ask what to review.
|
||||
|
||||
#### Option A: Single File
|
||||
|
||||
Review one test file:
|
||||
|
||||
```
|
||||
tests/e2e/checkout.spec.ts
|
||||
```
|
||||
|
||||
**Best for:**
|
||||
- Reviewing specific failing tests
|
||||
- Quick feedback on new tests
|
||||
- Learning from specific examples
|
||||
|
||||
#### Option B: Directory
|
||||
|
||||
Review all tests in a directory:
|
||||
|
||||
```
|
||||
tests/e2e/
|
||||
```
|
||||
|
||||
**Best for:**
|
||||
- Reviewing E2E test suite
|
||||
- Comparing test quality across files
|
||||
- Finding patterns of issues
|
||||
|
||||
#### Option C: Entire Suite
|
||||
|
||||
Review all tests:
|
||||
|
||||
```
|
||||
tests/
|
||||
```
|
||||
|
||||
**Best for:**
|
||||
- Release gate quality check
|
||||
- Comprehensive audit
|
||||
- Establishing baseline metrics
|
||||
|
||||
### 4. Review the Quality Report
|
||||
|
||||
TEA generates a comprehensive quality report with scoring.
|
||||
|
||||
#### Report Structure (`test-review.md`):
|
||||
|
||||
```markdown
|
||||
# Test Quality Review Report
|
||||
|
||||
**Date:** 2026-01-13
|
||||
**Scope:** tests/e2e/
|
||||
**Overall Score:** 76/100
|
||||
|
||||
## Summary
|
||||
|
||||
- **Tests Reviewed:** 12
|
||||
- **Passing Quality:** 9 tests (75%)
|
||||
- **Needs Improvement:** 3 tests (25%)
|
||||
- **Critical Issues:** 2
|
||||
- **Recommendations:** 6
|
||||
|
||||
## Critical Issues
|
||||
|
||||
### 1. Hard Waits Detected
|
||||
|
||||
**File:** `tests/e2e/checkout.spec.ts:45`
|
||||
**Issue:** Using `page.waitForTimeout(3000)`
|
||||
**Impact:** Test is flaky and unnecessarily slow
|
||||
**Severity:** Critical
|
||||
|
||||
**Current Code:**
|
||||
```typescript
|
||||
await page.click('button[type="submit"]');
|
||||
await page.waitForTimeout(3000); // ❌ Hard wait
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
```
|
||||
|
||||
**Fix:**
|
||||
```typescript
|
||||
await page.click('button[type="submit"]');
|
||||
// Wait for the API response that triggers success message
|
||||
await page.waitForResponse(resp =>
|
||||
resp.url().includes('/api/checkout') && resp.ok()
|
||||
);
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
```
|
||||
|
||||
**Why This Matters:**
|
||||
- Hard waits are fixed timeouts that don't wait for actual conditions
|
||||
- Tests fail intermittently on slower machines
|
||||
- Wastes time waiting even when response is fast
|
||||
- Network-first patterns are more reliable
|
||||
|
||||
---
|
||||
|
||||
### 2. Conditional Flow Control
|
||||
|
||||
**File:** `tests/e2e/profile.spec.ts:28`
|
||||
**Issue:** Using if/else to handle optional elements
|
||||
**Impact:** Non-deterministic test behavior
|
||||
**Severity:** Critical
|
||||
|
||||
**Current Code:**
|
||||
```typescript
|
||||
if (await page.locator('.banner').isVisible()) {
|
||||
await page.click('.dismiss');
|
||||
}
|
||||
// ❌ Test behavior changes based on banner presence
|
||||
```
|
||||
|
||||
**Fix:**
|
||||
```typescript
|
||||
// Option 1: Make banner presence deterministic
|
||||
await expect(page.locator('.banner')).toBeVisible();
|
||||
await page.click('.dismiss');
|
||||
|
||||
// Option 2: Test both scenarios separately
|
||||
test('should show banner for new users', async ({ page }) => {
|
||||
// Test with banner
|
||||
});
|
||||
|
||||
test('should not show banner for returning users', async ({ page }) => {
|
||||
// Test without banner
|
||||
});
|
||||
```
|
||||
|
||||
**Why This Matters:**
|
||||
- Tests should be deterministic (same result every run)
|
||||
- Conditionals hide bugs (what if banner should always show?)
|
||||
- Makes debugging harder
|
||||
- Violates test isolation principle
|
||||
|
||||
## Recommendations
|
||||
|
||||
### 1. Extract Repeated Setup
|
||||
|
||||
**File:** `tests/e2e/profile.spec.ts`
|
||||
**Issue:** Login code duplicated in every test
|
||||
**Severity:** Medium
|
||||
**Impact:** Maintenance burden, test verbosity
|
||||
|
||||
**Current:**
|
||||
```typescript
|
||||
test('test 1', async ({ page }) => {
|
||||
await page.goto('/login');
|
||||
await page.fill('[name="email"]', 'test@example.com');
|
||||
await page.fill('[name="password"]', 'password');
|
||||
await page.click('button[type="submit"]');
|
||||
// Test logic...
|
||||
});
|
||||
|
||||
test('test 2', async ({ page }) => {
|
||||
// Same login code repeated
|
||||
});
|
||||
```
|
||||
|
||||
**Fix (Vanilla Playwright):**
|
||||
```typescript
|
||||
// Create fixture in tests/support/fixtures/auth.ts
|
||||
import { test as base, Page } from '@playwright/test';
|
||||
|
||||
export const test = base.extend<{ authenticatedPage: Page }>({
|
||||
authenticatedPage: async ({ page }, use) => {
|
||||
await page.goto('/login');
|
||||
await page.getByLabel('Email').fill('test@example.com');
|
||||
await page.getByLabel('Password').fill('password');
|
||||
await page.getByRole('button', { name: 'Sign in' }).click();
|
||||
await page.waitForURL(/\/dashboard/);
|
||||
await use(page);
|
||||
}
|
||||
});
|
||||
|
||||
// Use in tests
|
||||
test('test 1', async ({ authenticatedPage }) => {
|
||||
// Already logged in
|
||||
});
|
||||
```
|
||||
|
||||
**Better (With Playwright Utils):**
|
||||
```typescript
|
||||
// Use built-in auth-session fixture
|
||||
import { test as base } from '@playwright/test';
|
||||
import { createAuthFixtures } from '@seontechnologies/playwright-utils/auth-session';
|
||||
|
||||
export const test = base.extend(createAuthFixtures());
|
||||
|
||||
// Use in tests - even simpler
|
||||
test('test 1', async ({ page, authToken }) => {
|
||||
// authToken already available (persisted, reused)
|
||||
await page.goto('/dashboard');
|
||||
// Already authenticated via authToken
|
||||
});
|
||||
```
|
||||
|
||||
**Playwright Utils Benefits:**
|
||||
- Token persisted to disk (faster subsequent runs)
|
||||
- Multi-user support out of the box
|
||||
- Automatic token renewal if expired
|
||||
- No manual login flow needed
|
||||
|
||||
---
|
||||
|
||||
### 2. Add Network Assertions
|
||||
|
||||
**File:** `tests/e2e/api-calls.spec.ts`
|
||||
**Issue:** No verification of API responses
|
||||
**Severity:** Low
|
||||
**Impact:** Tests don't catch API errors
|
||||
|
||||
**Current:**
|
||||
```typescript
|
||||
await page.click('button[name="save"]');
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
// ❌ What if API returned 500 but UI shows cached success?
|
||||
```
|
||||
|
||||
**Enhancement:**
|
||||
```typescript
|
||||
const responsePromise = page.waitForResponse(
|
||||
resp => resp.url().includes('/api/profile') && resp.status() === 200
|
||||
);
|
||||
await page.click('button[name="save"]');
|
||||
const response = await responsePromise;
|
||||
|
||||
// Verify API response
|
||||
const data = await response.json();
|
||||
expect(data.success).toBe(true);
|
||||
|
||||
// Verify UI
|
||||
await expect(page.locator('.success')).toBeVisible();
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### 3. Improve Test Names
|
||||
|
||||
**File:** `tests/e2e/checkout.spec.ts`
|
||||
**Issue:** Vague test names
|
||||
**Severity:** Low
|
||||
**Impact:** Hard to understand test purpose
|
||||
|
||||
**Current:**
|
||||
```typescript
|
||||
test('should work', async ({ page }) => { });
|
||||
test('test checkout', async ({ page }) => { });
|
||||
```
|
||||
|
||||
**Better:**
|
||||
```typescript
|
||||
test('should complete checkout with valid credit card', async ({ page }) => { });
|
||||
test('should show validation error for expired card', async ({ page }) => { });
|
||||
```
|
||||
|
||||
## Quality Scores by Category
|
||||
|
||||
| Category | Score | Target | Status |
|
||||
|----------|-------|--------|--------|
|
||||
| **Determinism** | 26/35 | 30/35 | ⚠️ Needs Improvement |
|
||||
| **Isolation** | 22/25 | 20/25 | ✅ Good |
|
||||
| **Assertions** | 18/20 | 16/20 | ✅ Good |
|
||||
| **Structure** | 7/10 | 8/10 | ⚠️ Minor Issues |
|
||||
| **Performance** | 3/10 | 8/10 | ❌ Critical |
|
||||
|
||||
### Scoring Breakdown
|
||||
|
||||
**Determinism (35 points max):**
|
||||
- No hard waits: 0/10 ❌ (found 3 instances)
|
||||
- No conditionals: 8/10 ⚠️ (found 2 instances)
|
||||
- No try-catch flow control: 10/10 ✅
|
||||
- Network-first patterns: 8/15 ⚠️ (some tests missing)
|
||||
|
||||
**Isolation (25 points max):**
|
||||
- Self-cleaning: 20/20 ✅
|
||||
- No global state: 5/5 ✅
|
||||
- Parallel-safe: 0/0 ✅ (not tested)
|
||||
|
||||
**Assertions (20 points max):**
|
||||
- Explicit in test body: 15/15 ✅
|
||||
- Specific and meaningful: 3/5 ⚠️ (some weak assertions)
|
||||
|
||||
**Structure (10 points max):**
|
||||
- Test size < 300 lines: 5/5 ✅
|
||||
- Clear names: 2/5 ⚠️ (some vague names)
|
||||
|
||||
**Performance (10 points max):**
|
||||
- Execution time < 1.5 min: 3/10 ❌ (3 tests exceed limit)
|
||||
|
||||
## Files Reviewed
|
||||
|
||||
| File | Score | Issues | Status |
|
||||
|------|-------|--------|--------|
|
||||
| `tests/e2e/checkout.spec.ts` | 65/100 | 4 | ❌ Needs Work |
|
||||
| `tests/e2e/profile.spec.ts` | 72/100 | 3 | ⚠️ Needs Improvement |
|
||||
| `tests/e2e/search.spec.ts` | 88/100 | 1 | ✅ Good |
|
||||
| `tests/api/profile.spec.ts` | 92/100 | 0 | ✅ Excellent |
|
||||
|
||||
## Next Steps
|
||||
|
||||
### Immediate (Fix Critical Issues)
|
||||
1. Remove hard waits in `checkout.spec.ts` (line 45, 67, 89)
|
||||
2. Fix conditional in `profile.spec.ts` (line 28)
|
||||
3. Optimize slow tests in `checkout.spec.ts`
|
||||
|
||||
### Short-term (Apply Recommendations)
|
||||
4. Extract login fixture from `profile.spec.ts`
|
||||
5. Add network assertions to `api-calls.spec.ts`
|
||||
6. Improve test names in `checkout.spec.ts`
|
||||
|
||||
### Long-term (Continuous Improvement)
|
||||
7. Re-run `*test-review` after fixes (target: 85/100)
|
||||
8. Add performance budgets to CI
|
||||
9. Document test patterns for team
|
||||
|
||||
## Knowledge Base References
|
||||
|
||||
TEA reviewed against these patterns:
|
||||
- [test-quality.md](/docs/reference/tea/knowledge-base.md#test-quality) - Execution limits, isolation
|
||||
- [network-first.md](/docs/reference/tea/knowledge-base.md#network-first) - Deterministic waits
|
||||
- [timing-debugging.md](/docs/reference/tea/knowledge-base.md#timing-debugging) - Race conditions
|
||||
- [selector-resilience.md](/docs/reference/tea/knowledge-base.md#selector-resilience) - Robust selectors
|
||||
```
|
||||
|
||||
## Understanding the Scores
|
||||
|
||||
### What Do Scores Mean?
|
||||
|
||||
| Score Range | Interpretation | Action |
|
||||
|-------------|----------------|--------|
|
||||
| **90-100** | Excellent | Minimal changes needed, production-ready |
|
||||
| **80-89** | Good | Minor improvements recommended |
|
||||
| **70-79** | Acceptable | Address recommendations before release |
|
||||
| **60-69** | Needs Improvement | Fix critical issues, apply recommendations |
|
||||
| **< 60** | Critical | Significant refactoring needed |
|
||||
|
||||
### Scoring Criteria
|
||||
|
||||
**Determinism (35 points):**
|
||||
- Tests produce same result every run
|
||||
- No random failures (flakiness)
|
||||
- No environment-dependent behavior
|
||||
|
||||
**Isolation (25 points):**
|
||||
- Tests don't depend on each other
|
||||
- Can run in any order
|
||||
- Clean up after themselves
|
||||
|
||||
**Assertions (20 points):**
|
||||
- Verify actual behavior
|
||||
- Specific and meaningful
|
||||
- Not abstracted away in helpers
|
||||
|
||||
**Structure (10 points):**
|
||||
- Readable and maintainable
|
||||
- Appropriate size
|
||||
- Clear naming
|
||||
|
||||
**Performance (10 points):**
|
||||
- Fast execution
|
||||
- Efficient selectors
|
||||
- No unnecessary waits
|
||||
|
||||
## What You Get
|
||||
|
||||
### Quality Report
|
||||
- Overall score (0-100)
|
||||
- Category scores (Determinism, Isolation, etc.)
|
||||
- File-by-file breakdown
|
||||
|
||||
### Critical Issues
|
||||
- Specific line numbers
|
||||
- Code examples (current vs fixed)
|
||||
- Why it matters explanation
|
||||
- Impact assessment
|
||||
|
||||
### Recommendations
|
||||
- Actionable improvements
|
||||
- Code examples
|
||||
- Priority/severity levels
|
||||
|
||||
### Next Steps
|
||||
- Immediate actions (fix critical)
|
||||
- Short-term improvements
|
||||
- Long-term quality goals
|
||||
|
||||
## Tips
|
||||
|
||||
### Review Before Release
|
||||
|
||||
Make test review part of release checklist:
|
||||
|
||||
```markdown
|
||||
## Release Checklist
|
||||
- [ ] All tests passing
|
||||
- [ ] Test review score > 80
|
||||
- [ ] Critical issues resolved
|
||||
- [ ] Performance within budget
|
||||
```
|
||||
|
||||
### Review After AI Generation
|
||||
|
||||
Always review AI-generated tests:
|
||||
|
||||
```
|
||||
1. Run *atdd or *automate
|
||||
2. Run *test-review on generated tests
|
||||
3. Fix critical issues
|
||||
4. Commit tests
|
||||
```
|
||||
|
||||
### Set Quality Gates
|
||||
|
||||
Use scores as quality gates:
|
||||
|
||||
```yaml
|
||||
# .github/workflows/test.yml
|
||||
- name: Review test quality
|
||||
run: |
|
||||
# Run test review
|
||||
# Parse score from report
|
||||
if [ $SCORE -lt 80 ]; then
|
||||
echo "Test quality below threshold"
|
||||
exit 1
|
||||
fi
|
||||
```
|
||||
|
||||
### Review Regularly
|
||||
|
||||
Schedule periodic reviews:
|
||||
|
||||
- **Per story:** Optional (spot check new tests)
|
||||
- **Per epic:** Recommended (ensure consistency)
|
||||
- **Per release:** Recommended for quality gates (required if using formal gate process)
|
||||
- **Quarterly:** Audit entire suite
|
||||
|
||||
### Focus Reviews
|
||||
|
||||
For large suites, review incrementally:
|
||||
|
||||
**Week 1:** Review E2E tests
|
||||
**Week 2:** Review API tests
|
||||
**Week 3:** Review component tests (Cypress CT or Vitest)
|
||||
**Week 4:** Apply fixes across all suites
|
||||
|
||||
**Component Testing Note:** TEA reviews component tests using framework-specific knowledge:
|
||||
- **Cypress:** Reviews Cypress Component Testing specs (*.cy.tsx)
|
||||
- **Playwright:** Reviews Vitest component tests (*.test.tsx)
|
||||
|
||||
### Use Reviews for Learning
|
||||
|
||||
Share reports with team:
|
||||
|
||||
```
|
||||
Team Meeting:
|
||||
- Review test-review.md
|
||||
- Discuss critical issues
|
||||
- Agree on patterns
|
||||
- Update team guidelines
|
||||
```
|
||||
|
||||
### Compare Over Time
|
||||
|
||||
Track improvement:
|
||||
|
||||
```markdown
|
||||
## Quality Trend
|
||||
|
||||
| Date | Score | Critical Issues | Notes |
|
||||
|------|-------|-----------------|-------|
|
||||
| 2026-01-01 | 65 | 5 | Baseline |
|
||||
| 2026-01-15 | 72 | 2 | Fixed hard waits |
|
||||
| 2026-02-01 | 84 | 0 | All critical resolved |
|
||||
```
|
||||
|
||||
## Common Issues
|
||||
|
||||
### Low Determinism Score
|
||||
|
||||
**Symptoms:**
|
||||
- Tests fail randomly
|
||||
- "Works on my machine"
|
||||
- CI failures that don't reproduce locally
|
||||
|
||||
**Common Causes:**
|
||||
- Hard waits (`waitForTimeout`)
|
||||
- Conditional flow control (`if/else`)
|
||||
- Try-catch for flow control
|
||||
- Missing network-first patterns
|
||||
|
||||
**Fix:** Review determinism section, apply network-first patterns
|
||||
|
||||
### Low Performance Score
|
||||
|
||||
**Symptoms:**
|
||||
- Tests take > 1.5 minutes each
|
||||
- Test suite takes hours
|
||||
- CI times out
|
||||
|
||||
**Common Causes:**
|
||||
- Unnecessary waits (hard timeouts)
|
||||
- Inefficient selectors (XPath, complex CSS)
|
||||
- Not using parallelization
|
||||
- Heavy setup in every test
|
||||
|
||||
**Fix:** Optimize waits, improve selectors, use fixtures
|
||||
|
||||
### Low Isolation Score
|
||||
|
||||
**Symptoms:**
|
||||
- Tests fail when run in different order
|
||||
- Tests fail in parallel
|
||||
- Test data conflicts
|
||||
|
||||
**Common Causes:**
|
||||
- Shared global state
|
||||
- Tests don't clean up
|
||||
- Hard-coded test data
|
||||
- Database not reset between tests
|
||||
|
||||
**Fix:** Use fixtures, clean up in afterEach, use unique test data
|
||||
|
||||
### "Too Many Issues to Fix"
|
||||
|
||||
**Problem:** Report shows 50+ issues, overwhelming.
|
||||
|
||||
**Solution:** Prioritize:
|
||||
1. Fix all critical issues first
|
||||
2. Apply top 3 recommendations
|
||||
3. Re-run review
|
||||
4. Iterate
|
||||
|
||||
Don't try to fix everything at once.
|
||||
|
||||
### Reviews Take Too Long
|
||||
|
||||
**Problem:** Reviewing entire suite takes hours.
|
||||
|
||||
**Solution:** Review incrementally:
|
||||
- Review new tests in PR review
|
||||
- Schedule directory reviews weekly
|
||||
- Full suite review quarterly
|
||||
|
||||
## Related Guides
|
||||
|
||||
- [How to Run ATDD](/docs/how-to/workflows/run-atdd.md) - Generate tests to review
|
||||
- [How to Run Automate](/docs/how-to/workflows/run-automate.md) - Expand coverage to review
|
||||
- [How to Run Trace](/docs/how-to/workflows/run-trace.md) - Coverage complements quality
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - What makes tests good
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Avoiding flakiness
|
||||
- [Fixture Architecture](/docs/explanation/tea/fixture-architecture.md) - Reusable patterns
|
||||
|
||||
## Reference
|
||||
|
||||
- [Command: *test-review](/docs/reference/tea/commands.md#test-review) - Full command reference
|
||||
- [Knowledge Base Index](/docs/reference/tea/knowledge-base.md) - Patterns TEA reviews against
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
883
docs/how-to/workflows/run-trace.md
Normal file
883
docs/how-to/workflows/run-trace.md
Normal file
@@ -0,0 +1,883 @@
|
||||
---
|
||||
title: "How to Run Trace with TEA"
|
||||
description: Map requirements to tests and make quality gate decisions using TEA's trace workflow
|
||||
---
|
||||
|
||||
# How to Run Trace with TEA
|
||||
|
||||
Use TEA's `*trace` workflow for requirements traceability and quality gate decisions. This is a two-phase workflow: Phase 1 analyzes coverage, Phase 2 makes the go/no-go decision.
|
||||
|
||||
## When to Use This
|
||||
|
||||
### Phase 1: Requirements Traceability
|
||||
- Map acceptance criteria to implemented tests
|
||||
- Identify coverage gaps
|
||||
- Prioritize missing tests
|
||||
- Refresh coverage after each story/epic
|
||||
|
||||
### Phase 2: Quality Gate Decision
|
||||
- Make go/no-go decision for release
|
||||
- Validate coverage meets thresholds
|
||||
- Document gate decision with evidence
|
||||
- Support business-approved waivers
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Requirements defined (stories, acceptance criteria, test design)
|
||||
- Tests implemented
|
||||
- For brownfield: Existing codebase with tests
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Run the Trace Workflow
|
||||
|
||||
```
|
||||
*trace
|
||||
```
|
||||
|
||||
### 2. Specify Phase
|
||||
|
||||
TEA will ask which phase you're running.
|
||||
|
||||
**Phase 1: Requirements Traceability**
|
||||
- Analyze coverage
|
||||
- Identify gaps
|
||||
- Generate recommendations
|
||||
|
||||
**Phase 2: Quality Gate Decision**
|
||||
- Make PASS/CONCERNS/FAIL/WAIVED decision
|
||||
- Requires Phase 1 complete
|
||||
|
||||
**Typical flow:** Run Phase 1 first, review gaps, then run Phase 2 for gate decision.
|
||||
|
||||
---
|
||||
|
||||
## Phase 1: Requirements Traceability
|
||||
|
||||
### 3. Provide Requirements Source
|
||||
|
||||
TEA will ask where requirements are defined.
|
||||
|
||||
**Options:**
|
||||
|
||||
| Source | Example | Best For |
|
||||
| --------------- | ----------------------------- | ---------------------- |
|
||||
| **Story file** | `story-profile-management.md` | Single story coverage |
|
||||
| **Test design** | `test-design-epic-1.md` | Epic coverage |
|
||||
| **PRD** | `PRD.md` | System-level coverage |
|
||||
| **Multiple** | All of the above | Comprehensive analysis |
|
||||
|
||||
**Example Response:**
|
||||
```
|
||||
Requirements:
|
||||
- story-profile-management.md (acceptance criteria)
|
||||
- test-design-epic-1.md (test priorities)
|
||||
```
|
||||
|
||||
### 4. Specify Test Location
|
||||
|
||||
TEA will ask where tests are located.
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Test location: tests/
|
||||
Include:
|
||||
- tests/api/
|
||||
- tests/e2e/
|
||||
```
|
||||
|
||||
### 5. Specify Focus Areas (Optional)
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Focus on:
|
||||
- Profile CRUD operations
|
||||
- Validation scenarios
|
||||
- Authorization checks
|
||||
```
|
||||
|
||||
### 6. Review Coverage Matrix
|
||||
|
||||
TEA generates a comprehensive traceability matrix.
|
||||
|
||||
#### Traceability Matrix (`traceability-matrix.md`):
|
||||
|
||||
```markdown
|
||||
# Requirements Traceability Matrix
|
||||
|
||||
**Date:** 2026-01-13
|
||||
**Scope:** Epic 1 - User Profile Management
|
||||
**Phase:** Phase 1 (Traceability Analysis)
|
||||
|
||||
## Coverage Summary
|
||||
|
||||
| Metric | Count | Percentage |
|
||||
| ---------------------- | ----- | ---------- |
|
||||
| **Total Requirements** | 15 | 100% |
|
||||
| **Full Coverage** | 11 | 73% |
|
||||
| **Partial Coverage** | 3 | 20% |
|
||||
| **No Coverage** | 1 | 7% |
|
||||
|
||||
### By Priority
|
||||
|
||||
| Priority | Total | Covered | Percentage |
|
||||
| -------- | ----- | ------- | ----------------- |
|
||||
| **P0** | 5 | 5 | 100% ✅ |
|
||||
| **P1** | 6 | 5 | 83% ⚠️ |
|
||||
| **P2** | 3 | 1 | 33% ⚠️ |
|
||||
| **P3** | 1 | 0 | 0% ✅ (acceptable) |
|
||||
|
||||
---
|
||||
|
||||
## Detailed Traceability
|
||||
|
||||
### ✅ Requirement 1: User can view their profile (P0)
|
||||
|
||||
**Acceptance Criteria:**
|
||||
- User navigates to /profile
|
||||
- Profile displays name, email, avatar
|
||||
- Data is current (not cached)
|
||||
|
||||
**Test Coverage:** FULL ✅
|
||||
|
||||
**Tests:**
|
||||
- `tests/e2e/profile-view.spec.ts:15` - "should display profile page with current data"
|
||||
- ✅ Navigates to /profile
|
||||
- ✅ Verifies name, email visible
|
||||
- ✅ Verifies avatar displayed
|
||||
- ✅ Validates data freshness via API assertion
|
||||
|
||||
- `tests/api/profile.spec.ts:8` - "should fetch user profile via API"
|
||||
- ✅ Calls GET /api/profile
|
||||
- ✅ Validates response schema
|
||||
- ✅ Confirms all fields present
|
||||
|
||||
---
|
||||
|
||||
### ⚠️ Requirement 2: User can edit profile (P0)
|
||||
|
||||
**Acceptance Criteria:**
|
||||
- User clicks "Edit Profile"
|
||||
- Can modify name, email, bio
|
||||
- Can upload avatar
|
||||
- Changes are persisted
|
||||
- Success message shown
|
||||
|
||||
**Test Coverage:** PARTIAL ⚠️
|
||||
|
||||
**Tests:**
|
||||
- `tests/e2e/profile-edit.spec.ts:22` - "should edit and save profile"
|
||||
- ✅ Clicks edit button
|
||||
- ✅ Modifies name and email
|
||||
- ⚠️ **Does NOT test bio field**
|
||||
- ❌ **Does NOT test avatar upload**
|
||||
- ✅ Verifies persistence
|
||||
- ✅ Verifies success message
|
||||
|
||||
- `tests/api/profile.spec.ts:25` - "should update profile via PATCH"
|
||||
- ✅ Calls PATCH /api/profile
|
||||
- ✅ Validates update response
|
||||
- ⚠️ **Only tests name/email, not bio/avatar**
|
||||
|
||||
**Missing Coverage:**
|
||||
- Bio field not tested in E2E or API
|
||||
- Avatar upload not tested
|
||||
|
||||
**Gap Severity:** HIGH (P0 requirement, critical path)
|
||||
|
||||
---
|
||||
|
||||
### ✅ Requirement 3: Invalid email shows validation error (P1)
|
||||
|
||||
**Acceptance Criteria:**
|
||||
- Enter invalid email format
|
||||
- See error message
|
||||
- Cannot save changes
|
||||
|
||||
**Test Coverage:** FULL ✅
|
||||
|
||||
**Tests:**
|
||||
- `tests/e2e/profile-edit.spec.ts:45` - "should show validation error for invalid email"
|
||||
- `tests/api/profile.spec.ts:50` - "should return 400 for invalid email"
|
||||
|
||||
---
|
||||
|
||||
### ❌ Requirement 15: Profile export as PDF (P2)
|
||||
|
||||
**Acceptance Criteria:**
|
||||
- User clicks "Export Profile"
|
||||
- PDF downloads with profile data
|
||||
|
||||
**Test Coverage:** NONE ❌
|
||||
|
||||
**Gap Analysis:**
|
||||
- **Priority:** P2 (medium)
|
||||
- **Risk:** Low (non-critical feature)
|
||||
- **Recommendation:** Add in next iteration (not blocking for release)
|
||||
|
||||
---
|
||||
|
||||
## Gap Prioritization
|
||||
|
||||
### Critical Gaps (Must Fix Before Release)
|
||||
|
||||
| Gap | Requirement | Priority | Risk | Recommendation |
|
||||
| --- | ------------------------ | -------- | ---- | ------------------- |
|
||||
| 1 | Bio field not tested | P0 | High | Add E2E + API tests |
|
||||
| 2 | Avatar upload not tested | P0 | High | Add E2E + API tests |
|
||||
|
||||
**Estimated Effort:** 3 hours
|
||||
**Owner:** QA team
|
||||
**Deadline:** Before release
|
||||
|
||||
### Non-Critical Gaps (Can Defer)
|
||||
|
||||
| Gap | Requirement | Priority | Risk | Recommendation |
|
||||
| --- | ------------------------- | -------- | ---- | ------------------- |
|
||||
| 3 | Profile export not tested | P2 | Low | Add in v1.3 release |
|
||||
|
||||
**Estimated Effort:** 2 hours
|
||||
**Owner:** QA team
|
||||
**Deadline:** Next release (February)
|
||||
|
||||
---
|
||||
|
||||
## Recommendations
|
||||
|
||||
### 1. Add Bio Field Tests
|
||||
|
||||
**Tests Needed (Vanilla Playwright):**
|
||||
```typescript
|
||||
// tests/e2e/profile-edit.spec.ts
|
||||
test('should edit bio field', async ({ page }) => {
|
||||
await page.goto('/profile');
|
||||
await page.getByRole('button', { name: 'Edit' }).click();
|
||||
await page.getByLabel('Bio').fill('New bio text');
|
||||
await page.getByRole('button', { name: 'Save' }).click();
|
||||
await expect(page.getByText('New bio text')).toBeVisible();
|
||||
});
|
||||
|
||||
// tests/api/profile.spec.ts
|
||||
test('should update bio via API', async ({ request }) => {
|
||||
const response = await request.patch('/api/profile', {
|
||||
data: { bio: 'Updated bio' }
|
||||
});
|
||||
expect(response.ok()).toBeTruthy();
|
||||
const { bio } = await response.json();
|
||||
expect(bio).toBe('Updated bio');
|
||||
});
|
||||
```
|
||||
|
||||
**With Playwright Utils:**
|
||||
```typescript
|
||||
// tests/e2e/profile-edit.spec.ts
|
||||
import { test } from '../support/fixtures'; // Composed with authToken
|
||||
|
||||
test('should edit bio field', async ({ page, authToken }) => {
|
||||
await page.goto('/profile');
|
||||
await page.getByRole('button', { name: 'Edit' }).click();
|
||||
await page.getByLabel('Bio').fill('New bio text');
|
||||
await page.getByRole('button', { name: 'Save' }).click();
|
||||
await expect(page.getByText('New bio text')).toBeVisible();
|
||||
});
|
||||
|
||||
// tests/api/profile.spec.ts
|
||||
import { test as base, expect } from '@playwright/test';
|
||||
import { test as apiRequestFixture } from '@seontechnologies/playwright-utils/api-request/fixtures';
|
||||
import { createAuthFixtures } from '@seontechnologies/playwright-utils/auth-session';
|
||||
import { mergeTests } from '@playwright/test';
|
||||
|
||||
// Merge API request + auth fixtures
|
||||
const authFixtureTest = base.extend(createAuthFixtures());
|
||||
const test = mergeTests(apiRequestFixture, authFixtureTest);
|
||||
|
||||
test('should update bio via API', async ({ apiRequest, authToken }) => {
|
||||
const { status, body } = await apiRequest({
|
||||
method: 'PATCH',
|
||||
path: '/api/profile',
|
||||
body: { bio: 'Updated bio' },
|
||||
headers: { Authorization: `Bearer ${authToken}` }
|
||||
});
|
||||
|
||||
expect(status).toBe(200);
|
||||
expect(body.bio).toBe('Updated bio');
|
||||
});
|
||||
```
|
||||
|
||||
**Note:** `authToken` requires auth-session fixture setup. See [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md#auth-session).
|
||||
|
||||
### 2. Add Avatar Upload Tests
|
||||
|
||||
**Tests Needed:**
|
||||
```typescript
|
||||
// tests/e2e/profile-edit.spec.ts
|
||||
test('should upload avatar image', async ({ page }) => {
|
||||
await page.goto('/profile');
|
||||
await page.getByRole('button', { name: 'Edit' }).click();
|
||||
|
||||
// Upload file
|
||||
await page.setInputFiles('[type="file"]', 'fixtures/avatar.png');
|
||||
await page.getByRole('button', { name: 'Save' }).click();
|
||||
|
||||
// Verify uploaded image displays
|
||||
await expect(page.locator('img[alt="Profile avatar"]')).toBeVisible();
|
||||
});
|
||||
|
||||
// tests/api/profile.spec.ts
|
||||
import { test, expect } from '@playwright/test';
|
||||
import fs from 'fs/promises';
|
||||
|
||||
test('should accept valid image upload', async ({ request }) => {
|
||||
const response = await request.post('/api/profile/avatar', {
|
||||
multipart: {
|
||||
file: {
|
||||
name: 'avatar.png',
|
||||
mimeType: 'image/png',
|
||||
buffer: await fs.readFile('fixtures/avatar.png')
|
||||
}
|
||||
}
|
||||
});
|
||||
expect(response.ok()).toBeTruthy();
|
||||
});
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Next Steps
|
||||
|
||||
After reviewing traceability:
|
||||
|
||||
1. **Fix critical gaps** - Add tests for P0/P1 requirements
|
||||
2. **Run *test-review** - Ensure new tests meet quality standards
|
||||
3. **Run Phase 2** - Make gate decision after gaps addressed
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Phase 2: Quality Gate Decision
|
||||
|
||||
After Phase 1 coverage analysis is complete, run Phase 2 for the gate decision.
|
||||
|
||||
**Prerequisites:**
|
||||
- Phase 1 traceability matrix complete
|
||||
- Test execution results available (must have test results)
|
||||
|
||||
**Note:** Phase 2 will skip if test execution results aren't provided. The workflow requires actual test run results to make gate decisions.
|
||||
|
||||
### 7. Run Phase 2
|
||||
|
||||
```
|
||||
*trace
|
||||
```
|
||||
|
||||
Select "Phase 2: Quality Gate Decision"
|
||||
|
||||
### 8. Provide Additional Context
|
||||
|
||||
TEA will ask for:
|
||||
|
||||
**Gate Type:**
|
||||
- Story gate (small release)
|
||||
- Epic gate (larger release)
|
||||
- Release gate (production deployment)
|
||||
- Hotfix gate (emergency fix)
|
||||
|
||||
**Decision Mode:**
|
||||
- **Deterministic** - Rule-based (coverage %, quality scores)
|
||||
- **Manual** - Team decision with TEA guidance
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Gate type: Epic gate
|
||||
Decision mode: Deterministic
|
||||
```
|
||||
|
||||
### 9. Provide Supporting Evidence
|
||||
|
||||
TEA will request:
|
||||
|
||||
**Phase 1 Results:**
|
||||
```
|
||||
traceability-matrix.md (from Phase 1)
|
||||
```
|
||||
|
||||
**Test Quality (Optional):**
|
||||
```
|
||||
test-review.md (from *test-review)
|
||||
```
|
||||
|
||||
**NFR Assessment (Optional):**
|
||||
```
|
||||
nfr-assessment.md (from *nfr-assess)
|
||||
```
|
||||
|
||||
### 10. Review Gate Decision
|
||||
|
||||
TEA makes evidence-based gate decision and writes to separate file.
|
||||
|
||||
#### Gate Decision (`gate-decision-{gate_type}-{story_id}.md`):
|
||||
|
||||
```markdown
|
||||
---
|
||||
|
||||
# Phase 2: Quality Gate Decision
|
||||
|
||||
**Gate Type:** Epic Gate
|
||||
**Decision:** PASS ✅
|
||||
**Date:** 2026-01-13
|
||||
**Approvers:** Product Manager, Tech Lead, QA Lead
|
||||
|
||||
## Decision Summary
|
||||
|
||||
**Verdict:** Ready to release
|
||||
|
||||
**Evidence:**
|
||||
- P0 coverage: 100% (5/5 requirements)
|
||||
- P1 coverage: 100% (6/6 requirements)
|
||||
- P2 coverage: 33% (1/3 requirements) - acceptable
|
||||
- Test quality score: 84/100
|
||||
- NFR assessment: PASS
|
||||
|
||||
## Coverage Analysis
|
||||
|
||||
| Priority | Required Coverage | Actual Coverage | Status |
|
||||
| -------- | ----------------- | --------------- | --------------------- |
|
||||
| **P0** | 100% | 100% | ✅ PASS |
|
||||
| **P1** | 90% | 100% | ✅ PASS |
|
||||
| **P2** | 50% | 33% | ⚠️ Below (acceptable) |
|
||||
| **P3** | 20% | 0% | ✅ PASS (low priority) |
|
||||
|
||||
**Rationale:**
|
||||
- All critical path (P0) requirements fully tested
|
||||
- All high-value (P1) requirements fully tested
|
||||
- P2 gap (profile export) is low risk and deferred to next release
|
||||
|
||||
## Quality Metrics
|
||||
|
||||
| Metric | Threshold | Actual | Status |
|
||||
| ------------------ | --------- | ------ | ------ |
|
||||
| P0/P1 Coverage | >95% | 100% | ✅ |
|
||||
| Test Quality Score | >80 | 84 | ✅ |
|
||||
| NFR Status | PASS | PASS | ✅ |
|
||||
|
||||
## Risks and Mitigations
|
||||
|
||||
### Accepted Risks
|
||||
|
||||
**Risk 1: Profile export not tested (P2)**
|
||||
- **Impact:** Medium (users can't export profile)
|
||||
- **Mitigation:** Feature flag disabled by default
|
||||
- **Plan:** Add tests in v1.3 release (February)
|
||||
- **Monitoring:** Track feature flag usage
|
||||
|
||||
## Approvals
|
||||
|
||||
- [x] **Product Manager** - Business requirements met (Approved: 2026-01-13)
|
||||
- [x] **Tech Lead** - Technical quality acceptable (Approved: 2026-01-13)
|
||||
- [x] **QA Lead** - Test coverage sufficient (Approved: 2026-01-13)
|
||||
|
||||
## Next Steps
|
||||
|
||||
### Deployment
|
||||
1. Merge to main branch
|
||||
2. Deploy to staging
|
||||
3. Run smoke tests in staging
|
||||
4. Deploy to production
|
||||
5. Monitor for 24 hours
|
||||
|
||||
### Monitoring
|
||||
- Set alerts for profile endpoint (P99 > 200ms)
|
||||
- Track error rates (target: <0.1%)
|
||||
- Monitor profile export feature flag usage
|
||||
|
||||
### Future Work
|
||||
- Add profile export tests (v1.3)
|
||||
- Expand P2 coverage to 50%
|
||||
```
|
||||
|
||||
### Gate Decision Rules
|
||||
|
||||
TEA uses deterministic rules when decision_mode = "deterministic":
|
||||
|
||||
| P0 Coverage | P1 Coverage | Overall Coverage | Decision |
|
||||
| ----------- | ----------- | ---------------- | ---------------------------- |
|
||||
| 100% | ≥90% | ≥80% | **PASS** ✅ |
|
||||
| 100% | 80-89% | ≥80% | **CONCERNS** ⚠️ |
|
||||
| <100% | Any | Any | **FAIL** ❌ |
|
||||
| Any | <80% | Any | **FAIL** ❌ |
|
||||
| Any | Any | <80% | **FAIL** ❌ |
|
||||
| Any | Any | Any | **WAIVED** ⏭️ (with approval) |
|
||||
|
||||
**Detailed Rules:**
|
||||
- **PASS:** P0=100%, P1≥90%, Overall≥80%
|
||||
- **CONCERNS:** P0=100%, P1 80-89%, Overall≥80% (below threshold but not critical)
|
||||
- **FAIL:** P0<100% OR P1<80% OR Overall<80% (critical gaps)
|
||||
|
||||
**PASS** ✅: All criteria met, ready to release
|
||||
|
||||
**CONCERNS** ⚠️: Some criteria not met, but:
|
||||
- Mitigation plan exists
|
||||
- Risk is acceptable
|
||||
- Team approves proceeding
|
||||
- Monitoring in place
|
||||
|
||||
**FAIL** ❌: Critical criteria not met:
|
||||
- P0 requirements not tested
|
||||
- Critical security vulnerabilities
|
||||
- System is broken
|
||||
- Cannot deploy
|
||||
|
||||
**WAIVED** ⏭️: Business approves proceeding despite concerns:
|
||||
- Documented business justification
|
||||
- Accepted risks quantified
|
||||
- Approver signatures
|
||||
- Future plans documented
|
||||
|
||||
### Example CONCERNS Decision
|
||||
|
||||
```markdown
|
||||
## Decision Summary
|
||||
|
||||
**Verdict:** CONCERNS ⚠️ - Proceed with monitoring
|
||||
|
||||
**Evidence:**
|
||||
- P0 coverage: 100%
|
||||
- P1 coverage: 85% (below 90% target)
|
||||
- Test quality: 78/100 (below 80 target)
|
||||
|
||||
**Gaps:**
|
||||
- 1 P1 requirement not tested (avatar upload)
|
||||
- Test quality score slightly below threshold
|
||||
|
||||
**Mitigation:**
|
||||
- Avatar upload not critical for v1.2 launch
|
||||
- Test quality issues are minor (no flakiness)
|
||||
- Monitoring alerts configured
|
||||
|
||||
**Approvals:**
|
||||
- Product Manager: APPROVED (business priority to launch)
|
||||
- Tech Lead: APPROVED (technical risk acceptable)
|
||||
```
|
||||
|
||||
### Example FAIL Decision
|
||||
|
||||
```markdown
|
||||
## Decision Summary
|
||||
|
||||
**Verdict:** FAIL ❌ - Cannot release
|
||||
|
||||
**Evidence:**
|
||||
- P0 coverage: 60% (below 95% threshold)
|
||||
- Critical security vulnerability (CVE-2024-12345)
|
||||
- Test quality: 55/100
|
||||
|
||||
**Blockers:**
|
||||
1. **Login flow not tested** (P0 requirement)
|
||||
- Critical path completely untested
|
||||
- Must add E2E and API tests
|
||||
|
||||
2. **SQL injection vulnerability**
|
||||
- Critical security issue
|
||||
- Must fix before deployment
|
||||
|
||||
**Actions Required:**
|
||||
1. Add login tests (QA team, 2 days)
|
||||
2. Fix SQL injection (backend team, 1 day)
|
||||
3. Re-run security scan (DevOps, 1 hour)
|
||||
4. Re-run *trace after fixes
|
||||
|
||||
**Cannot proceed until all blockers resolved.**
|
||||
```
|
||||
|
||||
## What You Get
|
||||
|
||||
### Phase 1: Traceability Matrix
|
||||
- Requirement-to-test mapping
|
||||
- Coverage classification (FULL/PARTIAL/NONE)
|
||||
- Gap identification with priorities
|
||||
- Actionable recommendations
|
||||
|
||||
### Phase 2: Gate Decision
|
||||
- Go/no-go verdict (PASS/CONCERNS/FAIL/WAIVED)
|
||||
- Evidence summary
|
||||
- Approval signatures
|
||||
- Next steps and monitoring plan
|
||||
|
||||
## Usage Patterns
|
||||
|
||||
### Greenfield Projects
|
||||
|
||||
**Phase 3:**
|
||||
```
|
||||
After architecture complete:
|
||||
1. Run *test-design (system-level)
|
||||
2. Run *trace Phase 1 (baseline)
|
||||
3. Use for implementation-readiness gate
|
||||
```
|
||||
|
||||
**Phase 4:**
|
||||
```
|
||||
After each epic/story:
|
||||
1. Run *trace Phase 1 (refresh coverage)
|
||||
2. Identify gaps
|
||||
3. Add missing tests
|
||||
```
|
||||
|
||||
**Release Gate:**
|
||||
```
|
||||
Before deployment:
|
||||
1. Run *trace Phase 1 (final coverage check)
|
||||
2. Run *trace Phase 2 (make gate decision)
|
||||
3. Get approvals
|
||||
4. Deploy (if PASS or WAIVED)
|
||||
```
|
||||
|
||||
### Brownfield Projects
|
||||
|
||||
**Phase 2:**
|
||||
```
|
||||
Before planning new work:
|
||||
1. Run *trace Phase 1 (establish baseline)
|
||||
2. Understand existing coverage
|
||||
3. Plan testing strategy
|
||||
```
|
||||
|
||||
**Phase 4:**
|
||||
```
|
||||
After each epic/story:
|
||||
1. Run *trace Phase 1 (refresh)
|
||||
2. Compare to baseline
|
||||
3. Track coverage improvement
|
||||
```
|
||||
|
||||
**Release Gate:**
|
||||
```
|
||||
Before deployment:
|
||||
1. Run *trace Phase 1 (final check)
|
||||
2. Run *trace Phase 2 (gate decision)
|
||||
3. Compare to baseline
|
||||
4. Deploy if coverage maintained or improved
|
||||
```
|
||||
|
||||
## Tips
|
||||
|
||||
### Run Phase 1 Frequently
|
||||
|
||||
Don't wait until release gate:
|
||||
|
||||
```
|
||||
After Story 1: *trace Phase 1 (identify gaps early)
|
||||
After Story 2: *trace Phase 1 (refresh)
|
||||
After Story 3: *trace Phase 1 (refresh)
|
||||
Before Release: *trace Phase 1 + Phase 2 (final gate)
|
||||
```
|
||||
|
||||
**Benefit:** Catch gaps early when they're cheap to fix.
|
||||
|
||||
### Use Coverage Trends
|
||||
|
||||
Track improvement over time:
|
||||
|
||||
```markdown
|
||||
## Coverage Trend
|
||||
|
||||
| Date | Epic | P0/P1 Coverage | Quality Score | Status |
|
||||
| ---------- | -------- | -------------- | ------------- | -------------- |
|
||||
| 2026-01-01 | Baseline | 45% | - | Starting point |
|
||||
| 2026-01-08 | Epic 1 | 78% | 72 | Improving |
|
||||
| 2026-01-15 | Epic 2 | 92% | 84 | Near target |
|
||||
| 2026-01-20 | Epic 3 | 100% | 88 | Ready! |
|
||||
```
|
||||
|
||||
### Set Coverage Targets by Priority
|
||||
|
||||
Don't aim for 100% across all priorities:
|
||||
|
||||
**Recommended Targets:**
|
||||
- **P0:** 100% (critical path must be tested)
|
||||
- **P1:** 90% (high-value scenarios)
|
||||
- **P2:** 50% (nice-to-have features)
|
||||
- **P3:** 20% (low-value edge cases)
|
||||
|
||||
### Use Classification Strategically
|
||||
|
||||
**FULL** ✅: Requirement completely tested
|
||||
- E2E test covers full user workflow
|
||||
- API test validates backend behavior
|
||||
- All acceptance criteria covered
|
||||
|
||||
**PARTIAL** ⚠️: Some aspects tested
|
||||
- E2E test exists but missing scenarios
|
||||
- API test exists but incomplete
|
||||
- Some acceptance criteria not covered
|
||||
|
||||
**NONE** ❌: No tests exist
|
||||
- Requirement identified but not tested
|
||||
- May be intentional (low priority) or oversight
|
||||
|
||||
**Classification helps prioritize:**
|
||||
- Fix NONE coverage for P0/P1 requirements first
|
||||
- Enhance PARTIAL coverage for P0 requirements
|
||||
- Accept PARTIAL or NONE for P2/P3 if time-constrained
|
||||
|
||||
### Automate Gate Decisions
|
||||
|
||||
Use traceability in CI:
|
||||
|
||||
```yaml
|
||||
# .github/workflows/gate-check.yml
|
||||
- name: Check coverage
|
||||
run: |
|
||||
# Run trace Phase 1
|
||||
# Parse coverage percentages
|
||||
if [ $P0_COVERAGE -lt 95 ]; then
|
||||
echo "P0 coverage below 95%"
|
||||
exit 1
|
||||
fi
|
||||
```
|
||||
|
||||
### Document Waivers Clearly
|
||||
|
||||
If proceeding with WAIVED:
|
||||
|
||||
**Required:**
|
||||
```markdown
|
||||
## Waiver Documentation
|
||||
|
||||
**Waived By:** VP Engineering, Product Lead
|
||||
**Date:** 2026-01-15
|
||||
**Gate Type:** Release Gate v1.2
|
||||
|
||||
**Justification:**
|
||||
Business critical to launch by Q1 for investor demo.
|
||||
Performance concerns acceptable for initial user base.
|
||||
|
||||
**Conditions:**
|
||||
- Set monitoring alerts for P99 > 300ms
|
||||
- Plan optimization for v1.3 (due February 28)
|
||||
- Monitor user feedback closely
|
||||
|
||||
**Accepted Risks:**
|
||||
- 1% of users may experience 350ms latency
|
||||
- Avatar upload feature incomplete
|
||||
- Profile export deferred to next release
|
||||
|
||||
**Quantified Impact:**
|
||||
- Affects <100 users at current scale
|
||||
- Workaround exists (manual export)
|
||||
- Monitoring will catch issues early
|
||||
|
||||
**Approvals:**
|
||||
- VP Engineering: [Signature] Date: 2026-01-15
|
||||
- Product Lead: [Signature] Date: 2026-01-15
|
||||
- QA Lead: [Signature] Date: 2026-01-15
|
||||
```
|
||||
|
||||
## Common Issues
|
||||
|
||||
### Too Many Gaps to Fix
|
||||
|
||||
**Problem:** Phase 1 shows 50 uncovered requirements.
|
||||
|
||||
**Solution:** Prioritize ruthlessly:
|
||||
1. Fix all P0 gaps (critical path)
|
||||
2. Fix high-risk P1 gaps
|
||||
3. Accept low-risk P1 gaps with mitigation
|
||||
4. Defer all P2/P3 gaps
|
||||
|
||||
**Don't try to fix everything** - focus on what matters for release.
|
||||
|
||||
### Can't Find Test Coverage
|
||||
|
||||
**Problem:** Tests exist but TEA can't map them to requirements.
|
||||
|
||||
**Cause:** Tests don't reference requirements.
|
||||
|
||||
**Solution:** Add traceability comments:
|
||||
```typescript
|
||||
test('should display profile', async ({ page }) => {
|
||||
// Covers: Requirement 1 - User can view profile
|
||||
// Acceptance criteria: Navigate to /profile, see name/email
|
||||
await page.goto('/profile');
|
||||
await expect(page.getByText('Test User')).toBeVisible();
|
||||
});
|
||||
```
|
||||
|
||||
Or use test IDs:
|
||||
```typescript
|
||||
test('[REQ-1] should display profile', async ({ page }) => {
|
||||
// Test code...
|
||||
});
|
||||
```
|
||||
|
||||
### Unclear What "FULL" vs "PARTIAL" Means
|
||||
|
||||
**FULL** ✅: All acceptance criteria tested
|
||||
```
|
||||
Requirement: User can edit profile
|
||||
Acceptance criteria:
|
||||
- Can modify name ✅ Tested
|
||||
- Can modify email ✅ Tested
|
||||
- Can upload avatar ✅ Tested
|
||||
- Changes persist ✅ Tested
|
||||
Result: FULL coverage
|
||||
```
|
||||
|
||||
**PARTIAL** ⚠️: Some criteria tested, some not
|
||||
```
|
||||
Requirement: User can edit profile
|
||||
Acceptance criteria:
|
||||
- Can modify name ✅ Tested
|
||||
- Can modify email ✅ Tested
|
||||
- Can upload avatar ❌ Not tested
|
||||
- Changes persist ✅ Tested
|
||||
Result: PARTIAL coverage (3/4 criteria)
|
||||
```
|
||||
|
||||
### Gate Decision Unclear
|
||||
|
||||
**Problem:** Not sure if PASS or CONCERNS is appropriate.
|
||||
|
||||
**Guideline:**
|
||||
|
||||
**Use PASS** ✅ if:
|
||||
- All P0 requirements 100% covered
|
||||
- P1 requirements >90% covered
|
||||
- No critical issues
|
||||
- NFRs met
|
||||
|
||||
**Use CONCERNS** ⚠️ if:
|
||||
- P1 coverage 85-90% (close to threshold)
|
||||
- Minor quality issues (score 70-79)
|
||||
- NFRs have mitigation plans
|
||||
- Team agrees risk is acceptable
|
||||
|
||||
**Use FAIL** ❌ if:
|
||||
- P0 coverage <100% (critical path gaps)
|
||||
- P1 coverage <85%
|
||||
- Critical security/performance issues
|
||||
- No mitigation possible
|
||||
|
||||
**When in doubt, use CONCERNS** and document the risk.
|
||||
|
||||
## Related Guides
|
||||
|
||||
- [How to Run Test Design](/docs/how-to/workflows/run-test-design.md) - Provides requirements for traceability
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Quality scores feed gate
|
||||
- [How to Run NFR Assessment](/docs/how-to/workflows/run-nfr-assess.md) - NFR status feeds gate
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Risk-Based Testing](/docs/explanation/tea/risk-based-testing.md) - Why P0 vs P3 matters
|
||||
- [TEA Overview](/docs/explanation/features/tea-overview.md) - Gate decisions in context
|
||||
|
||||
## Reference
|
||||
|
||||
- [Command: *trace](/docs/reference/tea/commands.md#trace) - Full command reference
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - Config options
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
712
docs/how-to/workflows/setup-ci.md
Normal file
712
docs/how-to/workflows/setup-ci.md
Normal file
@@ -0,0 +1,712 @@
|
||||
---
|
||||
title: "How to Set Up CI Pipeline with TEA"
|
||||
description: Configure automated test execution with selective testing and burn-in loops using TEA
|
||||
---
|
||||
|
||||
# How to Set Up CI Pipeline with TEA
|
||||
|
||||
Use TEA's `*ci` workflow to scaffold production-ready CI/CD configuration for automated test execution with selective testing, parallel sharding, and flakiness detection.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Need to automate test execution in CI/CD
|
||||
- Want selective testing (only run affected tests)
|
||||
- Need parallel execution for faster feedback
|
||||
- Want burn-in loops for flakiness detection
|
||||
- Setting up new CI/CD pipeline
|
||||
- Optimizing existing CI/CD workflow
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- BMad Method installed
|
||||
- TEA agent available
|
||||
- Test framework configured (run `*framework` first)
|
||||
- Tests written (have something to run in CI)
|
||||
- CI/CD platform access (GitHub Actions, GitLab CI, etc.)
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load TEA Agent
|
||||
|
||||
Start a fresh chat and load TEA:
|
||||
|
||||
```
|
||||
*tea
|
||||
```
|
||||
|
||||
### 2. Run the CI Workflow
|
||||
|
||||
```
|
||||
*ci
|
||||
```
|
||||
|
||||
### 3. Select CI/CD Platform
|
||||
|
||||
TEA will ask which platform you're using.
|
||||
|
||||
**Supported Platforms:**
|
||||
- **GitHub Actions** (most common)
|
||||
- **GitLab CI**
|
||||
- **Circle CI**
|
||||
- **Jenkins**
|
||||
- **Other** (TEA provides generic template)
|
||||
|
||||
**Example:**
|
||||
```
|
||||
GitHub Actions
|
||||
```
|
||||
|
||||
### 4. Configure Test Strategy
|
||||
|
||||
TEA will ask about your test execution strategy.
|
||||
|
||||
#### Repository Structure
|
||||
|
||||
**Question:** "What's your repository structure?"
|
||||
|
||||
**Options:**
|
||||
- **Single app** - One application in root
|
||||
- **Monorepo** - Multiple apps/packages
|
||||
- **Monorepo with affected detection** - Only test changed packages
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Monorepo with multiple apps
|
||||
Need selective testing for changed packages only
|
||||
```
|
||||
|
||||
#### Parallel Execution
|
||||
|
||||
**Question:** "Want to shard tests for parallel execution?"
|
||||
|
||||
**Options:**
|
||||
- **No sharding** - Run tests sequentially
|
||||
- **Shard by workers** - Split across N workers
|
||||
- **Shard by file** - Each file runs in parallel
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Yes, shard across 4 workers for faster execution
|
||||
```
|
||||
|
||||
**Why Shard?**
|
||||
- **4 workers:** 20-minute suite → 5 minutes
|
||||
- **Better resource usage:** Utilize CI runners efficiently
|
||||
- **Faster feedback:** Developers wait less
|
||||
|
||||
#### Burn-In Loops
|
||||
|
||||
**Question:** "Want burn-in loops for flakiness detection?"
|
||||
|
||||
**Options:**
|
||||
- **No burn-in** - Run tests once
|
||||
- **PR burn-in** - Run tests multiple times on PRs
|
||||
- **Nightly burn-in** - Dedicated flakiness detection job
|
||||
|
||||
**Example:**
|
||||
```
|
||||
Yes, run tests 5 times on PRs to catch flaky tests early
|
||||
```
|
||||
|
||||
**Why Burn-In?**
|
||||
- Catches flaky tests before they merge
|
||||
- Prevents intermittent CI failures
|
||||
- Builds confidence in test suite
|
||||
|
||||
### 5. Review Generated CI Configuration
|
||||
|
||||
TEA generates platform-specific workflow files.
|
||||
|
||||
#### GitHub Actions (`.github/workflows/test.yml`):
|
||||
|
||||
```yaml
|
||||
name: Test Suite
|
||||
|
||||
on:
|
||||
pull_request:
|
||||
push:
|
||||
branches: [main, develop]
|
||||
schedule:
|
||||
- cron: '0 2 * * *' # Nightly at 2 AM
|
||||
|
||||
jobs:
|
||||
# Main test job with sharding
|
||||
test:
|
||||
name: Test (Shard ${{ matrix.shard }})
|
||||
runs-on: ubuntu-latest
|
||||
timeout-minutes: 15
|
||||
|
||||
strategy:
|
||||
fail-fast: false
|
||||
matrix:
|
||||
shard: [1, 2, 3, 4]
|
||||
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Setup Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version-file: '.nvmrc'
|
||||
cache: 'npm'
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Install Playwright browsers
|
||||
run: npx playwright install --with-deps
|
||||
|
||||
- name: Run tests
|
||||
run: npx playwright test --shard=${{ matrix.shard }}/4
|
||||
|
||||
- name: Upload test results
|
||||
if: always()
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: test-results-${{ matrix.shard }}
|
||||
path: test-results/
|
||||
retention-days: 7
|
||||
|
||||
- name: Upload test report
|
||||
if: always()
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: playwright-report-${{ matrix.shard }}
|
||||
path: playwright-report/
|
||||
retention-days: 7
|
||||
|
||||
# Burn-in job for flakiness detection (PRs only)
|
||||
burn-in:
|
||||
name: Burn-In (Flakiness Detection)
|
||||
runs-on: ubuntu-latest
|
||||
if: github.event_name == 'pull_request'
|
||||
timeout-minutes: 30
|
||||
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Setup Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version-file: '.nvmrc'
|
||||
cache: 'npm'
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Install Playwright browsers
|
||||
run: npx playwright install --with-deps
|
||||
|
||||
- name: Run burn-in loop
|
||||
run: |
|
||||
for i in {1..5}; do
|
||||
echo "=== Burn-in iteration $i/5 ==="
|
||||
npx playwright test --grep-invert "@skip" || exit 1
|
||||
done
|
||||
|
||||
- name: Upload burn-in results
|
||||
if: failure()
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: burn-in-failures
|
||||
path: test-results/
|
||||
|
||||
# Selective testing (changed files only)
|
||||
selective:
|
||||
name: Selective Tests
|
||||
runs-on: ubuntu-latest
|
||||
if: github.event_name == 'pull_request'
|
||||
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 0 # Full history for git diff
|
||||
|
||||
- name: Setup Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version-file: '.nvmrc'
|
||||
cache: 'npm'
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Install Playwright browsers
|
||||
run: npx playwright install --with-deps
|
||||
|
||||
- name: Run selective tests
|
||||
run: npm run test:changed
|
||||
```
|
||||
|
||||
#### GitLab CI (`.gitlab-ci.yml`):
|
||||
|
||||
```yaml
|
||||
variables:
|
||||
NODE_VERSION: "18"
|
||||
|
||||
stages:
|
||||
- test
|
||||
- burn-in
|
||||
|
||||
# Test job with parallel execution
|
||||
test:
|
||||
stage: test
|
||||
image: node:$NODE_VERSION
|
||||
parallel: 4
|
||||
script:
|
||||
- npm ci
|
||||
- npx playwright install --with-deps
|
||||
- npx playwright test --shard=$CI_NODE_INDEX/$CI_NODE_TOTAL
|
||||
artifacts:
|
||||
when: always
|
||||
paths:
|
||||
- test-results/
|
||||
- playwright-report/
|
||||
expire_in: 7 days
|
||||
rules:
|
||||
- if: $CI_PIPELINE_SOURCE == "merge_request_event"
|
||||
- if: $CI_COMMIT_BRANCH == $CI_DEFAULT_BRANCH
|
||||
|
||||
# Burn-in job for flakiness detection
|
||||
burn-in:
|
||||
stage: burn-in
|
||||
image: node:$NODE_VERSION
|
||||
script:
|
||||
- npm ci
|
||||
- npx playwright install --with-deps
|
||||
- |
|
||||
for i in {1..5}; do
|
||||
echo "=== Burn-in iteration $i/5 ==="
|
||||
npx playwright test || exit 1
|
||||
done
|
||||
artifacts:
|
||||
when: on_failure
|
||||
paths:
|
||||
- test-results/
|
||||
rules:
|
||||
- if: $CI_PIPELINE_SOURCE == "merge_request_event"
|
||||
```
|
||||
|
||||
#### Burn-In Testing
|
||||
|
||||
**Option 1: Classic Burn-In (Playwright Built-In)**
|
||||
|
||||
```json
|
||||
{
|
||||
"scripts": {
|
||||
"test": "playwright test",
|
||||
"test:burn-in": "playwright test --repeat-each=5 --retries=0"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**How it works:**
|
||||
- Runs every test 5 times
|
||||
- Fails if any iteration fails
|
||||
- Detects flakiness before merge
|
||||
|
||||
**Use when:** Small test suite, want to run everything multiple times
|
||||
|
||||
---
|
||||
|
||||
**Option 2: Smart Burn-In (Playwright Utils)**
|
||||
|
||||
If `tea_use_playwright_utils: true`:
|
||||
|
||||
**scripts/burn-in-changed.ts:**
|
||||
```typescript
|
||||
import { runBurnIn } from '@seontechnologies/playwright-utils/burn-in';
|
||||
|
||||
await runBurnIn({
|
||||
configPath: 'playwright.burn-in.config.ts',
|
||||
baseBranch: 'main'
|
||||
});
|
||||
```
|
||||
|
||||
**playwright.burn-in.config.ts:**
|
||||
```typescript
|
||||
import type { BurnInConfig } from '@seontechnologies/playwright-utils/burn-in';
|
||||
|
||||
const config: BurnInConfig = {
|
||||
skipBurnInPatterns: ['**/config/**', '**/*.md', '**/*types*'],
|
||||
burnInTestPercentage: 0.3,
|
||||
burnIn: { repeatEach: 5, retries: 0 }
|
||||
};
|
||||
|
||||
export default config;
|
||||
```
|
||||
|
||||
**package.json:**
|
||||
```json
|
||||
{
|
||||
"scripts": {
|
||||
"test:burn-in": "tsx scripts/burn-in-changed.ts"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**How it works:**
|
||||
- Git diff analysis (only affected tests)
|
||||
- Smart filtering (skip configs, docs, types)
|
||||
- Volume control (run 30% of affected tests)
|
||||
- Each test runs 5 times
|
||||
|
||||
**Use when:** Large test suite, want intelligent selection
|
||||
|
||||
---
|
||||
|
||||
**Comparison:**
|
||||
|
||||
| Feature | Classic Burn-In | Smart Burn-In (PW-Utils) |
|
||||
|---------|----------------|--------------------------|
|
||||
| Changed 1 file | Runs all 500 tests × 5 = 2500 runs | Runs 3 affected tests × 5 = 15 runs |
|
||||
| Config change | Runs all tests | Skips (no tests affected) |
|
||||
| Type change | Runs all tests | Skips (no runtime impact) |
|
||||
| Setup | Zero config | Requires config file |
|
||||
|
||||
**Recommendation:** Start with classic (simple), upgrade to smart (faster) when suite grows.
|
||||
|
||||
### 6. Configure Secrets
|
||||
|
||||
TEA provides a secrets checklist.
|
||||
|
||||
**Required Secrets** (add to CI/CD platform):
|
||||
|
||||
```markdown
|
||||
## GitHub Actions Secrets
|
||||
|
||||
Repository Settings → Secrets and variables → Actions
|
||||
|
||||
### Required
|
||||
- None (tests run without external auth)
|
||||
|
||||
### Optional
|
||||
- `TEST_USER_EMAIL` - Test user credentials
|
||||
- `TEST_USER_PASSWORD` - Test user password
|
||||
- `API_BASE_URL` - API endpoint for tests
|
||||
- `DATABASE_URL` - Test database (if needed)
|
||||
```
|
||||
|
||||
**How to Add Secrets:**
|
||||
|
||||
**GitHub Actions:**
|
||||
1. Go to repo Settings → Secrets → Actions
|
||||
2. Click "New repository secret"
|
||||
3. Add name and value
|
||||
4. Use in workflow: `${{ secrets.TEST_USER_EMAIL }}`
|
||||
|
||||
**GitLab CI:**
|
||||
1. Go to Project Settings → CI/CD → Variables
|
||||
2. Add variable name and value
|
||||
3. Use in workflow: `$TEST_USER_EMAIL`
|
||||
|
||||
### 7. Test the CI Pipeline
|
||||
|
||||
#### Push and Verify
|
||||
|
||||
**Commit the workflow file:**
|
||||
```bash
|
||||
git add .github/workflows/test.yml
|
||||
git commit -m "ci: add automated test pipeline"
|
||||
git push
|
||||
```
|
||||
|
||||
**Watch the CI run:**
|
||||
- GitHub Actions: Go to Actions tab
|
||||
- GitLab CI: Go to CI/CD → Pipelines
|
||||
- Circle CI: Go to Pipelines
|
||||
|
||||
**Expected Result:**
|
||||
```
|
||||
✓ test (shard 1/4) - 3m 24s
|
||||
✓ test (shard 2/4) - 3m 18s
|
||||
✓ test (shard 3/4) - 3m 31s
|
||||
✓ test (shard 4/4) - 3m 15s
|
||||
✓ burn-in - 15m 42s
|
||||
```
|
||||
|
||||
#### Test on Pull Request
|
||||
|
||||
**Create test PR:**
|
||||
```bash
|
||||
git checkout -b test-ci-setup
|
||||
echo "# Test" > test.md
|
||||
git add test.md
|
||||
git commit -m "test: verify CI setup"
|
||||
git push -u origin test-ci-setup
|
||||
```
|
||||
|
||||
**Open PR and verify:**
|
||||
- Tests run automatically
|
||||
- Burn-in runs (if configured for PRs)
|
||||
- Selective tests run (if applicable)
|
||||
- All checks pass ✓
|
||||
|
||||
## What You Get
|
||||
|
||||
### Automated Test Execution
|
||||
- **On every PR** - Catch issues before merge
|
||||
- **On every push to main** - Protect production
|
||||
- **Nightly** - Comprehensive regression testing
|
||||
|
||||
### Parallel Execution
|
||||
- **4x faster feedback** - Shard across multiple workers
|
||||
- **Efficient resource usage** - Maximize CI runner utilization
|
||||
|
||||
### Selective Testing
|
||||
- **Run only affected tests** - Git diff-based selection
|
||||
- **Faster PR feedback** - Don't run entire suite every time
|
||||
|
||||
### Flakiness Detection
|
||||
- **Burn-in loops** - Run tests multiple times
|
||||
- **Early detection** - Catch flaky tests in PRs
|
||||
- **Confidence building** - Know tests are reliable
|
||||
|
||||
### Artifact Collection
|
||||
- **Test results** - Saved for 7 days
|
||||
- **Screenshots** - On test failures
|
||||
- **Videos** - Full test recordings
|
||||
- **Traces** - Playwright trace files for debugging
|
||||
|
||||
## Tips
|
||||
|
||||
### Start Simple, Add Complexity
|
||||
|
||||
**Week 1:** Basic pipeline
|
||||
```yaml
|
||||
- Run tests on PR
|
||||
- Single worker (no sharding)
|
||||
```
|
||||
|
||||
**Week 2:** Add parallelization
|
||||
```yaml
|
||||
- Shard across 4 workers
|
||||
- Faster feedback
|
||||
```
|
||||
|
||||
**Week 3:** Add selective testing
|
||||
```yaml
|
||||
- Git diff-based selection
|
||||
- Skip unaffected tests
|
||||
```
|
||||
|
||||
**Week 4:** Add burn-in
|
||||
```yaml
|
||||
- Detect flaky tests
|
||||
- Run on PR and nightly
|
||||
```
|
||||
|
||||
### Optimize for Feedback Speed
|
||||
|
||||
**Goal:** PR feedback in < 5 minutes
|
||||
|
||||
**Strategies:**
|
||||
- Shard tests across workers (4 workers = 4x faster)
|
||||
- Use selective testing (run 20% of tests, not 100%)
|
||||
- Cache dependencies (`actions/cache`, `cache: 'npm'`)
|
||||
- Run smoke tests first, full suite after
|
||||
|
||||
**Example fast workflow:**
|
||||
```yaml
|
||||
jobs:
|
||||
smoke:
|
||||
# Run critical path tests (2 min)
|
||||
run: npm run test:smoke
|
||||
|
||||
full:
|
||||
needs: smoke
|
||||
# Run full suite only if smoke passes (10 min)
|
||||
run: npm test
|
||||
```
|
||||
|
||||
### Use Test Tags
|
||||
|
||||
Tag tests for selective execution:
|
||||
|
||||
```typescript
|
||||
// Critical path tests (always run)
|
||||
test('@critical should login', async ({ page }) => { });
|
||||
|
||||
// Smoke tests (run first)
|
||||
test('@smoke should load homepage', async ({ page }) => { });
|
||||
|
||||
// Slow tests (run nightly only)
|
||||
test('@slow should process large file', async ({ page }) => { });
|
||||
|
||||
// Skip in CI
|
||||
test('@local-only should use local service', async ({ page }) => { });
|
||||
```
|
||||
|
||||
**In CI:**
|
||||
```bash
|
||||
# PR: Run critical and smoke only
|
||||
npx playwright test --grep "@critical|@smoke"
|
||||
|
||||
# Nightly: Run everything except local-only
|
||||
npx playwright test --grep-invert "@local-only"
|
||||
```
|
||||
|
||||
### Monitor CI Performance
|
||||
|
||||
Track metrics:
|
||||
|
||||
```markdown
|
||||
## CI Metrics
|
||||
|
||||
| Metric | Target | Current | Status |
|
||||
|--------|--------|---------|--------|
|
||||
| PR feedback time | < 5 min | 3m 24s | ✅ |
|
||||
| Full suite time | < 15 min | 12m 18s | ✅ |
|
||||
| Flakiness rate | < 1% | 0.3% | ✅ |
|
||||
| CI cost/month | < $100 | $75 | ✅ |
|
||||
```
|
||||
|
||||
### Handle Flaky Tests
|
||||
|
||||
When burn-in detects flakiness:
|
||||
|
||||
1. **Quarantine flaky test:**
|
||||
```typescript
|
||||
test.skip('flaky test - investigating', async ({ page }) => {
|
||||
// TODO: Fix flakiness
|
||||
});
|
||||
```
|
||||
|
||||
2. **Investigate with trace viewer:**
|
||||
```bash
|
||||
npx playwright show-trace test-results/trace.zip
|
||||
```
|
||||
|
||||
3. **Fix root cause:**
|
||||
- Add network-first patterns
|
||||
- Remove hard waits
|
||||
- Fix race conditions
|
||||
|
||||
4. **Verify fix:**
|
||||
```bash
|
||||
npm run test:burn-in -- tests/flaky.spec.ts --repeat 20
|
||||
```
|
||||
|
||||
### Secure Secrets
|
||||
|
||||
**Don't commit secrets to code:**
|
||||
```yaml
|
||||
# ❌ Bad
|
||||
- run: API_KEY=sk-1234... npm test
|
||||
|
||||
# ✅ Good
|
||||
- run: npm test
|
||||
env:
|
||||
API_KEY: ${{ secrets.API_KEY }}
|
||||
```
|
||||
|
||||
**Use environment-specific secrets:**
|
||||
- `STAGING_API_URL`
|
||||
- `PROD_API_URL`
|
||||
- `TEST_API_URL`
|
||||
|
||||
### Cache Aggressively
|
||||
|
||||
Speed up CI with caching:
|
||||
|
||||
```yaml
|
||||
# Cache node_modules
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
cache: 'npm'
|
||||
|
||||
# Cache Playwright browsers
|
||||
- name: Cache Playwright browsers
|
||||
uses: actions/cache@v4
|
||||
with:
|
||||
path: ~/.cache/ms-playwright
|
||||
key: playwright-${{ hashFiles('package-lock.json') }}
|
||||
```
|
||||
|
||||
## Common Issues
|
||||
|
||||
### Tests Pass Locally, Fail in CI
|
||||
|
||||
**Symptoms:**
|
||||
- Green locally, red in CI
|
||||
- "Works on my machine"
|
||||
|
||||
**Common Causes:**
|
||||
- Different Node version
|
||||
- Different browser version
|
||||
- Missing environment variables
|
||||
- Timezone differences
|
||||
- Race conditions (CI slower)
|
||||
|
||||
**Solutions:**
|
||||
```yaml
|
||||
# Pin Node version
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version-file: '.nvmrc'
|
||||
|
||||
# Pin browser versions
|
||||
- run: npx playwright install --with-deps chromium@1.40.0
|
||||
|
||||
# Set timezone
|
||||
env:
|
||||
TZ: 'America/New_York'
|
||||
```
|
||||
|
||||
### CI Takes Too Long
|
||||
|
||||
**Problem:** CI takes 30+ minutes, developers wait too long.
|
||||
|
||||
**Solutions:**
|
||||
1. **Shard tests:** 4 workers = 4x faster
|
||||
2. **Selective testing:** Only run affected tests on PR
|
||||
3. **Smoke tests first:** Run critical path (2 min), full suite after
|
||||
4. **Cache dependencies:** `npm ci` with cache
|
||||
5. **Optimize tests:** Remove slow tests, hard waits
|
||||
|
||||
### Burn-In Always Fails
|
||||
|
||||
**Problem:** Burn-in job fails every time.
|
||||
|
||||
**Cause:** Test suite is flaky.
|
||||
|
||||
**Solution:**
|
||||
1. Identify flaky tests (check which iteration fails)
|
||||
2. Fix flaky tests using `*test-review`
|
||||
3. Re-run burn-in on specific files:
|
||||
```bash
|
||||
npm run test:burn-in tests/flaky.spec.ts
|
||||
```
|
||||
|
||||
### Out of CI Minutes
|
||||
|
||||
**Problem:** Using too many CI minutes, hitting plan limit.
|
||||
|
||||
**Solutions:**
|
||||
1. Run full suite only on main branch
|
||||
2. Use selective testing on PRs
|
||||
3. Run expensive tests nightly only
|
||||
4. Self-host runners (for GitHub Actions)
|
||||
|
||||
## Related Guides
|
||||
|
||||
- [How to Set Up Test Framework](/docs/how-to/workflows/setup-test-framework.md) - Run first
|
||||
- [How to Run Test Review](/docs/how-to/workflows/run-test-review.md) - Audit CI tests
|
||||
- [Integrate Playwright Utils](/docs/how-to/customization/integrate-playwright-utils.md) - Burn-in utility
|
||||
|
||||
## Understanding the Concepts
|
||||
|
||||
- [Test Quality Standards](/docs/explanation/tea/test-quality-standards.md) - Why determinism matters
|
||||
- [Network-First Patterns](/docs/explanation/tea/network-first-patterns.md) - Avoid CI flakiness
|
||||
|
||||
## Reference
|
||||
|
||||
- [Command: *ci](/docs/reference/tea/commands.md#ci) - Full command reference
|
||||
- [TEA Configuration](/docs/reference/tea/configuration.md) - CI-related config options
|
||||
|
||||
---
|
||||
|
||||
Generated with [BMad Method](https://bmad-method.org) - TEA (Test Architect)
|
||||
89
docs/how-to/workflows/setup-party-mode.md
Normal file
89
docs/how-to/workflows/setup-party-mode.md
Normal file
@@ -0,0 +1,89 @@
|
||||
---
|
||||
title: "How to Set Up Party Mode"
|
||||
description: How to set up and use Party Mode for multi-agent collaboration
|
||||
---
|
||||
|
||||
Use Party Mode to orchestrate dynamic multi-agent conversations with your entire BMad team.
|
||||
|
||||
## When to Use This
|
||||
|
||||
- Exploring complex topics that benefit from diverse expert perspectives
|
||||
- Brainstorming with agents who can build on each other's ideas
|
||||
- Getting comprehensive views across multiple domains
|
||||
- Strategic decisions with trade-offs
|
||||
|
||||
:::note[Prerequisites]
|
||||
- BMad Method installed with multiple agents
|
||||
- Any agent loaded that supports party mode
|
||||
:::
|
||||
|
||||
## Steps
|
||||
|
||||
### 1. Load Any Agent
|
||||
|
||||
Start with any agent that supports party mode (most do).
|
||||
|
||||
### 2. Start Party Mode
|
||||
|
||||
```
|
||||
*party-mode
|
||||
```
|
||||
|
||||
Or use the full path:
|
||||
```
|
||||
/bmad:core:workflows:party-mode
|
||||
```
|
||||
|
||||
### 3. Introduce Your Topic
|
||||
|
||||
Present a topic or question for the group to discuss:
|
||||
|
||||
```
|
||||
I'm trying to decide between a monolithic architecture
|
||||
and microservices for our new platform.
|
||||
```
|
||||
|
||||
### 4. Engage with the Discussion
|
||||
|
||||
The facilitator will:
|
||||
- Select 2-3 most relevant agents based on expertise
|
||||
- Let agents respond in character
|
||||
- Enable natural cross-talk and debate
|
||||
- Continue until you choose to exit
|
||||
|
||||
### 5. Exit When Ready
|
||||
|
||||
Type "exit" or "done" to conclude the session. Participating agents will say personalized farewells.
|
||||
|
||||
## What Happens
|
||||
|
||||
1. **Agent Roster** — Party Mode loads your complete agent roster
|
||||
2. **Introduction** — Available team members are introduced
|
||||
3. **Topic Analysis** — The facilitator analyzes your topic
|
||||
4. **Agent Selection** — 2-3 most relevant agents are selected
|
||||
5. **Discussion** — Agents respond, reference each other, engage in cross-talk
|
||||
6. **Exit** — Session concludes with farewells
|
||||
|
||||
## Example Party Compositions
|
||||
|
||||
| Topic | Typical Agents |
|
||||
| ---------------------- | ------------------------------------------------------------- |
|
||||
| **Product Strategy** | PM + Innovation Strategist (CIS) + Analyst |
|
||||
| **Technical Design** | Architect + Creative Problem Solver (CIS) + Game Architect |
|
||||
| **User Experience** | UX Designer + Design Thinking Coach (CIS) + Storyteller (CIS) |
|
||||
| **Quality Assessment** | TEA + DEV + Architect |
|
||||
|
||||
## Key Features
|
||||
|
||||
- **Intelligent agent selection** — Selects based on expertise needed
|
||||
- **Authentic personalities** — Each agent maintains their unique voice
|
||||
- **Natural cross-talk** — Agents reference and build on each other
|
||||
- **Graceful exit** — Personalized farewells
|
||||
|
||||
## Tips
|
||||
|
||||
- **Be specific about your topic** — Better agent selection
|
||||
- **Let the conversation flow** — Don't over-direct
|
||||
- **Ask follow-up questions** — Go deeper on interesting points
|
||||
- **Take notes on key insights** — Capture valuable perspectives
|
||||
- **Use for strategic decisions** — Not routine tasks
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user