test: phase 0 - fix failing tests and setup CI/CD

- Fixed 6 failing tests across http-server-auth.test.ts and single-session.test.ts
- All tests now pass (68 passing, 0 failing)
- Added GitHub Actions workflow for automated testing
- Added comprehensive testing documentation and strategy
- Tests fixed without changing application behavior
This commit is contained in:
czlonkowski
2025-07-28 12:04:38 +02:00
parent 5450bc35c3
commit cf960ed2ac
8 changed files with 3008 additions and 89 deletions

211
docs/testing-checklist.md Normal file
View File

@@ -0,0 +1,211 @@
# n8n-MCP Testing Implementation Checklist
## Immediate Actions (Day 1)
- [ ] Install Vitest and remove Jest
- [ ] Create vitest.config.ts
- [ ] Setup global test configuration
- [ ] Migrate existing tests to Vitest syntax
- [ ] Create GitHub Actions workflow file
- [ ] Setup coverage reporting with Codecov
## Week 1: Foundation
### Testing Infrastructure
- [ ] Create test directory structure
- [ ] Setup mock infrastructure for better-sqlite3
- [ ] Create mock for n8n-nodes-base package
- [ ] Setup test database utilities
- [ ] Create factory pattern for nodes
- [ ] Create builder pattern for workflows
- [ ] Setup global test utilities
- [ ] Configure test environment variables
### CI/CD Pipeline
- [ ] GitHub Actions for test execution
- [ ] Coverage reporting integration
- [ ] Performance benchmark tracking
- [ ] Test result artifacts
- [ ] Branch protection rules
- [ ] Required status checks
## Week 2: Mock Infrastructure
### Database Mocking
- [ ] Complete better-sqlite3 mock implementation
- [ ] Mock prepared statements
- [ ] Mock transactions
- [ ] Mock FTS5 search functionality
- [ ] Test data seeding utilities
### External Dependencies
- [ ] Mock axios for API calls
- [ ] Mock file system operations
- [ ] Mock MCP SDK
- [ ] Mock Express server
- [ ] Mock WebSocket connections
## Week 3-4: Unit Tests
### Core Services (Priority 1)
- [ ] `config-validator.ts` - 95% coverage
- [ ] `enhanced-config-validator.ts` - 95% coverage
- [ ] `workflow-validator.ts` - 90% coverage
- [ ] `expression-validator.ts` - 90% coverage
- [ ] `property-filter.ts` - 90% coverage
- [ ] `example-generator.ts` - 85% coverage
### Parsers (Priority 2)
- [ ] `node-parser.ts` - 90% coverage
- [ ] `property-extractor.ts` - 90% coverage
### MCP Layer (Priority 3)
- [ ] `tools.ts` - 90% coverage
- [ ] `handlers-n8n-manager.ts` - 85% coverage
- [ ] `handlers-workflow-diff.ts` - 85% coverage
- [ ] `tools-documentation.ts` - 80% coverage
### Database Layer (Priority 4)
- [ ] `node-repository.ts` - 85% coverage
- [ ] `database-adapter.ts` - 85% coverage
- [ ] `template-repository.ts` - 80% coverage
### Loaders and Mappers (Priority 5)
- [ ] `node-loader.ts` - 85% coverage
- [ ] `docs-mapper.ts` - 80% coverage
## Week 5-6: Integration Tests
### MCP Protocol Tests
- [ ] Full MCP server initialization
- [ ] Tool invocation flow
- [ ] Error handling and recovery
- [ ] Concurrent request handling
- [ ] Session management
### n8n API Integration
- [ ] Workflow CRUD operations
- [ ] Webhook triggering
- [ ] Execution monitoring
- [ ] Authentication handling
- [ ] Error scenarios
### Database Integration
- [ ] SQLite operations with real DB
- [ ] FTS5 search functionality
- [ ] Transaction handling
- [ ] Migration testing
- [ ] Performance under load
## Week 7-8: E2E & Performance
### End-to-End Scenarios
- [ ] Complete workflow creation flow
- [ ] AI agent workflow setup
- [ ] Template import and validation
- [ ] Workflow execution monitoring
- [ ] Error recovery scenarios
### Performance Benchmarks
- [ ] Node loading speed (< 50ms per node)
- [ ] Search performance (< 100ms for 1000 nodes)
- [ ] Validation speed (< 10ms simple, < 100ms complex)
- [ ] Database query performance
- [ ] Memory usage profiling
- [ ] Concurrent request handling
### Load Testing
- [ ] 100 concurrent MCP requests
- [ ] 10,000 nodes in database
- [ ] 1,000 workflow validations/minute
- [ ] Memory leak detection
- [ ] Resource cleanup verification
## Testing Quality Gates
### Coverage Requirements
- [ ] Overall: 80%+
- [ ] Core services: 90%+
- [ ] MCP tools: 90%+
- [ ] Critical paths: 95%+
- [ ] New code: 90%+
### Performance Requirements
- [ ] All unit tests < 10ms
- [ ] Integration tests < 1s
- [ ] E2E tests < 10s
- [ ] Full suite < 5 minutes
- [ ] No memory leaks
### Code Quality
- [ ] No ESLint errors
- [ ] No TypeScript errors
- [ ] No console.log in tests
- [ ] All tests have descriptions
- [ ] No hardcoded values
## Monitoring & Maintenance
### Daily
- [ ] Check CI pipeline status
- [ ] Review failed tests
- [ ] Monitor flaky tests
### Weekly
- [ ] Review coverage reports
- [ ] Update test documentation
- [ ] Performance benchmark review
- [ ] Team sync on testing progress
### Monthly
- [ ] Update baseline benchmarks
- [ ] Review and refactor tests
- [ ] Update testing strategy
- [ ] Training/knowledge sharing
## Risk Mitigation
### Technical Risks
- [ ] Mock complexity - Use simple, maintainable mocks
- [ ] Test brittleness - Focus on behavior, not implementation
- [ ] Performance impact - Run heavy tests in parallel
- [ ] Flaky tests - Proper async handling and isolation
### Process Risks
- [ ] Slow adoption - Provide training and examples
- [ ] Coverage gaming - Review test quality, not just numbers
- [ ] Maintenance burden - Automate what's possible
- [ ] Integration complexity - Use test containers
## Success Criteria
### Technical Metrics
- Coverage: 80%+ overall, 90%+ critical paths
- Performance: All benchmarks within limits
- Reliability: Zero flaky tests
- Speed: CI pipeline < 5 minutes
### Team Metrics
- All developers writing tests
- Tests reviewed in PRs
- No production bugs from tested code
- Improved development velocity
## Resources & Tools
### Documentation
- Vitest: https://vitest.dev/
- Testing Library: https://testing-library.com/
- MSW: https://mswjs.io/
- Testcontainers: https://www.testcontainers.com/
### Monitoring
- Codecov: https://codecov.io/
- GitHub Actions: https://github.com/features/actions
- Benchmark Action: https://github.com/benchmark-action/github-action-benchmark
### Team Resources
- Testing best practices guide
- Example test implementations
- Mock usage patterns
- Performance optimization tips

View File

@@ -0,0 +1,472 @@
# n8n-MCP Testing Implementation Guide
## Phase 1: Foundation Setup (Week 1-2)
### 1.1 Install Vitest and Dependencies
```bash
# Remove Jest
npm uninstall jest ts-jest @types/jest
# Install Vitest and related packages
npm install -D vitest @vitest/ui @vitest/coverage-v8
npm install -D @testing-library/jest-dom
npm install -D msw # For API mocking
npm install -D @faker-js/faker # For test data
npm install -D fishery # For factories
```
### 1.2 Update package.json Scripts
```json
{
"scripts": {
// Testing
"test": "vitest",
"test:ui": "vitest --ui",
"test:unit": "vitest run tests/unit",
"test:integration": "vitest run tests/integration",
"test:e2e": "vitest run tests/e2e",
"test:watch": "vitest watch",
"test:coverage": "vitest run --coverage",
"test:coverage:check": "vitest run --coverage --coverage.thresholdAutoUpdate=false",
// Benchmarks
"bench": "vitest bench",
"bench:compare": "vitest bench --compare",
// CI specific
"test:ci": "vitest run --reporter=junit --reporter=default",
"test:ci:coverage": "vitest run --coverage --reporter=junit --reporter=default"
}
}
```
### 1.3 Migrate Existing Tests
```typescript
// Before (Jest)
import { describe, test, expect } from '@jest/globals';
// After (Vitest)
import { describe, it, expect, vi } from 'vitest';
// Update mock syntax
// Jest: jest.mock('module')
// Vitest: vi.mock('module')
// Update timer mocks
// Jest: jest.useFakeTimers()
// Vitest: vi.useFakeTimers()
```
### 1.4 Create Test Database Setup
```typescript
// tests/setup/test-database.ts
import Database from 'better-sqlite3';
import { readFileSync } from 'fs';
import { join } from 'path';
export class TestDatabase {
private db: Database.Database;
constructor() {
this.db = new Database(':memory:');
this.initialize();
}
private initialize() {
const schema = readFileSync(
join(__dirname, '../../src/database/schema.sql'),
'utf8'
);
this.db.exec(schema);
}
seedNodes(nodes: any[]) {
const stmt = this.db.prepare(`
INSERT INTO nodes (type, displayName, name, group, version, description, properties)
VALUES (?, ?, ?, ?, ?, ?, ?)
`);
const insertMany = this.db.transaction((nodes) => {
for (const node of nodes) {
stmt.run(
node.type,
node.displayName,
node.name,
node.group,
node.version,
node.description,
JSON.stringify(node.properties)
);
}
});
insertMany(nodes);
}
close() {
this.db.close();
}
getDb() {
return this.db;
}
}
```
## Phase 2: Core Unit Tests (Week 3-4)
### 2.1 Test Organization Template
```typescript
// tests/unit/services/[service-name].test.ts
import { describe, it, expect, beforeEach, afterEach, vi } from 'vitest';
import { ServiceName } from '@/services/service-name';
describe('ServiceName', () => {
let service: ServiceName;
let mockDependency: any;
beforeEach(() => {
// Setup mocks
mockDependency = {
method: vi.fn()
};
// Create service instance
service = new ServiceName(mockDependency);
});
afterEach(() => {
vi.clearAllMocks();
});
describe('methodName', () => {
it('should handle happy path', async () => {
// Arrange
const input = { /* test data */ };
mockDependency.method.mockResolvedValue({ /* mock response */ });
// Act
const result = await service.methodName(input);
// Assert
expect(result).toEqual(/* expected output */);
expect(mockDependency.method).toHaveBeenCalledWith(/* expected args */);
});
it('should handle errors gracefully', async () => {
// Arrange
mockDependency.method.mockRejectedValue(new Error('Test error'));
// Act & Assert
await expect(service.methodName({})).rejects.toThrow('Expected error message');
});
});
});
```
### 2.2 Mock Strategies by Layer
#### Database Layer
```typescript
// tests/unit/database/node-repository.test.ts
import { vi } from 'vitest';
vi.mock('better-sqlite3', () => ({
default: vi.fn(() => ({
prepare: vi.fn(() => ({
all: vi.fn(() => mockData),
get: vi.fn((id) => mockData.find(d => d.id === id)),
run: vi.fn(() => ({ changes: 1 }))
})),
exec: vi.fn(),
close: vi.fn()
}))
}));
```
#### External APIs
```typescript
// tests/unit/services/__mocks__/axios.ts
export default {
create: vi.fn(() => ({
get: vi.fn(() => Promise.resolve({ data: {} })),
post: vi.fn(() => Promise.resolve({ data: { id: '123' } })),
put: vi.fn(() => Promise.resolve({ data: {} })),
delete: vi.fn(() => Promise.resolve({ data: {} }))
}))
};
```
#### File System
```typescript
// Use memfs for file system mocking
import { vol } from 'memfs';
vi.mock('fs', () => vol);
beforeEach(() => {
vol.reset();
vol.fromJSON({
'/test/file.json': JSON.stringify({ test: 'data' })
});
});
```
### 2.3 Critical Path Tests
```typescript
// Priority 1: Node Loading and Parsing
// tests/unit/loaders/node-loader.test.ts
// Priority 2: Configuration Validation
// tests/unit/services/config-validator.test.ts
// Priority 3: MCP Tools
// tests/unit/mcp/tools.test.ts
// Priority 4: Database Operations
// tests/unit/database/node-repository.test.ts
// Priority 5: Workflow Validation
// tests/unit/services/workflow-validator.test.ts
```
## Phase 3: Integration Tests (Week 5-6)
### 3.1 Test Container Setup
```typescript
// tests/setup/test-containers.ts
import { GenericContainer, StartedTestContainer } from 'testcontainers';
export class N8nTestContainer {
private container: StartedTestContainer;
async start() {
this.container = await new GenericContainer('n8nio/n8n:latest')
.withExposedPorts(5678)
.withEnv('N8N_BASIC_AUTH_ACTIVE', 'false')
.withEnv('N8N_ENCRYPTION_KEY', 'test-key')
.start();
return {
url: `http://localhost:${this.container.getMappedPort(5678)}`,
stop: () => this.container.stop()
};
}
}
```
### 3.2 Integration Test Pattern
```typescript
// tests/integration/n8n-api/workflow-crud.test.ts
import { N8nTestContainer } from '@tests/setup/test-containers';
import { N8nAPIClient } from '@/services/n8n-api-client';
describe('n8n API Integration', () => {
let container: any;
let apiClient: N8nAPIClient;
beforeAll(async () => {
container = await new N8nTestContainer().start();
apiClient = new N8nAPIClient(container.url);
}, 30000);
afterAll(async () => {
await container.stop();
});
it('should create and retrieve workflow', async () => {
// Create workflow
const workflow = createTestWorkflow();
const created = await apiClient.createWorkflow(workflow);
expect(created.id).toBeDefined();
// Retrieve workflow
const retrieved = await apiClient.getWorkflow(created.id);
expect(retrieved.name).toBe(workflow.name);
});
});
```
## Phase 4: E2E & Performance (Week 7-8)
### 4.1 E2E Test Setup
```typescript
// tests/e2e/workflows/complete-workflow.test.ts
import { MCPClient } from '@tests/utils/mcp-client';
import { N8nTestContainer } from '@tests/setup/test-containers';
describe('Complete Workflow E2E', () => {
let mcpServer: any;
let n8nContainer: any;
let mcpClient: MCPClient;
beforeAll(async () => {
// Start n8n
n8nContainer = await new N8nTestContainer().start();
// Start MCP server
mcpServer = await startMCPServer({
n8nUrl: n8nContainer.url
});
// Create MCP client
mcpClient = new MCPClient(mcpServer.url);
}, 60000);
it('should execute complete workflow creation flow', async () => {
// 1. Search for nodes
const searchResult = await mcpClient.call('search_nodes', {
query: 'webhook http slack'
});
// 2. Get node details
const webhookInfo = await mcpClient.call('get_node_info', {
nodeType: 'nodes-base.webhook'
});
// 3. Create workflow
const workflow = new WorkflowBuilder('E2E Test')
.addWebhookNode()
.addHttpRequestNode()
.addSlackNode()
.connectSequentially()
.build();
// 4. Validate workflow
const validation = await mcpClient.call('validate_workflow', {
workflow
});
expect(validation.isValid).toBe(true);
// 5. Deploy to n8n
const deployed = await mcpClient.call('n8n_create_workflow', {
...workflow
});
expect(deployed.id).toBeDefined();
expect(deployed.active).toBe(false);
});
});
```
### 4.2 Performance Benchmarks
```typescript
// vitest.benchmark.config.ts
export default {
test: {
benchmark: {
// Output benchmark results
outputFile: './benchmark-results.json',
// Compare with baseline
compare: './benchmark-baseline.json',
// Fail if performance degrades by more than 10%
threshold: {
p95: 1.1, // 110% of baseline
p99: 1.2 // 120% of baseline
}
}
}
};
```
## Testing Best Practices
### 1. Test Naming Convention
```typescript
// Format: should [expected behavior] when [condition]
it('should return user data when valid ID is provided')
it('should throw ValidationError when email is invalid')
it('should retry 3 times when network fails')
```
### 2. Test Data Builders
```typescript
// Use builders for complex test data
const user = new UserBuilder()
.withEmail('test@example.com')
.withRole('admin')
.build();
```
### 3. Custom Matchers
```typescript
// tests/utils/matchers.ts
export const toBeValidNode = (received: any) => {
const pass =
received.type &&
received.displayName &&
received.properties &&
Array.isArray(received.properties);
return {
pass,
message: () => `expected ${received} to be a valid node`
};
};
// Usage
expect(node).toBeValidNode();
```
### 4. Snapshot Testing
```typescript
// For complex structures
it('should generate correct node schema', () => {
const schema = generateNodeSchema(node);
expect(schema).toMatchSnapshot();
});
```
### 5. Test Isolation
```typescript
// Always clean up after tests
afterEach(async () => {
await cleanup();
vi.clearAllMocks();
vi.restoreAllMocks();
});
```
## Coverage Goals by Module
| Module | Target | Priority | Notes |
|--------|--------|----------|-------|
| services/config-validator | 95% | High | Critical for reliability |
| services/workflow-validator | 90% | High | Core functionality |
| mcp/tools | 90% | High | User-facing API |
| database/node-repository | 85% | Medium | Well-tested DB layer |
| loaders/node-loader | 85% | Medium | External dependencies |
| parsers/* | 90% | High | Data transformation |
| utils/* | 80% | Low | Helper functions |
| scripts/* | 50% | Low | One-time scripts |
## Continuous Improvement
1. **Weekly Reviews**: Review test coverage and identify gaps
2. **Performance Baselines**: Update benchmarks monthly
3. **Flaky Test Detection**: Monitor and fix within 48 hours
4. **Test Documentation**: Keep examples updated
5. **Developer Training**: Pair programming on tests
## Success Metrics
- [ ] All tests pass in CI (0 failures)
- [ ] Coverage > 80% overall
- [ ] No flaky tests
- [ ] CI runs < 5 minutes
- [ ] Performance benchmarks stable
- [ ] Zero production bugs from tested code

View File

@@ -0,0 +1,920 @@
# n8n-MCP Testing Strategy - AI/LLM Optimized
## Overview for AI Implementation
This testing strategy is optimized for implementation by AI agents like Claude Code. Each section contains explicit instructions, file paths, and complete code examples to minimize ambiguity.
## Key Principles for AI Implementation
1. **Explicit Over Implicit**: Every instruction includes exact file paths and complete code
2. **Sequential Dependencies**: Tasks are ordered to avoid forward references
3. **Atomic Tasks**: Each task can be completed independently
4. **Verification Steps**: Each task includes verification commands
5. **Error Recovery**: Each section includes troubleshooting steps
## Phase 0: Immediate Fixes (Day 1)
### Task 0.1: Fix Failing Tests
**Files to modify:**
- `/tests/src/tests/single-session.test.ts`
- `/tests/http-server-auth.test.ts`
**Step 1: Fix TypeScript errors in single-session.test.ts**
```typescript
// FIND these lines (around line 147, 188, 189):
expect(resNoAuth.body).toEqual({
// REPLACE with:
expect((resNoAuth as any).body).toEqual({
```
**Step 2: Fix auth test issues**
```typescript
// In tests/http-server-auth.test.ts
// FIND the mockExit setup
const mockExit = jest.spyOn(process, 'exit').mockImplementation();
// REPLACE with:
const mockExit = vi.spyOn(process, 'exit').mockImplementation(() => {
throw new Error('Process exited');
});
```
**Verification:**
```bash
npm test
# Should show 4 passing test suites instead of 2
```
### Task 0.2: Setup GitHub Actions
**Create file:** `.github/workflows/test.yml`
```yaml
name: Test Suite
on:
push:
branches: [main]
pull_request:
branches: [main]
jobs:
test:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: actions/setup-node@v4
with:
node-version: 20
cache: 'npm'
- run: npm ci
- run: npm test
- run: npm run lint
- run: npm run typecheck || true # Allow to fail initially
```
**Verification:**
```bash
git add .github/workflows/test.yml
git commit -m "chore: add GitHub Actions for testing"
git push
# Check Actions tab on GitHub - should see workflow running
```
## Phase 1: Vitest Migration (Week 1)
### Task 1.1: Install Vitest
**Execute these commands in order:**
```bash
# Remove Jest
npm uninstall jest ts-jest @types/jest
# Install Vitest
npm install -D vitest @vitest/ui @vitest/coverage-v8
# Install testing utilities
npm install -D @testing-library/jest-dom
npm install -D msw
npm install -D @faker-js/faker
npm install -D fishery
```
**Verification:**
```bash
npm list vitest # Should show vitest version
```
### Task 1.2: Create Vitest Configuration
**Create file:** `vitest.config.ts`
```typescript
import { defineConfig } from 'vitest/config';
import path from 'path';
export default defineConfig({
test: {
globals: true,
environment: 'node',
setupFiles: ['./tests/setup/global-setup.ts'],
coverage: {
provider: 'v8',
reporter: ['text', 'json', 'html', 'lcov'],
exclude: [
'node_modules/',
'tests/',
'**/*.d.ts',
'**/*.test.ts',
'scripts/',
'dist/'
],
thresholds: {
lines: 80,
functions: 80,
branches: 75,
statements: 80
}
}
},
resolve: {
alias: {
'@': path.resolve(__dirname, './src'),
'@tests': path.resolve(__dirname, './tests')
}
}
});
```
### Task 1.3: Create Global Setup
**Create file:** `tests/setup/global-setup.ts`
```typescript
import { beforeEach, afterEach, vi } from 'vitest';
// Reset mocks between tests
beforeEach(() => {
vi.clearAllMocks();
});
// Clean up after each test
afterEach(() => {
vi.restoreAllMocks();
});
// Global test timeout
vi.setConfig({ testTimeout: 10000 });
// Silence console during tests unless DEBUG=true
if (process.env.DEBUG !== 'true') {
global.console = {
...console,
log: vi.fn(),
debug: vi.fn(),
info: vi.fn(),
warn: vi.fn(),
error: vi.fn(),
};
}
```
### Task 1.4: Update package.json Scripts
**Modify file:** `package.json`
```json
{
"scripts": {
"test": "vitest",
"test:ui": "vitest --ui",
"test:run": "vitest run",
"test:coverage": "vitest run --coverage",
"test:watch": "vitest watch",
"test:unit": "vitest run tests/unit",
"test:integration": "vitest run tests/integration",
"test:e2e": "vitest run tests/e2e"
}
}
```
### Task 1.5: Migrate First Test File
**Modify file:** `tests/logger.test.ts`
```typescript
// Change line 1 FROM:
import { jest } from '@jest/globals';
// TO:
import { describe, it, expect, vi, beforeEach } from 'vitest';
// Replace all occurrences:
// FIND: jest.fn()
// REPLACE: vi.fn()
// FIND: jest.spyOn
// REPLACE: vi.spyOn
```
**Verification:**
```bash
npm test tests/logger.test.ts
# Should pass with Vitest
```
## Phase 2: Test Infrastructure (Week 2)
### Task 2.1: Create Directory Structure
**Execute these commands:**
```bash
# Create test directories
mkdir -p tests/unit/{services,database,mcp,utils,loaders,parsers}
mkdir -p tests/integration/{mcp-protocol,n8n-api,database}
mkdir -p tests/e2e/{workflows,setup,fixtures}
mkdir -p tests/performance/{node-loading,search,validation}
mkdir -p tests/fixtures/{factories,nodes,workflows}
mkdir -p tests/utils/{builders,mocks,assertions}
mkdir -p tests/setup
```
### Task 2.2: Create Database Mock
**Create file:** `tests/unit/database/__mocks__/better-sqlite3.ts`
```typescript
import { vi } from 'vitest';
export class MockDatabase {
private data = new Map<string, any[]>();
private prepared = new Map<string, any>();
constructor() {
this.data.set('nodes', []);
this.data.set('templates', []);
this.data.set('tools_documentation', []);
}
prepare(sql: string) {
const key = this.extractTableName(sql);
return {
all: vi.fn(() => this.data.get(key) || []),
get: vi.fn((id: string) => {
const items = this.data.get(key) || [];
return items.find(item => item.id === id);
}),
run: vi.fn((params: any) => {
const items = this.data.get(key) || [];
items.push(params);
this.data.set(key, items);
return { changes: 1, lastInsertRowid: items.length };
})
};
}
exec(sql: string) {
// Mock schema creation
return true;
}
close() {
// Mock close
return true;
}
// Helper to extract table name from SQL
private extractTableName(sql: string): string {
const match = sql.match(/FROM\s+(\w+)|INTO\s+(\w+)|UPDATE\s+(\w+)/i);
return match ? (match[1] || match[2] || match[3]) : 'nodes';
}
// Test helper to seed data
_seedData(table: string, data: any[]) {
this.data.set(table, data);
}
}
export default vi.fn(() => new MockDatabase());
```
### Task 2.3: Create Node Factory
**Create file:** `tests/fixtures/factories/node.factory.ts`
```typescript
import { Factory } from 'fishery';
import { faker } from '@faker-js/faker';
interface NodeDefinition {
name: string;
displayName: string;
description: string;
version: number;
defaults: { name: string };
inputs: string[];
outputs: string[];
properties: any[];
credentials?: any[];
group?: string[];
}
export const nodeFactory = Factory.define<NodeDefinition>(() => ({
name: faker.helpers.slugify(faker.word.noun()),
displayName: faker.company.name(),
description: faker.lorem.sentence(),
version: faker.number.int({ min: 1, max: 5 }),
defaults: {
name: faker.word.noun()
},
inputs: ['main'],
outputs: ['main'],
group: [faker.helpers.arrayElement(['transform', 'trigger', 'output'])],
properties: [
{
displayName: 'Resource',
name: 'resource',
type: 'options',
default: 'user',
options: [
{ name: 'User', value: 'user' },
{ name: 'Post', value: 'post' }
]
}
],
credentials: []
}));
// Specific node factories
export const webhookNodeFactory = nodeFactory.params({
name: 'webhook',
displayName: 'Webhook',
description: 'Starts the workflow when a webhook is called',
group: ['trigger'],
properties: [
{
displayName: 'Path',
name: 'path',
type: 'string',
default: 'webhook',
required: true
},
{
displayName: 'Method',
name: 'method',
type: 'options',
default: 'GET',
options: [
{ name: 'GET', value: 'GET' },
{ name: 'POST', value: 'POST' }
]
}
]
});
export const slackNodeFactory = nodeFactory.params({
name: 'slack',
displayName: 'Slack',
description: 'Send messages to Slack',
group: ['output'],
credentials: [
{
name: 'slackApi',
required: true
}
],
properties: [
{
displayName: 'Resource',
name: 'resource',
type: 'options',
default: 'message',
options: [
{ name: 'Message', value: 'message' },
{ name: 'Channel', value: 'channel' }
]
},
{
displayName: 'Operation',
name: 'operation',
type: 'options',
displayOptions: {
show: {
resource: ['message']
}
},
default: 'post',
options: [
{ name: 'Post', value: 'post' },
{ name: 'Update', value: 'update' }
]
},
{
displayName: 'Channel',
name: 'channel',
type: 'string',
required: true,
displayOptions: {
show: {
resource: ['message'],
operation: ['post']
}
},
default: ''
}
]
});
```
### Task 2.4: Create Workflow Builder
**Create file:** `tests/utils/builders/workflow.builder.ts`
```typescript
interface INode {
id: string;
name: string;
type: string;
typeVersion: number;
position: [number, number];
parameters: any;
}
interface IConnection {
node: string;
type: string;
index: number;
}
interface IConnections {
[key: string]: {
[key: string]: IConnection[][];
};
}
interface IWorkflow {
name: string;
nodes: INode[];
connections: IConnections;
active: boolean;
settings?: any;
}
export class WorkflowBuilder {
private workflow: IWorkflow;
private nodeCounter = 0;
constructor(name: string) {
this.workflow = {
name,
nodes: [],
connections: {},
active: false,
settings: {}
};
}
addNode(params: Partial<INode>): this {
const node: INode = {
id: params.id || `node_${this.nodeCounter++}`,
name: params.name || params.type?.split('.').pop() || 'Node',
type: params.type || 'n8n-nodes-base.noOp',
typeVersion: params.typeVersion || 1,
position: params.position || [250 + this.nodeCounter * 200, 300],
parameters: params.parameters || {}
};
this.workflow.nodes.push(node);
return this;
}
addWebhookNode(path: string = 'test-webhook'): this {
return this.addNode({
type: 'n8n-nodes-base.webhook',
name: 'Webhook',
parameters: {
path,
method: 'POST'
}
});
}
addSlackNode(channel: string = '#general'): this {
return this.addNode({
type: 'n8n-nodes-base.slack',
name: 'Slack',
typeVersion: 2.2,
parameters: {
resource: 'message',
operation: 'post',
channel,
text: '={{ $json.message }}'
}
});
}
connect(fromId: string, toId: string, outputIndex = 0): this {
if (!this.workflow.connections[fromId]) {
this.workflow.connections[fromId] = { main: [] };
}
if (!this.workflow.connections[fromId].main[outputIndex]) {
this.workflow.connections[fromId].main[outputIndex] = [];
}
this.workflow.connections[fromId].main[outputIndex].push({
node: toId,
type: 'main',
index: 0
});
return this;
}
connectSequentially(): this {
for (let i = 0; i < this.workflow.nodes.length - 1; i++) {
this.connect(
this.workflow.nodes[i].id,
this.workflow.nodes[i + 1].id
);
}
return this;
}
activate(): this {
this.workflow.active = true;
return this;
}
build(): IWorkflow {
return JSON.parse(JSON.stringify(this.workflow));
}
}
// Usage example:
// const workflow = new WorkflowBuilder('Test Workflow')
// .addWebhookNode()
// .addSlackNode()
// .connectSequentially()
// .build();
```
## Phase 3: Unit Tests (Week 3-4)
### Task 3.1: Test Config Validator
**Create file:** `tests/unit/services/config-validator.test.ts`
```typescript
import { describe, it, expect, beforeEach, vi } from 'vitest';
import { ConfigValidator } from '@/services/config-validator';
import { nodeFactory, slackNodeFactory } from '@tests/fixtures/factories/node.factory';
// Mock the database
vi.mock('better-sqlite3');
describe('ConfigValidator', () => {
let validator: ConfigValidator;
let mockDb: any;
beforeEach(() => {
// Setup mock database with test data
mockDb = {
prepare: vi.fn().mockReturnValue({
get: vi.fn().mockReturnValue({
properties: JSON.stringify(slackNodeFactory.build().properties)
})
})
};
validator = new ConfigValidator(mockDb);
});
describe('validate', () => {
it('should validate required fields for Slack message post', () => {
const config = {
resource: 'message',
operation: 'post'
// Missing required 'channel' field
};
const result = validator.validate('n8n-nodes-base.slack', config);
expect(result.isValid).toBe(false);
expect(result.errors).toContain('channel is required');
});
it('should pass validation with all required fields', () => {
const config = {
resource: 'message',
operation: 'post',
channel: '#general'
};
const result = validator.validate('n8n-nodes-base.slack', config);
expect(result.isValid).toBe(true);
expect(result.errors).toHaveLength(0);
});
it('should handle unknown node types', () => {
const result = validator.validate('unknown.node', {});
expect(result.isValid).toBe(false);
expect(result.errors).toContain('Unknown node type: unknown.node');
});
});
});
```
**Verification:**
```bash
npm test tests/unit/services/config-validator.test.ts
# Should create and pass the test
```
### Task 3.2: Create Test Template for Each Service
**For each service in `src/services/`, create a test file using this template:**
```typescript
// tests/unit/services/[service-name].test.ts
import { describe, it, expect, beforeEach, vi } from 'vitest';
import { ServiceName } from '@/services/[service-name]';
describe('ServiceName', () => {
let service: ServiceName;
beforeEach(() => {
service = new ServiceName();
});
describe('mainMethod', () => {
it('should handle basic case', () => {
// Arrange
const input = {};
// Act
const result = service.mainMethod(input);
// Assert
expect(result).toBeDefined();
});
});
});
```
**Files to create tests for:**
1. `tests/unit/services/enhanced-config-validator.test.ts`
2. `tests/unit/services/workflow-validator.test.ts`
3. `tests/unit/services/expression-validator.test.ts`
4. `tests/unit/services/property-filter.test.ts`
5. `tests/unit/services/example-generator.test.ts`
## Phase 4: Integration Tests (Week 5-6)
### Task 4.1: MCP Protocol Test
**Create file:** `tests/integration/mcp-protocol/protocol-compliance.test.ts`
```typescript
import { describe, it, expect, beforeEach } from 'vitest';
import { MCPServer } from '@/mcp/server';
import { InMemoryTransport } from '@modelcontextprotocol/sdk/inMemory.js';
describe('MCP Protocol Compliance', () => {
let server: MCPServer;
let clientTransport: any;
let serverTransport: any;
beforeEach(async () => {
[clientTransport, serverTransport] = InMemoryTransport.createLinkedPair();
server = new MCPServer();
await server.connect(serverTransport);
});
it('should reject requests without jsonrpc version', async () => {
const response = await clientTransport.send({
id: 1,
method: 'tools/list'
// Missing jsonrpc: "2.0"
});
expect(response.error).toBeDefined();
expect(response.error.code).toBe(-32600); // Invalid Request
});
it('should handle tools/list request', async () => {
const response = await clientTransport.send({
jsonrpc: '2.0',
id: 1,
method: 'tools/list'
});
expect(response.result).toBeDefined();
expect(response.result.tools).toBeInstanceOf(Array);
expect(response.result.tools.length).toBeGreaterThan(0);
});
});
```
## Phase 5: E2E Tests (Week 7-8)
### Task 5.1: E2E Test Setup without Playwright
**Create file:** `tests/e2e/setup/n8n-test-setup.ts`
```typescript
import { execSync } from 'child_process';
import { readFileSync, writeFileSync } from 'fs';
import path from 'path';
export class N8nTestSetup {
private containerName = 'n8n-test';
private dataPath = path.join(__dirname, '../fixtures/n8n-test-data');
async setup(): Promise<{ url: string; cleanup: () => void }> {
// Stop any existing container
try {
execSync(`docker stop ${this.containerName}`, { stdio: 'ignore' });
execSync(`docker rm ${this.containerName}`, { stdio: 'ignore' });
} catch (e) {
// Container doesn't exist, continue
}
// Start n8n with pre-configured database
execSync(`
docker run -d \
--name ${this.containerName} \
-p 5678:5678 \
-e N8N_BASIC_AUTH_ACTIVE=false \
-e N8N_ENCRYPTION_KEY=test-key \
-e DB_TYPE=sqlite \
-e N8N_USER_MANAGEMENT_DISABLED=true \
-v ${this.dataPath}:/home/node/.n8n \
n8nio/n8n:latest
`);
// Wait for n8n to be ready
await this.waitForN8n();
return {
url: 'http://localhost:5678',
cleanup: () => this.cleanup()
};
}
private async waitForN8n(maxRetries = 30) {
for (let i = 0; i < maxRetries; i++) {
try {
execSync('curl -f http://localhost:5678/healthz', { stdio: 'ignore' });
return;
} catch (e) {
await new Promise(resolve => setTimeout(resolve, 2000));
}
}
throw new Error('n8n failed to start');
}
private cleanup() {
execSync(`docker stop ${this.containerName}`, { stdio: 'ignore' });
execSync(`docker rm ${this.containerName}`, { stdio: 'ignore' });
}
}
```
### Task 5.2: Create Pre-configured Database
**Create file:** `tests/e2e/fixtures/setup-test-db.sql`
```sql
-- Create initial user (bypasses setup wizard)
INSERT INTO user (email, password, personalizationAnswers, settings, createdAt, updatedAt)
VALUES (
'test@example.com',
'$2a$10$mockHashedPassword',
'{}',
'{"userManagement":{"showSetupOnFirstLoad":false}}',
datetime('now'),
datetime('now')
);
-- Create API key for testing
INSERT INTO api_keys (userId, label, apiKey, createdAt, updatedAt)
VALUES (
1,
'Test API Key',
'test-api-key-for-e2e-testing',
datetime('now'),
datetime('now')
);
```
## AI Implementation Guidelines
### 1. Task Execution Order
Always execute tasks in this sequence:
1. Fix failing tests (Phase 0)
2. Set up CI/CD (Phase 0)
3. Migrate to Vitest (Phase 1)
4. Create test infrastructure (Phase 2)
5. Write unit tests (Phase 3)
6. Write integration tests (Phase 4)
7. Write E2E tests (Phase 5)
### 2. File Creation Pattern
When creating a new test file:
1. Create the file with the exact path specified
2. Copy the provided template exactly
3. Run the verification command
4. If it fails, check imports and file paths
5. Commit after each successful test file
### 3. Error Recovery
If a test fails:
1. Check the exact error message
2. Verify all imports are correct
3. Ensure mocks are properly set up
4. Check that the source file exists
5. Run with DEBUG=true for more information
### 4. Coverage Tracking
After each phase:
```bash
npm run test:coverage
# Check coverage/index.html for detailed report
# Ensure coverage is increasing
```
### 5. Commit Strategy
Make atomic commits:
```bash
# After each successful task
git add [specific files]
git commit -m "test: [phase] - [specific task completed]"
# Examples:
git commit -m "test: phase 0 - fix failing tests"
git commit -m "test: phase 1 - migrate to vitest"
git commit -m "test: phase 2 - create test infrastructure"
```
## Verification Checklist
After each phase, verify:
**Phase 0:**
- [ ] All 6 test suites pass
- [ ] GitHub Actions workflow runs
**Phase 1:**
- [ ] Vitest installed and configured
- [ ] npm test runs Vitest
- [ ] At least one test migrated
**Phase 2:**
- [ ] Directory structure created
- [ ] Database mock works
- [ ] Factories generate valid data
- [ ] Builders create valid workflows
**Phase 3:**
- [ ] Config validator tests pass
- [ ] Coverage > 50%
**Phase 4:**
- [ ] MCP protocol tests pass
- [ ] Coverage > 70%
**Phase 5:**
- [ ] E2E tests run without Playwright
- [ ] Coverage > 80%
## Common Issues and Solutions
### Issue: Cannot find module '@/services/...'
**Solution:** Check tsconfig.json has path aliases configured
### Issue: Mock not working
**Solution:** Ensure vi.mock() is at top of file, outside describe blocks
### Issue: Test timeout
**Solution:** Increase timeout for specific test:
```typescript
it('should handle slow operation', async () => {
// test code
}, 30000); // 30 second timeout
```
### Issue: Coverage not updating
**Solution:**
```bash
rm -rf coverage/
npm run test:coverage
```
## Success Criteria
The implementation is successful when:
1. All tests pass (0 failures)
2. Coverage exceeds 80%
3. CI/CD pipeline is green
4. No TypeScript errors
5. All phases completed
This AI-optimized plan provides explicit, step-by-step instructions that can be followed sequentially without ambiguity.

1227
docs/testing-strategy.md Normal file

File diff suppressed because it is too large Load Diff