- Create benchmark test suites for critical operations: - Node loading performance - Database query performance - Search operations performance - Validation performance - MCP tool execution performance - Add GitHub Actions workflow for benchmark tracking: - Runs on push to main and PRs - Uses github-action-benchmark for historical tracking - Comments on PRs with performance results - Alerts on >10% performance regressions - Stores results in GitHub Pages - Create benchmark infrastructure: - Custom Vitest benchmark configuration - JSON reporter for CI results - Result formatter for github-action-benchmark - Performance threshold documentation - Add supporting utilities: - SQLiteStorageService for benchmark database setup - MCPEngine wrapper for testing MCP tools - Test factories for generating benchmark data - Enhanced NodeRepository with benchmark methods - Document benchmark system: - Comprehensive benchmark guide in docs/BENCHMARKS.md - Performance thresholds in .github/BENCHMARK_THRESHOLDS.md - README for benchmarks directory - Integration with existing test suite The benchmark system will help monitor performance over time and catch regressions before they reach production. 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
170 lines
5.1 KiB
TypeScript
170 lines
5.1 KiB
TypeScript
/**
|
|
* Example test demonstrating test environment configuration usage
|
|
*/
|
|
|
|
import { describe, it, expect, beforeAll, afterAll } from 'vitest';
|
|
import {
|
|
getTestConfig,
|
|
getTestTimeout,
|
|
isFeatureEnabled,
|
|
isTestMode
|
|
} from '@tests/setup/test-env';
|
|
import {
|
|
withEnvOverrides,
|
|
createTestDatabasePath,
|
|
getMockApiUrl,
|
|
measurePerformance,
|
|
createTestLogger,
|
|
waitForCondition
|
|
} from '@tests/helpers/env-helpers';
|
|
|
|
describe('Test Environment Configuration Example', () => {
|
|
const config = getTestConfig();
|
|
const logger = createTestLogger('test-env-example');
|
|
|
|
beforeAll(() => {
|
|
logger.info('Test suite starting with configuration:', {
|
|
environment: config.nodeEnv,
|
|
database: config.database.path,
|
|
apiUrl: config.api.url
|
|
});
|
|
});
|
|
|
|
afterAll(() => {
|
|
logger.info('Test suite completed');
|
|
});
|
|
|
|
it('should be in test mode', () => {
|
|
expect(isTestMode()).toBe(true);
|
|
expect(config.nodeEnv).toBe('test');
|
|
expect(config.isTest).toBe(true);
|
|
});
|
|
|
|
it('should have proper database configuration', () => {
|
|
expect(config.database.path).toBeDefined();
|
|
expect(config.database.rebuildOnStart).toBe(false);
|
|
expect(config.database.seedData).toBe(true);
|
|
});
|
|
|
|
it('should have mock API configuration', () => {
|
|
expect(config.api.url).toMatch(/mock-api/);
|
|
expect(config.api.key).toBe('test-api-key-12345');
|
|
});
|
|
|
|
it('should respect test timeouts', { timeout: getTestTimeout('unit') }, async () => {
|
|
const timeout = getTestTimeout('unit');
|
|
expect(timeout).toBe(5000);
|
|
|
|
// Simulate async operation
|
|
await new Promise(resolve => setTimeout(resolve, 100));
|
|
});
|
|
|
|
it('should support environment overrides', () => {
|
|
const originalLogLevel = config.logging.level;
|
|
|
|
const result = withEnvOverrides({
|
|
LOG_LEVEL: 'debug',
|
|
DEBUG: 'true'
|
|
}, () => {
|
|
const newConfig = getTestConfig();
|
|
expect(newConfig.logging.level).toBe('debug');
|
|
expect(newConfig.logging.debug).toBe(true);
|
|
return 'success';
|
|
});
|
|
|
|
expect(result).toBe('success');
|
|
expect(config.logging.level).toBe(originalLogLevel);
|
|
});
|
|
|
|
it('should generate unique test database paths', () => {
|
|
const path1 = createTestDatabasePath('feature1');
|
|
const path2 = createTestDatabasePath('feature1');
|
|
|
|
if (path1 !== ':memory:') {
|
|
expect(path1).not.toBe(path2);
|
|
expect(path1).toMatch(/test-feature1-\d+-\w+\.db$/);
|
|
}
|
|
});
|
|
|
|
it('should construct mock API URLs', () => {
|
|
const baseUrl = getMockApiUrl();
|
|
const endpointUrl = getMockApiUrl('/nodes');
|
|
|
|
expect(baseUrl).toBe(config.api.url);
|
|
expect(endpointUrl).toBe(`${config.api.url}/nodes`);
|
|
});
|
|
|
|
it.skipIf(!isFeatureEnabled('mockExternalApis'))('should check feature flags', () => {
|
|
expect(config.features.mockExternalApis).toBe(true);
|
|
expect(isFeatureEnabled('mockExternalApis')).toBe(true);
|
|
});
|
|
|
|
it('should measure performance', async () => {
|
|
const measure = measurePerformance('test-operation');
|
|
|
|
// Simulate some work
|
|
measure.mark('start-processing');
|
|
await new Promise(resolve => setTimeout(resolve, 50));
|
|
measure.mark('mid-processing');
|
|
await new Promise(resolve => setTimeout(resolve, 50));
|
|
|
|
const results = measure.end();
|
|
|
|
expect(results.total).toBeGreaterThan(100);
|
|
expect(results.marks['start-processing']).toBeLessThan(results.marks['mid-processing']);
|
|
});
|
|
|
|
it('should wait for conditions', async () => {
|
|
let counter = 0;
|
|
const incrementCounter = setInterval(() => counter++, 100);
|
|
|
|
try {
|
|
await waitForCondition(
|
|
() => counter >= 3,
|
|
{
|
|
timeout: 1000,
|
|
interval: 50,
|
|
message: 'Counter did not reach 3'
|
|
}
|
|
);
|
|
|
|
expect(counter).toBeGreaterThanOrEqual(3);
|
|
} finally {
|
|
clearInterval(incrementCounter);
|
|
}
|
|
});
|
|
|
|
it('should have proper logging configuration', () => {
|
|
expect(config.logging.level).toBe('error');
|
|
expect(config.logging.debug).toBe(false);
|
|
expect(config.logging.showStack).toBe(true);
|
|
|
|
// Logger should respect configuration
|
|
logger.debug('This should not appear in test output');
|
|
logger.error('This should appear in test output');
|
|
});
|
|
|
|
it('should have performance thresholds', () => {
|
|
expect(config.performance.thresholds.apiResponse).toBe(100);
|
|
expect(config.performance.thresholds.dbQuery).toBe(50);
|
|
expect(config.performance.thresholds.nodeParse).toBe(200);
|
|
});
|
|
|
|
it('should disable caching and rate limiting in tests', () => {
|
|
expect(config.cache.enabled).toBe(false);
|
|
expect(config.cache.ttl).toBe(0);
|
|
expect(config.rateLimiting.max).toBe(0);
|
|
expect(config.rateLimiting.window).toBe(0);
|
|
});
|
|
|
|
it('should configure test paths', () => {
|
|
expect(config.paths.fixtures).toBe('./tests/fixtures');
|
|
expect(config.paths.data).toBe('./tests/data');
|
|
expect(config.paths.snapshots).toBe('./tests/__snapshots__');
|
|
});
|
|
|
|
it('should support MSW configuration', () => {
|
|
expect(config.mocking.msw.enabled).toBe(true);
|
|
expect(config.mocking.msw.apiDelay).toBe(0);
|
|
});
|
|
}); |