- Remove AI agent coordination files and progress tracking
- Remove temporary test results and generated artifacts
- Remove diagnostic test scripts from src/scripts/
- Remove development planning documents
- Update .gitignore to exclude test artifacts
- Clean up 53 temporary files total
- Add test result artifacts storage with multiple formats (JUnit, JSON, HTML)
- Configure GitHub Actions to upload and preserve test outputs
- Add PR comment integration with test summaries
- Create benchmark comparison workflow for PR performance tracking
- Add detailed test report generation scripts
- Configure artifact retention policies (30 days for tests, 90 for combined)
- Set up test metadata collection for better debugging
This completes all remaining test infrastructure tasks and provides
comprehensive visibility into test results across CI/CD pipeline.
- Create benchmark test suites for critical operations:
- Node loading performance
- Database query performance
- Search operations performance
- Validation performance
- MCP tool execution performance
- Add GitHub Actions workflow for benchmark tracking:
- Runs on push to main and PRs
- Uses github-action-benchmark for historical tracking
- Comments on PRs with performance results
- Alerts on >10% performance regressions
- Stores results in GitHub Pages
- Create benchmark infrastructure:
- Custom Vitest benchmark configuration
- JSON reporter for CI results
- Result formatter for github-action-benchmark
- Performance threshold documentation
- Add supporting utilities:
- SQLiteStorageService for benchmark database setup
- MCPEngine wrapper for testing MCP tools
- Test factories for generating benchmark data
- Enhanced NodeRepository with benchmark methods
- Document benchmark system:
- Comprehensive benchmark guide in docs/BENCHMARKS.md
- Performance thresholds in .github/BENCHMARK_THRESHOLDS.md
- README for benchmarks directory
- Integration with existing test suite
The benchmark system will help monitor performance over time and catch regressions before they reach production.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
- Add bin configuration to package.json for npx execution
- Implement smart database path resolution for npx/global/local installs
- Create dedicated npm publish script using runtime-only dependencies
- Add .npmignore to control published package contents
- Update README with npx as primary installation method
- Add n8n version badge to README
- Sync version between package.json and package.runtime.json
- Update CHANGELOG for v2.7.8 release
This allows users to run 'npx n8n-mcp' without installing the package,
reducing friction and making it easier to get started with n8n-MCP.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
- Removed Git LFS tracking for nodes.db to fix Docker builds
- Docker was getting LFS pointer file instead of actual database
- Database is only 20MB, safe to track directly
- Updated .gitignore to allow nodes.db tracking
- This ensures Docker always gets the actual database file
- Includes trigger detection fix (webhook, cron, interval, emailReadImap)
Resolves issues where Docker images showed isTrigger: false for trigger nodes
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
- Created complete CHANGELOG.md from v1.0.0 to v2.7.0 based on git history
- Updated README.md to fix incorrect CHANGELOG path reference
- Added n8n-docs/ directory to .gitignore (was accidentally tracked)
- Removed n8n-docs from git tracking
The CHANGELOG now follows Keep a Changelog format and includes all major features,
fixes, and improvements for each version with proper categorization.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
## Root cause
- Docker buildx memory options were incorrectly formatted
- Database build during Docker image creation was failing on multi-platform builds
- n8n native dependencies caused issues across different architectures
## Solution
- Removed invalid buildx driver-opts configuration
- Eliminated database build stage from Dockerfile
- Now using pre-built nodes.db file (11MB) from repository
- Fixed .dockerignore to include nodes.db in build context
- Added .dockerignore to version control (was incorrectly gitignored)
## Benefits
- Faster builds (no n8n package installation during build)
- More reliable multi-platform builds (amd64 + arm64)
- Simpler Dockerfile (3 stages instead of 4)
Database can still be rebuilt locally using 'npm run rebuild' when needed.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
- Remove all references to workflow execution/management features
- Delete legacy scripts for bidirectional n8n integration
- Update documentation to focus on node documentation serving only
- Remove old docker-compose files for workflow management
- Add simplified docker-compose.yml for documentation server
- Update CHANGELOG.md to reflect v2.0.0 and v2.1.0 changes
- Update Dockerfile to use v2 paths and database
The project is now clearly focused on serving n8n node documentation
to AI assistants, with no workflow execution capabilities.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
Major refactoring to align with actual requirements:
- Purpose: Serve n8n node code/documentation to AI agents only
- No workflow execution or management features
- Complete node information including source code, docs, and examples
New features:
- Node documentation service with SQLite FTS5 search
- Documentation fetcher from n8n-docs repository
- Example workflow generator for each node type
- Simplified MCP tools focused on node information
- Complete database rebuild with all node data
MCP Tools:
- list_nodes: List available nodes
- get_node_info: Get complete node information
- search_nodes: Full-text search across nodes
- get_node_example: Get usage examples
- get_node_source_code: Get source code only
- get_node_documentation: Get documentation only
- rebuild_database: Rebuild entire database
- get_database_statistics: Database stats
Database schema includes:
- Node source code and metadata
- Official documentation from n8n-docs
- Generated usage examples
- Full-text search capabilities
- Category and type filtering
Updated README with:
- Clear purpose statement
- Claude Desktop installation instructions
- Complete tool documentation
- Troubleshooting guide
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>