mirror of
https://github.com/czlonkowski/n8n-mcp.git
synced 2026-01-30 14:32:04 +00:00
Compare commits
414 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
ab6b554692 | ||
|
|
32264da107 | ||
|
|
ef1cf747a3 | ||
|
|
dbdc88d629 | ||
|
|
538618b1bc | ||
|
|
41830c88fe | ||
|
|
0d2d9bdd52 | ||
|
|
05f68b8ea1 | ||
|
|
5881304ed8 | ||
|
|
0f5b0d9463 | ||
|
|
4399899255 | ||
|
|
8d20c64f5c | ||
|
|
fe1309151a | ||
|
|
dd62040155 | ||
|
|
112b40119c | ||
|
|
318986f546 | ||
|
|
aa8a6a7069 | ||
|
|
e11a885b0d | ||
|
|
ee99cb7ba1 | ||
|
|
66cb66b31b | ||
|
|
b67d6ba353 | ||
|
|
3ba5584df9 | ||
|
|
be0211d826 | ||
|
|
0d71a16f83 | ||
|
|
085f6db7a2 | ||
|
|
b6bc3b732e | ||
|
|
c16c9a2398 | ||
|
|
1d34ad81d5 | ||
|
|
4566253bdc | ||
|
|
54c598717c | ||
|
|
8b5b01de98 | ||
|
|
275e573d8d | ||
|
|
6256105053 | ||
|
|
1f43784315 | ||
|
|
80e3391773 | ||
|
|
c580a3dde4 | ||
|
|
fc8fb66900 | ||
|
|
4625ebf64d | ||
|
|
43dea68f0b | ||
|
|
dc62fd66cb | ||
|
|
a94ff0586c | ||
|
|
29b2b1d4c1 | ||
|
|
fa6ff89516 | ||
|
|
34811eaf69 | ||
|
|
52c9902efd | ||
|
|
fba8b2a490 | ||
|
|
275e4f8cef | ||
|
|
4016ac42ef | ||
|
|
b8227ff775 | ||
|
|
f61fd9b429 | ||
|
|
4b36ed6a95 | ||
|
|
f072b2e003 | ||
|
|
cfd2325ca4 | ||
|
|
978347e8d0 | ||
|
|
1b7dd3b517 | ||
|
|
c52bbcbb83 | ||
|
|
5fb63cd725 | ||
|
|
36eb8e3864 | ||
|
|
51278f52e9 | ||
|
|
6479ac2bf5 | ||
|
|
08d43bd7fb | ||
|
|
914805f5ea | ||
|
|
08a1d42f09 | ||
|
|
ae11738ac7 | ||
|
|
6e365714e2 | ||
|
|
a2cc37bdf7 | ||
|
|
cf3c66c0ea | ||
|
|
f33b626179 | ||
|
|
2113714ec2 | ||
|
|
49757e3c22 | ||
|
|
dd521d0d87 | ||
|
|
331883f944 | ||
|
|
f3164e202f | ||
|
|
8e2e1dce62 | ||
|
|
b986beef2c | ||
|
|
943f5862a3 | ||
|
|
2c536a25fd | ||
|
|
e95ac7c335 | ||
|
|
e2c8fd0125 | ||
|
|
3332eb09fc | ||
|
|
bd03412fc8 | ||
|
|
73fa494735 | ||
|
|
67d8f5d4d4 | ||
|
|
d2a250e23d | ||
|
|
710f054b93 | ||
|
|
fd65727632 | ||
|
|
5d9936a909 | ||
|
|
de95fb21ba | ||
|
|
2bcd7c757b | ||
|
|
50439e2aa1 | ||
|
|
96cb9eca0f | ||
|
|
36dc8b489c | ||
|
|
cffd5e8b2e | ||
|
|
1ad2c6f6d2 | ||
|
|
28cff8c77b | ||
|
|
0818b4d56c | ||
|
|
5e2a6bdb9c | ||
|
|
ec9d8fdb7e | ||
|
|
ddc4de8c3e | ||
|
|
c67659a7c3 | ||
|
|
4cf8bb5c98 | ||
|
|
53b5dc312d | ||
|
|
1eedb43e9f | ||
|
|
81dfbbbd77 | ||
|
|
3ba3f101b3 | ||
|
|
92eb4ef34f | ||
|
|
ccbe04f007 | ||
|
|
91ad08493c | ||
|
|
7bb021163f | ||
|
|
59ae78f03a | ||
|
|
cb224de01f | ||
|
|
fd9ea985f2 | ||
|
|
225bb06cd5 | ||
|
|
2627028be3 | ||
|
|
cc9fe69449 | ||
|
|
0144484f96 | ||
|
|
2b7bc48699 | ||
|
|
0ec02fa0da | ||
|
|
d207cc3723 | ||
|
|
eeb4b6ac3e | ||
|
|
06cbb40213 | ||
|
|
9a00a99011 | ||
|
|
36aedd5050 | ||
|
|
59f49c47ab | ||
|
|
b106550520 | ||
|
|
e1be4473a3 | ||
|
|
b12a927a10 | ||
|
|
08abdb7937 | ||
|
|
95bb002577 | ||
|
|
36e02c68d3 | ||
|
|
3078273d93 | ||
|
|
aeb74102e5 | ||
|
|
af949b09a5 | ||
|
|
44568a6edd | ||
|
|
59e4cb85ac | ||
|
|
f78f53e731 | ||
|
|
c6e0e528d1 | ||
|
|
34bafe240d | ||
|
|
f139d38c81 | ||
|
|
aeaba3b9ca | ||
|
|
a7bfa73479 | ||
|
|
ee125c52f8 | ||
|
|
f9194ee74c | ||
|
|
2a85000411 | ||
|
|
653f395666 | ||
|
|
cfe3c5e584 | ||
|
|
67c3c9c9c8 | ||
|
|
6d50cf93f0 | ||
|
|
de9f222cfe | ||
|
|
da593400d2 | ||
|
|
126d09c66b | ||
|
|
4f81962953 | ||
|
|
9e7a0e0487 | ||
|
|
a7dc07abab | ||
|
|
1c56eb0daa | ||
|
|
fcf778c79d | ||
|
|
c519cd5060 | ||
|
|
69f3a31d41 | ||
|
|
bd8a7f68ac | ||
|
|
abc6a31302 | ||
|
|
57459c27e3 | ||
|
|
9380602439 | ||
|
|
a696af8cfa | ||
|
|
b467bec93e | ||
|
|
6e042467b2 | ||
|
|
287b9aa819 | ||
|
|
3331b72df4 | ||
|
|
c0d7145a5a | ||
|
|
08e906739f | ||
|
|
ae329c3bb6 | ||
|
|
1cfbdc3bdf | ||
|
|
b3d42b3390 | ||
|
|
4feb905bd0 | ||
|
|
ad1f611d2a | ||
|
|
02574e5555 | ||
|
|
b27d245dab | ||
|
|
ecf0d50a63 | ||
|
|
1db9ecf33f | ||
|
|
fc973d83db | ||
|
|
2e19eaa309 | ||
|
|
73db3dfdfe | ||
|
|
7fcfa8f696 | ||
|
|
c8cdd3c0b5 | ||
|
|
62d01ab237 | ||
|
|
00289e90d7 | ||
|
|
5c01624c3a | ||
|
|
dad3a442d9 | ||
|
|
7a402bc7ad | ||
|
|
88e288f8f6 | ||
|
|
12a7f1e8bf | ||
|
|
2f18a2bb9a | ||
|
|
9b94e3be9c | ||
|
|
9e1a4129c0 | ||
|
|
4b764c6110 | ||
|
|
c3b691cedf | ||
|
|
4bf8f7006d | ||
|
|
2a9a3b9410 | ||
|
|
cd27d78bfd | ||
|
|
8d1ae278ee | ||
|
|
a84dbd6a15 | ||
|
|
1728495146 | ||
|
|
2305aaab9e | ||
|
|
f74427bdb5 | ||
|
|
fe59688e03 | ||
|
|
675989971c | ||
|
|
d875ac1e0c | ||
|
|
5bf1bc46e9 | ||
|
|
3bab53a3be | ||
|
|
8ffda534be | ||
|
|
0bf0e1cd74 | ||
|
|
9fb847a16f | ||
|
|
bf999232a3 | ||
|
|
59e476fdf0 | ||
|
|
711cecb90d | ||
|
|
582c9aac53 | ||
|
|
997cc93a0a | ||
|
|
2f234780dd | ||
|
|
99518f71cf | ||
|
|
fe1e3640af | ||
|
|
aef9d983e2 | ||
|
|
e252a36e3f | ||
|
|
39e13c451f | ||
|
|
a8e0b1ed34 | ||
|
|
ed7de10fd2 | ||
|
|
b7fa12667b | ||
|
|
4854a50854 | ||
|
|
cb5691f17d | ||
|
|
6d45ff8bcb | ||
|
|
64b9cf47a7 | ||
|
|
f4dff6b8e1 | ||
|
|
ec0d2e8a6e | ||
|
|
a1db133a50 | ||
|
|
d8bab6e667 | ||
|
|
3728a9cc67 | ||
|
|
47e6a7846c | ||
|
|
cabda2a0f8 | ||
|
|
34cb8f8c44 | ||
|
|
48df87f76c | ||
|
|
540c5270c6 | ||
|
|
6210378687 | ||
|
|
8c2b1cfbbe | ||
|
|
d862f4961d | ||
|
|
2057f98e76 | ||
|
|
fff47f9f9d | ||
|
|
87cc84f593 | ||
|
|
8405497263 | ||
|
|
7a66f71c23 | ||
|
|
9cbbc6bb67 | ||
|
|
fbce712714 | ||
|
|
f13685fcd7 | ||
|
|
89b1ef2354 | ||
|
|
951d5b7e1b | ||
|
|
263753254a | ||
|
|
2896e393d3 | ||
|
|
9fa1c44149 | ||
|
|
e217d022d6 | ||
|
|
ca150287c9 | ||
|
|
5825a85ccc | ||
|
|
fecc584145 | ||
|
|
09bbcd7001 | ||
|
|
c2195d7da6 | ||
|
|
d8c5c7d4df | ||
|
|
2716207d72 | ||
|
|
a5cf4193e4 | ||
|
|
a1a9ff63d2 | ||
|
|
676c693885 | ||
|
|
e14c647b7d | ||
|
|
481d74c249 | ||
|
|
6f21a717cd | ||
|
|
75b55776f2 | ||
|
|
fa04ece8ea | ||
|
|
acfffbb0f2 | ||
|
|
3b2be46119 | ||
|
|
671c175d71 | ||
|
|
09e69df5a7 | ||
|
|
f150802bed | ||
|
|
5960d2826e | ||
|
|
78abda601a | ||
|
|
2491caecdc | ||
|
|
5e45fe299a | ||
|
|
f6ee6349a0 | ||
|
|
370b063fe4 | ||
|
|
3506497412 | ||
|
|
247c8d74af | ||
|
|
f6160d43a0 | ||
|
|
c23442249a | ||
|
|
3981b9108a | ||
|
|
60f78d5783 | ||
|
|
ceb082efca | ||
|
|
27339ec78d | ||
|
|
eb28bf0f2a | ||
|
|
4390b72d2a | ||
|
|
3b469d0afe | ||
|
|
0c31f12372 | ||
|
|
77b454d8ca | ||
|
|
627c0144a4 | ||
|
|
11df329e0f | ||
|
|
9a13b977dc | ||
|
|
dd36735a1a | ||
|
|
c1fb3db568 | ||
|
|
149976323c | ||
|
|
14bd0f55d3 | ||
|
|
3f8acb7e4a | ||
|
|
1a926630b8 | ||
|
|
c5aebc1450 | ||
|
|
60305cde74 | ||
|
|
3f719ac174 | ||
|
|
594d4975cb | ||
|
|
f237fad1e8 | ||
|
|
bc1cc109b5 | ||
|
|
424f8ae1ff | ||
|
|
f0338ea5ce | ||
|
|
8ed66208e6 | ||
|
|
f6a1b62590 | ||
|
|
34c7f756e1 | ||
|
|
b366d40d67 | ||
|
|
05eec1cc81 | ||
|
|
7e76369d2a | ||
|
|
a5ac4297bc | ||
|
|
4823bd53bc | ||
|
|
32e434fb76 | ||
|
|
bc7bd8e2c0 | ||
|
|
34fbdc30fe | ||
|
|
27b89f4c92 | ||
|
|
70653b16bd | ||
|
|
e6f1d6bcf0 | ||
|
|
44f92063c3 | ||
|
|
17530c0f72 | ||
|
|
0ef69fbf75 | ||
|
|
f39c9a5389 | ||
|
|
92d7577f22 | ||
|
|
874aea6920 | ||
|
|
19caa7bbb4 | ||
|
|
dff0387ae2 | ||
|
|
469cc1720d | ||
|
|
99cdae7655 | ||
|
|
abc226f111 | ||
|
|
16e6a1fc44 | ||
|
|
a7a6d64931 | ||
|
|
03c4e3b9a5 | ||
|
|
297acb039e | ||
|
|
aaf7c83301 | ||
|
|
7147f5ef05 | ||
|
|
2ae0d559bf | ||
|
|
55be451f11 | ||
|
|
28a369deb4 | ||
|
|
0199bcd44d | ||
|
|
6b886acaca | ||
|
|
5f30643406 | ||
|
|
a7846c4ee9 | ||
|
|
0c4a2199f5 | ||
|
|
c18c4e7584 | ||
|
|
1e586c0b23 | ||
|
|
6e24da722b | ||
|
|
d49416fc58 | ||
|
|
b4021acd14 | ||
|
|
61b54266b3 | ||
|
|
319f22f26e | ||
|
|
ea650bc767 | ||
|
|
3b767c798c | ||
|
|
e7895d2e01 | ||
|
|
f35097ed46 | ||
|
|
10c29dd585 | ||
|
|
696f461cab | ||
|
|
1441508c00 | ||
|
|
6b4bb7ff66 | ||
|
|
9e79b53465 | ||
|
|
8ce7c62299 | ||
|
|
15e6e97fd9 | ||
|
|
984af0a72f | ||
|
|
2df1f1b32b | ||
|
|
45fac6fe5e | ||
|
|
b65a2f8f3d | ||
|
|
f3658a4cab | ||
|
|
182016d932 | ||
|
|
36839a1c30 | ||
|
|
cac43ed384 | ||
|
|
8fd8c082ee | ||
|
|
baab3a02dc | ||
|
|
b2a5cf49f7 | ||
|
|
640e758c24 | ||
|
|
685171e9b7 | ||
|
|
567b54eaf7 | ||
|
|
bb774f8c70 | ||
|
|
fddc363221 | ||
|
|
13c1663489 | ||
|
|
48986263bf | ||
|
|
00f3f1fbfd | ||
|
|
a77379b40b | ||
|
|
680ccce47c | ||
|
|
c320eb4b35 | ||
|
|
f508d9873b | ||
|
|
9e322ad590 | ||
|
|
a4e711a4e8 | ||
|
|
bb39af3d9d | ||
|
|
999e31b13a | ||
|
|
72d90a2584 | ||
|
|
9003c24808 | ||
|
|
b944afa1bb | ||
|
|
ba3d1b35f2 | ||
|
|
6d95786938 | ||
|
|
21d4b9b9fb | ||
|
|
f3b777d8e8 | ||
|
|
035c4a349e | ||
|
|
08f3d8120d | ||
|
|
4b1aaa936d | ||
|
|
e94bb5479c | ||
|
|
1a99e9c6c7 | ||
|
|
7dc938065f | ||
|
|
8022ee1f65 | ||
|
|
9e71c71698 | ||
|
|
df4066022f | ||
|
|
7a71c3c3f8 | ||
|
|
3bfad51519 |
102
.claude/agents/n8n-mcp-tester.md
Normal file
102
.claude/agents/n8n-mcp-tester.md
Normal file
@@ -0,0 +1,102 @@
|
||||
---
|
||||
name: n8n-mcp-tester
|
||||
description: Use this agent when you need to test n8n-mcp functionality after implementing fixes or new features. This agent should be invoked proactively after code changes to validate that the MCP tools work correctly with real n8n workflows. The agent tests the actual MCP server tools, not just unit tests. Examples:\n\n<example>\nContext: Developer has just fixed a validation bug in the n8n-mcp server\nuser: "I've fixed the validation issue for HTTP Request nodes"\nassistant: "I've completed the fix. Now let me build the project and then test it."\n<builds project>\nassistant: "Project built successfully. Please reload the MCP server in Claude Desktop to apply the changes."\nuser: "Reloaded"\nassistant: "Great! Now let me invoke the n8n-mcp-tester agent to verify the fix works correctly."\n<commentary>\nSince a fix has been implemented and the MCP server reloaded, use the n8n-mcp-tester agent to validate the fix works in real-world usage.\n</commentary>\n</example>\n\n<example>\nContext: Developer has added a new MCP tool for workflow validation\nuser: "I've implemented the new workflow validation tool"\nassistant: "Implementation complete. Let me build and prepare for testing."\n<builds project>\nassistant: "Build successful. Please reload the MCP server to load the new tool."\nuser: "Done"\nassistant: "Perfect! I'll now use the n8n-mcp-tester agent to test the new workflow validation tool."\n<commentary>\nAfter implementing new MCP functionality and reloading the server, invoke n8n-mcp-tester to verify it works correctly.\n</commentary>\n</example>
|
||||
tools: Glob, Grep, LS, Read, WebFetch, TodoWrite, WebSearch, mcp__puppeteer__puppeteer_navigate, mcp__puppeteer__puppeteer_screenshot, mcp__puppeteer__puppeteer_click, mcp__puppeteer__puppeteer_fill, mcp__puppeteer__puppeteer_select, mcp__puppeteer__puppeteer_hover, mcp__puppeteer__puppeteer_evaluate, ListMcpResourcesTool, ReadMcpResourceTool, mcp__supabase__list_organizations, mcp__supabase__get_organization, mcp__supabase__list_projects, mcp__supabase__get_project, mcp__supabase__get_cost, mcp__supabase__confirm_cost, mcp__supabase__create_project, mcp__supabase__pause_project, mcp__supabase__restore_project, mcp__supabase__create_branch, mcp__supabase__list_branches, mcp__supabase__delete_branch, mcp__supabase__merge_branch, mcp__supabase__reset_branch, mcp__supabase__rebase_branch, mcp__supabase__list_tables, mcp__supabase__list_extensions, mcp__supabase__list_migrations, mcp__supabase__apply_migration, mcp__supabase__execute_sql, mcp__supabase__get_logs, mcp__supabase__get_advisors, mcp__supabase__get_project_url, mcp__supabase__get_anon_key, mcp__supabase__generate_typescript_types, mcp__supabase__search_docs, mcp__supabase__list_edge_functions, mcp__supabase__deploy_edge_function, mcp__n8n-mcp__tools_documentation, mcp__n8n-mcp__list_nodes, mcp__n8n-mcp__get_node_info, mcp__n8n-mcp__search_nodes, mcp__n8n-mcp__list_ai_tools, mcp__n8n-mcp__get_node_documentation, mcp__n8n-mcp__get_database_statistics, mcp__n8n-mcp__get_node_essentials, mcp__n8n-mcp__search_node_properties, mcp__n8n-mcp__get_node_for_task, mcp__n8n-mcp__list_tasks, mcp__n8n-mcp__validate_node_operation, mcp__n8n-mcp__validate_node_minimal, mcp__n8n-mcp__get_property_dependencies, mcp__n8n-mcp__get_node_as_tool_info, mcp__n8n-mcp__list_node_templates, mcp__n8n-mcp__get_template, mcp__n8n-mcp__search_templates, mcp__n8n-mcp__get_templates_for_task, mcp__n8n-mcp__validate_workflow, mcp__n8n-mcp__validate_workflow_connections, mcp__n8n-mcp__validate_workflow_expressions, mcp__n8n-mcp__n8n_create_workflow, mcp__n8n-mcp__n8n_get_workflow, mcp__n8n-mcp__n8n_get_workflow_details, mcp__n8n-mcp__n8n_get_workflow_structure, mcp__n8n-mcp__n8n_get_workflow_minimal, mcp__n8n-mcp__n8n_update_full_workflow, mcp__n8n-mcp__n8n_update_partial_workflow, mcp__n8n-mcp__n8n_delete_workflow, mcp__n8n-mcp__n8n_list_workflows, mcp__n8n-mcp__n8n_validate_workflow, mcp__n8n-mcp__n8n_trigger_webhook_workflow, mcp__n8n-mcp__n8n_get_execution, mcp__n8n-mcp__n8n_list_executions, mcp__n8n-mcp__n8n_delete_execution, mcp__n8n-mcp__n8n_health_check, mcp__n8n-mcp__n8n_list_available_tools, mcp__n8n-mcp__n8n_diagnostic
|
||||
model: sonnet
|
||||
---
|
||||
|
||||
You are n8n-mcp-tester, a specialized testing agent for the n8n Model Context Protocol (MCP) server. You validate that MCP tools and functionality work correctly in real-world scenarios after fixes or new features are implemented.
|
||||
|
||||
## Your Core Responsibilities
|
||||
|
||||
You test the n8n-mcp server by:
|
||||
1. Using MCP tools to build, validate, and manipulate n8n workflows
|
||||
2. Verifying that recent fixes resolve the reported issues
|
||||
3. Testing new functionality works as designed
|
||||
4. Reporting clear, actionable results back to the invoking agent
|
||||
|
||||
## Testing Methodology
|
||||
|
||||
When invoked with a test request, you will:
|
||||
|
||||
1. **Understand the Context**: Identify what was fixed or added based on the instructions from the invoking agent
|
||||
|
||||
2. **Design Test Scenarios**: Create specific test cases that:
|
||||
- Target the exact functionality that was changed
|
||||
- Include both positive and negative test cases
|
||||
- Test edge cases and boundary conditions
|
||||
- Use realistic n8n workflow configurations
|
||||
|
||||
3. **Execute Tests Using MCP Tools**: You have access to all n8n-mcp tools including:
|
||||
- `search_nodes`: Find relevant n8n nodes
|
||||
- `get_node_info`: Get detailed node configuration
|
||||
- `get_node_essentials`: Get simplified node information
|
||||
- `validate_node_config`: Validate node configurations
|
||||
- `n8n_validate_workflow`: Validate complete workflows
|
||||
- `get_node_example`: Get working examples
|
||||
- `search_templates`: Find workflow templates
|
||||
- Additional tools as available in the MCP server
|
||||
|
||||
4. **Verify Expected Behavior**:
|
||||
- Confirm fixes resolve the original issue
|
||||
- Verify new features work as documented
|
||||
- Check for regressions in related functionality
|
||||
- Test error handling and edge cases
|
||||
|
||||
5. **Report Results**: Provide clear feedback including:
|
||||
- What was tested (specific tools and scenarios)
|
||||
- Whether the fix/feature works as expected
|
||||
- Any unexpected behaviors or issues discovered
|
||||
- Specific error messages if failures occur
|
||||
- Recommendations for additional testing if needed
|
||||
|
||||
## Testing Guidelines
|
||||
|
||||
- **Be Thorough**: Test multiple variations and edge cases
|
||||
- **Be Specific**: Use exact node types, properties, and configurations mentioned in the fix
|
||||
- **Be Realistic**: Create test scenarios that mirror actual n8n usage
|
||||
- **Be Clear**: Report results in a structured, easy-to-understand format
|
||||
- **Be Efficient**: Focus testing on the changed functionality first
|
||||
|
||||
## Example Test Execution
|
||||
|
||||
If testing a validation fix for HTTP Request nodes:
|
||||
1. Call `tools_documentation` to get a list of available tools and get documentation on `search_nodes` tool.
|
||||
2. Search for HTTP Request node using `search_nodes`
|
||||
3. Get node configuration with `get_node_info` or `get_node_essentials`
|
||||
4. Create test configurations that previously failed
|
||||
5. Validate using `validate_node_config` with different profiles
|
||||
6. Test in a complete workflow using `n8n_validate_workflow`
|
||||
6. Report whether validation now works correctly
|
||||
|
||||
## Important Constraints
|
||||
|
||||
- You can only test using the MCP tools available in the server
|
||||
- You cannot modify code or files - only test existing functionality
|
||||
- You must work with the current state of the MCP server (already reloaded)
|
||||
- Focus on functional testing, not unit testing
|
||||
- Report issues objectively without attempting to fix them
|
||||
|
||||
## Response Format
|
||||
|
||||
Structure your test results as:
|
||||
|
||||
```
|
||||
### Test Report: [Feature/Fix Name]
|
||||
|
||||
**Test Objective**: [What was being tested]
|
||||
|
||||
**Test Scenarios**:
|
||||
1. [Scenario 1]: ✅/❌ [Result]
|
||||
2. [Scenario 2]: ✅/❌ [Result]
|
||||
|
||||
**Findings**:
|
||||
- [Key finding 1]
|
||||
- [Key finding 2]
|
||||
|
||||
**Conclusion**: [Overall assessment - works as expected / issues found]
|
||||
|
||||
**Details**: [Any error messages, unexpected behaviors, or additional context]
|
||||
```
|
||||
|
||||
Remember: Your role is to validate that the n8n-mcp server works correctly in practice, providing confidence that fixes and new features function as intended before deployment.
|
||||
114
.env.example
114
.env.example
@@ -69,6 +69,55 @@ AUTH_TOKEN=your-secure-token-here
|
||||
# Default: 0 (disabled)
|
||||
# TRUST_PROXY=0
|
||||
|
||||
# =========================
|
||||
# SECURITY CONFIGURATION
|
||||
# =========================
|
||||
|
||||
# Rate Limiting Configuration
|
||||
# Protects authentication endpoint from brute force attacks
|
||||
# Window: Time period in milliseconds (default: 900000 = 15 minutes)
|
||||
# Max: Maximum authentication attempts per IP within window (default: 20)
|
||||
# AUTH_RATE_LIMIT_WINDOW=900000
|
||||
# AUTH_RATE_LIMIT_MAX=20
|
||||
|
||||
# SSRF Protection Mode
|
||||
# Prevents webhooks from accessing internal networks and cloud metadata
|
||||
#
|
||||
# Modes:
|
||||
# - strict (default): Block localhost + private IPs + cloud metadata
|
||||
# Use for: Production deployments, cloud environments
|
||||
# Security: Maximum
|
||||
#
|
||||
# - moderate: Allow localhost, block private IPs + cloud metadata
|
||||
# Use for: Local development with local n8n instance
|
||||
# Security: Good balance
|
||||
# Example: n8n running on http://localhost:5678 or http://host.docker.internal:5678
|
||||
#
|
||||
# - permissive: Allow localhost + private IPs, block cloud metadata
|
||||
# Use for: Internal network testing, private cloud (NOT for production)
|
||||
# Security: Minimal - use with caution
|
||||
#
|
||||
# Default: strict
|
||||
# WEBHOOK_SECURITY_MODE=strict
|
||||
#
|
||||
# For local development with local n8n:
|
||||
# WEBHOOK_SECURITY_MODE=moderate
|
||||
|
||||
# =========================
|
||||
# MULTI-TENANT CONFIGURATION
|
||||
# =========================
|
||||
# Enable multi-tenant mode for dynamic instance support
|
||||
# When enabled, n8n API tools will be available for all sessions,
|
||||
# and instance configuration will be determined from HTTP headers
|
||||
# Default: false (single-tenant mode using environment variables)
|
||||
ENABLE_MULTI_TENANT=false
|
||||
|
||||
# Session isolation strategy for multi-tenant mode
|
||||
# - "instance": Create separate sessions per instance ID (recommended)
|
||||
# - "shared": Share sessions but switch contexts (advanced)
|
||||
# Default: instance
|
||||
# MULTI_TENANT_SESSION_STRATEGY=instance
|
||||
|
||||
# =========================
|
||||
# N8N API CONFIGURATION
|
||||
# =========================
|
||||
@@ -86,4 +135,67 @@ AUTH_TOKEN=your-secure-token-here
|
||||
# N8N_API_TIMEOUT=30000
|
||||
|
||||
# Maximum number of API request retries (default: 3)
|
||||
# N8N_API_MAX_RETRIES=3
|
||||
# N8N_API_MAX_RETRIES=3
|
||||
|
||||
# =========================
|
||||
# CACHE CONFIGURATION
|
||||
# =========================
|
||||
# Optional: Configure instance cache settings for flexible instance support
|
||||
|
||||
# Maximum number of cached instances (default: 100, min: 1, max: 10000)
|
||||
# INSTANCE_CACHE_MAX=100
|
||||
|
||||
# Cache TTL in minutes (default: 30, min: 1, max: 1440/24 hours)
|
||||
# INSTANCE_CACHE_TTL_MINUTES=30
|
||||
|
||||
# =========================
|
||||
# OPENAI API CONFIGURATION
|
||||
# =========================
|
||||
# Optional: Enable AI-powered template metadata generation
|
||||
# Provides structured metadata for improved template discovery
|
||||
|
||||
# OpenAI API Key (get from https://platform.openai.com/api-keys)
|
||||
# OPENAI_API_KEY=
|
||||
|
||||
# OpenAI Model for metadata generation (default: gpt-4o-mini)
|
||||
# OPENAI_MODEL=gpt-4o-mini
|
||||
|
||||
# Batch size for metadata generation (default: 100)
|
||||
# Templates are processed in batches using OpenAI's Batch API for 50% cost savings
|
||||
# OPENAI_BATCH_SIZE=100
|
||||
|
||||
# Enable metadata generation during template fetch (default: false)
|
||||
# Set to true to automatically generate metadata when running fetch:templates
|
||||
# METADATA_GENERATION_ENABLED=false
|
||||
|
||||
# ========================================
|
||||
# INTEGRATION TESTING CONFIGURATION
|
||||
# ========================================
|
||||
# Configuration for integration tests that call real n8n instance API
|
||||
|
||||
# n8n API Configuration for Integration Tests
|
||||
# For local development: Use your local n8n instance
|
||||
# For CI: These will be provided by GitHub secrets
|
||||
# N8N_API_URL=http://localhost:5678
|
||||
# N8N_API_KEY=
|
||||
|
||||
# Pre-activated Webhook Workflows for Testing
|
||||
# These workflows must be created manually in n8n and activated
|
||||
# because n8n API doesn't support workflow activation.
|
||||
#
|
||||
# Setup Instructions:
|
||||
# 1. Create 4 workflows in n8n UI (one for each HTTP method)
|
||||
# 2. Each workflow should have a single Webhook node
|
||||
# 3. Configure webhook paths: mcp-test-get, mcp-test-post, mcp-test-put, mcp-test-delete
|
||||
# 4. ACTIVATE each workflow in n8n UI
|
||||
# 5. Copy the workflow IDs here
|
||||
#
|
||||
# N8N_TEST_WEBHOOK_GET_ID= # Workflow ID for GET method webhook
|
||||
# N8N_TEST_WEBHOOK_POST_ID= # Workflow ID for POST method webhook
|
||||
# N8N_TEST_WEBHOOK_PUT_ID= # Workflow ID for PUT method webhook
|
||||
# N8N_TEST_WEBHOOK_DELETE_ID= # Workflow ID for DELETE method webhook
|
||||
|
||||
# Test Configuration
|
||||
N8N_TEST_CLEANUP_ENABLED=true # Enable automatic cleanup of test workflows
|
||||
N8N_TEST_TAG=mcp-integration-test # Tag applied to all test workflows
|
||||
N8N_TEST_NAME_PREFIX=[MCP-TEST] # Name prefix for test workflows
|
||||
135
.github/workflows/benchmark-pr.yml
vendored
135
.github/workflows/benchmark-pr.yml
vendored
@@ -2,11 +2,19 @@ name: Benchmark PR Comparison
|
||||
on:
|
||||
pull_request:
|
||||
branches: [main]
|
||||
paths:
|
||||
- 'src/**'
|
||||
- 'tests/benchmarks/**'
|
||||
- 'package.json'
|
||||
- 'vitest.config.benchmark.ts'
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
|
||||
permissions:
|
||||
pull-requests: write
|
||||
@@ -85,71 +93,84 @@ jobs:
|
||||
- name: Post benchmark comparison to PR
|
||||
if: always()
|
||||
uses: actions/github-script@v7
|
||||
continue-on-error: true
|
||||
with:
|
||||
script: |
|
||||
const fs = require('fs');
|
||||
let comment = '## ⚡ Benchmark Comparison\n\n';
|
||||
|
||||
try {
|
||||
if (fs.existsSync('benchmark-comparison.md')) {
|
||||
const comparison = fs.readFileSync('benchmark-comparison.md', 'utf8');
|
||||
comment += comparison;
|
||||
} else {
|
||||
comment += 'Benchmark comparison could not be generated.';
|
||||
const fs = require('fs');
|
||||
let comment = '## ⚡ Benchmark Comparison\n\n';
|
||||
|
||||
try {
|
||||
if (fs.existsSync('benchmark-comparison.md')) {
|
||||
const comparison = fs.readFileSync('benchmark-comparison.md', 'utf8');
|
||||
comment += comparison;
|
||||
} else {
|
||||
comment += 'Benchmark comparison could not be generated.';
|
||||
}
|
||||
} catch (error) {
|
||||
comment += `Error reading benchmark comparison: ${error.message}`;
|
||||
}
|
||||
} catch (error) {
|
||||
comment += `Error reading benchmark comparison: ${error.message}`;
|
||||
}
|
||||
|
||||
comment += '\n\n---\n';
|
||||
comment += `*[View full benchmark results](https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }})*`;
|
||||
|
||||
// Find existing comment
|
||||
const { data: comments } = await github.rest.issues.listComments({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
issue_number: context.issue.number,
|
||||
});
|
||||
|
||||
const botComment = comments.find(comment =>
|
||||
comment.user.type === 'Bot' &&
|
||||
comment.body.includes('## ⚡ Benchmark Comparison')
|
||||
);
|
||||
|
||||
if (botComment) {
|
||||
await github.rest.issues.updateComment({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
comment_id: botComment.id,
|
||||
body: comment
|
||||
});
|
||||
} else {
|
||||
await github.rest.issues.createComment({
|
||||
|
||||
comment += '\n\n---\n';
|
||||
comment += `*[View full benchmark results](https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }})*`;
|
||||
|
||||
// Find existing comment
|
||||
const { data: comments } = await github.rest.issues.listComments({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
issue_number: context.issue.number,
|
||||
body: comment
|
||||
});
|
||||
|
||||
const botComment = comments.find(comment =>
|
||||
comment.user.type === 'Bot' &&
|
||||
comment.body.includes('## ⚡ Benchmark Comparison')
|
||||
);
|
||||
|
||||
if (botComment) {
|
||||
await github.rest.issues.updateComment({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
comment_id: botComment.id,
|
||||
body: comment
|
||||
});
|
||||
} else {
|
||||
await github.rest.issues.createComment({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
issue_number: context.issue.number,
|
||||
body: comment
|
||||
});
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Failed to create/update PR comment:', error.message);
|
||||
console.log('This is likely due to insufficient permissions for external PRs.');
|
||||
console.log('Benchmark comparison has been saved to artifacts instead.');
|
||||
}
|
||||
|
||||
# Add status check
|
||||
- name: Set benchmark status
|
||||
if: always()
|
||||
uses: actions/github-script@v7
|
||||
continue-on-error: true
|
||||
with:
|
||||
script: |
|
||||
const hasRegression = '${{ steps.compare.outputs.REGRESSION }}' === 'true';
|
||||
const state = hasRegression ? 'failure' : 'success';
|
||||
const description = hasRegression
|
||||
? 'Performance regressions detected'
|
||||
: 'No performance regressions';
|
||||
|
||||
await github.rest.repos.createCommitStatus({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
sha: context.sha,
|
||||
state: state,
|
||||
target_url: `https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}`,
|
||||
description: description,
|
||||
context: 'benchmarks/regression-check'
|
||||
});
|
||||
try {
|
||||
const hasRegression = '${{ steps.compare.outputs.REGRESSION }}' === 'true';
|
||||
const state = hasRegression ? 'failure' : 'success';
|
||||
const description = hasRegression
|
||||
? 'Performance regressions detected'
|
||||
: 'No performance regressions';
|
||||
|
||||
await github.rest.repos.createCommitStatus({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
sha: context.sha,
|
||||
state: state,
|
||||
target_url: `https://github.com/${{ github.repository }}/actions/runs/${{ github.run_id }}`,
|
||||
description: description,
|
||||
context: 'benchmarks/regression-check'
|
||||
});
|
||||
} catch (error) {
|
||||
console.error('Failed to create commit status:', error.message);
|
||||
console.log('This is likely due to insufficient permissions for external PRs.');
|
||||
}
|
||||
110
.github/workflows/benchmark.yml
vendored
110
.github/workflows/benchmark.yml
vendored
@@ -3,8 +3,34 @@ name: Performance Benchmarks
|
||||
on:
|
||||
push:
|
||||
branches: [main, feat/comprehensive-testing-suite]
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
pull_request:
|
||||
branches: [main]
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
workflow_dispatch:
|
||||
|
||||
permissions:
|
||||
@@ -77,12 +103,14 @@ jobs:
|
||||
# Store benchmark results and compare
|
||||
- name: Store benchmark result
|
||||
uses: benchmark-action/github-action-benchmark@v1
|
||||
continue-on-error: true
|
||||
id: benchmark
|
||||
with:
|
||||
name: n8n-mcp Benchmarks
|
||||
tool: 'customSmallerIsBetter'
|
||||
output-file-path: benchmark-results-formatted.json
|
||||
github-token: ${{ secrets.GITHUB_TOKEN }}
|
||||
auto-push: true
|
||||
auto-push: ${{ github.event_name == 'push' && github.ref == 'refs/heads/main' }}
|
||||
# Where to store benchmark data
|
||||
benchmark-data-dir-path: 'benchmarks'
|
||||
# Alert when performance regresses by 10%
|
||||
@@ -94,52 +122,60 @@ jobs:
|
||||
summary-always: true
|
||||
# Max number of data points to retain
|
||||
max-items-in-chart: 50
|
||||
fail-on-alert: false
|
||||
|
||||
# Comment on PR with benchmark results
|
||||
- name: Comment PR with results
|
||||
uses: actions/github-script@v7
|
||||
if: github.event_name == 'pull_request'
|
||||
continue-on-error: true
|
||||
with:
|
||||
github-token: ${{ secrets.GITHUB_TOKEN }}
|
||||
script: |
|
||||
const fs = require('fs');
|
||||
const summary = JSON.parse(fs.readFileSync('benchmark-summary.json', 'utf8'));
|
||||
|
||||
// Format results for PR comment
|
||||
let comment = '## 📊 Performance Benchmark Results\n\n';
|
||||
comment += `🕐 Run at: ${new Date(summary.timestamp).toLocaleString()}\n\n`;
|
||||
comment += '| Benchmark | Time | Ops/sec | Range |\n';
|
||||
comment += '|-----------|------|---------|-------|\n';
|
||||
|
||||
// Group benchmarks by category
|
||||
const categories = {};
|
||||
for (const benchmark of summary.benchmarks) {
|
||||
const [category, ...nameParts] = benchmark.name.split(' - ');
|
||||
if (!categories[category]) categories[category] = [];
|
||||
categories[category].push({
|
||||
...benchmark,
|
||||
shortName: nameParts.join(' - ')
|
||||
});
|
||||
}
|
||||
|
||||
// Display by category
|
||||
for (const [category, benchmarks] of Object.entries(categories)) {
|
||||
comment += `\n### ${category}\n`;
|
||||
for (const benchmark of benchmarks) {
|
||||
comment += `| ${benchmark.shortName} | ${benchmark.time} | ${benchmark.opsPerSec} | ${benchmark.range} |\n`;
|
||||
try {
|
||||
const fs = require('fs');
|
||||
const summary = JSON.parse(fs.readFileSync('benchmark-summary.json', 'utf8'));
|
||||
|
||||
// Format results for PR comment
|
||||
let comment = '## 📊 Performance Benchmark Results\n\n';
|
||||
comment += `🕐 Run at: ${new Date(summary.timestamp).toLocaleString()}\n\n`;
|
||||
comment += '| Benchmark | Time | Ops/sec | Range |\n';
|
||||
comment += '|-----------|------|---------|-------|\n';
|
||||
|
||||
// Group benchmarks by category
|
||||
const categories = {};
|
||||
for (const benchmark of summary.benchmarks) {
|
||||
const [category, ...nameParts] = benchmark.name.split(' - ');
|
||||
if (!categories[category]) categories[category] = [];
|
||||
categories[category].push({
|
||||
...benchmark,
|
||||
shortName: nameParts.join(' - ')
|
||||
});
|
||||
}
|
||||
|
||||
// Display by category
|
||||
for (const [category, benchmarks] of Object.entries(categories)) {
|
||||
comment += `\n### ${category}\n`;
|
||||
for (const benchmark of benchmarks) {
|
||||
comment += `| ${benchmark.shortName} | ${benchmark.time} | ${benchmark.opsPerSec} | ${benchmark.range} |\n`;
|
||||
}
|
||||
}
|
||||
|
||||
// Add comparison link
|
||||
comment += '\n\n📈 [View historical benchmark trends](https://czlonkowski.github.io/n8n-mcp/benchmarks/)\n';
|
||||
comment += '\n⚡ Performance regressions >10% will be flagged automatically.\n';
|
||||
|
||||
await github.rest.issues.createComment({
|
||||
issue_number: context.issue.number,
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
body: comment
|
||||
});
|
||||
} catch (error) {
|
||||
console.error('Failed to create PR comment:', error.message);
|
||||
console.log('This is likely due to insufficient permissions for external PRs.');
|
||||
console.log('Benchmark results have been saved to artifacts instead.');
|
||||
}
|
||||
|
||||
// Add comparison link
|
||||
comment += '\n\n📈 [View historical benchmark trends](https://czlonkowski.github.io/n8n-mcp/benchmarks/)\n';
|
||||
comment += '\n⚡ Performance regressions >10% will be flagged automatically.\n';
|
||||
|
||||
github.rest.issues.createComment({
|
||||
issue_number: context.issue.number,
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
body: comment
|
||||
});
|
||||
|
||||
# Deploy benchmark results to GitHub Pages
|
||||
deploy:
|
||||
|
||||
26
.github/workflows/docker-build-n8n.yml
vendored
26
.github/workflows/docker-build-n8n.yml
vendored
@@ -6,9 +6,35 @@ on:
|
||||
- main
|
||||
tags:
|
||||
- 'v*'
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
pull_request:
|
||||
branches:
|
||||
- main
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
workflow_dispatch:
|
||||
|
||||
env:
|
||||
|
||||
70
.github/workflows/docker-build.yml
vendored
70
.github/workflows/docker-build.yml
vendored
@@ -5,29 +5,43 @@ on:
|
||||
push:
|
||||
branches:
|
||||
- main
|
||||
tags:
|
||||
- 'v*'
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- 'LICENSE'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'docs/**'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
pull_request:
|
||||
branches:
|
||||
- main
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- 'LICENSE'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'docs/**'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
workflow_dispatch:
|
||||
|
||||
# Prevent concurrent Docker pushes across all workflows (shared with release.yml)
|
||||
# This ensures docker-build.yml and release.yml never push to 'latest' simultaneously
|
||||
concurrency:
|
||||
group: docker-push-${{ github.ref }}
|
||||
cancel-in-progress: false
|
||||
|
||||
env:
|
||||
REGISTRY: ghcr.io
|
||||
IMAGE_NAME: ${{ github.repository }}
|
||||
@@ -79,16 +93,54 @@ jobs:
|
||||
uses: docker/build-push-action@v5
|
||||
with:
|
||||
context: .
|
||||
no-cache: true
|
||||
no-cache: false
|
||||
platforms: linux/amd64,linux/arm64
|
||||
push: ${{ github.event_name != 'pull_request' }}
|
||||
tags: ${{ steps.meta.outputs.tags }}
|
||||
labels: ${{ steps.meta.outputs.labels }}
|
||||
cache-from: type=gha
|
||||
cache-to: type=gha,mode=max
|
||||
provenance: false
|
||||
|
||||
- name: Verify multi-arch manifest for latest tag
|
||||
if: github.event_name != 'pull_request' && github.ref == 'refs/heads/main'
|
||||
run: |
|
||||
echo "Verifying multi-arch manifest for latest tag..."
|
||||
|
||||
# Retry with exponential backoff (registry propagation can take time)
|
||||
MAX_ATTEMPTS=5
|
||||
ATTEMPT=1
|
||||
WAIT_TIME=2
|
||||
|
||||
while [ $ATTEMPT -le $MAX_ATTEMPTS ]; do
|
||||
echo "Attempt $ATTEMPT of $MAX_ATTEMPTS..."
|
||||
|
||||
MANIFEST=$(docker buildx imagetools inspect ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:latest 2>&1 || true)
|
||||
|
||||
# Check for both platforms
|
||||
if echo "$MANIFEST" | grep -q "linux/amd64" && echo "$MANIFEST" | grep -q "linux/arm64"; then
|
||||
echo "✅ Multi-arch manifest verified: both amd64 and arm64 present"
|
||||
echo "$MANIFEST"
|
||||
exit 0
|
||||
fi
|
||||
|
||||
if [ $ATTEMPT -lt $MAX_ATTEMPTS ]; then
|
||||
echo "⏳ Registry still propagating, waiting ${WAIT_TIME}s before retry..."
|
||||
sleep $WAIT_TIME
|
||||
WAIT_TIME=$((WAIT_TIME * 2)) # Exponential backoff: 2s, 4s, 8s, 16s
|
||||
fi
|
||||
|
||||
ATTEMPT=$((ATTEMPT + 1))
|
||||
done
|
||||
|
||||
echo "❌ ERROR: Multi-arch manifest incomplete after $MAX_ATTEMPTS attempts!"
|
||||
echo "$MANIFEST"
|
||||
exit 1
|
||||
|
||||
build-railway:
|
||||
name: Build Railway Docker Image
|
||||
runs-on: ubuntu-latest
|
||||
needs: build
|
||||
permissions:
|
||||
contents: read
|
||||
packages: write
|
||||
@@ -133,11 +185,13 @@ jobs:
|
||||
with:
|
||||
context: .
|
||||
file: ./Dockerfile.railway
|
||||
no-cache: true
|
||||
no-cache: false
|
||||
platforms: linux/amd64
|
||||
push: ${{ github.event_name != 'pull_request' }}
|
||||
tags: ${{ steps.meta-railway.outputs.tags }}
|
||||
labels: ${{ steps.meta-railway.outputs.labels }}
|
||||
cache-from: type=gha
|
||||
cache-to: type=gha,mode=max
|
||||
provenance: false
|
||||
|
||||
# Nginx build commented out until Phase 2
|
||||
|
||||
636
.github/workflows/release.yml
vendored
Normal file
636
.github/workflows/release.yml
vendored
Normal file
@@ -0,0 +1,636 @@
|
||||
name: Automated Release
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
paths:
|
||||
- 'package.json'
|
||||
- 'package.runtime.json'
|
||||
|
||||
permissions:
|
||||
contents: write
|
||||
packages: write
|
||||
issues: write
|
||||
pull-requests: write
|
||||
|
||||
# Prevent concurrent Docker pushes across all workflows (shared with docker-build.yml)
|
||||
# This ensures release.yml and docker-build.yml never push to 'latest' simultaneously
|
||||
concurrency:
|
||||
group: docker-push-${{ github.ref }}
|
||||
cancel-in-progress: false
|
||||
|
||||
env:
|
||||
REGISTRY: ghcr.io
|
||||
IMAGE_NAME: ${{ github.repository }}
|
||||
|
||||
jobs:
|
||||
detect-version-change:
|
||||
name: Detect Version Change
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
version-changed: ${{ steps.check.outputs.changed }}
|
||||
new-version: ${{ steps.check.outputs.version }}
|
||||
previous-version: ${{ steps.check.outputs.previous-version }}
|
||||
is-prerelease: ${{ steps.check.outputs.is-prerelease }}
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 2
|
||||
|
||||
- name: Check for version change
|
||||
id: check
|
||||
run: |
|
||||
# Get current version from package.json
|
||||
CURRENT_VERSION=$(node -e "console.log(require('./package.json').version)")
|
||||
|
||||
# Get previous version from git history safely
|
||||
PREVIOUS_VERSION=$(git show HEAD~1:package.json 2>/dev/null | node -e "
|
||||
try {
|
||||
const data = require('fs').readFileSync(0, 'utf8');
|
||||
const pkg = JSON.parse(data);
|
||||
console.log(pkg.version || '0.0.0');
|
||||
} catch (e) {
|
||||
console.log('0.0.0');
|
||||
}
|
||||
" || echo "0.0.0")
|
||||
|
||||
echo "Previous version: $PREVIOUS_VERSION"
|
||||
echo "Current version: $CURRENT_VERSION"
|
||||
|
||||
# Check if version changed
|
||||
if [ "$CURRENT_VERSION" != "$PREVIOUS_VERSION" ]; then
|
||||
echo "changed=true" >> $GITHUB_OUTPUT
|
||||
echo "version=$CURRENT_VERSION" >> $GITHUB_OUTPUT
|
||||
echo "previous-version=$PREVIOUS_VERSION" >> $GITHUB_OUTPUT
|
||||
|
||||
# Check if it's a prerelease (contains alpha, beta, rc, dev)
|
||||
if echo "$CURRENT_VERSION" | grep -E "(alpha|beta|rc|dev)" > /dev/null; then
|
||||
echo "is-prerelease=true" >> $GITHUB_OUTPUT
|
||||
else
|
||||
echo "is-prerelease=false" >> $GITHUB_OUTPUT
|
||||
fi
|
||||
|
||||
echo "🎉 Version changed from $PREVIOUS_VERSION to $CURRENT_VERSION"
|
||||
else
|
||||
echo "changed=false" >> $GITHUB_OUTPUT
|
||||
echo "version=$CURRENT_VERSION" >> $GITHUB_OUTPUT
|
||||
echo "previous-version=$PREVIOUS_VERSION" >> $GITHUB_OUTPUT
|
||||
echo "is-prerelease=false" >> $GITHUB_OUTPUT
|
||||
echo "ℹ️ No version change detected"
|
||||
fi
|
||||
|
||||
- name: Validate version against npm registry
|
||||
if: steps.check.outputs.changed == 'true'
|
||||
run: |
|
||||
CURRENT_VERSION="${{ steps.check.outputs.version }}"
|
||||
|
||||
# Get latest version from npm (handle package not found)
|
||||
NPM_VERSION=$(npm view n8n-mcp version 2>/dev/null || echo "0.0.0")
|
||||
|
||||
echo "Current version: $CURRENT_VERSION"
|
||||
echo "NPM registry version: $NPM_VERSION"
|
||||
|
||||
# Check if version already exists in npm
|
||||
if [ "$CURRENT_VERSION" = "$NPM_VERSION" ]; then
|
||||
echo "❌ Error: Version $CURRENT_VERSION already published to npm"
|
||||
echo "Please bump the version in package.json before releasing"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Simple semver comparison (assumes format: major.minor.patch)
|
||||
# Compare if current version is greater than npm version
|
||||
if [ "$NPM_VERSION" != "0.0.0" ]; then
|
||||
# Sort versions and check if current is not the highest
|
||||
HIGHEST=$(printf '%s\n%s' "$NPM_VERSION" "$CURRENT_VERSION" | sort -V | tail -n1)
|
||||
if [ "$HIGHEST" != "$CURRENT_VERSION" ]; then
|
||||
echo "❌ Error: Version $CURRENT_VERSION is not greater than npm version $NPM_VERSION"
|
||||
echo "Please use a higher version number"
|
||||
exit 1
|
||||
fi
|
||||
fi
|
||||
|
||||
echo "✅ Version $CURRENT_VERSION is valid (higher than npm version $NPM_VERSION)"
|
||||
|
||||
extract-changelog:
|
||||
name: Extract Changelog
|
||||
runs-on: ubuntu-latest
|
||||
needs: detect-version-change
|
||||
if: needs.detect-version-change.outputs.version-changed == 'true'
|
||||
outputs:
|
||||
release-notes: ${{ steps.extract.outputs.notes }}
|
||||
has-notes: ${{ steps.extract.outputs.has-notes }}
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Extract changelog for version
|
||||
id: extract
|
||||
run: |
|
||||
VERSION="${{ needs.detect-version-change.outputs.new-version }}"
|
||||
CHANGELOG_FILE="docs/CHANGELOG.md"
|
||||
|
||||
if [ ! -f "$CHANGELOG_FILE" ]; then
|
||||
echo "Changelog file not found at $CHANGELOG_FILE"
|
||||
echo "has-notes=false" >> $GITHUB_OUTPUT
|
||||
echo "notes=No changelog entries found for version $VERSION" >> $GITHUB_OUTPUT
|
||||
exit 0
|
||||
fi
|
||||
|
||||
# Use the extracted changelog script
|
||||
if NOTES=$(node scripts/extract-changelog.js "$VERSION" "$CHANGELOG_FILE" 2>/dev/null); then
|
||||
echo "has-notes=true" >> $GITHUB_OUTPUT
|
||||
|
||||
# Use heredoc to properly handle multiline content
|
||||
{
|
||||
echo "notes<<EOF"
|
||||
echo "$NOTES"
|
||||
echo "EOF"
|
||||
} >> $GITHUB_OUTPUT
|
||||
|
||||
echo "✅ Successfully extracted changelog for version $VERSION"
|
||||
else
|
||||
echo "has-notes=false" >> $GITHUB_OUTPUT
|
||||
echo "notes=No changelog entries found for version $VERSION" >> $GITHUB_OUTPUT
|
||||
echo "⚠️ Could not extract changelog for version $VERSION"
|
||||
fi
|
||||
|
||||
create-release:
|
||||
name: Create GitHub Release
|
||||
runs-on: ubuntu-latest
|
||||
needs: [detect-version-change, extract-changelog]
|
||||
if: needs.detect-version-change.outputs.version-changed == 'true'
|
||||
outputs:
|
||||
release-id: ${{ steps.create.outputs.id }}
|
||||
upload-url: ${{ steps.create.outputs.upload_url }}
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Create Git Tag
|
||||
run: |
|
||||
VERSION="${{ needs.detect-version-change.outputs.new-version }}"
|
||||
git config user.name "github-actions[bot]"
|
||||
git config user.email "github-actions[bot]@users.noreply.github.com"
|
||||
|
||||
# Create annotated tag
|
||||
git tag -a "v$VERSION" -m "Release v$VERSION"
|
||||
git push origin "v$VERSION"
|
||||
|
||||
- name: Create GitHub Release
|
||||
id: create
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
run: |
|
||||
VERSION="${{ needs.detect-version-change.outputs.new-version }}"
|
||||
IS_PRERELEASE="${{ needs.detect-version-change.outputs.is-prerelease }}"
|
||||
|
||||
# Create release body
|
||||
cat > release_body.md << 'EOF'
|
||||
# Release v${{ needs.detect-version-change.outputs.new-version }}
|
||||
|
||||
${{ needs.extract-changelog.outputs.release-notes }}
|
||||
|
||||
---
|
||||
|
||||
## Installation
|
||||
|
||||
### NPM Package
|
||||
```bash
|
||||
# Install globally
|
||||
npm install -g n8n-mcp
|
||||
|
||||
# Or run directly
|
||||
npx n8n-mcp
|
||||
```
|
||||
|
||||
### Docker
|
||||
```bash
|
||||
# Standard image
|
||||
docker run -p 3000:3000 ghcr.io/czlonkowski/n8n-mcp:v${{ needs.detect-version-change.outputs.new-version }}
|
||||
|
||||
# Railway optimized
|
||||
docker run -p 3000:3000 ghcr.io/czlonkowski/n8n-mcp-railway:v${{ needs.detect-version-change.outputs.new-version }}
|
||||
```
|
||||
|
||||
## Documentation
|
||||
- [Installation Guide](https://github.com/czlonkowski/n8n-mcp#installation)
|
||||
- [Docker Deployment](https://github.com/czlonkowski/n8n-mcp/blob/main/docs/DOCKER_README.md)
|
||||
- [n8n Integration](https://github.com/czlonkowski/n8n-mcp/blob/main/docs/N8N_DEPLOYMENT.md)
|
||||
- [Complete Changelog](https://github.com/czlonkowski/n8n-mcp/blob/main/docs/CHANGELOG.md)
|
||||
|
||||
🤖 *Generated with [Claude Code](https://claude.ai/code)*
|
||||
EOF
|
||||
|
||||
# Create release using gh CLI
|
||||
if [ "$IS_PRERELEASE" = "true" ]; then
|
||||
PRERELEASE_FLAG="--prerelease"
|
||||
else
|
||||
PRERELEASE_FLAG=""
|
||||
fi
|
||||
|
||||
gh release create "v$VERSION" \
|
||||
--title "Release v$VERSION" \
|
||||
--notes-file release_body.md \
|
||||
$PRERELEASE_FLAG
|
||||
|
||||
# Output release info for next jobs
|
||||
RELEASE_ID=$(gh release view "v$VERSION" --json id --jq '.id')
|
||||
echo "id=$RELEASE_ID" >> $GITHUB_OUTPUT
|
||||
echo "upload_url=https://uploads.github.com/repos/${{ github.repository }}/releases/$RELEASE_ID/assets{?name,label}" >> $GITHUB_OUTPUT
|
||||
|
||||
build-and-verify:
|
||||
name: Build and Verify
|
||||
runs-on: ubuntu-latest
|
||||
needs: detect-version-change
|
||||
if: needs.detect-version-change.outputs.version-changed == 'true'
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Setup Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: 20
|
||||
cache: 'npm'
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Build project
|
||||
run: npm run build
|
||||
|
||||
# Database is already built and committed during development
|
||||
# Rebuilding here causes segfault due to memory pressure (exit code 139)
|
||||
- name: Verify database exists
|
||||
run: |
|
||||
if [ ! -f "data/nodes.db" ]; then
|
||||
echo "❌ Error: data/nodes.db not found"
|
||||
echo "Please run 'npm run rebuild' locally and commit the database"
|
||||
exit 1
|
||||
fi
|
||||
echo "✅ Database exists ($(du -h data/nodes.db | cut -f1))"
|
||||
|
||||
# Skip tests - they already passed in PR before merge
|
||||
# Running them again on the same commit adds no safety, only time (~6-7 min)
|
||||
|
||||
- name: Run type checking
|
||||
run: npm run typecheck
|
||||
|
||||
publish-npm:
|
||||
name: Publish to NPM
|
||||
runs-on: ubuntu-latest
|
||||
needs: [detect-version-change, build-and-verify, create-release]
|
||||
if: needs.detect-version-change.outputs.version-changed == 'true'
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Setup Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: 20
|
||||
cache: 'npm'
|
||||
registry-url: 'https://registry.npmjs.org'
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Build project
|
||||
run: npm run build
|
||||
|
||||
# Database is already built and committed during development
|
||||
- name: Verify database exists
|
||||
run: |
|
||||
if [ ! -f "data/nodes.db" ]; then
|
||||
echo "❌ Error: data/nodes.db not found"
|
||||
exit 1
|
||||
fi
|
||||
echo "✅ Database exists ($(du -h data/nodes.db | cut -f1))"
|
||||
|
||||
- name: Sync runtime version
|
||||
run: npm run sync:runtime-version
|
||||
|
||||
- name: Prepare package for publishing
|
||||
run: |
|
||||
# Create publish directory
|
||||
PUBLISH_DIR="npm-publish-temp"
|
||||
rm -rf $PUBLISH_DIR
|
||||
mkdir -p $PUBLISH_DIR
|
||||
|
||||
# Copy necessary files
|
||||
cp -r dist $PUBLISH_DIR/
|
||||
cp -r data $PUBLISH_DIR/
|
||||
cp README.md $PUBLISH_DIR/
|
||||
cp LICENSE $PUBLISH_DIR/
|
||||
cp .env.example $PUBLISH_DIR/
|
||||
|
||||
# Use runtime package.json as base
|
||||
cp package.runtime.json $PUBLISH_DIR/package.json
|
||||
|
||||
cd $PUBLISH_DIR
|
||||
|
||||
# Update package.json with complete metadata
|
||||
node -e "
|
||||
const pkg = require('./package.json');
|
||||
pkg.name = 'n8n-mcp';
|
||||
pkg.description = 'Integration between n8n workflow automation and Model Context Protocol (MCP)';
|
||||
pkg.main = 'dist/index.js';
|
||||
pkg.types = 'dist/index.d.ts';
|
||||
pkg.exports = {
|
||||
'.': {
|
||||
types: './dist/index.d.ts',
|
||||
require: './dist/index.js',
|
||||
import: './dist/index.js'
|
||||
}
|
||||
};
|
||||
pkg.bin = { 'n8n-mcp': './dist/mcp/index.js' };
|
||||
pkg.repository = { type: 'git', url: 'git+https://github.com/czlonkowski/n8n-mcp.git' };
|
||||
pkg.keywords = ['n8n', 'mcp', 'model-context-protocol', 'ai', 'workflow', 'automation'];
|
||||
pkg.author = 'Romuald Czlonkowski @ www.aiadvisors.pl/en';
|
||||
pkg.license = 'MIT';
|
||||
pkg.bugs = { url: 'https://github.com/czlonkowski/n8n-mcp/issues' };
|
||||
pkg.homepage = 'https://github.com/czlonkowski/n8n-mcp#readme';
|
||||
pkg.files = ['dist/**/*', 'data/nodes.db', '.env.example', 'README.md', 'LICENSE'];
|
||||
delete pkg.private;
|
||||
require('fs').writeFileSync('./package.json', JSON.stringify(pkg, null, 2));
|
||||
"
|
||||
|
||||
echo "Package prepared for publishing:"
|
||||
echo "Name: $(node -e "console.log(require('./package.json').name)")"
|
||||
echo "Version: $(node -e "console.log(require('./package.json').version)")"
|
||||
|
||||
- name: Publish to NPM with retry
|
||||
uses: nick-invision/retry@v2
|
||||
with:
|
||||
timeout_minutes: 5
|
||||
max_attempts: 3
|
||||
command: |
|
||||
cd npm-publish-temp
|
||||
npm publish --access public
|
||||
env:
|
||||
NODE_AUTH_TOKEN: ${{ secrets.NPM_TOKEN }}
|
||||
|
||||
- name: Clean up
|
||||
if: always()
|
||||
run: rm -rf npm-publish-temp
|
||||
|
||||
build-docker:
|
||||
name: Build and Push Docker Images
|
||||
runs-on: ubuntu-latest
|
||||
needs: [detect-version-change, build-and-verify]
|
||||
if: needs.detect-version-change.outputs.version-changed == 'true'
|
||||
permissions:
|
||||
contents: read
|
||||
packages: write
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
lfs: true
|
||||
|
||||
- name: Check disk space
|
||||
run: |
|
||||
echo "Disk usage before Docker build:"
|
||||
df -h
|
||||
|
||||
# Check available space (require at least 2GB)
|
||||
AVAILABLE_GB=$(df / --output=avail --block-size=1G | tail -1)
|
||||
if [ "$AVAILABLE_GB" -lt 2 ]; then
|
||||
echo "❌ Insufficient disk space: ${AVAILABLE_GB}GB available, 2GB required"
|
||||
exit 1
|
||||
fi
|
||||
echo "✅ Sufficient disk space: ${AVAILABLE_GB}GB available"
|
||||
|
||||
- name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v3
|
||||
|
||||
- name: Set up Docker Buildx
|
||||
uses: docker/setup-buildx-action@v3
|
||||
|
||||
- name: Log in to GitHub Container Registry
|
||||
uses: docker/login-action@v3
|
||||
with:
|
||||
registry: ${{ env.REGISTRY }}
|
||||
username: ${{ github.actor }}
|
||||
password: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
- name: Extract metadata for standard image
|
||||
id: meta
|
||||
uses: docker/metadata-action@v5
|
||||
with:
|
||||
images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
||||
tags: |
|
||||
type=semver,pattern={{version}},value=v${{ needs.detect-version-change.outputs.new-version }}
|
||||
type=semver,pattern={{major}}.{{minor}},value=v${{ needs.detect-version-change.outputs.new-version }}
|
||||
type=semver,pattern={{major}},value=v${{ needs.detect-version-change.outputs.new-version }}
|
||||
type=raw,value=latest,enable={{is_default_branch}}
|
||||
|
||||
- name: Build and push standard Docker image
|
||||
uses: docker/build-push-action@v5
|
||||
with:
|
||||
context: .
|
||||
platforms: linux/amd64,linux/arm64
|
||||
push: true
|
||||
tags: ${{ steps.meta.outputs.tags }}
|
||||
labels: ${{ steps.meta.outputs.labels }}
|
||||
cache-from: type=gha
|
||||
cache-to: type=gha,mode=max
|
||||
|
||||
- name: Verify multi-arch manifest for latest tag
|
||||
run: |
|
||||
echo "Verifying multi-arch manifest for latest tag..."
|
||||
|
||||
# Retry with exponential backoff (registry propagation can take time)
|
||||
MAX_ATTEMPTS=5
|
||||
ATTEMPT=1
|
||||
WAIT_TIME=2
|
||||
|
||||
while [ $ATTEMPT -le $MAX_ATTEMPTS ]; do
|
||||
echo "Attempt $ATTEMPT of $MAX_ATTEMPTS..."
|
||||
|
||||
MANIFEST=$(docker buildx imagetools inspect ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:latest 2>&1 || true)
|
||||
|
||||
# Check for both platforms
|
||||
if echo "$MANIFEST" | grep -q "linux/amd64" && echo "$MANIFEST" | grep -q "linux/arm64"; then
|
||||
echo "✅ Multi-arch manifest verified: both amd64 and arm64 present"
|
||||
echo "$MANIFEST"
|
||||
exit 0
|
||||
fi
|
||||
|
||||
if [ $ATTEMPT -lt $MAX_ATTEMPTS ]; then
|
||||
echo "⏳ Registry still propagating, waiting ${WAIT_TIME}s before retry..."
|
||||
sleep $WAIT_TIME
|
||||
WAIT_TIME=$((WAIT_TIME * 2)) # Exponential backoff: 2s, 4s, 8s, 16s
|
||||
fi
|
||||
|
||||
ATTEMPT=$((ATTEMPT + 1))
|
||||
done
|
||||
|
||||
echo "❌ ERROR: Multi-arch manifest incomplete after $MAX_ATTEMPTS attempts!"
|
||||
echo "$MANIFEST"
|
||||
exit 1
|
||||
|
||||
- name: Verify multi-arch manifest for version tag
|
||||
run: |
|
||||
VERSION="${{ needs.detect-version-change.outputs.new-version }}"
|
||||
echo "Verifying multi-arch manifest for version tag :$VERSION (without 'v' prefix)..."
|
||||
|
||||
# Retry with exponential backoff (registry propagation can take time)
|
||||
MAX_ATTEMPTS=5
|
||||
ATTEMPT=1
|
||||
WAIT_TIME=2
|
||||
|
||||
while [ $ATTEMPT -le $MAX_ATTEMPTS ]; do
|
||||
echo "Attempt $ATTEMPT of $MAX_ATTEMPTS..."
|
||||
|
||||
MANIFEST=$(docker buildx imagetools inspect ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}:$VERSION 2>&1 || true)
|
||||
|
||||
# Check for both platforms
|
||||
if echo "$MANIFEST" | grep -q "linux/amd64" && echo "$MANIFEST" | grep -q "linux/arm64"; then
|
||||
echo "✅ Multi-arch manifest verified for $VERSION: both amd64 and arm64 present"
|
||||
echo "$MANIFEST"
|
||||
exit 0
|
||||
fi
|
||||
|
||||
if [ $ATTEMPT -lt $MAX_ATTEMPTS ]; then
|
||||
echo "⏳ Registry still propagating, waiting ${WAIT_TIME}s before retry..."
|
||||
sleep $WAIT_TIME
|
||||
WAIT_TIME=$((WAIT_TIME * 2)) # Exponential backoff: 2s, 4s, 8s, 16s
|
||||
fi
|
||||
|
||||
ATTEMPT=$((ATTEMPT + 1))
|
||||
done
|
||||
|
||||
echo "❌ ERROR: Multi-arch manifest incomplete for version $VERSION after $MAX_ATTEMPTS attempts!"
|
||||
echo "$MANIFEST"
|
||||
exit 1
|
||||
|
||||
- name: Extract metadata for Railway image
|
||||
id: meta-railway
|
||||
uses: docker/metadata-action@v5
|
||||
with:
|
||||
images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}-railway
|
||||
tags: |
|
||||
type=semver,pattern={{version}},value=v${{ needs.detect-version-change.outputs.new-version }}
|
||||
type=semver,pattern={{major}}.{{minor}},value=v${{ needs.detect-version-change.outputs.new-version }}
|
||||
type=semver,pattern={{major}},value=v${{ needs.detect-version-change.outputs.new-version }}
|
||||
type=raw,value=latest,enable={{is_default_branch}}
|
||||
|
||||
- name: Build and push Railway Docker image
|
||||
uses: docker/build-push-action@v5
|
||||
with:
|
||||
context: .
|
||||
file: ./Dockerfile.railway
|
||||
platforms: linux/amd64
|
||||
push: true
|
||||
tags: ${{ steps.meta-railway.outputs.tags }}
|
||||
labels: ${{ steps.meta-railway.outputs.labels }}
|
||||
cache-from: type=gha
|
||||
cache-to: type=gha,mode=max
|
||||
|
||||
update-documentation:
|
||||
name: Update Documentation
|
||||
runs-on: ubuntu-latest
|
||||
needs: [detect-version-change, create-release, publish-npm, build-docker]
|
||||
if: needs.detect-version-change.outputs.version-changed == 'true' && !failure()
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
token: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
- name: Setup Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: 20
|
||||
|
||||
- name: Update version badges in README
|
||||
run: |
|
||||
VERSION="${{ needs.detect-version-change.outputs.new-version }}"
|
||||
|
||||
# Update README version badges
|
||||
if [ -f "README.md" ]; then
|
||||
# Update npm version badge
|
||||
sed -i.bak "s|npm/v/n8n-mcp/[^)]*|npm/v/n8n-mcp/$VERSION|g" README.md
|
||||
|
||||
# Update any other version references
|
||||
sed -i.bak "s|version-[0-9][0-9]*\.[0-9][0-9]*\.[0-9][0-9]*|version-$VERSION|g" README.md
|
||||
|
||||
# Clean up backup file
|
||||
rm -f README.md.bak
|
||||
|
||||
echo "✅ Updated version badges in README.md to $VERSION"
|
||||
fi
|
||||
|
||||
- name: Commit documentation updates
|
||||
env:
|
||||
VERSION: ${{ needs.detect-version-change.outputs.new-version }}
|
||||
run: |
|
||||
git config user.name "github-actions[bot]"
|
||||
git config user.email "github-actions[bot]@users.noreply.github.com"
|
||||
|
||||
if git diff --quiet; then
|
||||
echo "No documentation changes to commit"
|
||||
else
|
||||
git add README.md
|
||||
git commit -m "docs: update version badges to v${VERSION}"
|
||||
git push
|
||||
echo "✅ Committed documentation updates"
|
||||
fi
|
||||
|
||||
notify-completion:
|
||||
name: Notify Release Completion
|
||||
runs-on: ubuntu-latest
|
||||
needs: [detect-version-change, create-release, publish-npm, build-docker, update-documentation]
|
||||
if: always() && needs.detect-version-change.outputs.version-changed == 'true'
|
||||
steps:
|
||||
- name: Create release summary
|
||||
run: |
|
||||
VERSION="${{ needs.detect-version-change.outputs.new-version }}"
|
||||
RELEASE_URL="https://github.com/${{ github.repository }}/releases/tag/v$VERSION"
|
||||
|
||||
echo "## 🎉 Release v$VERSION Published Successfully!" >> $GITHUB_STEP_SUMMARY
|
||||
echo "" >> $GITHUB_STEP_SUMMARY
|
||||
echo "### ✅ Completed Tasks:" >> $GITHUB_STEP_SUMMARY
|
||||
echo "" >> $GITHUB_STEP_SUMMARY
|
||||
|
||||
# Check job statuses
|
||||
if [ "${{ needs.create-release.result }}" = "success" ]; then
|
||||
echo "- ✅ GitHub Release created: [$RELEASE_URL]($RELEASE_URL)" >> $GITHUB_STEP_SUMMARY
|
||||
else
|
||||
echo "- ❌ GitHub Release creation failed" >> $GITHUB_STEP_SUMMARY
|
||||
fi
|
||||
|
||||
if [ "${{ needs.publish-npm.result }}" = "success" ]; then
|
||||
echo "- ✅ NPM package published: [npmjs.com/package/n8n-mcp](https://www.npmjs.com/package/n8n-mcp)" >> $GITHUB_STEP_SUMMARY
|
||||
else
|
||||
echo "- ❌ NPM publishing failed" >> $GITHUB_STEP_SUMMARY
|
||||
fi
|
||||
|
||||
if [ "${{ needs.build-docker.result }}" = "success" ]; then
|
||||
echo "- ✅ Docker images built and pushed" >> $GITHUB_STEP_SUMMARY
|
||||
echo " - Standard: \`ghcr.io/czlonkowski/n8n-mcp:v$VERSION\`" >> $GITHUB_STEP_SUMMARY
|
||||
echo " - Railway: \`ghcr.io/czlonkowski/n8n-mcp-railway:v$VERSION\`" >> $GITHUB_STEP_SUMMARY
|
||||
else
|
||||
echo "- ❌ Docker image building failed" >> $GITHUB_STEP_SUMMARY
|
||||
fi
|
||||
|
||||
if [ "${{ needs.update-documentation.result }}" = "success" ]; then
|
||||
echo "- ✅ Documentation updated" >> $GITHUB_STEP_SUMMARY
|
||||
else
|
||||
echo "- ⚠️ Documentation update skipped or failed" >> $GITHUB_STEP_SUMMARY
|
||||
fi
|
||||
|
||||
echo "" >> $GITHUB_STEP_SUMMARY
|
||||
echo "### 📦 Installation:" >> $GITHUB_STEP_SUMMARY
|
||||
echo "" >> $GITHUB_STEP_SUMMARY
|
||||
echo "\`\`\`bash" >> $GITHUB_STEP_SUMMARY
|
||||
echo "# NPM" >> $GITHUB_STEP_SUMMARY
|
||||
echo "npx n8n-mcp" >> $GITHUB_STEP_SUMMARY
|
||||
echo "" >> $GITHUB_STEP_SUMMARY
|
||||
echo "# Docker" >> $GITHUB_STEP_SUMMARY
|
||||
echo "docker run -p 3000:3000 ghcr.io/czlonkowski/n8n-mcp:v$VERSION" >> $GITHUB_STEP_SUMMARY
|
||||
echo "\`\`\`" >> $GITHUB_STEP_SUMMARY
|
||||
|
||||
echo "🎉 Release automation completed for v$VERSION!"
|
||||
89
.github/workflows/test.yml
vendored
89
.github/workflows/test.yml
vendored
@@ -2,8 +2,34 @@ name: Test Suite
|
||||
on:
|
||||
push:
|
||||
branches: [main, feat/comprehensive-testing-suite]
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
pull_request:
|
||||
branches: [main]
|
||||
paths-ignore:
|
||||
- '**.md'
|
||||
- '**.txt'
|
||||
- 'docs/**'
|
||||
- 'examples/**'
|
||||
- '.github/FUNDING.yml'
|
||||
- '.github/ISSUE_TEMPLATE/**'
|
||||
- '.github/pull_request_template.md'
|
||||
- '.gitignore'
|
||||
- 'LICENSE*'
|
||||
- 'ATTRIBUTION.md'
|
||||
- 'SECURITY.md'
|
||||
- 'CODE_OF_CONDUCT.md'
|
||||
|
||||
permissions:
|
||||
contents: read
|
||||
@@ -46,6 +72,12 @@ jobs:
|
||||
run: npm run test:integration -- --reporter=default --reporter=junit
|
||||
env:
|
||||
CI: true
|
||||
N8N_API_URL: ${{ secrets.N8N_API_URL }}
|
||||
N8N_API_KEY: ${{ secrets.N8N_API_KEY }}
|
||||
N8N_TEST_WEBHOOK_GET_URL: ${{ secrets.N8N_TEST_WEBHOOK_GET_URL }}
|
||||
N8N_TEST_WEBHOOK_POST_URL: ${{ secrets.N8N_TEST_WEBHOOK_POST_URL }}
|
||||
N8N_TEST_WEBHOOK_PUT_URL: ${{ secrets.N8N_TEST_WEBHOOK_PUT_URL }}
|
||||
N8N_TEST_WEBHOOK_DELETE_URL: ${{ secrets.N8N_TEST_WEBHOOK_DELETE_URL }}
|
||||
|
||||
# Generate test summary
|
||||
- name: Generate test summary
|
||||
@@ -122,6 +154,7 @@ jobs:
|
||||
- name: Create test report comment
|
||||
if: github.event_name == 'pull_request' && always()
|
||||
uses: actions/github-script@v7
|
||||
continue-on-error: true
|
||||
with:
|
||||
script: |
|
||||
const fs = require('fs');
|
||||
@@ -135,34 +168,40 @@ jobs:
|
||||
console.error('Error reading test summary:', error);
|
||||
}
|
||||
|
||||
// Find existing comment
|
||||
const { data: comments } = await github.rest.issues.listComments({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
issue_number: context.issue.number,
|
||||
});
|
||||
|
||||
const botComment = comments.find(comment =>
|
||||
comment.user.type === 'Bot' &&
|
||||
comment.body.includes('## Test Results')
|
||||
);
|
||||
|
||||
if (botComment) {
|
||||
// Update existing comment
|
||||
await github.rest.issues.updateComment({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
comment_id: botComment.id,
|
||||
body: summary
|
||||
});
|
||||
} else {
|
||||
// Create new comment
|
||||
await github.rest.issues.createComment({
|
||||
try {
|
||||
// Find existing comment
|
||||
const { data: comments } = await github.rest.issues.listComments({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
issue_number: context.issue.number,
|
||||
body: summary
|
||||
});
|
||||
|
||||
const botComment = comments.find(comment =>
|
||||
comment.user.type === 'Bot' &&
|
||||
comment.body.includes('## Test Results')
|
||||
);
|
||||
|
||||
if (botComment) {
|
||||
// Update existing comment
|
||||
await github.rest.issues.updateComment({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
comment_id: botComment.id,
|
||||
body: summary
|
||||
});
|
||||
} else {
|
||||
// Create new comment
|
||||
await github.rest.issues.createComment({
|
||||
owner: context.repo.owner,
|
||||
repo: context.repo.repo,
|
||||
issue_number: context.issue.number,
|
||||
body: summary
|
||||
});
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Failed to create/update PR comment:', error.message);
|
||||
console.log('This is likely due to insufficient permissions for external PRs.');
|
||||
console.log('Test results have been saved to the job summary instead.');
|
||||
}
|
||||
|
||||
# Generate job summary
|
||||
@@ -234,11 +273,13 @@ jobs:
|
||||
- name: Publish test results
|
||||
uses: dorny/test-reporter@v1
|
||||
if: always()
|
||||
continue-on-error: true
|
||||
with:
|
||||
name: Test Results
|
||||
path: 'artifacts/test-results-*/test-results/junit.xml'
|
||||
reporter: java-junit
|
||||
fail-on-error: false
|
||||
fail-on-empty: false
|
||||
|
||||
# Create a combined artifact with all results
|
||||
- name: Create combined results artifact
|
||||
|
||||
14
.gitignore
vendored
14
.gitignore
vendored
@@ -89,11 +89,19 @@ docker-compose.override.yml
|
||||
temp/
|
||||
tmp/
|
||||
|
||||
# Batch processing error files (may contain API tokens from templates)
|
||||
docs/batch_*.jsonl
|
||||
**/batch_*_error.jsonl
|
||||
|
||||
# Local documentation and analysis files
|
||||
docs/local/
|
||||
|
||||
# Database files
|
||||
# Database files - nodes.db is now tracked directly
|
||||
# data/*.db
|
||||
data/*.db-journal
|
||||
data/*.db.bak
|
||||
data/*.db.backup
|
||||
!data/.gitkeep
|
||||
!data/nodes.db
|
||||
|
||||
@@ -126,3 +134,9 @@ n8n-mcp-wrapper.sh
|
||||
|
||||
# Package tarballs
|
||||
*.tgz
|
||||
|
||||
# MCP configuration files
|
||||
.mcp.json
|
||||
|
||||
# Telemetry configuration (user-specific)
|
||||
~/.n8n-mcp/
|
||||
|
||||
3319
CHANGELOG.md
Normal file
3319
CHANGELOG.md
Normal file
File diff suppressed because it is too large
Load Diff
@@ -180,6 +180,9 @@ The MCP server exposes tools in several categories:
|
||||
- Sub-agents are not allowed to spawn further sub-agents
|
||||
- When you use sub-agents, do not allow them to commit and push. That should be done by you
|
||||
|
||||
### Development Best Practices
|
||||
- Run typecheck and lint after every code change
|
||||
|
||||
# important-instruction-reminders
|
||||
Do what has been asked; nothing more, nothing less.
|
||||
NEVER create files unless they're absolutely necessary for achieving your goal.
|
||||
@@ -188,4 +191,5 @@ NEVER proactively create documentation files (*.md) or README files. Only create
|
||||
- When you make changes to MCP server, you need to ask the user to reload it before you test
|
||||
- When the user asks to review issues, you should use GH CLI to get the issue and all the comments
|
||||
- When the task can be divided into separated subtasks, you should spawn separate sub-agents to handle them in paralel
|
||||
- Use the best sub-agent for the task as per their descriptions
|
||||
- Use the best sub-agent for the task as per their descriptions
|
||||
- Do not use hyperbolic or dramatic language in comments and documentation
|
||||
16
Dockerfile
16
Dockerfile
@@ -9,11 +9,13 @@ WORKDIR /app
|
||||
COPY tsconfig*.json ./
|
||||
|
||||
# Create minimal package.json and install ONLY build dependencies
|
||||
# Note: openai and zod are needed for TypeScript compilation of template metadata modules
|
||||
RUN --mount=type=cache,target=/root/.npm \
|
||||
echo '{}' > package.json && \
|
||||
npm install --no-save typescript@^5.8.3 @types/node@^22.15.30 @types/express@^5.0.3 \
|
||||
@modelcontextprotocol/sdk@^1.12.1 dotenv@^16.5.0 express@^5.1.0 axios@^1.10.0 \
|
||||
n8n-workflow@^1.96.0 uuid@^11.0.5 @types/uuid@^10.0.0
|
||||
n8n-workflow@^1.96.0 uuid@^11.0.5 @types/uuid@^10.0.0 \
|
||||
openai@^4.77.0 zod@^3.24.1 lru-cache@^11.2.1 @supabase/supabase-js@^2.57.4
|
||||
|
||||
# Copy source and build
|
||||
COPY src ./src
|
||||
@@ -32,9 +34,13 @@ RUN apk add --no-cache curl su-exec && \
|
||||
# Copy runtime-only package.json
|
||||
COPY package.runtime.json package.json
|
||||
|
||||
# Install runtime dependencies with cache mount
|
||||
# Install runtime dependencies with better-sqlite3 compilation
|
||||
# Build tools (python3, make, g++) are installed, used for compilation, then removed
|
||||
# This enables native SQLite (better-sqlite3) instead of sql.js, preventing memory leaks
|
||||
RUN --mount=type=cache,target=/root/.npm \
|
||||
npm install --production --no-audit --no-fund
|
||||
apk add --no-cache python3 make g++ && \
|
||||
npm install --production --no-audit --no-fund && \
|
||||
apk del python3 make g++
|
||||
|
||||
# Copy built application
|
||||
COPY --from=builder /app/dist ./dist
|
||||
@@ -72,6 +78,10 @@ USER nodejs
|
||||
# Set Docker environment flag
|
||||
ENV IS_DOCKER=true
|
||||
|
||||
# Telemetry: Anonymous usage statistics are ENABLED by default
|
||||
# To opt-out, uncomment the following line:
|
||||
# ENV N8N_MCP_TELEMETRY_DISABLED=true
|
||||
|
||||
# Expose HTTP port
|
||||
EXPOSE 3000
|
||||
|
||||
|
||||
@@ -25,16 +25,20 @@ RUN npm run build
|
||||
FROM node:22-alpine AS runtime
|
||||
WORKDIR /app
|
||||
|
||||
# Install system dependencies
|
||||
RUN apk add --no-cache curl python3 make g++ && \
|
||||
# Install runtime dependencies
|
||||
RUN apk add --no-cache curl && \
|
||||
rm -rf /var/cache/apk/*
|
||||
|
||||
# Copy runtime-only package.json
|
||||
COPY package.runtime.json package.json
|
||||
|
||||
# Install only production dependencies
|
||||
RUN npm install --production --no-audit --no-fund && \
|
||||
npm cache clean --force
|
||||
# Install production dependencies with temporary build tools
|
||||
# Build tools (python3, make, g++) enable better-sqlite3 compilation (native SQLite)
|
||||
# They are removed after installation to reduce image size and attack surface
|
||||
RUN apk add --no-cache python3 make g++ && \
|
||||
npm install --production --no-audit --no-fund && \
|
||||
npm cache clean --force && \
|
||||
apk del python3 make g++
|
||||
|
||||
# Copy built application from builder stage
|
||||
COPY --from=builder /app/dist ./dist
|
||||
|
||||
336
MEMORY_TEMPLATE_UPDATE.md
Normal file
336
MEMORY_TEMPLATE_UPDATE.md
Normal file
@@ -0,0 +1,336 @@
|
||||
# Template Update Process - Quick Reference
|
||||
|
||||
## Overview
|
||||
|
||||
The n8n-mcp project maintains a database of workflow templates from n8n.io. This guide explains how to update the template database incrementally without rebuilding from scratch.
|
||||
|
||||
## Current Database State
|
||||
|
||||
As of the last update:
|
||||
- **2,598 templates** in database
|
||||
- Templates from the last 12 months
|
||||
- Latest template: September 12, 2025
|
||||
|
||||
## Quick Commands
|
||||
|
||||
### Incremental Update (Recommended)
|
||||
```bash
|
||||
# Build if needed
|
||||
npm run build
|
||||
|
||||
# Fetch only NEW templates (5-10 minutes)
|
||||
npm run fetch:templates:update
|
||||
```
|
||||
|
||||
### Full Rebuild (Rare)
|
||||
```bash
|
||||
# Rebuild entire database from scratch (30-40 minutes)
|
||||
npm run fetch:templates
|
||||
```
|
||||
|
||||
## How It Works
|
||||
|
||||
### Incremental Update Mode (`--update`)
|
||||
|
||||
The incremental update is **smart and efficient**:
|
||||
|
||||
1. **Loads existing template IDs** from database (~2,598 templates)
|
||||
2. **Fetches template list** from n8n.io API (all templates from last 12 months)
|
||||
3. **Filters** to find only NEW templates not in database
|
||||
4. **Fetches details** for new templates only (saves time and API calls)
|
||||
5. **Saves** new templates to database (existing ones untouched)
|
||||
6. **Rebuilds FTS5** search index for new templates
|
||||
|
||||
### Key Benefits
|
||||
|
||||
✅ **Non-destructive**: All existing templates preserved
|
||||
✅ **Fast**: Only fetches new templates (5-10 min vs 30-40 min)
|
||||
✅ **API friendly**: Reduces load on n8n.io API
|
||||
✅ **Safe**: Preserves AI-generated metadata
|
||||
✅ **Smart**: Automatically skips duplicates
|
||||
|
||||
## Performance Comparison
|
||||
|
||||
| Mode | Templates Fetched | Time | Use Case |
|
||||
|------|------------------|------|----------|
|
||||
| **Update** | Only new (~50-200) | 5-10 min | Regular updates |
|
||||
| **Rebuild** | All (~8000+) | 30-40 min | Initial setup or corruption |
|
||||
|
||||
## Command Options
|
||||
|
||||
### Basic Update
|
||||
```bash
|
||||
npm run fetch:templates:update
|
||||
```
|
||||
|
||||
### Full Rebuild
|
||||
```bash
|
||||
npm run fetch:templates
|
||||
```
|
||||
|
||||
### With Metadata Generation
|
||||
```bash
|
||||
# Update templates and generate AI metadata
|
||||
npm run fetch:templates -- --update --generate-metadata
|
||||
|
||||
# Or just generate metadata for existing templates
|
||||
npm run fetch:templates -- --metadata-only
|
||||
```
|
||||
|
||||
### Help
|
||||
```bash
|
||||
npm run fetch:templates -- --help
|
||||
```
|
||||
|
||||
## Update Frequency
|
||||
|
||||
Recommended update schedule:
|
||||
- **Weekly**: Run incremental update to get latest templates
|
||||
- **Monthly**: Review database statistics
|
||||
- **As needed**: Rebuild only if database corruption suspected
|
||||
|
||||
## Template Filtering
|
||||
|
||||
The fetcher automatically filters templates:
|
||||
- ✅ **Includes**: Templates from last 12 months
|
||||
- ✅ **Includes**: Templates with >10 views
|
||||
- ❌ **Excludes**: Templates with ≤10 views (too niche)
|
||||
- ❌ **Excludes**: Templates older than 12 months
|
||||
|
||||
## Workflow
|
||||
|
||||
### Regular Update Workflow
|
||||
|
||||
```bash
|
||||
# 1. Check current state
|
||||
sqlite3 data/nodes.db "SELECT COUNT(*) FROM templates"
|
||||
|
||||
# 2. Build project (if code changed)
|
||||
npm run build
|
||||
|
||||
# 3. Run incremental update
|
||||
npm run fetch:templates:update
|
||||
|
||||
# 4. Verify new templates added
|
||||
sqlite3 data/nodes.db "SELECT COUNT(*) FROM templates"
|
||||
```
|
||||
|
||||
### After n8n Dependency Update
|
||||
|
||||
When you update n8n dependencies, templates remain compatible:
|
||||
```bash
|
||||
# 1. Update n8n (from MEMORY_N8N_UPDATE.md)
|
||||
npm run update:all
|
||||
|
||||
# 2. Fetch new templates incrementally
|
||||
npm run fetch:templates:update
|
||||
|
||||
# 3. Check how many templates were added
|
||||
sqlite3 data/nodes.db "SELECT COUNT(*) FROM templates"
|
||||
|
||||
# 4. Generate AI metadata for new templates (optional, requires OPENAI_API_KEY)
|
||||
npm run fetch:templates -- --metadata-only
|
||||
|
||||
# 5. IMPORTANT: Sanitize templates before pushing database
|
||||
npm run build
|
||||
npm run sanitize:templates
|
||||
```
|
||||
|
||||
Templates are independent of n8n version - they're just workflow JSON data.
|
||||
|
||||
**CRITICAL**: Always run `npm run sanitize:templates` before pushing the database to remove API tokens from template workflows.
|
||||
|
||||
**Note**: New templates fetched via `--update` mode will NOT have AI-generated metadata by default. You need to run `--metadata-only` separately to generate metadata for templates that don't have it yet.
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### No New Templates Found
|
||||
|
||||
This is normal! It means:
|
||||
- All recent templates are already in your database
|
||||
- n8n.io hasn't published many new templates recently
|
||||
- Your database is up to date
|
||||
|
||||
```bash
|
||||
📊 Update mode: 0 new templates to fetch (skipping 2598 existing)
|
||||
✅ All templates already have metadata
|
||||
```
|
||||
|
||||
### API Rate Limiting
|
||||
|
||||
If you hit rate limits:
|
||||
- The fetcher includes built-in delays (150ms between requests)
|
||||
- Wait a few minutes and try again
|
||||
- Use `--update` mode instead of full rebuild
|
||||
|
||||
### Database Corruption
|
||||
|
||||
If you suspect corruption:
|
||||
```bash
|
||||
# Full rebuild from scratch
|
||||
npm run fetch:templates
|
||||
|
||||
# This will:
|
||||
# - Drop and recreate templates table
|
||||
# - Fetch all templates fresh
|
||||
# - Rebuild search indexes
|
||||
```
|
||||
|
||||
## Database Schema
|
||||
|
||||
Templates are stored with:
|
||||
- Basic info (id, name, description, author, views, created_at)
|
||||
- Node types used (JSON array)
|
||||
- Complete workflow (gzip compressed, base64 encoded)
|
||||
- AI-generated metadata (optional, requires OpenAI API key)
|
||||
- FTS5 search index for fast text search
|
||||
|
||||
## Metadata Generation
|
||||
|
||||
Generate AI metadata for templates:
|
||||
```bash
|
||||
# Requires OPENAI_API_KEY in .env
|
||||
export OPENAI_API_KEY="sk-..."
|
||||
|
||||
# Generate for templates without metadata (recommended after incremental update)
|
||||
npm run fetch:templates -- --metadata-only
|
||||
|
||||
# Generate during template fetch (slower, but automatic)
|
||||
npm run fetch:templates:update -- --generate-metadata
|
||||
```
|
||||
|
||||
**Important**: Incremental updates (`--update`) do NOT generate metadata by default. After running `npm run fetch:templates:update`, you'll have new templates without metadata. Run `--metadata-only` separately to generate metadata for them.
|
||||
|
||||
### Check Metadata Coverage
|
||||
|
||||
```bash
|
||||
# See how many templates have metadata
|
||||
sqlite3 data/nodes.db "SELECT
|
||||
COUNT(*) as total,
|
||||
SUM(CASE WHEN metadata_json IS NOT NULL THEN 1 ELSE 0 END) as with_metadata,
|
||||
SUM(CASE WHEN metadata_json IS NULL THEN 1 ELSE 0 END) as without_metadata
|
||||
FROM templates"
|
||||
|
||||
# See recent templates without metadata
|
||||
sqlite3 data/nodes.db "SELECT id, name, created_at
|
||||
FROM templates
|
||||
WHERE metadata_json IS NULL
|
||||
ORDER BY created_at DESC
|
||||
LIMIT 10"
|
||||
```
|
||||
|
||||
Metadata includes:
|
||||
- Categories
|
||||
- Complexity level (simple/medium/complex)
|
||||
- Use cases
|
||||
- Estimated setup time
|
||||
- Required services
|
||||
- Key features
|
||||
- Target audience
|
||||
|
||||
### Metadata Generation Troubleshooting
|
||||
|
||||
If metadata generation fails:
|
||||
|
||||
1. **Check error file**: Errors are saved to `temp/batch/batch_*_error.jsonl`
|
||||
2. **Common issues**:
|
||||
- `"Unsupported value: 'temperature'"` - Model doesn't support custom temperature
|
||||
- `"Invalid request"` - Check OPENAI_API_KEY is valid
|
||||
- Model availability issues
|
||||
3. **Model**: Uses `gpt-5-mini-2025-08-07` by default
|
||||
4. **Token limit**: 3000 tokens per request for detailed metadata
|
||||
|
||||
The system will automatically:
|
||||
- Process error files and assign default metadata to failed templates
|
||||
- Save error details for debugging
|
||||
- Continue processing even if some templates fail
|
||||
|
||||
**Example error handling**:
|
||||
```bash
|
||||
# If you see: "No output file available for batch job"
|
||||
# Check: temp/batch/batch_*_error.jsonl for error details
|
||||
# The system now automatically processes errors and generates default metadata
|
||||
```
|
||||
|
||||
## Environment Variables
|
||||
|
||||
Optional configuration:
|
||||
```bash
|
||||
# OpenAI for metadata generation
|
||||
OPENAI_API_KEY=sk-...
|
||||
OPENAI_MODEL=gpt-4o-mini # Default model
|
||||
OPENAI_BATCH_SIZE=50 # Batch size for metadata generation
|
||||
|
||||
# Metadata generation limits
|
||||
METADATA_LIMIT=100 # Max templates to process (0 = all)
|
||||
```
|
||||
|
||||
## Statistics
|
||||
|
||||
After update, check stats:
|
||||
```bash
|
||||
# Template count
|
||||
sqlite3 data/nodes.db "SELECT COUNT(*) FROM templates"
|
||||
|
||||
# Most recent template
|
||||
sqlite3 data/nodes.db "SELECT MAX(created_at) FROM templates"
|
||||
|
||||
# Templates by view count
|
||||
sqlite3 data/nodes.db "SELECT COUNT(*),
|
||||
CASE
|
||||
WHEN views < 50 THEN '<50'
|
||||
WHEN views < 100 THEN '50-100'
|
||||
WHEN views < 500 THEN '100-500'
|
||||
ELSE '500+'
|
||||
END as view_range
|
||||
FROM templates GROUP BY view_range"
|
||||
```
|
||||
|
||||
## Integration with n8n-mcp
|
||||
|
||||
Templates are available through MCP tools:
|
||||
- `list_templates`: List all templates
|
||||
- `get_template`: Get specific template with workflow
|
||||
- `search_templates`: Search by keyword
|
||||
- `list_node_templates`: Templates using specific nodes
|
||||
- `get_templates_for_task`: Templates for common tasks
|
||||
- `search_templates_by_metadata`: Advanced filtering
|
||||
|
||||
See `npm run test:templates` for usage examples.
|
||||
|
||||
## Time Estimates
|
||||
|
||||
Typical incremental update:
|
||||
- Loading existing IDs: 1-2 seconds
|
||||
- Fetching template list: 2-3 minutes
|
||||
- Filtering new templates: instant
|
||||
- Fetching details for 100 new templates: ~15 seconds (0.15s each)
|
||||
- Saving and indexing: 5-10 seconds
|
||||
- **Total: 3-5 minutes**
|
||||
|
||||
Full rebuild:
|
||||
- Fetching 8000+ templates: 25-30 minutes
|
||||
- Saving and indexing: 5-10 minutes
|
||||
- **Total: 30-40 minutes**
|
||||
|
||||
## Best Practices
|
||||
|
||||
1. **Use incremental updates** for regular maintenance
|
||||
2. **Rebuild only when necessary** (corruption, major changes)
|
||||
3. **Generate metadata incrementally** to avoid OpenAI costs
|
||||
4. **Monitor template count** to verify updates working
|
||||
5. **Keep database backed up** before major operations
|
||||
|
||||
## Next Steps
|
||||
|
||||
After updating templates:
|
||||
1. Test template search: `npm run test:templates`
|
||||
2. Verify MCP tools work: Test in Claude Desktop
|
||||
3. Check statistics in database
|
||||
4. Commit changes if desired (database changes)
|
||||
|
||||
## Related Documentation
|
||||
|
||||
- `MEMORY_N8N_UPDATE.md` - Updating n8n dependencies
|
||||
- `CLAUDE.md` - Project overview and architecture
|
||||
- `README.md` - User documentation
|
||||
484
P0-R3-TEST-PLAN.md
Normal file
484
P0-R3-TEST-PLAN.md
Normal file
@@ -0,0 +1,484 @@
|
||||
# P0-R3 Feature Test Coverage Plan
|
||||
|
||||
## Executive Summary
|
||||
|
||||
This document outlines comprehensive test coverage for the P0-R3 feature (Template-based Configuration Examples). The feature adds real-world configuration examples from popular templates to node search and essentials tools.
|
||||
|
||||
**Feature Overview:**
|
||||
- New database table: `template_node_configs` (197 pre-extracted configurations)
|
||||
- Enhanced tools: `search_nodes({includeExamples: true})` and `get_node_essentials({includeExamples: true})`
|
||||
- Breaking changes: Removed `get_node_for_task` tool
|
||||
|
||||
## Test Files Created
|
||||
|
||||
### Unit Tests
|
||||
|
||||
#### 1. `/tests/unit/scripts/fetch-templates-extraction.test.ts` ✅
|
||||
**Purpose:** Test template extraction logic from `fetch-templates.ts`
|
||||
|
||||
**Coverage:**
|
||||
- `extractNodeConfigs()` - 90%+ coverage
|
||||
- Valid workflows with multiple nodes
|
||||
- Empty workflows
|
||||
- Malformed compressed data
|
||||
- Invalid JSON
|
||||
- Nodes without parameters
|
||||
- Sticky note filtering
|
||||
- Credential handling
|
||||
- Expression detection
|
||||
- Special characters
|
||||
- Large workflows (100 nodes)
|
||||
|
||||
- `detectExpressions()` - 100% coverage
|
||||
- `={{...}}` syntax detection
|
||||
- `$json` references
|
||||
- `$node` references
|
||||
- Nested objects
|
||||
- Arrays
|
||||
- Null/undefined handling
|
||||
- Multiple expression types
|
||||
|
||||
**Test Count:** 27 tests
|
||||
**Expected Coverage:** 92%+
|
||||
|
||||
---
|
||||
|
||||
#### 2. `/tests/unit/mcp/search-nodes-examples.test.ts` ✅
|
||||
**Purpose:** Test `search_nodes` tool with includeExamples parameter
|
||||
|
||||
**Coverage:**
|
||||
- includeExamples parameter behavior
|
||||
- false: no examples returned
|
||||
- undefined: no examples returned (default)
|
||||
- true: examples returned
|
||||
- Example data structure validation
|
||||
- Top 2 limit enforcement
|
||||
- Backward compatibility
|
||||
- Performance (<100ms)
|
||||
- Error handling (malformed JSON, database errors)
|
||||
- searchNodesLIKE integration
|
||||
- searchNodesFTS integration
|
||||
|
||||
**Test Count:** 12 tests
|
||||
**Expected Coverage:** 85%+
|
||||
|
||||
---
|
||||
|
||||
#### 3. `/tests/unit/mcp/get-node-essentials-examples.test.ts` ✅
|
||||
**Purpose:** Test `get_node_essentials` tool with includeExamples parameter
|
||||
|
||||
**Coverage:**
|
||||
- includeExamples parameter behavior
|
||||
- Full metadata structure
|
||||
- configuration object
|
||||
- source (template, views, complexity)
|
||||
- useCases (limited to 2)
|
||||
- metadata (hasCredentials, hasExpressions)
|
||||
- Cache key differentiation
|
||||
- Backward compatibility
|
||||
- Performance (<100ms)
|
||||
- Error handling
|
||||
- Top 3 limit enforcement
|
||||
|
||||
**Test Count:** 13 tests
|
||||
**Expected Coverage:** 88%+
|
||||
|
||||
---
|
||||
|
||||
### Integration Tests
|
||||
|
||||
#### 4. `/tests/integration/database/template-node-configs.test.ts` ✅
|
||||
**Purpose:** Test database schema, migrations, and operations
|
||||
|
||||
**Coverage:**
|
||||
- Schema validation
|
||||
- Table creation
|
||||
- All columns present
|
||||
- Correct types and constraints
|
||||
- CHECK constraint on complexity
|
||||
- Indexes
|
||||
- idx_config_node_type_rank
|
||||
- idx_config_complexity
|
||||
- idx_config_auth
|
||||
- View: ranked_node_configs
|
||||
- Top 5 per node_type
|
||||
- Correct ordering
|
||||
- Foreign key constraints
|
||||
- CASCADE delete
|
||||
- Referential integrity
|
||||
- Data operations
|
||||
- INSERT with all fields
|
||||
- Nullable fields
|
||||
- Rank updates
|
||||
- Delete rank > 10
|
||||
- Performance
|
||||
- 1000 records < 10ms queries
|
||||
- Migration idempotency
|
||||
|
||||
**Test Count:** 19 tests
|
||||
**Expected Coverage:** 95%+
|
||||
|
||||
---
|
||||
|
||||
#### 5. `/tests/integration/mcp/template-examples-e2e.test.ts` ✅
|
||||
**Purpose:** End-to-end integration testing
|
||||
|
||||
**Coverage:**
|
||||
- Direct SQL queries
|
||||
- Top 2 examples for search_nodes
|
||||
- Top 3 examples with metadata for get_node_essentials
|
||||
- Data structure validation
|
||||
- Valid JSON in all fields
|
||||
- Credentials when has_credentials=1
|
||||
- Ranked view functionality
|
||||
- Performance with 100+ configs
|
||||
- Query performance < 5ms
|
||||
- Complexity filtering
|
||||
- Edge cases
|
||||
- Non-existent node types
|
||||
- Long parameters_json (100 params)
|
||||
- Special characters (Unicode, emojis, symbols)
|
||||
- Data integrity
|
||||
- Foreign key constraints
|
||||
- Cascade deletes
|
||||
|
||||
**Test Count:** 14 tests
|
||||
**Expected Coverage:** 90%+
|
||||
|
||||
---
|
||||
|
||||
### Test Fixtures
|
||||
|
||||
#### 6. `/tests/fixtures/template-configs.ts` ✅
|
||||
**Purpose:** Reusable test data
|
||||
|
||||
**Provides:**
|
||||
- `sampleConfigs`: 7 realistic node configurations
|
||||
- simpleWebhook
|
||||
- webhookWithAuth
|
||||
- httpRequestBasic
|
||||
- httpRequestWithExpressions
|
||||
- slackMessage
|
||||
- codeNodeTransform
|
||||
- codeNodeWithExpressions
|
||||
|
||||
- `sampleWorkflows`: 3 complete workflows
|
||||
- webhookToSlack
|
||||
- apiWorkflow
|
||||
- complexWorkflow
|
||||
|
||||
- **Helper Functions:**
|
||||
- `compressWorkflow()` - Compress to base64
|
||||
- `createTemplateMetadata()` - Generate metadata
|
||||
- `createConfigBatch()` - Batch create configs
|
||||
- `getConfigByComplexity()` - Filter by complexity
|
||||
- `getConfigsWithExpressions()` - Filter with expressions
|
||||
- `getConfigsWithCredentials()` - Filter with credentials
|
||||
- `createInsertStatement()` - SQL insert helper
|
||||
|
||||
---
|
||||
|
||||
## Existing Tests Requiring Updates
|
||||
|
||||
### High Priority
|
||||
|
||||
#### 1. `tests/unit/mcp/parameter-validation.test.ts`
|
||||
**Line 480:** Remove `get_node_for_task` from legacyValidationTools array
|
||||
|
||||
```typescript
|
||||
// REMOVE THIS:
|
||||
{ name: 'get_node_for_task', args: {}, expected: 'Missing required parameters for get_node_for_task: task' },
|
||||
```
|
||||
|
||||
**Status:** ⚠️ BREAKING CHANGE - Tool removed
|
||||
|
||||
---
|
||||
|
||||
#### 2. `tests/unit/mcp/tools.test.ts`
|
||||
**Update:** Remove `get_node_for_task` from templates category
|
||||
|
||||
```typescript
|
||||
// BEFORE:
|
||||
templates: ['list_tasks', 'get_node_for_task', 'search_templates', ...]
|
||||
|
||||
// AFTER:
|
||||
templates: ['list_tasks', 'search_templates', ...]
|
||||
```
|
||||
|
||||
**Add:** Tests for new includeExamples parameter in tool definitions
|
||||
|
||||
```typescript
|
||||
it('should have includeExamples parameter in search_nodes', () => {
|
||||
const searchNodesTool = tools.find(t => t.name === 'search_nodes');
|
||||
expect(searchNodesTool.inputSchema.properties.includeExamples).toBeDefined();
|
||||
expect(searchNodesTool.inputSchema.properties.includeExamples.type).toBe('boolean');
|
||||
expect(searchNodesTool.inputSchema.properties.includeExamples.default).toBe(false);
|
||||
});
|
||||
|
||||
it('should have includeExamples parameter in get_node_essentials', () => {
|
||||
const essentialsTool = tools.find(t => t.name === 'get_node_essentials');
|
||||
expect(essentialsTool.inputSchema.properties.includeExamples).toBeDefined();
|
||||
});
|
||||
```
|
||||
|
||||
**Status:** ⚠️ REQUIRED UPDATE
|
||||
|
||||
---
|
||||
|
||||
#### 3. `tests/integration/mcp-protocol/session-management.test.ts`
|
||||
**Remove:** Test case calling `get_node_for_task` with invalid task
|
||||
|
||||
```typescript
|
||||
// REMOVE THIS TEST:
|
||||
client.callTool({ name: 'get_node_for_task', arguments: { task: 'invalid_task' } }).catch(e => e)
|
||||
```
|
||||
|
||||
**Status:** ⚠️ BREAKING CHANGE
|
||||
|
||||
---
|
||||
|
||||
#### 4. `tests/integration/mcp-protocol/tool-invocation.test.ts`
|
||||
**Remove:** Entire `get_node_for_task` describe block
|
||||
|
||||
**Add:** Tests for new includeExamples functionality
|
||||
|
||||
```typescript
|
||||
describe('search_nodes with includeExamples', () => {
|
||||
it('should return examples when includeExamples is true', async () => {
|
||||
const response = await client.callTool({
|
||||
name: 'search_nodes',
|
||||
arguments: { query: 'webhook', includeExamples: true }
|
||||
});
|
||||
|
||||
expect(response.results).toBeDefined();
|
||||
// Examples may or may not be present depending on database
|
||||
});
|
||||
|
||||
it('should not return examples when includeExamples is false', async () => {
|
||||
const response = await client.callTool({
|
||||
name: 'search_nodes',
|
||||
arguments: { query: 'webhook', includeExamples: false }
|
||||
});
|
||||
|
||||
expect(response.results).toBeDefined();
|
||||
response.results.forEach(node => {
|
||||
expect(node.examples).toBeUndefined();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('get_node_essentials with includeExamples', () => {
|
||||
it('should return examples with metadata when includeExamples is true', async () => {
|
||||
const response = await client.callTool({
|
||||
name: 'get_node_essentials',
|
||||
arguments: { nodeType: 'nodes-base.webhook', includeExamples: true }
|
||||
});
|
||||
|
||||
expect(response.nodeType).toBeDefined();
|
||||
// Examples may or may not be present depending on database
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
**Status:** ⚠️ REQUIRED UPDATE
|
||||
|
||||
---
|
||||
|
||||
### Medium Priority
|
||||
|
||||
#### 5. `tests/unit/services/task-templates.test.ts`
|
||||
**Status:** ✅ No changes needed (TaskTemplates marked as deprecated but not removed)
|
||||
|
||||
**Note:** TaskTemplates remains for backward compatibility. Tests should continue to pass.
|
||||
|
||||
---
|
||||
|
||||
## Test Execution Plan
|
||||
|
||||
### Phase 1: Unit Tests
|
||||
```bash
|
||||
# Run new unit tests
|
||||
npm test tests/unit/scripts/fetch-templates-extraction.test.ts
|
||||
npm test tests/unit/mcp/search-nodes-examples.test.ts
|
||||
npm test tests/unit/mcp/get-node-essentials-examples.test.ts
|
||||
|
||||
# Expected: All pass, 52 tests
|
||||
```
|
||||
|
||||
### Phase 2: Integration Tests
|
||||
```bash
|
||||
# Run new integration tests
|
||||
npm test tests/integration/database/template-node-configs.test.ts
|
||||
npm test tests/integration/mcp/template-examples-e2e.test.ts
|
||||
|
||||
# Expected: All pass, 33 tests
|
||||
```
|
||||
|
||||
### Phase 3: Update Existing Tests
|
||||
```bash
|
||||
# Update files as outlined above, then run:
|
||||
npm test tests/unit/mcp/parameter-validation.test.ts
|
||||
npm test tests/unit/mcp/tools.test.ts
|
||||
npm test tests/integration/mcp-protocol/session-management.test.ts
|
||||
npm test tests/integration/mcp-protocol/tool-invocation.test.ts
|
||||
|
||||
# Expected: All pass after updates
|
||||
```
|
||||
|
||||
### Phase 4: Full Test Suite
|
||||
```bash
|
||||
# Run all tests
|
||||
npm test
|
||||
|
||||
# Run with coverage
|
||||
npm run test:coverage
|
||||
|
||||
# Expected coverage improvements:
|
||||
# - src/scripts/fetch-templates.ts: +20% (60% → 80%)
|
||||
# - src/mcp/server.ts: +5% (75% → 80%)
|
||||
# - Overall project: +2% (current → current+2%)
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Coverage Expectations
|
||||
|
||||
### New Code Coverage
|
||||
|
||||
| File | Function | Target | Tests |
|
||||
|------|----------|--------|-------|
|
||||
| fetch-templates.ts | extractNodeConfigs | 95% | 15 tests |
|
||||
| fetch-templates.ts | detectExpressions | 100% | 12 tests |
|
||||
| server.ts | searchNodes (with examples) | 90% | 8 tests |
|
||||
| server.ts | getNodeEssentials (with examples) | 90% | 10 tests |
|
||||
| Database migration | template_node_configs | 100% | 19 tests |
|
||||
|
||||
### Overall Coverage Goals
|
||||
|
||||
- **Unit Tests:** 90%+ coverage for new code
|
||||
- **Integration Tests:** All happy paths + critical error paths
|
||||
- **E2E Tests:** Complete feature workflows
|
||||
- **Performance:** All queries <10ms (database), <100ms (MCP)
|
||||
|
||||
---
|
||||
|
||||
## Test Infrastructure
|
||||
|
||||
### Dependencies Required
|
||||
All dependencies already present in `package.json`:
|
||||
- vitest (test runner)
|
||||
- better-sqlite3 (database)
|
||||
- @vitest/coverage-v8 (coverage)
|
||||
|
||||
### Test Utilities Used
|
||||
- TestDatabase helper (from existing test utils)
|
||||
- createTestDatabaseAdapter (from existing test utils)
|
||||
- Standard vitest matchers
|
||||
|
||||
### No New Dependencies Required ✅
|
||||
|
||||
---
|
||||
|
||||
## Regression Prevention
|
||||
|
||||
### Critical Paths Protected
|
||||
|
||||
1. **Backward Compatibility**
|
||||
- Tools work without includeExamples parameter
|
||||
- Existing workflows unchanged
|
||||
- Cache keys differentiated
|
||||
|
||||
2. **Performance**
|
||||
- No degradation when includeExamples=false
|
||||
- Indexed queries <10ms
|
||||
- Example fetch errors don't break responses
|
||||
|
||||
3. **Data Integrity**
|
||||
- Foreign key constraints enforced
|
||||
- JSON validation in all fields
|
||||
- Rank calculations correct
|
||||
|
||||
---
|
||||
|
||||
## CI/CD Integration
|
||||
|
||||
### GitHub Actions Updates
|
||||
No changes required. Existing test commands will run new tests:
|
||||
|
||||
```yaml
|
||||
- run: npm test
|
||||
- run: npm run test:coverage
|
||||
```
|
||||
|
||||
### Coverage Thresholds
|
||||
Current thresholds maintained. Expected improvements:
|
||||
- Lines: +2%
|
||||
- Functions: +3%
|
||||
- Branches: +2%
|
||||
|
||||
---
|
||||
|
||||
## Manual Testing Checklist
|
||||
|
||||
### Pre-Deployment Verification
|
||||
|
||||
- [ ] Run `npm run rebuild` - Verify migration applies cleanly
|
||||
- [ ] Run `npm run fetch:templates --extract-only` - Verify extraction works
|
||||
- [ ] Check database: `SELECT COUNT(*) FROM template_node_configs` - Should be ~197
|
||||
- [ ] Test MCP tool: `search_nodes({query: "webhook", includeExamples: true})`
|
||||
- [ ] Test MCP tool: `get_node_essentials({nodeType: "nodes-base.webhook", includeExamples: true})`
|
||||
- [ ] Verify backward compatibility: Tools work without includeExamples parameter
|
||||
- [ ] Performance test: Query 100 nodes with examples < 200ms
|
||||
|
||||
---
|
||||
|
||||
## Rollback Plan
|
||||
|
||||
If issues are detected:
|
||||
|
||||
1. **Database Rollback:**
|
||||
```sql
|
||||
DROP TABLE IF EXISTS template_node_configs;
|
||||
DROP VIEW IF EXISTS ranked_node_configs;
|
||||
```
|
||||
|
||||
2. **Code Rollback:**
|
||||
- Revert server.ts changes
|
||||
- Revert tools.ts changes
|
||||
- Restore get_node_for_task tool (if critical)
|
||||
|
||||
3. **Test Rollback:**
|
||||
- Revert parameter-validation.test.ts
|
||||
- Revert tools.test.ts
|
||||
- Revert tool-invocation.test.ts
|
||||
|
||||
---
|
||||
|
||||
## Success Metrics
|
||||
|
||||
### Test Metrics
|
||||
- ✅ 85+ new tests added
|
||||
- ✅ 0 tests failing after updates
|
||||
- ✅ Coverage increase 2%+
|
||||
- ✅ All performance tests pass
|
||||
|
||||
### Feature Metrics
|
||||
- ✅ 197 template configs extracted
|
||||
- ✅ Top 2/3 examples returned correctly
|
||||
- ✅ Query performance <10ms
|
||||
- ✅ No backward compatibility breaks
|
||||
|
||||
---
|
||||
|
||||
## Conclusion
|
||||
|
||||
This test plan provides **comprehensive coverage** for the P0-R3 feature with:
|
||||
- **85+ new tests** across unit, integration, and E2E levels
|
||||
- **Complete coverage** of extraction, storage, and retrieval
|
||||
- **Backward compatibility** protection
|
||||
- **Performance validation** (<10ms queries)
|
||||
- **Clear migration path** for existing tests
|
||||
|
||||
**All test files are ready for execution.** Update the 4 existing test files as outlined, then run the full test suite.
|
||||
|
||||
**Estimated Total Implementation Time:** 2-3 hours for updating existing tests + validation
|
||||
69
PRIVACY.md
Normal file
69
PRIVACY.md
Normal file
@@ -0,0 +1,69 @@
|
||||
# Privacy Policy for n8n-mcp Telemetry
|
||||
|
||||
## Overview
|
||||
n8n-mcp collects anonymous usage statistics to help improve the tool. This data collection is designed to respect user privacy while providing valuable insights into how the tool is used.
|
||||
|
||||
## What We Collect
|
||||
- **Anonymous User ID**: A hashed identifier derived from your machine characteristics (no personal information)
|
||||
- **Tool Usage**: Which MCP tools are used and their performance metrics
|
||||
- **Workflow Patterns**: Sanitized workflow structures (all sensitive data removed)
|
||||
- **Error Types**: Categories of errors encountered (no error messages with user data)
|
||||
- **System Information**: Platform, architecture, Node.js version, and n8n-mcp version
|
||||
|
||||
## What We DON'T Collect
|
||||
- Personal information or usernames
|
||||
- API keys, tokens, or credentials
|
||||
- URLs, endpoints, or hostnames
|
||||
- Email addresses or contact information
|
||||
- File paths or directory structures
|
||||
- Actual workflow data or parameters
|
||||
- Database connection strings
|
||||
- Any authentication information
|
||||
|
||||
## Data Sanitization
|
||||
All collected data undergoes automatic sanitization:
|
||||
- URLs are replaced with `[URL]` or `[REDACTED]`
|
||||
- Long alphanumeric strings (potential keys) are replaced with `[KEY]`
|
||||
- Email addresses are replaced with `[EMAIL]`
|
||||
- Authentication-related fields are completely removed
|
||||
|
||||
## Data Storage
|
||||
- Data is stored securely using Supabase
|
||||
- Anonymous users have write-only access (cannot read data back)
|
||||
- Row Level Security (RLS) policies prevent data access by anonymous users
|
||||
|
||||
## Opt-Out
|
||||
You can disable telemetry at any time:
|
||||
```bash
|
||||
npx n8n-mcp telemetry disable
|
||||
```
|
||||
|
||||
To re-enable:
|
||||
```bash
|
||||
npx n8n-mcp telemetry enable
|
||||
```
|
||||
|
||||
To check status:
|
||||
```bash
|
||||
npx n8n-mcp telemetry status
|
||||
```
|
||||
|
||||
## Data Usage
|
||||
Collected data is used solely to:
|
||||
- Understand which features are most used
|
||||
- Identify common error patterns
|
||||
- Improve tool performance and reliability
|
||||
- Guide development priorities
|
||||
|
||||
## Data Retention
|
||||
- Data is retained for analysis purposes
|
||||
- No personal identification is possible from the collected data
|
||||
|
||||
## Changes to This Policy
|
||||
We may update this privacy policy from time to time. Updates will be reflected in this document.
|
||||
|
||||
## Contact
|
||||
For questions about telemetry or privacy, please open an issue on GitHub:
|
||||
https://github.com/czlonkowski/n8n-mcp/issues
|
||||
|
||||
Last updated: 2025-09-25
|
||||
713
README.md
713
README.md
@@ -2,13 +2,12 @@
|
||||
|
||||
[](https://opensource.org/licenses/MIT)
|
||||
[](https://github.com/czlonkowski/n8n-mcp)
|
||||
[](https://github.com/czlonkowski/n8n-mcp)
|
||||
[](https://www.npmjs.com/package/n8n-mcp)
|
||||
[](https://codecov.io/gh/czlonkowski/n8n-mcp)
|
||||
[](https://github.com/czlonkowski/n8n-mcp/actions)
|
||||
[](https://github.com/n8n-io/n8n)
|
||||
[](https://github.com/czlonkowski/n8n-mcp/actions)
|
||||
[](https://github.com/n8n-io/n8n)
|
||||
[](https://github.com/czlonkowski/n8n-mcp/pkgs/container/n8n-mcp)
|
||||
[](https://railway.com/deploy/VY6UOG?referralCode=n8n-mcp)
|
||||
[](https://railway.com/deploy/n8n-mcp?referralCode=n8n-mcp)
|
||||
|
||||
A Model Context Protocol (MCP) server that provides AI assistants with comprehensive access to n8n node documentation, properties, and operations. Deploy in minutes to give Claude and other AI assistants deep knowledge about n8n's 525+ workflow automation nodes.
|
||||
|
||||
@@ -16,11 +15,13 @@ A Model Context Protocol (MCP) server that provides AI assistants with comprehen
|
||||
|
||||
n8n-MCP serves as a bridge between n8n's workflow automation platform and AI models, enabling them to understand and work with n8n nodes effectively. It provides structured access to:
|
||||
|
||||
- 📚 **532 n8n nodes** from both n8n-nodes-base and @n8n/n8n-nodes-langchain
|
||||
- 📚 **536 n8n nodes** from both n8n-nodes-base and @n8n/n8n-nodes-langchain
|
||||
- 🔧 **Node properties** - 99% coverage with detailed schemas
|
||||
- ⚡ **Node operations** - 63.6% coverage of available actions
|
||||
- 📄 **Documentation** - 90% coverage from official n8n docs (including AI nodes)
|
||||
- 🤖 **AI tools** - 263 AI-capable nodes detected with full documentation
|
||||
- 💡 **Real-world examples** - 2,646 pre-extracted configurations from popular templates
|
||||
- 🎯 **Template library** - 2,500+ workflow templates with smart filtering
|
||||
|
||||
|
||||
## ⚠️ Important Safety Warning
|
||||
@@ -197,10 +198,36 @@ Add to Claude Desktop config:
|
||||
}
|
||||
```
|
||||
|
||||
>💡 Tip: If you’re running n8n locally on the same machine (e.g., via Docker), use http://host.docker.internal:5678 as the N8N_API_URL.
|
||||
>💡 Tip: If you're running n8n locally on the same machine (e.g., via Docker), use http://host.docker.internal:5678 as the N8N_API_URL.
|
||||
|
||||
> **Note**: The n8n API credentials are optional. Without them, you'll have access to all documentation and validation tools. With them, you'll additionally get workflow management capabilities (create, update, execute workflows).
|
||||
|
||||
### 🏠 Local n8n Instance Configuration
|
||||
|
||||
If you're running n8n locally (e.g., `http://localhost:5678` or Docker), you need to allow localhost webhooks:
|
||||
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"n8n-mcp": {
|
||||
"command": "docker",
|
||||
"args": [
|
||||
"run", "-i", "--rm", "--init",
|
||||
"-e", "MCP_MODE=stdio",
|
||||
"-e", "LOG_LEVEL=error",
|
||||
"-e", "DISABLE_CONSOLE_OUTPUT=true",
|
||||
"-e", "N8N_API_URL=http://host.docker.internal:5678",
|
||||
"-e", "N8N_API_KEY=your-api-key",
|
||||
"-e", "WEBHOOK_SECURITY_MODE=moderate",
|
||||
"ghcr.io/czlonkowski/n8n-mcp:latest"
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
> ⚠️ **Important:** Set `WEBHOOK_SECURITY_MODE=moderate` to allow webhooks to your local n8n instance. This is safe for local development while still blocking private networks and cloud metadata.
|
||||
|
||||
**Important:** The `-i` flag is required for MCP stdio communication.
|
||||
|
||||
> 🔧 If you encounter any issues with Docker, check our [Docker Troubleshooting Guide](./docs/DOCKER_TROUBLESHOOTING.md).
|
||||
@@ -212,6 +239,131 @@ Add to Claude Desktop config:
|
||||
|
||||
**Restart Claude Desktop after updating configuration** - That's it! 🎉
|
||||
|
||||
## 🔐 Privacy & Telemetry
|
||||
|
||||
n8n-mcp collects anonymous usage statistics to improve the tool. [View our privacy policy](./PRIVACY.md).
|
||||
|
||||
### Opting Out
|
||||
|
||||
**For npx users:**
|
||||
```bash
|
||||
npx n8n-mcp telemetry disable
|
||||
```
|
||||
|
||||
**For Docker users:**
|
||||
Add the following environment variable to your Docker configuration:
|
||||
```json
|
||||
"-e", "N8N_MCP_TELEMETRY_DISABLED=true"
|
||||
```
|
||||
|
||||
Example in Claude Desktop config:
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"n8n-mcp": {
|
||||
"command": "docker",
|
||||
"args": [
|
||||
"run",
|
||||
"-i",
|
||||
"--rm",
|
||||
"--init",
|
||||
"-e", "MCP_MODE=stdio",
|
||||
"-e", "LOG_LEVEL=error",
|
||||
"-e", "N8N_MCP_TELEMETRY_DISABLED=true",
|
||||
"ghcr.io/czlonkowski/n8n-mcp:latest"
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**For docker-compose users:**
|
||||
Set in your environment file or docker-compose.yml:
|
||||
```yaml
|
||||
environment:
|
||||
N8N_MCP_TELEMETRY_DISABLED: "true"
|
||||
```
|
||||
|
||||
## ⚙️ Database & Memory Configuration
|
||||
|
||||
### Database Adapters
|
||||
|
||||
n8n-mcp uses SQLite for storing node documentation. Two adapters are available:
|
||||
|
||||
1. **better-sqlite3** (Default in Docker)
|
||||
- Native C++ bindings for best performance
|
||||
- Direct disk writes (no memory overhead)
|
||||
- **Now enabled by default** in Docker images (v2.20.2+)
|
||||
- Memory usage: ~100-120 MB stable
|
||||
|
||||
2. **sql.js** (Fallback)
|
||||
- Pure JavaScript implementation
|
||||
- In-memory database with periodic saves
|
||||
- Used when better-sqlite3 compilation fails
|
||||
- Memory usage: ~150-200 MB stable
|
||||
|
||||
### Memory Optimization (sql.js)
|
||||
|
||||
If using sql.js fallback, you can configure the save interval to balance between data safety and memory efficiency:
|
||||
|
||||
**Environment Variable:**
|
||||
```bash
|
||||
SQLJS_SAVE_INTERVAL_MS=5000 # Default: 5000ms (5 seconds)
|
||||
```
|
||||
|
||||
**Usage:**
|
||||
- Controls how long to wait after database changes before saving to disk
|
||||
- Lower values = more frequent saves = higher memory churn
|
||||
- Higher values = less frequent saves = lower memory usage
|
||||
- Minimum: 100ms
|
||||
- Recommended: 5000-10000ms for production
|
||||
|
||||
**Docker Configuration:**
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"n8n-mcp": {
|
||||
"command": "docker",
|
||||
"args": [
|
||||
"run",
|
||||
"-i",
|
||||
"--rm",
|
||||
"--init",
|
||||
"-e", "SQLJS_SAVE_INTERVAL_MS=10000",
|
||||
"ghcr.io/czlonkowski/n8n-mcp:latest"
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**docker-compose:**
|
||||
```yaml
|
||||
environment:
|
||||
SQLJS_SAVE_INTERVAL_MS: "10000"
|
||||
```
|
||||
|
||||
### Memory Leak Fix (v2.20.2)
|
||||
|
||||
**Issue #330** identified a critical memory leak in long-running Docker/Kubernetes deployments:
|
||||
- **Before:** 100 MB → 2.2 GB over 72 hours (OOM kills)
|
||||
- **After:** Stable at 100-200 MB indefinitely
|
||||
|
||||
**Fixes Applied:**
|
||||
- ✅ Docker images now use better-sqlite3 by default (eliminates leak entirely)
|
||||
- ✅ sql.js fallback optimized (98% reduction in save frequency)
|
||||
- ✅ Removed unnecessary memory allocations (50% reduction per save)
|
||||
- ✅ Configurable save interval via `SQLJS_SAVE_INTERVAL_MS`
|
||||
|
||||
For Kubernetes deployments with memory limits:
|
||||
```yaml
|
||||
resources:
|
||||
requests:
|
||||
memory: 256Mi
|
||||
limits:
|
||||
memory: 512Mi
|
||||
```
|
||||
|
||||
## 💖 Support This Project
|
||||
|
||||
<div align="center">
|
||||
@@ -296,7 +448,7 @@ Add to Claude Desktop config:
|
||||
|
||||
Deploy n8n-MCP to Railway's cloud platform with zero configuration:
|
||||
|
||||
[](https://railway.com/deploy/VY6UOG?referralCode=n8n-mcp)
|
||||
[](https://railway.com/deploy/n8n-mcp?referralCode=n8n-mcp)
|
||||
|
||||
**Benefits:**
|
||||
- ☁️ **Instant cloud hosting** - No server setup required
|
||||
@@ -346,133 +498,394 @@ Step-by-step tutorial for connecting n8n-MCP to Cursor IDE with custom rules.
|
||||
### [Windsurf](./docs/WINDSURF_SETUP.md)
|
||||
Complete guide for integrating n8n-MCP with Windsurf using project rules.
|
||||
|
||||
### [Codex](./docs/CODEX_SETUP.md)
|
||||
Complete guide for integrating n8n-MCP with Codex.
|
||||
|
||||
## 🎓 Add Claude Skills (Optional)
|
||||
|
||||
Supercharge your n8n workflow building with specialized skills that teach AI how to build production-ready workflows!
|
||||
|
||||
[](https://www.youtube.com/watch?v=e6VvRqmUY2Y)
|
||||
|
||||
Learn more: [n8n-skills repository](https://github.com/czlonkowski/n8n-skills)
|
||||
|
||||
## 🤖 Claude Project Setup
|
||||
|
||||
For the best results when using n8n-MCP with Claude Projects, use these enhanced system instructions:
|
||||
|
||||
```markdown
|
||||
````markdown
|
||||
You are an expert in n8n automation software using n8n-MCP tools. Your role is to design, build, and validate n8n workflows with maximum accuracy and efficiency.
|
||||
|
||||
## Core Workflow Process
|
||||
## Core Principles
|
||||
|
||||
1. **ALWAYS start new conversation with**: `tools_documentation()` to understand best practices and available tools.
|
||||
### 1. Silent Execution
|
||||
CRITICAL: Execute tools without commentary. Only respond AFTER all tools complete.
|
||||
|
||||
2. **Discovery Phase** - Find the right nodes:
|
||||
- Think deeply about user request and the logic you are going to build to fulfill it. Ask follow-up questions to clarify the user's intent, if something is unclear. Then, proceed with the rest of your instructions.
|
||||
- `search_nodes({query: 'keyword'})` - Search by functionality
|
||||
❌ BAD: "Let me search for Slack nodes... Great! Now let me get details..."
|
||||
✅ GOOD: [Execute search_nodes and get_node_essentials in parallel, then respond]
|
||||
|
||||
### 2. Parallel Execution
|
||||
When operations are independent, execute them in parallel for maximum performance.
|
||||
|
||||
✅ GOOD: Call search_nodes, list_nodes, and search_templates simultaneously
|
||||
❌ BAD: Sequential tool calls (await each one before the next)
|
||||
|
||||
### 3. Templates First
|
||||
ALWAYS check templates before building from scratch (2,500+ available).
|
||||
|
||||
### 4. Multi-Level Validation
|
||||
Use validate_node_minimal → validate_node_operation → validate_workflow pattern.
|
||||
|
||||
### 5. Never Trust Defaults
|
||||
⚠️ CRITICAL: Default parameter values are the #1 source of runtime failures.
|
||||
ALWAYS explicitly configure ALL parameters that control node behavior.
|
||||
|
||||
## Workflow Process
|
||||
|
||||
1. **Start**: Call `tools_documentation()` for best practices
|
||||
|
||||
2. **Template Discovery Phase** (FIRST - parallel when searching multiple)
|
||||
- `search_templates_by_metadata({complexity: "simple"})` - Smart filtering
|
||||
- `get_templates_for_task('webhook_processing')` - Curated by task
|
||||
- `search_templates('slack notification')` - Text search
|
||||
- `list_node_templates(['n8n-nodes-base.slack'])` - By node type
|
||||
|
||||
**Filtering strategies**:
|
||||
- Beginners: `complexity: "simple"` + `maxSetupMinutes: 30`
|
||||
- By role: `targetAudience: "marketers"` | `"developers"` | `"analysts"`
|
||||
- By time: `maxSetupMinutes: 15` for quick wins
|
||||
- By service: `requiredService: "openai"` for compatibility
|
||||
|
||||
3. **Node Discovery** (if no suitable template - parallel execution)
|
||||
- Think deeply about requirements. Ask clarifying questions if unclear.
|
||||
- `search_nodes({query: 'keyword', includeExamples: true})` - Parallel for multiple nodes
|
||||
- `list_nodes({category: 'trigger'})` - Browse by category
|
||||
- `list_ai_tools()` - See AI-capable nodes (remember: ANY node can be an AI tool!)
|
||||
- `list_ai_tools()` - AI-capable nodes
|
||||
|
||||
3. **Configuration Phase** - Get node details efficiently:
|
||||
- `get_node_essentials(nodeType)` - Start here! Only 10-20 essential properties
|
||||
4. **Configuration Phase** (parallel for multiple nodes)
|
||||
- `get_node_essentials(nodeType, {includeExamples: true})` - 10-20 key properties
|
||||
- `search_node_properties(nodeType, 'auth')` - Find specific properties
|
||||
- `get_node_for_task('send_email')` - Get pre-configured templates
|
||||
- `get_node_documentation(nodeType)` - Human-readable docs when needed
|
||||
- It is good common practice to show a visual representation of the workflow architecture to the user and asking for opinion, before moving forward.
|
||||
- `get_node_documentation(nodeType)` - Human-readable docs
|
||||
- Show workflow architecture to user for approval before proceeding
|
||||
|
||||
4. **Pre-Validation Phase** - Validate BEFORE building:
|
||||
5. **Validation Phase** (parallel for multiple nodes)
|
||||
- `validate_node_minimal(nodeType, config)` - Quick required fields check
|
||||
- `validate_node_operation(nodeType, config, profile)` - Full operation-aware validation
|
||||
- Fix any validation errors before proceeding
|
||||
- `validate_node_operation(nodeType, config, 'runtime')` - Full validation with fixes
|
||||
- Fix ALL errors before proceeding
|
||||
|
||||
5. **Building Phase** - Create the workflow:
|
||||
- Use validated configurations from step 4
|
||||
6. **Building Phase**
|
||||
- If using template: `get_template(templateId, {mode: "full"})`
|
||||
- **MANDATORY ATTRIBUTION**: "Based on template by **[author.name]** (@[username]). View at: [url]"
|
||||
- Build from validated configurations
|
||||
- ⚠️ EXPLICITLY set ALL parameters - never rely on defaults
|
||||
- Connect nodes with proper structure
|
||||
- Add error handling where appropriate
|
||||
- Use expressions like $json, $node["NodeName"].json
|
||||
- Build the workflow in an artifact for easy editing downstream (unless the user asked to create in n8n instance)
|
||||
- Add error handling
|
||||
- Use n8n expressions: $json, $node["NodeName"].json
|
||||
- Build in artifact (unless deploying to n8n instance)
|
||||
|
||||
6. **Workflow Validation Phase** - Validate complete workflow:
|
||||
- `validate_workflow(workflow)` - Complete validation including connections
|
||||
- `validate_workflow_connections(workflow)` - Check structure and AI tool connections
|
||||
- `validate_workflow_expressions(workflow)` - Validate all n8n expressions
|
||||
- Fix any issues found before deployment
|
||||
7. **Workflow Validation** (before deployment)
|
||||
- `validate_workflow(workflow)` - Complete validation
|
||||
- `validate_workflow_connections(workflow)` - Structure check
|
||||
- `validate_workflow_expressions(workflow)` - Expression validation
|
||||
- Fix ALL issues before deployment
|
||||
|
||||
7. **Deployment Phase** (if n8n API configured):
|
||||
- `n8n_create_workflow(workflow)` - Deploy validated workflow
|
||||
- `n8n_validate_workflow({id: 'workflow-id'})` - Post-deployment validation
|
||||
- `n8n_update_partial_workflow()` - Make incremental updates using diffs
|
||||
- `n8n_trigger_webhook_workflow()` - Test webhook workflows
|
||||
8. **Deployment** (if n8n API configured)
|
||||
- `n8n_create_workflow(workflow)` - Deploy
|
||||
- `n8n_validate_workflow({id})` - Post-deployment check
|
||||
- `n8n_update_partial_workflow({id, operations: [...]})` - Batch updates
|
||||
- `n8n_trigger_webhook_workflow()` - Test webhooks
|
||||
|
||||
## Key Insights
|
||||
## Critical Warnings
|
||||
|
||||
- **USE CODE NODE ONLY WHEN IT IS NECESSARY** - always prefer to use standard nodes over code node. Use code node only when you are sure you need it.
|
||||
- **VALIDATE EARLY AND OFTEN** - Catch errors before they reach deployment
|
||||
- **USE DIFF UPDATES** - Use n8n_update_partial_workflow for 80-90% token savings
|
||||
- **ANY node can be an AI tool** - not just those with usableAsTool=true
|
||||
- **Pre-validate configurations** - Use validate_node_minimal before building
|
||||
- **Post-validate workflows** - Always validate complete workflows before deployment
|
||||
- **Incremental updates** - Use diff operations for existing workflows
|
||||
- **Test thoroughly** - Validate both locally and after deployment to n8n
|
||||
### ⚠️ Never Trust Defaults
|
||||
Default values cause runtime failures. Example:
|
||||
```json
|
||||
// ❌ FAILS at runtime
|
||||
{resource: "message", operation: "post", text: "Hello"}
|
||||
|
||||
// ✅ WORKS - all parameters explicit
|
||||
{resource: "message", operation: "post", select: "channel", channelId: "C123", text: "Hello"}
|
||||
```
|
||||
|
||||
### ⚠️ Example Availability
|
||||
`includeExamples: true` returns real configurations from workflow templates.
|
||||
- Coverage varies by node popularity
|
||||
- When no examples available, use `get_node_essentials` + `validate_node_minimal`
|
||||
|
||||
## Validation Strategy
|
||||
|
||||
### Before Building:
|
||||
1. validate_node_minimal() - Check required fields
|
||||
2. validate_node_operation() - Full configuration validation
|
||||
3. Fix all errors before proceeding
|
||||
### Level 1 - Quick Check (before building)
|
||||
`validate_node_minimal(nodeType, config)` - Required fields only (<100ms)
|
||||
|
||||
### After Building:
|
||||
1. validate_workflow() - Complete workflow validation
|
||||
2. validate_workflow_connections() - Structure validation
|
||||
3. validate_workflow_expressions() - Expression syntax check
|
||||
### Level 2 - Comprehensive (before building)
|
||||
`validate_node_operation(nodeType, config, 'runtime')` - Full validation with fixes
|
||||
|
||||
### After Deployment:
|
||||
1. n8n_validate_workflow({id}) - Validate deployed workflow
|
||||
2. n8n_list_executions() - Monitor execution status
|
||||
3. n8n_update_partial_workflow() - Fix issues using diffs
|
||||
### Level 3 - Complete (after building)
|
||||
`validate_workflow(workflow)` - Connections, expressions, AI tools
|
||||
|
||||
## Response Structure
|
||||
### Level 4 - Post-Deployment
|
||||
1. `n8n_validate_workflow({id})` - Validate deployed workflow
|
||||
2. `n8n_autofix_workflow({id})` - Auto-fix common errors
|
||||
3. `n8n_list_executions()` - Monitor execution status
|
||||
|
||||
1. **Discovery**: Show available nodes and options
|
||||
2. **Pre-Validation**: Validate node configurations first
|
||||
3. **Configuration**: Show only validated, working configs
|
||||
4. **Building**: Construct workflow with validated components
|
||||
5. **Workflow Validation**: Full workflow validation results
|
||||
6. **Deployment**: Deploy only after all validations pass
|
||||
7. **Post-Validation**: Verify deployment succeeded
|
||||
## Response Format
|
||||
|
||||
### Initial Creation
|
||||
```
|
||||
[Silent tool execution in parallel]
|
||||
|
||||
Created workflow:
|
||||
- Webhook trigger → Slack notification
|
||||
- Configured: POST /webhook → #general channel
|
||||
|
||||
Validation: ✅ All checks passed
|
||||
```
|
||||
|
||||
### Modifications
|
||||
```
|
||||
[Silent tool execution]
|
||||
|
||||
Updated workflow:
|
||||
- Added error handling to HTTP node
|
||||
- Fixed required Slack parameters
|
||||
|
||||
Changes validated successfully.
|
||||
```
|
||||
|
||||
## Batch Operations
|
||||
|
||||
Use `n8n_update_partial_workflow` with multiple operations in a single call:
|
||||
|
||||
✅ GOOD - Batch multiple operations:
|
||||
```json
|
||||
n8n_update_partial_workflow({
|
||||
id: "wf-123",
|
||||
operations: [
|
||||
{type: "updateNode", nodeId: "slack-1", changes: {...}},
|
||||
{type: "updateNode", nodeId: "http-1", changes: {...}},
|
||||
{type: "cleanStaleConnections"}
|
||||
]
|
||||
})
|
||||
```
|
||||
|
||||
❌ BAD - Separate calls:
|
||||
```json
|
||||
n8n_update_partial_workflow({id: "wf-123", operations: [{...}]})
|
||||
n8n_update_partial_workflow({id: "wf-123", operations: [{...}]})
|
||||
```
|
||||
|
||||
### CRITICAL: addConnection Syntax
|
||||
|
||||
The `addConnection` operation requires **four separate string parameters**. Common mistakes cause misleading errors.
|
||||
|
||||
❌ WRONG - Object format (fails with "Expected string, received object"):
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"connection": {
|
||||
"source": {"nodeId": "node-1", "outputIndex": 0},
|
||||
"destination": {"nodeId": "node-2", "inputIndex": 0}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
❌ WRONG - Combined string (fails with "Source node not found"):
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "node-1:main:0",
|
||||
"target": "node-2:main:0"
|
||||
}
|
||||
```
|
||||
|
||||
✅ CORRECT - Four separate string parameters:
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "node-id-string",
|
||||
"target": "target-node-id-string",
|
||||
"sourcePort": "main",
|
||||
"targetPort": "main"
|
||||
}
|
||||
```
|
||||
|
||||
**Reference**: [GitHub Issue #327](https://github.com/czlonkowski/n8n-mcp/issues/327)
|
||||
|
||||
### ⚠️ CRITICAL: IF Node Multi-Output Routing
|
||||
|
||||
IF nodes have **two outputs** (TRUE and FALSE). Use the **`branch` parameter** to route to the correct output:
|
||||
|
||||
✅ CORRECT - Route to TRUE branch (when condition is met):
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "if-node-id",
|
||||
"target": "success-handler-id",
|
||||
"sourcePort": "main",
|
||||
"targetPort": "main",
|
||||
"branch": "true"
|
||||
}
|
||||
```
|
||||
|
||||
✅ CORRECT - Route to FALSE branch (when condition is NOT met):
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "if-node-id",
|
||||
"target": "failure-handler-id",
|
||||
"sourcePort": "main",
|
||||
"targetPort": "main",
|
||||
"branch": "false"
|
||||
}
|
||||
```
|
||||
|
||||
**Common Pattern** - Complete IF node routing:
|
||||
```json
|
||||
n8n_update_partial_workflow({
|
||||
id: "workflow-id",
|
||||
operations: [
|
||||
{type: "addConnection", source: "If Node", target: "True Handler", sourcePort: "main", targetPort: "main", branch: "true"},
|
||||
{type: "addConnection", source: "If Node", target: "False Handler", sourcePort: "main", targetPort: "main", branch: "false"}
|
||||
]
|
||||
})
|
||||
```
|
||||
|
||||
**Note**: Without the `branch` parameter, both connections may end up on the same output, causing logic errors!
|
||||
|
||||
### removeConnection Syntax
|
||||
|
||||
Use the same four-parameter format:
|
||||
```json
|
||||
{
|
||||
"type": "removeConnection",
|
||||
"source": "source-node-id",
|
||||
"target": "target-node-id",
|
||||
"sourcePort": "main",
|
||||
"targetPort": "main"
|
||||
}
|
||||
```
|
||||
|
||||
## Example Workflow
|
||||
|
||||
### 1. Discovery & Configuration
|
||||
search_nodes({query: 'slack'})
|
||||
get_node_essentials('n8n-nodes-base.slack')
|
||||
### Template-First Approach
|
||||
|
||||
### 2. Pre-Validation
|
||||
validate_node_minimal('n8n-nodes-base.slack', {resource:'message', operation:'send'})
|
||||
```
|
||||
// STEP 1: Template Discovery (parallel execution)
|
||||
[Silent execution]
|
||||
search_templates_by_metadata({
|
||||
requiredService: 'slack',
|
||||
complexity: 'simple',
|
||||
targetAudience: 'marketers'
|
||||
})
|
||||
get_templates_for_task('slack_integration')
|
||||
|
||||
// STEP 2: Use template
|
||||
get_template(templateId, {mode: 'full'})
|
||||
validate_workflow(workflow)
|
||||
|
||||
// Response after all tools complete:
|
||||
"Found template by **David Ashby** (@cfomodz).
|
||||
View at: https://n8n.io/workflows/2414
|
||||
|
||||
Validation: ✅ All checks passed"
|
||||
```
|
||||
|
||||
### Building from Scratch (if no template)
|
||||
|
||||
```
|
||||
// STEP 1: Discovery (parallel execution)
|
||||
[Silent execution]
|
||||
search_nodes({query: 'slack', includeExamples: true})
|
||||
list_nodes({category: 'communication'})
|
||||
|
||||
// STEP 2: Configuration (parallel execution)
|
||||
[Silent execution]
|
||||
get_node_essentials('n8n-nodes-base.slack', {includeExamples: true})
|
||||
get_node_essentials('n8n-nodes-base.webhook', {includeExamples: true})
|
||||
|
||||
// STEP 3: Validation (parallel execution)
|
||||
[Silent execution]
|
||||
validate_node_minimal('n8n-nodes-base.slack', config)
|
||||
validate_node_operation('n8n-nodes-base.slack', fullConfig, 'runtime')
|
||||
|
||||
### 3. Build Workflow
|
||||
// Create workflow JSON with validated configs
|
||||
// STEP 4: Build
|
||||
// Construct workflow with validated configs
|
||||
// ⚠️ Set ALL parameters explicitly
|
||||
|
||||
### 4. Workflow Validation
|
||||
// STEP 5: Validate
|
||||
[Silent execution]
|
||||
validate_workflow(workflowJson)
|
||||
validate_workflow_connections(workflowJson)
|
||||
validate_workflow_expressions(workflowJson)
|
||||
|
||||
### 5. Deploy (if configured)
|
||||
n8n_create_workflow(validatedWorkflow)
|
||||
n8n_validate_workflow({id: createdWorkflowId})
|
||||
// Response after all tools complete:
|
||||
"Created workflow: Webhook → Slack
|
||||
Validation: ✅ Passed"
|
||||
```
|
||||
|
||||
### 6. Update Using Diffs
|
||||
### Batch Updates
|
||||
|
||||
```json
|
||||
// ONE call with multiple operations
|
||||
n8n_update_partial_workflow({
|
||||
workflowId: id,
|
||||
id: "wf-123",
|
||||
operations: [
|
||||
{type: 'updateNode', nodeId: 'slack1', changes: {position: [100, 200]}}
|
||||
{type: "updateNode", nodeId: "slack-1", changes: {position: [100, 200]}},
|
||||
{type: "updateNode", nodeId: "http-1", changes: {position: [300, 200]}},
|
||||
{type: "cleanStaleConnections"}
|
||||
]
|
||||
})
|
||||
```
|
||||
|
||||
## Important Rules
|
||||
|
||||
- ALWAYS validate before building
|
||||
- ALWAYS validate after building
|
||||
- NEVER deploy unvalidated workflows
|
||||
- USE diff operations for updates (80-90% token savings)
|
||||
- STATE validation results clearly
|
||||
- FIX all errors before proceeding
|
||||
```
|
||||
### Core Behavior
|
||||
1. **Silent execution** - No commentary between tools
|
||||
2. **Parallel by default** - Execute independent operations simultaneously
|
||||
3. **Templates first** - Always check before building (2,500+ available)
|
||||
4. **Multi-level validation** - Quick check → Full validation → Workflow validation
|
||||
5. **Never trust defaults** - Explicitly configure ALL parameters
|
||||
|
||||
Save these instructions in your Claude Project for optimal n8n workflow assistance with comprehensive validation.
|
||||
### Attribution & Credits
|
||||
- **MANDATORY TEMPLATE ATTRIBUTION**: Share author name, username, and n8n.io link
|
||||
- **Template validation** - Always validate before deployment (may need updates)
|
||||
|
||||
### Performance
|
||||
- **Batch operations** - Use diff operations with multiple changes in one call
|
||||
- **Parallel execution** - Search, validate, and configure simultaneously
|
||||
- **Template metadata** - Use smart filtering for faster discovery
|
||||
|
||||
### Code Node Usage
|
||||
- **Avoid when possible** - Prefer standard nodes
|
||||
- **Only when necessary** - Use code node as last resort
|
||||
- **AI tool capability** - ANY node can be an AI tool (not just marked ones)
|
||||
|
||||
### Most Popular n8n Nodes (for get_node_essentials):
|
||||
|
||||
1. **n8n-nodes-base.code** - JavaScript/Python scripting
|
||||
2. **n8n-nodes-base.httpRequest** - HTTP API calls
|
||||
3. **n8n-nodes-base.webhook** - Event-driven triggers
|
||||
4. **n8n-nodes-base.set** - Data transformation
|
||||
5. **n8n-nodes-base.if** - Conditional routing
|
||||
6. **n8n-nodes-base.manualTrigger** - Manual workflow execution
|
||||
7. **n8n-nodes-base.respondToWebhook** - Webhook responses
|
||||
8. **n8n-nodes-base.scheduleTrigger** - Time-based triggers
|
||||
9. **@n8n/n8n-nodes-langchain.agent** - AI agents
|
||||
10. **n8n-nodes-base.googleSheets** - Spreadsheet integration
|
||||
11. **n8n-nodes-base.merge** - Data merging
|
||||
12. **n8n-nodes-base.switch** - Multi-branch routing
|
||||
13. **n8n-nodes-base.telegram** - Telegram bot integration
|
||||
14. **@n8n/n8n-nodes-langchain.lmChatOpenAi** - OpenAI chat models
|
||||
15. **n8n-nodes-base.splitInBatches** - Batch processing
|
||||
16. **n8n-nodes-base.openAi** - OpenAI legacy node
|
||||
17. **n8n-nodes-base.gmail** - Email automation
|
||||
18. **n8n-nodes-base.function** - Custom functions
|
||||
19. **n8n-nodes-base.stickyNote** - Workflow documentation
|
||||
20. **n8n-nodes-base.executeWorkflowTrigger** - Sub-workflow calls
|
||||
|
||||
**Note:** LangChain nodes use the `@n8n/n8n-nodes-langchain.` prefix, core nodes use `n8n-nodes-base.`
|
||||
|
||||
````
|
||||
|
||||
Save these instructions in your Claude Project for optimal n8n workflow assistance with intelligent template discovery.
|
||||
|
||||
## 🚨 Important: Sharing Guidelines
|
||||
|
||||
@@ -488,11 +901,16 @@ This tool was created to benefit everyone in the n8n community without friction.
|
||||
## Features
|
||||
|
||||
- **🔍 Smart Node Search**: Find nodes by name, category, or functionality
|
||||
- **📖 Essential Properties**: Get only the 10-20 properties that matter (NEW in v2.4.0)
|
||||
- **🎯 Task Templates**: Pre-configured settings for common automation tasks
|
||||
- **📖 Essential Properties**: Get only the 10-20 properties that matter
|
||||
- **💡 Real-World Examples**: 2,646 pre-extracted configurations from popular templates
|
||||
- **✅ Config Validation**: Validate node configurations before deployment
|
||||
- **🤖 AI Workflow Validation**: Comprehensive validation for AI Agent workflows (NEW in v2.17.0!)
|
||||
- Missing language model detection
|
||||
- AI tool connection validation
|
||||
- Streaming mode constraints
|
||||
- Memory and output parser checks
|
||||
- **🔗 Dependency Analysis**: Understand property relationships and conditions
|
||||
- **💡 Working Examples**: Real-world examples for immediate use
|
||||
- **🎯 Template Discovery**: 2,500+ workflow templates with smart filtering
|
||||
- **⚡ Fast Response**: Average query time ~12ms with optimized SQLite
|
||||
- **🌐 Universal Compatibility**: Works with any Node.js version
|
||||
|
||||
@@ -518,20 +936,32 @@ Once connected, Claude can use these powerful tools:
|
||||
- **`tools_documentation`** - Get documentation for any MCP tool (START HERE!)
|
||||
- **`list_nodes`** - List all n8n nodes with filtering options
|
||||
- **`get_node_info`** - Get comprehensive information about a specific node
|
||||
- **`get_node_essentials`** - Get only essential properties with examples (10-20 properties instead of 200+)
|
||||
- **`search_nodes`** - Full-text search across all node documentation
|
||||
- **`get_node_essentials`** - Get only essential properties (10-20 instead of 200+). Use `includeExamples: true` to get top 3 real-world configurations from popular templates
|
||||
- **`search_nodes`** - Full-text search across all node documentation. Use `includeExamples: true` to get top 2 real-world configurations per node from templates
|
||||
- **`search_node_properties`** - Find specific properties within nodes
|
||||
- **`list_ai_tools`** - List all AI-capable nodes (ANY node can be used as AI tool!)
|
||||
- **`get_node_as_tool_info`** - Get guidance on using any node as an AI tool
|
||||
|
||||
### Advanced Tools
|
||||
- **`get_node_for_task`** - Pre-configured node settings for common tasks
|
||||
- **`list_tasks`** - Discover available task templates
|
||||
- **`validate_node_operation`** - Validate node configurations (operation-aware, profiles support)
|
||||
- **`validate_node_minimal`** - Quick validation for just required fields
|
||||
- **`validate_workflow`** - Complete workflow validation including AI tool connections
|
||||
### Template Tools
|
||||
- **`list_templates`** - Browse all templates with descriptions and optional metadata (2,500+ templates)
|
||||
- **`search_templates`** - Text search across template names and descriptions
|
||||
- **`search_templates_by_metadata`** - Advanced filtering by complexity, setup time, services, audience
|
||||
- **`list_node_templates`** - Find templates using specific nodes
|
||||
- **`get_template`** - Get complete workflow JSON for import
|
||||
- **`get_templates_for_task`** - Curated templates for common automation tasks
|
||||
|
||||
### Validation Tools
|
||||
- **`validate_workflow`** - Complete workflow validation including **AI Agent validation** (NEW in v2.17.0!)
|
||||
- Detects missing language model connections
|
||||
- Validates AI tool connections (no false warnings)
|
||||
- Enforces streaming mode constraints
|
||||
- Checks memory and output parser configurations
|
||||
- **`validate_workflow_connections`** - Check workflow structure and AI tool connections
|
||||
- **`validate_workflow_expressions`** - Validate n8n expressions including $fromAI()
|
||||
- **`validate_node_operation`** - Validate node configurations (operation-aware, profiles support)
|
||||
- **`validate_node_minimal`** - Quick validation for just required fields
|
||||
|
||||
### Advanced Tools
|
||||
- **`get_property_dependencies`** - Analyze property visibility conditions
|
||||
- **`get_node_documentation`** - Get parsed documentation from n8n-docs
|
||||
- **`get_database_statistics`** - View database metrics and coverage
|
||||
@@ -550,6 +980,7 @@ These powerful tools allow you to manage n8n workflows directly from Claude. The
|
||||
- **`n8n_delete_workflow`** - Delete workflows permanently
|
||||
- **`n8n_list_workflows`** - List workflows with filtering and pagination
|
||||
- **`n8n_validate_workflow`** - Validate workflows already in n8n by ID (NEW in v2.6.3)
|
||||
- **`n8n_autofix_workflow`** - Automatically fix common workflow errors (NEW in v2.13.0!)
|
||||
|
||||
#### Execution Management
|
||||
- **`n8n_trigger_webhook_workflow`** - Trigger workflows via webhook URL
|
||||
@@ -565,14 +996,17 @@ These powerful tools allow you to manage n8n workflows directly from Claude. The
|
||||
### Example Usage
|
||||
|
||||
```typescript
|
||||
// Get essentials for quick configuration
|
||||
get_node_essentials("nodes-base.httpRequest")
|
||||
// Get essentials with real-world examples from templates
|
||||
get_node_essentials({
|
||||
nodeType: "nodes-base.httpRequest",
|
||||
includeExamples: true // Returns top 3 configs from popular templates
|
||||
})
|
||||
|
||||
// Find nodes for a specific task
|
||||
search_nodes({ query: "send email gmail" })
|
||||
|
||||
// Get pre-configured settings
|
||||
get_node_for_task("send_email")
|
||||
// Search nodes with configuration examples
|
||||
search_nodes({
|
||||
query: "send email gmail",
|
||||
includeExamples: true // Returns top 2 configs per node
|
||||
})
|
||||
|
||||
// Validate before deployment
|
||||
validate_node_operation({
|
||||
@@ -663,12 +1097,14 @@ npm run dev:http # HTTP dev mode
|
||||
|
||||
## 📊 Metrics & Coverage
|
||||
|
||||
Current database coverage (n8n v1.103.2):
|
||||
Current database coverage (n8n v1.113.3):
|
||||
|
||||
- ✅ **532/532** nodes loaded (100%)
|
||||
- ✅ **525** nodes with properties (98.7%)
|
||||
- ✅ **536/536** nodes loaded (100%)
|
||||
- ✅ **528** nodes with properties (98.7%)
|
||||
- ✅ **470** nodes with documentation (88%)
|
||||
- ✅ **267** AI-capable tools detected
|
||||
- ✅ **2,646** pre-extracted template configurations
|
||||
- ✅ **2,500+** workflow templates available
|
||||
- ✅ **AI Agent & LangChain nodes** fully documented
|
||||
- ⚡ **Average response time**: ~12ms
|
||||
- 💾 **Database size**: ~15MB (optimized)
|
||||
@@ -708,7 +1144,7 @@ docker run --rm ghcr.io/czlonkowski/n8n-mcp:latest --version
|
||||
|
||||
## 🧪 Testing
|
||||
|
||||
The project includes a comprehensive test suite with **1,356 tests** ensuring code quality and reliability:
|
||||
The project includes a comprehensive test suite with **2,883 tests** ensuring code quality and reliability:
|
||||
|
||||
```bash
|
||||
# Run all tests
|
||||
@@ -728,9 +1164,9 @@ npm run test:bench # Performance benchmarks
|
||||
|
||||
### Test Suite Overview
|
||||
|
||||
- **Total Tests**: 1,356 (100% passing)
|
||||
- **Unit Tests**: 1,107 tests across 44 files
|
||||
- **Integration Tests**: 249 tests across 14 files
|
||||
- **Total Tests**: 2,883 (100% passing)
|
||||
- **Unit Tests**: 2,526 tests across 99 files
|
||||
- **Integration Tests**: 357 tests across 20 files
|
||||
- **Execution Time**: ~2.5 minutes in CI
|
||||
- **Test Framework**: Vitest (for speed and TypeScript support)
|
||||
- **Mocking**: MSW for API mocking, custom mocks for databases
|
||||
@@ -744,22 +1180,24 @@ npm run test:bench # Performance benchmarks
|
||||
|
||||
### Testing Architecture
|
||||
|
||||
- **Unit Tests**: Isolated component testing with mocks
|
||||
- Services layer: ~450 tests
|
||||
- Parsers: ~200 tests
|
||||
- Database repositories: ~100 tests
|
||||
- MCP tools: ~180 tests
|
||||
**Total: 3,336 tests** across unit and integration test suites
|
||||
|
||||
- **Integration Tests**: Full system behavior validation
|
||||
- MCP Protocol compliance: 72 tests
|
||||
- Database operations: 89 tests
|
||||
- Error handling: 44 tests
|
||||
- Performance: 44 tests
|
||||
- **Unit Tests** (2,766 tests): Isolated component testing with mocks
|
||||
- Services layer: Enhanced validation, property filtering, workflow validation
|
||||
- Parsers: Node parsing, property extraction, documentation mapping
|
||||
- Database: Repositories, adapters, migrations, FTS5 search
|
||||
- MCP tools: Tool definitions, documentation system
|
||||
- HTTP server: Multi-tenant support, security, configuration
|
||||
|
||||
- **Benchmarks**: Performance testing for critical paths
|
||||
- Database queries
|
||||
- Node loading
|
||||
- Search operations
|
||||
- **Integration Tests** (570 tests): Full system behavior validation
|
||||
- **n8n API Integration** (172 tests): All 18 MCP handler tools tested against real n8n instance
|
||||
- Workflow management: Create, read, update, delete, list, validate, autofix
|
||||
- Execution management: Trigger, retrieve, list, delete
|
||||
- System tools: Health check, tool listing, diagnostics
|
||||
- **MCP Protocol** (119 tests): Protocol compliance, session management, error handling
|
||||
- **Database** (226 tests): Repository operations, transactions, performance, FTS5 search
|
||||
- **Templates** (35 tests): Template fetching, storage, metadata operations
|
||||
- **Docker** (18 tests): Configuration, entrypoint, security validation
|
||||
|
||||
For detailed testing documentation, see [Testing Architecture](./docs/testing-architecture.md).
|
||||
|
||||
@@ -807,6 +1245,23 @@ See [Automated Release Guide](./docs/AUTOMATED_RELEASES.md) for complete details
|
||||
- [Anthropic](https://anthropic.com) for the Model Context Protocol
|
||||
- All contributors and users of this project
|
||||
|
||||
### Template Attribution
|
||||
|
||||
All workflow templates in this project are fetched from n8n's public template gallery at [n8n.io/workflows](https://n8n.io/workflows). Each template includes:
|
||||
- Full attribution to the original creator (name and username)
|
||||
- Direct link to the source template on n8n.io
|
||||
- Original workflow ID for reference
|
||||
|
||||
The AI agent instructions in this project contain mandatory attribution requirements. When using any template, the AI will automatically:
|
||||
- Share the template author's name and username
|
||||
- Provide a direct link to the original template on n8n.io
|
||||
- Display attribution in the format: "This workflow is based on a template by **[author]** (@[username]). View the original at: [url]"
|
||||
|
||||
Template creators retain all rights to their workflows. This project indexes templates to improve discoverability through AI assistants. If you're a template creator and have concerns about your template being indexed, please open an issue.
|
||||
|
||||
Special thanks to the prolific template contributors whose work helps thousands of users automate their workflows, including:
|
||||
**David Ashby** (@cfomodz), **Yaron Been** (@yaron-nofluff), **Jimleuk** (@jimleuk), **Davide** (@n3witalia), **David Olusola** (@dae221), **Ranjan Dailata** (@ranjancse), **Airtop** (@cesar-at-airtop), **Joseph LePage** (@joe), **Don Jayamaha Jr** (@don-the-gem-dealer), **Angel Menendez** (@djangelic), and the entire n8n community of creators!
|
||||
|
||||
---
|
||||
|
||||
<div align="center">
|
||||
|
||||
41
_config.yml
Normal file
41
_config.yml
Normal file
@@ -0,0 +1,41 @@
|
||||
# Jekyll configuration for GitHub Pages
|
||||
# This is only used for serving benchmark results
|
||||
|
||||
# Only process benchmark-related files
|
||||
include:
|
||||
- index.html
|
||||
- benchmarks/
|
||||
|
||||
# Exclude everything else to prevent Liquid syntax errors
|
||||
exclude:
|
||||
- "*.md"
|
||||
- "*.json"
|
||||
- "*.ts"
|
||||
- "*.js"
|
||||
- "*.yml"
|
||||
- src/
|
||||
- tests/
|
||||
- docs/
|
||||
- scripts/
|
||||
- dist/
|
||||
- node_modules/
|
||||
- package.json
|
||||
- package-lock.json
|
||||
- tsconfig.json
|
||||
- README.md
|
||||
- CHANGELOG.md
|
||||
- LICENSE
|
||||
- Dockerfile*
|
||||
- docker-compose*
|
||||
- .github/
|
||||
- .vscode/
|
||||
- .claude/
|
||||
- deploy/
|
||||
- examples/
|
||||
- data/
|
||||
|
||||
# Disable Jekyll processing for files we don't want processed
|
||||
plugins: []
|
||||
|
||||
# Use simple theme
|
||||
theme: null
|
||||
BIN
data/nodes.db
BIN
data/nodes.db
Binary file not shown.
0
data/templates.db
Normal file
0
data/templates.db
Normal file
@@ -22,7 +22,7 @@ services:
|
||||
networks:
|
||||
- n8n-network
|
||||
healthcheck:
|
||||
test: ["CMD", "curl", "-f", "http://localhost:5678/healthz"]
|
||||
test: ["CMD", "sh", "-c", "wget --quiet --spider --tries=1 --timeout=10 http://localhost:5678/healthz || exit 1"]
|
||||
interval: 30s
|
||||
timeout: 10s
|
||||
retries: 3
|
||||
|
||||
@@ -23,7 +23,11 @@ services:
|
||||
# Database
|
||||
NODE_DB_PATH: ${NODE_DB_PATH:-/app/data/nodes.db}
|
||||
REBUILD_ON_START: ${REBUILD_ON_START:-false}
|
||||
|
||||
|
||||
# Telemetry: Anonymous usage statistics are ENABLED by default
|
||||
# To opt-out, uncomment and set to 'true':
|
||||
# N8N_MCP_TELEMETRY_DISABLED: ${N8N_MCP_TELEMETRY_DISABLED:-true}
|
||||
|
||||
# Optional: n8n API configuration (enables 16 additional management tools)
|
||||
# Uncomment and configure to enable n8n workflow management
|
||||
# N8N_API_URL: ${N8N_API_URL}
|
||||
|
||||
384
docs/AUTOMATED_RELEASES.md
Normal file
384
docs/AUTOMATED_RELEASES.md
Normal file
@@ -0,0 +1,384 @@
|
||||
# Automated Release Process
|
||||
|
||||
This document describes the automated release system for n8n-mcp, which handles version detection, changelog parsing, and multi-artifact publishing.
|
||||
|
||||
## Overview
|
||||
|
||||
The automated release system is triggered when the version in `package.json` is updated and pushed to the main branch. It handles:
|
||||
|
||||
- 🏷️ **GitHub Releases**: Creates releases with changelog content
|
||||
- 📦 **NPM Publishing**: Publishes optimized runtime package
|
||||
- 🐳 **Docker Images**: Builds and pushes multi-platform images
|
||||
- 📚 **Documentation**: Updates version badges automatically
|
||||
|
||||
## Quick Start
|
||||
|
||||
### For Maintainers
|
||||
|
||||
Use the prepared release script for a guided experience:
|
||||
|
||||
```bash
|
||||
npm run prepare:release
|
||||
```
|
||||
|
||||
This script will:
|
||||
1. Prompt for the new version
|
||||
2. Update `package.json` and `package.runtime.json`
|
||||
3. Update the changelog
|
||||
4. Run tests and build
|
||||
5. Create a git commit
|
||||
6. Optionally push to trigger the release
|
||||
|
||||
### Manual Process
|
||||
|
||||
1. **Update the version**:
|
||||
```bash
|
||||
# Edit package.json version field
|
||||
vim package.json
|
||||
|
||||
# Sync to runtime package
|
||||
npm run sync:runtime-version
|
||||
```
|
||||
|
||||
2. **Update the changelog**:
|
||||
```bash
|
||||
# Edit docs/CHANGELOG.md
|
||||
vim docs/CHANGELOG.md
|
||||
```
|
||||
|
||||
3. **Test and commit**:
|
||||
```bash
|
||||
# Ensure everything works
|
||||
npm test
|
||||
npm run build
|
||||
npm run rebuild
|
||||
|
||||
# Commit changes
|
||||
git add package.json package.runtime.json docs/CHANGELOG.md
|
||||
git commit -m "chore: release vX.Y.Z"
|
||||
git push
|
||||
```
|
||||
|
||||
## Workflow Details
|
||||
|
||||
### Version Detection
|
||||
|
||||
The workflow monitors pushes to the main branch and detects when `package.json` version changes:
|
||||
|
||||
```yaml
|
||||
paths:
|
||||
- 'package.json'
|
||||
- 'package.runtime.json'
|
||||
```
|
||||
|
||||
### Changelog Parsing
|
||||
|
||||
Automatically extracts release notes from `docs/CHANGELOG.md` using the version header format:
|
||||
|
||||
```markdown
|
||||
## [2.10.0] - 2025-08-02
|
||||
|
||||
### Added
|
||||
- New feature descriptions
|
||||
|
||||
### Changed
|
||||
- Changed feature descriptions
|
||||
|
||||
### Fixed
|
||||
- Bug fix descriptions
|
||||
```
|
||||
|
||||
### Release Artifacts
|
||||
|
||||
#### GitHub Release
|
||||
- Created with extracted changelog content
|
||||
- Tagged with `vX.Y.Z` format
|
||||
- Includes installation instructions
|
||||
- Links to documentation
|
||||
|
||||
#### NPM Package
|
||||
- Published as `n8n-mcp` on npmjs.com
|
||||
- Uses runtime-only dependencies (8 packages vs 50+ dev deps)
|
||||
- Optimized for `npx` usage
|
||||
- ~50MB vs 1GB+ with dev dependencies
|
||||
|
||||
#### Docker Images
|
||||
- **Standard**: `ghcr.io/czlonkowski/n8n-mcp:vX.Y.Z`
|
||||
- **Railway**: `ghcr.io/czlonkowski/n8n-mcp-railway:vX.Y.Z`
|
||||
- Multi-platform: linux/amd64, linux/arm64
|
||||
- Semantic version tags: `vX.Y.Z`, `vX.Y`, `vX`, `latest`
|
||||
|
||||
## Configuration
|
||||
|
||||
### Required Secrets
|
||||
|
||||
Set these in GitHub repository settings → Secrets:
|
||||
|
||||
| Secret | Description | Required |
|
||||
|--------|-------------|----------|
|
||||
| `NPM_TOKEN` | NPM authentication token for publishing | ✅ Yes |
|
||||
| `GITHUB_TOKEN` | Automatically provided by GitHub Actions | ✅ Auto |
|
||||
|
||||
### NPM Token Setup
|
||||
|
||||
1. Login to [npmjs.com](https://www.npmjs.com)
|
||||
2. Go to Account Settings → Access Tokens
|
||||
3. Create a new **Automation** token
|
||||
4. Add as `NPM_TOKEN` secret in GitHub
|
||||
|
||||
## Testing
|
||||
|
||||
### Test Release Automation
|
||||
|
||||
Validate the release system without triggering a release:
|
||||
|
||||
```bash
|
||||
npm run test:release-automation
|
||||
```
|
||||
|
||||
This checks:
|
||||
- ✅ File existence and structure
|
||||
- ✅ Version detection logic
|
||||
- ✅ Changelog parsing
|
||||
- ✅ Build process
|
||||
- ✅ NPM package preparation
|
||||
- ✅ Docker configuration
|
||||
- ✅ Workflow syntax
|
||||
- ✅ Environment setup
|
||||
|
||||
### Local Testing
|
||||
|
||||
Test individual components:
|
||||
|
||||
```bash
|
||||
# Test version detection
|
||||
node -e "console.log(require('./package.json').version)"
|
||||
|
||||
# Test changelog parsing
|
||||
node scripts/test-release-automation.js
|
||||
|
||||
# Test npm package preparation
|
||||
npm run prepare:publish
|
||||
|
||||
# Test Docker build
|
||||
docker build -t test-image .
|
||||
```
|
||||
|
||||
## Workflow Jobs
|
||||
|
||||
### 1. Version Detection
|
||||
- Compares current vs previous version in git history
|
||||
- Determines if it's a prerelease (alpha, beta, rc, dev)
|
||||
- Outputs version information for other jobs
|
||||
|
||||
### 2. Changelog Extraction
|
||||
- Parses `docs/CHANGELOG.md` for the current version
|
||||
- Extracts content between version headers
|
||||
- Provides formatted release notes
|
||||
|
||||
### 3. GitHub Release Creation
|
||||
- Creates annotated git tag
|
||||
- Creates GitHub release with changelog content
|
||||
- Handles prerelease flag for alpha/beta versions
|
||||
|
||||
### 4. Build and Test
|
||||
- Installs dependencies
|
||||
- Runs full test suite
|
||||
- Builds TypeScript
|
||||
- Rebuilds node database
|
||||
- Type checking
|
||||
|
||||
### 5. NPM Publishing
|
||||
- Prepares optimized package structure
|
||||
- Uses `package.runtime.json` for dependencies
|
||||
- Publishes to npmjs.com registry
|
||||
- Automatic cleanup
|
||||
|
||||
### 6. Docker Building
|
||||
- Multi-platform builds (amd64, arm64)
|
||||
- Two image variants (standard, railway)
|
||||
- Semantic versioning tags
|
||||
- GitHub Container Registry
|
||||
|
||||
### 7. Documentation Updates
|
||||
- Updates version badges in README
|
||||
- Commits documentation changes
|
||||
- Automatic push back to repository
|
||||
|
||||
## Monitoring
|
||||
|
||||
### GitHub Actions
|
||||
Monitor releases at: https://github.com/czlonkowski/n8n-mcp/actions
|
||||
|
||||
### Release Status
|
||||
- **GitHub Releases**: https://github.com/czlonkowski/n8n-mcp/releases
|
||||
- **NPM Package**: https://www.npmjs.com/package/n8n-mcp
|
||||
- **Docker Images**: https://github.com/czlonkowski/n8n-mcp/pkgs/container/n8n-mcp
|
||||
|
||||
### Notifications
|
||||
|
||||
The workflow provides comprehensive summaries:
|
||||
- ✅ Success notifications with links
|
||||
- ❌ Failure notifications with error details
|
||||
- 📊 Artifact information and installation commands
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
#### NPM Publishing Fails
|
||||
```
|
||||
Error: 401 Unauthorized
|
||||
```
|
||||
**Solution**: Check NPM_TOKEN secret is valid and has publishing permissions.
|
||||
|
||||
#### Docker Build Fails
|
||||
```
|
||||
Error: failed to solve: could not read from registry
|
||||
```
|
||||
**Solution**: Check GitHub Container Registry permissions and GITHUB_TOKEN.
|
||||
|
||||
#### Changelog Parsing Fails
|
||||
```
|
||||
No changelog entries found for version X.Y.Z
|
||||
```
|
||||
**Solution**: Ensure changelog follows the correct format:
|
||||
```markdown
|
||||
## [X.Y.Z] - YYYY-MM-DD
|
||||
```
|
||||
|
||||
#### Version Detection Fails
|
||||
```
|
||||
Version not incremented
|
||||
```
|
||||
**Solution**: Ensure new version is greater than the previous version.
|
||||
|
||||
### Recovery Steps
|
||||
|
||||
#### Failed NPM Publish
|
||||
1. Check if version was already published
|
||||
2. If not, manually publish:
|
||||
```bash
|
||||
npm run prepare:publish
|
||||
cd npm-publish-temp
|
||||
npm publish
|
||||
```
|
||||
|
||||
#### Failed Docker Build
|
||||
1. Build locally to test:
|
||||
```bash
|
||||
docker build -t test-build .
|
||||
```
|
||||
2. Re-trigger workflow or push a fix
|
||||
|
||||
#### Incomplete Release
|
||||
1. Delete the created tag if needed:
|
||||
```bash
|
||||
git tag -d vX.Y.Z
|
||||
git push --delete origin vX.Y.Z
|
||||
```
|
||||
2. Fix issues and push again
|
||||
|
||||
## Security
|
||||
|
||||
### Secrets Management
|
||||
- NPM_TOKEN has limited scope (publish only)
|
||||
- GITHUB_TOKEN has automatic scoping
|
||||
- No secrets are logged or exposed
|
||||
|
||||
### Package Security
|
||||
- Runtime package excludes development dependencies
|
||||
- No build tools or test frameworks in published package
|
||||
- Minimal attack surface (~50MB vs 1GB+)
|
||||
|
||||
### Docker Security
|
||||
- Multi-stage builds
|
||||
- Non-root user execution
|
||||
- Minimal base images
|
||||
- Security scanning enabled
|
||||
|
||||
## Changelog Format
|
||||
|
||||
The automated system expects changelog entries in [Keep a Changelog](https://keepachangelog.com/) format:
|
||||
|
||||
```markdown
|
||||
# Changelog
|
||||
|
||||
All notable changes to this project will be documented in this file.
|
||||
|
||||
## [Unreleased]
|
||||
|
||||
### Added
|
||||
- New features for next release
|
||||
|
||||
## [2.10.0] - 2025-08-02
|
||||
|
||||
### Added
|
||||
- Automated release system
|
||||
- Multi-platform Docker builds
|
||||
|
||||
### Changed
|
||||
- Improved version detection
|
||||
- Enhanced error handling
|
||||
|
||||
### Fixed
|
||||
- Fixed changelog parsing edge cases
|
||||
- Fixed Docker build optimization
|
||||
|
||||
## [2.9.1] - 2025-08-01
|
||||
|
||||
...
|
||||
```
|
||||
|
||||
## Version Strategy
|
||||
|
||||
### Semantic Versioning
|
||||
- **MAJOR** (X.0.0): Breaking changes
|
||||
- **MINOR** (X.Y.0): New features, backward compatible
|
||||
- **PATCH** (X.Y.Z): Bug fixes, backward compatible
|
||||
|
||||
### Prerelease Versions
|
||||
- **Alpha**: `X.Y.Z-alpha.N` - Early development
|
||||
- **Beta**: `X.Y.Z-beta.N` - Feature complete, testing
|
||||
- **RC**: `X.Y.Z-rc.N` - Release candidate
|
||||
|
||||
Prerelease versions are automatically detected and marked appropriately.
|
||||
|
||||
## Best Practices
|
||||
|
||||
### Before Releasing
|
||||
1. ✅ Run `npm run test:release-automation`
|
||||
2. ✅ Update changelog with meaningful descriptions
|
||||
3. ✅ Test locally with `npm test && npm run build`
|
||||
4. ✅ Review breaking changes
|
||||
5. ✅ Consider impact on users
|
||||
|
||||
### Version Bumping
|
||||
- Use `npm run prepare:release` for guided process
|
||||
- Follow semantic versioning strictly
|
||||
- Document breaking changes clearly
|
||||
- Consider backward compatibility
|
||||
|
||||
### Changelog Writing
|
||||
- Be specific about changes
|
||||
- Include migration notes for breaking changes
|
||||
- Credit contributors
|
||||
- Use consistent formatting
|
||||
|
||||
## Contributing
|
||||
|
||||
### For Maintainers
|
||||
1. Use automated tools: `npm run prepare:release`
|
||||
2. Follow semantic versioning
|
||||
3. Update changelog thoroughly
|
||||
4. Test before releasing
|
||||
|
||||
### For Contributors
|
||||
- Breaking changes require MAJOR version bump
|
||||
- New features require MINOR version bump
|
||||
- Bug fixes require PATCH version bump
|
||||
- Update changelog in PR descriptions
|
||||
|
||||
---
|
||||
|
||||
🤖 *This automated release system was designed with [Claude Code](https://claude.ai/code)*
|
||||
@@ -5,7 +5,528 @@ All notable changes to this project will be documented in this file.
|
||||
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
|
||||
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
||||
|
||||
## [Unreleased]
|
||||
## [Unreleased] - Phase 0: Connection Operations Critical Fixes
|
||||
|
||||
### Fixed
|
||||
- **🐛 CRITICAL: Fixed `addConnection` sourceIndex handling (Issue #272, discovered in hands-on testing)**
|
||||
- Multi-output nodes (IF, Switch) now work correctly with sourceIndex parameter
|
||||
- Changed from `||` to `??` operator to properly handle explicit 0 values
|
||||
- Added defensive array validation before accessing indices
|
||||
- Improves rating from 3/10 to 8/10 for multi-output node scenarios
|
||||
- **Impact**: IF nodes, Switch nodes, and all conditional routing now reliable
|
||||
|
||||
- **🐛 CRITICAL: Added runtime validation for `updateConnection` (Issue #272, #204)**
|
||||
- Prevents server crashes when `updates` object is missing
|
||||
- Provides helpful error message with:
|
||||
- Clear explanation of what's wrong
|
||||
- Correct format example
|
||||
- Suggestion to use removeConnection + addConnection for rewiring
|
||||
- Validates `updates` is an object, not string or other type
|
||||
- **Impact**: No more cryptic "Cannot read properties of undefined" crashes
|
||||
|
||||
### Enhanced
|
||||
- **Error Messages**: `updateConnection` errors now include actionable guidance
|
||||
- Example format shown in error
|
||||
- Alternative approaches suggested (removeConnection + addConnection)
|
||||
- Clear explanation that updateConnection modifies properties, not targets
|
||||
|
||||
### Testing
|
||||
- Added 8 comprehensive tests for Phase 0 fixes
|
||||
- 2 tests for updateConnection validation (missing updates, invalid type)
|
||||
- 5 tests for sourceIndex handling (IF nodes, parallel execution, Switch nodes, explicit 0)
|
||||
- 1 test for complex multi-output routing scenarios
|
||||
- All 126 existing tests still passing
|
||||
|
||||
### Documentation
|
||||
- Updated tool documentation to clarify:
|
||||
- `addConnection` now properly handles sourceIndex (Phase 0 fix noted)
|
||||
- `updateConnection` REQUIRES 'updates' object (Phase 0 validation noted)
|
||||
- Added pitfalls about updateConnection limitations
|
||||
- Clarified that updateConnection modifies properties, NOT connection targets
|
||||
|
||||
### Developer Experience
|
||||
- More defensive programming throughout connection operations
|
||||
- Better use of nullish coalescing (??) vs. logical OR (||)
|
||||
- Clear inline comments explaining expected behavior
|
||||
- Improved type safety with runtime guards
|
||||
|
||||
### References
|
||||
- Comprehensive analysis: `docs/local/connection-operations-deep-dive-and-improvement-plan.md`
|
||||
- Based on hands-on testing with n8n-mcp-tester agent
|
||||
- Overall experience rating improved from 4.5/10 to estimated 6/10
|
||||
|
||||
## [2.14.4] - 2025-09-30
|
||||
|
||||
### Added
|
||||
- **Workflow Cleanup Operations**: Two new operations for `n8n_update_partial_workflow` to handle broken workflow recovery
|
||||
- `cleanStaleConnections`: Automatically removes all connections referencing non-existent nodes
|
||||
- Essential after node renames or deletions that leave dangling connection references
|
||||
- Supports `dryRun: true` mode to preview what would be removed
|
||||
- Removes both source and target stale connections
|
||||
- `replaceConnections`: Replace entire connections object in a single operation
|
||||
- Faster than crafting many individual connection operations
|
||||
- Useful for bulk connection rewiring
|
||||
|
||||
- **Graceful Error Handling for Connection Operations**: Enhanced `removeConnection` operation
|
||||
- New `ignoreErrors` flag: When `true`, operation succeeds even if connection doesn't exist
|
||||
- Perfect for cleanup scenarios where you're not sure if connections exist
|
||||
- Maintains backwards compatibility (defaults to `false` for strict validation)
|
||||
|
||||
- **Best-Effort Mode**: New `continueOnError` mode for `WorkflowDiffRequest`
|
||||
- Apply valid operations even if some fail
|
||||
- Returns detailed results with `applied` and `failed` operation indices
|
||||
- Breaks atomic guarantees intentionally for bulk cleanup scenarios
|
||||
- Maintains atomic mode as default for safety
|
||||
|
||||
### Enhanced
|
||||
- **Tool Documentation**: Updated `n8n_update_partial_workflow` documentation
|
||||
- Added examples for cleanup scenarios
|
||||
- Documented new operation types and modes
|
||||
- Added best practices for workflow recovery
|
||||
- Clarified atomic vs. best-effort behavior
|
||||
|
||||
- **Type System**: Extended workflow diff types
|
||||
- Added `CleanStaleConnectionsOperation` interface
|
||||
- Added `ReplaceConnectionsOperation` interface
|
||||
- Extended `WorkflowDiffResult` with `applied`, `failed`, and `staleConnectionsRemoved` fields
|
||||
- Updated type guards for new connection operations
|
||||
|
||||
### Testing
|
||||
- Added comprehensive test suite for v2.14.4 features
|
||||
- 15 new tests covering all new operations and modes
|
||||
- Tests for cleanStaleConnections with various stale scenarios
|
||||
- Tests for replaceConnections validation
|
||||
- Tests for ignoreErrors flag behavior
|
||||
- Tests for continueOnError mode with mixed success/failure
|
||||
- Backwards compatibility verification tests
|
||||
|
||||
### Impact
|
||||
- **Time Saved**: Reduces broken workflow fix time from 10-15 minutes to 30 seconds
|
||||
- **Token Efficiency**: `cleanStaleConnections` is 1 operation vs 10+ manual operations
|
||||
- **User Experience**: Dramatically improved workflow recovery capabilities
|
||||
- **Backwards Compatibility**: 100% - all additions are optional and default to existing behavior
|
||||
|
||||
## [2.13.2] - 2025-01-24
|
||||
|
||||
### Added
|
||||
- **Operation and Resource Validation with Intelligent Suggestions**: New similarity services for n8n node configuration validation
|
||||
- `OperationSimilarityService`: Validates operations and suggests similar alternatives using Levenshtein distance and pattern matching
|
||||
- `ResourceSimilarityService`: Validates resources with automatic plural/singular conversion and typo detection
|
||||
- Provides "Did you mean...?" suggestions when invalid operations or resources are used
|
||||
- Example: `operation: "listFiles"` suggests `"search"` for Google Drive nodes
|
||||
- Example: `resource: "files"` suggests singular `"file"` with 95% confidence
|
||||
- Confidence-based suggestions (minimum 30% threshold) with contextual fix messages
|
||||
- Resource-aware operation filtering ensures suggestions are contextually appropriate
|
||||
- 5-minute cache duration for performance optimization
|
||||
- Integrated into `EnhancedConfigValidator` for seamless validation flow
|
||||
|
||||
- **Custom Error Handling**: New `ValidationServiceError` class for better error management
|
||||
- Proper error chaining with cause tracking
|
||||
- Specialized factory methods for common error scenarios
|
||||
- Type-safe error propagation throughout the validation pipeline
|
||||
|
||||
### Enhanced
|
||||
- **Code Quality and Security Improvements** (based on code review feedback):
|
||||
- Safe JSON parsing with try-catch error boundaries
|
||||
- Type guards for safe property access (`getOperationValue`, `getResourceValue`)
|
||||
- Memory leak prevention with periodic cache cleanup
|
||||
- Performance optimization with early termination for exact matches
|
||||
- Replaced magic numbers with named constants for better maintainability
|
||||
- Comprehensive JSDoc documentation for all public methods
|
||||
- Improved confidence calculation for typos and transpositions
|
||||
|
||||
### Fixed
|
||||
- **Test Compatibility**: Updated test expectations to correctly handle exact match scenarios
|
||||
- **Cache Management**: Fixed cache cleanup to prevent unbounded memory growth
|
||||
- **Validation Deduplication**: Enhanced config validator now properly replaces base validator errors with detailed suggestions
|
||||
|
||||
### Testing
|
||||
- Added comprehensive test coverage for similarity services (37 new tests)
|
||||
- All unit tests passing with proper edge case handling
|
||||
- Integration confirmed via n8n-mcp-tester agent validation
|
||||
|
||||
## [2.13.1] - 2025-01-24
|
||||
|
||||
### Changed
|
||||
- **Removed 5-operation limit from n8n_update_partial_workflow**: The workflow diff engine now supports unlimited operations per request
|
||||
- Previously limited to 5 operations for "transactional integrity"
|
||||
- Analysis revealed the limit was unnecessary - the clone-validate-apply pattern already ensures atomicity
|
||||
- All operations are validated before any are applied, maintaining data integrity
|
||||
- Enables complex workflow refactoring in single API calls
|
||||
- Updated documentation and examples to demonstrate large batch operations (26+ operations)
|
||||
|
||||
## [2.13.0] - 2025-01-24
|
||||
|
||||
### Added
|
||||
- **Webhook Path Autofixer**: Automatically generates UUIDs for webhook nodes missing path configuration
|
||||
- Generates unique UUID for both `path` parameter and `webhookId` field
|
||||
- Conditionally updates typeVersion to 2.1 only when < 2.1 to ensure compatibility
|
||||
- High confidence fix (95%) as UUID generation is deterministic
|
||||
- Resolves webhook nodes showing "?" in the n8n UI
|
||||
|
||||
- **Enhanced Node Type Suggestions**: Intelligent node type correction with similarity matching
|
||||
- Multi-factor scoring system: name similarity, category match, package match, pattern match
|
||||
- Handles deprecated package prefixes (n8n-nodes-base. → nodes-base.)
|
||||
- Corrects capitalization mistakes (HttpRequest → httpRequest)
|
||||
- Suggests correct packages (nodes-base.openai → nodes-langchain.openAi)
|
||||
- Only auto-fixes suggestions with ≥90% confidence
|
||||
- 5-minute cache for performance optimization
|
||||
|
||||
- **n8n_autofix_workflow Tool**: New MCP tool for automatic workflow error correction
|
||||
- Comprehensive documentation with examples and best practices
|
||||
- Supports 5 fix types: expression-format, typeversion-correction, error-output-config, node-type-correction, webhook-missing-path
|
||||
- Confidence-based system (high/medium/low) for safe fixes
|
||||
- Preview mode to review changes before applying
|
||||
- Integrated with workflow validation pipeline
|
||||
|
||||
### Fixed
|
||||
- **Security**: Eliminated ReDoS vulnerability in NodeSimilarityService
|
||||
- Replaced all regex patterns with string-based matching
|
||||
- No performance impact while maintaining accuracy
|
||||
|
||||
- **Performance**: Optimized similarity matching algorithms
|
||||
- Levenshtein distance algorithm optimized from O(m*n) space to O(n)
|
||||
- Added early termination for performance improvement
|
||||
- Cache invalidation with version tracking prevents memory leaks
|
||||
|
||||
- **Code Quality**: Improved maintainability and type safety
|
||||
- Extracted magic numbers into named constants
|
||||
- Added proper type guards for runtime safety
|
||||
- Created centralized node-type-utils for consistent type normalization
|
||||
- Fixed silent failures in setNestedValue operations
|
||||
|
||||
### Changed
|
||||
- Template sanitizer now includes defensive null checks for runtime safety
|
||||
- Workflow validator uses centralized type normalization utility
|
||||
|
||||
## [2.12.2] - 2025-01-22
|
||||
|
||||
### Changed
|
||||
- Updated n8n dependencies to latest versions:
|
||||
- n8n: 1.111.0 → 1.112.3
|
||||
- n8n-core: 1.110.0 → 1.111.0
|
||||
- n8n-workflow: 1.108.0 → 1.109.0
|
||||
- @n8n/n8n-nodes-langchain: 1.110.0 → 1.111.1
|
||||
- Rebuilt node database with 536 nodes (438 from n8n-nodes-base, 98 from langchain)
|
||||
|
||||
## [2.12.1] - 2025-01-21
|
||||
|
||||
### Added
|
||||
- **Comprehensive Expression Format Validation System**: Three-tier validation strategy for n8n expressions
|
||||
- **Universal Expression Validator**: 100% reliable detection of expression format issues
|
||||
- Enforces required `=` prefix for all expressions `{{ }}`
|
||||
- Validates expression syntax (bracket matching, empty expressions)
|
||||
- Detects common mistakes (template literals, nested brackets, double prefixes)
|
||||
- Provides confidence score of 1.0 for universal rules
|
||||
- **Confidence-Based Node-Specific Recommendations**: Intelligent resource locator suggestions
|
||||
- Confidence scoring system (0.0 to 1.0) for field-specific recommendations
|
||||
- High confidence (≥0.8): Exact field matches for known nodes (GitHub owner/repository, Slack channels)
|
||||
- Medium confidence (≥0.5): Field pattern matches (fields ending in Id, Key, Name)
|
||||
- Factors: exact field match, field patterns, value patterns, node category
|
||||
- **Resource Locator Format Detection**: Identifies fields needing `__rl` structure
|
||||
- Validates resource locator mode (id, url, expression, name, list)
|
||||
- Auto-fixes missing prefixes in resource locator values
|
||||
- Provides clear JSON examples showing correct format
|
||||
- **Enhanced Safety Features**:
|
||||
- Recursion depth protection (MAX_RECURSION_DEPTH = 100) prevents infinite loops
|
||||
- Pattern matching precision using exact/prefix matching instead of includes()
|
||||
- Circular reference detection with WeakSet
|
||||
- **Separation of Concerns**: Clean architecture for maintainability
|
||||
- Universal rules separated from node-specific intelligence
|
||||
- Confidence-based application of suggestions
|
||||
- Future-proof design that works with any n8n node
|
||||
|
||||
## [2.12.1] - 2025-09-22
|
||||
|
||||
### Fixed
|
||||
- **Error Output Validation**: Enhanced workflow validator to detect incorrect error output configurations
|
||||
- Detects when multiple nodes are incorrectly placed in the same output array (main[0])
|
||||
- Validates that error handlers are properly connected to main[1] (error output) instead of main[0]
|
||||
- Cross-validates onError property ('continueErrorOutput') matches actual connection structure
|
||||
- Provides clear, actionable error messages with JSON examples showing correct configuration
|
||||
- Uses heuristic detection for error handler nodes (names containing "error", "fail", "catch", etc.)
|
||||
- Added comprehensive test coverage with 16+ test cases
|
||||
|
||||
### Improved
|
||||
- **Validation Messages**: Error messages now include detailed JSON examples showing both incorrect and correct configurations
|
||||
- **Pattern Detection**: Fixed `checkWorkflowPatterns` to check main[1] for error outputs instead of non-existent outputs.error
|
||||
- **Test Coverage**: Added new test file `workflow-validator-error-outputs.test.ts` with extensive error output validation scenarios
|
||||
|
||||
## [2.12.0] - 2025-09-19
|
||||
|
||||
### Added
|
||||
- **Flexible Instance Configuration**: Complete multi-instance support for serving multiple n8n instances dynamically
|
||||
- New `InstanceContext` interface for runtime configuration without multi-tenancy implications
|
||||
- Dual-mode API client supporting both singleton (env vars) and instance-specific configurations
|
||||
- LRU cache with SHA-256 hashing for secure client management (100 instances, 30-min TTL)
|
||||
- Comprehensive input validation preventing injection attacks and invalid configurations
|
||||
- Session context management in HTTP server for per-session instance configuration
|
||||
- 100% backward compatibility - existing deployments work unchanged
|
||||
- Full test coverage with 83 new tests covering security, caching, and validation
|
||||
|
||||
### Security
|
||||
- **SHA-256 Cache Key Hashing**: All instance identifiers are hashed before caching
|
||||
- **Input Validation**: Comprehensive validation for URLs, API keys, and numeric parameters
|
||||
- **Secure Logging**: Sensitive data never logged, only partial hashes for debugging
|
||||
- **Memory Management**: LRU eviction and TTL prevent unbounded growth
|
||||
- **URL Validation**: Blocks dangerous protocols (file://, javascript://, etc.)
|
||||
|
||||
### Performance
|
||||
- **Efficient Caching**: LRU cache with automatic cleanup reduces API client creation
|
||||
- **Fast Lookups**: SHA-256 hashed keys for O(1) cache access
|
||||
- **Memory Optimized**: Maximum 100 concurrent instances with 30-minute TTL
|
||||
- **Token Savings**: Reuses existing clients instead of recreating
|
||||
|
||||
### Documentation
|
||||
- Added comprehensive [Flexible Instance Configuration Guide](./FLEXIBLE_INSTANCE_CONFIGURATION.md)
|
||||
- Detailed architecture, usage examples, and security considerations
|
||||
- Migration guide for existing deployments
|
||||
- Complete API documentation for InstanceContext
|
||||
|
||||
## [2.11.3] - 2025-09-17
|
||||
|
||||
### Fixed
|
||||
- **n8n_update_partial_workflow Tool**: Fixed critical bug where updateNode and updateConnection operations were using incorrect property name
|
||||
- Changed from `changes` property to `updates` property to match documentation and expected behavior
|
||||
- Resolves issue where AI agents would break workflow connections when updating nodes
|
||||
- Fixes GitHub issues #159 (update_partial_workflow is invalid) and #168 (partial workflow update returns error)
|
||||
- All related tests updated to use correct property name
|
||||
|
||||
## [2.11.2] - 2025-09-16
|
||||
|
||||
### Updated
|
||||
- **n8n Dependencies**: Updated to latest versions for compatibility and new features
|
||||
- n8n: 1.110.1 → 1.111.0
|
||||
- n8n-core: 1.109.0 → 1.110.0
|
||||
- n8n-workflow: 1.107.0 → 1.108.0
|
||||
- @n8n/n8n-nodes-langchain: 1.109.1 → 1.110.0
|
||||
- **Node Database**: Rebuilt with 535 nodes from updated n8n packages
|
||||
- **Templates**: Preserved all 2,598 workflow templates with metadata intact
|
||||
- All critical nodes validated successfully (httpRequest, code, slack, agent)
|
||||
- Test suite: 1,911 tests passing, 5 flaky performance tests failing (99.7% pass rate)
|
||||
|
||||
## [2.11.1] - 2025-09-15
|
||||
|
||||
### Added
|
||||
- **Optional Fields Parameter for search_templates**: Enhanced search_templates tool with field filtering capability
|
||||
- New optional `fields` parameter accepts an array of field names to include in response
|
||||
- Supported fields: 'id', 'name', 'description', 'author', 'nodes', 'views', 'created', 'url', 'metadata'
|
||||
- Reduces response size by 70-98% when requesting only specific fields (e.g., just id and name)
|
||||
- Maintains full backward compatibility - existing calls without fields parameter work unchanged
|
||||
- Example: `search_templates({query: "slack", fields: ["id", "name"]})` returns minimal data
|
||||
- Significantly improves AI agent performance by reducing token usage
|
||||
|
||||
### Added
|
||||
- **Fuzzy Node Type Matching for Templates**: Improved template discovery with flexible node type resolution
|
||||
- Templates can now be found using simple node names: `["slack"]` instead of `["n8n-nodes-base.slack"]`
|
||||
- Accepts various input formats: bare names, partial prefixes, and case variations
|
||||
- Automatically expands related node types: `["email"]` finds Gmail, email send, and related templates
|
||||
- `["slack"]` also finds `slackTrigger` templates
|
||||
- Case-insensitive matching: `["Slack"]`, `["WEBHOOK"]`, `["HttpRequest"]` all work
|
||||
- Backward compatible - existing exact formats continue working
|
||||
- Reduces failed queries by approximately 50%
|
||||
- Added `template-node-resolver.ts` utility for node type resolution
|
||||
- Added 23 tests for template node resolution
|
||||
- **Structured Template Metadata System**: Comprehensive metadata for intelligent template discovery
|
||||
- Generated metadata for 2,534 templates (97.5% coverage) using OpenAI's batch API
|
||||
- Rich metadata structure: categories, complexity, use cases, setup time, required services, key features, target audience
|
||||
- New `search_templates_by_metadata` tool for advanced filtering by multiple criteria
|
||||
- Enhanced `list_templates` tool with optional `includeMetadata` parameter
|
||||
- Templates now always include descriptions in list responses
|
||||
- Metadata enables filtering by complexity level (simple/medium/complex)
|
||||
- Filter by estimated setup time ranges (5-480 minutes)
|
||||
- Filter by required external services (OpenAI, Slack, Google, etc.)
|
||||
- Filter by target audience (developers, marketers, analysts, etc.)
|
||||
- Multiple filter combinations supported for precise template discovery
|
||||
- SQLite JSON extraction for efficient metadata queries
|
||||
- Batch processing with OpenAI's gpt-4o-mini model for cost efficiency
|
||||
- Added comprehensive tool documentation for new metadata features
|
||||
- New database columns: metadata_json, metadata_generated_at
|
||||
- Repository methods for metadata search and filtering
|
||||
|
||||
## [2.11.0] - 2025-01-14
|
||||
|
||||
### Added
|
||||
- **Comprehensive Template Pagination**: All template search and list tools now return paginated responses
|
||||
- Consistent `PaginatedResponse` format with `items`, `total`, `limit`, `offset`, and `hasMore` fields
|
||||
- Customizable limits (1-100) and offset parameters for all template tools
|
||||
- Count methods for accurate pagination information across all template queries
|
||||
- **New `list_templates` Tool**: Efficient browsing of all available templates
|
||||
- Returns minimal data (id, name, views, nodeCount) for quick overview
|
||||
- Supports sorting by views, created_at, or name
|
||||
- Optimized for discovering templates without downloading full workflow data
|
||||
- **Flexible Template Retrieval Modes**: Enhanced `get_template` with three response modes
|
||||
- `nodes_only`: Returns just node types and names (minimal tokens)
|
||||
- `structure`: Returns nodes with positions and connections (moderate detail)
|
||||
- `full`: Returns complete workflow JSON (default, maximum detail)
|
||||
- Reduces token usage by 80-90% in minimal modes
|
||||
|
||||
### Enhanced
|
||||
- **Template Database Compression**: Implemented gzip compression for workflow JSONs
|
||||
- Workflow data compressed from ~75MB to 12.10MB (84% reduction)
|
||||
- Database size reduced from 117MB to 48MB despite 5x more templates
|
||||
- Transparent compression/decompression with base64 encoding
|
||||
- No API changes - compression is handled internally
|
||||
- **Template Quality Filtering**: Automatic filtering of low-quality templates
|
||||
- Templates with ≤10 views are excluded from the database
|
||||
- Expanded coverage from 499 to 2,596 high-quality templates (5x increase)
|
||||
- Filtered 4,505 raw templates down to 2,596 based on popularity
|
||||
- Ensures AI agents work with proven, valuable workflows
|
||||
- **Enhanced Database Statistics**: Template metrics now included
|
||||
- Shows total template count, average/min/max views
|
||||
- Provides complete database overview including template coverage
|
||||
|
||||
### Performance
|
||||
- **Database Optimization**: 59% size reduction while storing 5x more content
|
||||
- Previous: ~40MB database with 499 templates
|
||||
- Current: ~48MB database with 2,596 templates
|
||||
- Without compression would be ~120MB+
|
||||
- **Token Efficiency**: 80-90% reduction in response size for minimal queries
|
||||
- `list_templates`: ~10 tokens per template vs 100+ for full data
|
||||
- `get_template` with `nodes_only`: Returns just essential node information
|
||||
- Pagination prevents overwhelming responses for large result sets
|
||||
|
||||
### Fixed
|
||||
- **Test Suite Compatibility**: Updated all tests for new template system
|
||||
- Fixed parameter validation tests to expect new method signatures
|
||||
- Updated integration tests to use templates with >10 views
|
||||
- Removed redundant test files that were testing at wrong abstraction level
|
||||
- All 1,700+ tests now passing
|
||||
|
||||
## [2.10.9] - 2025-01-09
|
||||
|
||||
### Changed
|
||||
- **Dependencies**: Updated n8n packages to 1.110.1
|
||||
- n8n: 1.109.2 → 1.110.1
|
||||
- n8n-core: 1.108.0 → 1.109.0
|
||||
- n8n-workflow: 1.106.0 → 1.107.0
|
||||
- @n8n/n8n-nodes-langchain: 1.108.1 → 1.109.1
|
||||
|
||||
### Updated
|
||||
- **Node Database**: Rebuilt with 536 nodes from updated n8n packages
|
||||
- **Templates**: Refreshed workflow templates database with latest 499 templates from n8n.io
|
||||
|
||||
## [2.10.8] - 2025-09-04
|
||||
|
||||
### Updated
|
||||
- **n8n Dependencies**: Updated to latest versions for compatibility and new features
|
||||
- n8n: 1.107.4 → 1.109.2
|
||||
- @n8n/n8n-nodes-langchain: 1.106.2 → 1.109.1
|
||||
- n8n-nodes-base: 1.106.3 → 1.108.0 (via dependencies)
|
||||
- **Node Database**: Rebuilt with 535 nodes from updated n8n packages
|
||||
- **Node.js Compatibility**: Optimized for Node.js v22.17.0 LTS
|
||||
- Enhanced better-sqlite3 native binary compatibility
|
||||
- Fixed SQL.js fallback mode for environments without native binaries
|
||||
- **CI/CD Improvements**: Fixed Rollup native module compatibility for GitHub Actions
|
||||
- Added explicit platform-specific rollup binaries for cross-platform builds
|
||||
- Resolved npm ci failures in Linux CI environment
|
||||
- Fixed package-lock.json synchronization issues
|
||||
- **Platform Support**: Enhanced cross-platform deployment compatibility
|
||||
- macOS ARM64 and Linux x64 platform binaries included
|
||||
- Improved npm package distribution with proper dependency resolution
|
||||
- All 1,728+ tests passing with updated dependencies
|
||||
|
||||
### Fixed
|
||||
- **CI/CD Pipeline**: Resolved test failures in GitHub Actions
|
||||
- Fixed pyodide version conflicts between langchain dependencies
|
||||
- Regenerated package-lock.json with proper dependency resolution
|
||||
- Fixed Rollup native module loading in Linux CI environment
|
||||
- **Database Compatibility**: Enhanced SQL.js fallback reliability
|
||||
- Improved parameter binding and state management
|
||||
- Fixed statement cleanup to prevent memory leaks
|
||||
- **Deployment Reliability**: Better handling of platform-specific dependencies
|
||||
- npm ci now works consistently across development and CI environments
|
||||
|
||||
## [2.10.5] - 2025-08-20
|
||||
|
||||
### Updated
|
||||
- **n8n Dependencies**: Updated to latest versions for compatibility and new features
|
||||
- n8n: 1.106.3 → 1.107.4
|
||||
- n8n-core: 1.105.3 → 1.106.2
|
||||
- n8n-workflow: 1.103.3 → 1.104.1
|
||||
- @n8n/n8n-nodes-langchain: 1.105.3 → 1.106.2
|
||||
- **Node Database**: Rebuilt with 535 nodes from updated n8n packages
|
||||
- All tests passing with updated dependencies
|
||||
|
||||
## [2.10.4] - 2025-08-12
|
||||
|
||||
### Updated
|
||||
- **n8n Dependencies**: Updated to latest versions for compatibility and new features
|
||||
- n8n: 1.105.2 → 1.106.3
|
||||
- n8n-core: 1.104.1 → 1.105.3
|
||||
- n8n-workflow: 1.102.1 → 1.103.3
|
||||
- @n8n/n8n-nodes-langchain: 1.104.1 → 1.105.3
|
||||
- **Node Database**: Rebuilt with 535 nodes from updated n8n packages
|
||||
- All 1,728 tests passing with updated dependencies
|
||||
|
||||
## [2.10.3] - 2025-08-07
|
||||
|
||||
### Fixed
|
||||
- **Validation System Robustness**: Fixed multiple critical validation issues affecting AI agents and workflow validation (fixes #58, #68, #70, #73)
|
||||
- **Issue #73**: Fixed `validate_node_minimal` crash when config is undefined
|
||||
- Added safe property access with optional chaining (`config?.resource`)
|
||||
- Tool now handles undefined, null, and malformed configs gracefully
|
||||
- **Issue #58**: Fixed `validate_node_operation` crash on invalid nodeType
|
||||
- Added type checking before calling string methods
|
||||
- Prevents "Cannot read properties of undefined (reading 'replace')" error
|
||||
- **Issue #70**: Fixed validation profile settings being ignored
|
||||
- Extended profile parameter to all validation phases (nodes, connections, expressions)
|
||||
- Added Sticky Notes filtering to reduce false positives
|
||||
- Enhanced cycle detection to allow legitimate loops (SplitInBatches)
|
||||
- **Issue #68**: Added error recovery suggestions for AI agents
|
||||
- New `addErrorRecoverySuggestions()` method provides actionable recovery steps
|
||||
- Categorizes errors and suggests specific fixes for each type
|
||||
- Helps AI agents self-correct when validation fails
|
||||
|
||||
### Added
|
||||
- **Input Validation System**: Comprehensive validation for all MCP tool inputs
|
||||
- Created `validation-schemas.ts` with custom validation utilities
|
||||
- No external dependencies - pure TypeScript implementation
|
||||
- Tool-specific validation schemas for all MCP tools
|
||||
- Clear error messages with field-level details
|
||||
- **Enhanced Cycle Detection**: Improved detection of legitimate loops vs actual cycles
|
||||
- Recognizes SplitInBatches loop patterns as valid
|
||||
- Reduces false positive cycle warnings
|
||||
- **Comprehensive Test Suite**: Added 16 tests covering all validation fixes
|
||||
- Tests for crash prevention with malformed inputs
|
||||
- Tests for profile behavior across validation phases
|
||||
- Tests for error recovery suggestions
|
||||
- Tests for legitimate loop patterns
|
||||
|
||||
### Enhanced
|
||||
- **Validation Profiles**: Now consistently applied across all validation phases
|
||||
- `minimal`: Reduces warnings for basic validation
|
||||
- `runtime`: Standard validation for production workflows
|
||||
- `ai-friendly`: Optimized for AI agent workflow creation
|
||||
- `strict`: Maximum validation for critical workflows
|
||||
- **Error Messages**: More helpful and actionable for both humans and AI agents
|
||||
- Specific recovery suggestions for common errors
|
||||
- Clear guidance on fixing validation issues
|
||||
- Examples of correct configurations
|
||||
|
||||
## [2.10.2] - 2025-08-05
|
||||
|
||||
### Updated
|
||||
- **n8n Dependencies**: Updated to latest versions for compatibility and new features
|
||||
- n8n: 1.104.1 → 1.105.2
|
||||
- n8n-core: 1.103.1 → 1.104.1
|
||||
- n8n-workflow: 1.101.0 → 1.102.1
|
||||
- @n8n/n8n-nodes-langchain: 1.103.1 → 1.104.1
|
||||
- **Node Database**: Rebuilt with 534 nodes from updated n8n packages
|
||||
- **Template Library**: Fetched 499 workflow templates from the last 12 months
|
||||
- Templates are filtered to include only those created or updated within the past year
|
||||
- This ensures the template library contains fresh and actively maintained workflows
|
||||
- All 1,620 tests passing with updated dependencies
|
||||
|
||||
## [2.10.1] - 2025-08-02
|
||||
|
||||
### Fixed
|
||||
- **Memory Leak in SimpleCache**: Fixed critical memory leak causing MCP server connection loss after several hours (fixes #118)
|
||||
- Added proper timer cleanup in `SimpleCache.destroy()` method
|
||||
- Updated MCP server shutdown to clean up cache timers
|
||||
- Enhanced HTTP server error handling with transport error handlers
|
||||
- Fixed event listener cleanup to prevent accumulation
|
||||
- Added comprehensive test coverage for memory leak prevention
|
||||
|
||||
## [2.10.0] - 2025-08-02
|
||||
|
||||
@@ -1074,6 +1595,18 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
|
||||
- Basic n8n and MCP integration
|
||||
- Core workflow automation features
|
||||
|
||||
[2.12.0]: https://github.com/czlonkowski/n8n-mcp/compare/v2.11.3...v2.12.0
|
||||
[2.11.3]: https://github.com/czlonkowski/n8n-mcp/compare/v2.11.2...v2.11.3
|
||||
[2.11.2]: https://github.com/czlonkowski/n8n-mcp/compare/v2.11.1...v2.11.2
|
||||
[2.11.1]: https://github.com/czlonkowski/n8n-mcp/compare/v2.11.0...v2.11.1
|
||||
[2.11.0]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.9...v2.11.0
|
||||
[2.10.9]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.8...v2.10.9
|
||||
[2.10.8]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.5...v2.10.8
|
||||
[2.10.5]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.4...v2.10.5
|
||||
[2.10.4]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.3...v2.10.4
|
||||
[2.10.3]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.2...v2.10.3
|
||||
[2.10.2]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.1...v2.10.2
|
||||
[2.10.1]: https://github.com/czlonkowski/n8n-mcp/compare/v2.10.0...v2.10.1
|
||||
[2.10.0]: https://github.com/czlonkowski/n8n-mcp/compare/v2.9.1...v2.10.0
|
||||
[2.9.1]: https://github.com/czlonkowski/n8n-mcp/compare/v2.9.0...v2.9.1
|
||||
[2.9.0]: https://github.com/czlonkowski/n8n-mcp/compare/v2.8.3...v2.9.0
|
||||
|
||||
111
docs/CI_TEST_INFRASTRUCTURE.md
Normal file
111
docs/CI_TEST_INFRASTRUCTURE.md
Normal file
@@ -0,0 +1,111 @@
|
||||
# CI Test Infrastructure - Known Issues
|
||||
|
||||
## Integration Test Failures for External Contributor PRs
|
||||
|
||||
### Issue Summary
|
||||
|
||||
Integration tests fail for external contributor PRs with "No response from n8n server" errors, despite the code changes being correct. This is a **test infrastructure issue**, not a code quality issue.
|
||||
|
||||
### Root Cause
|
||||
|
||||
1. **GitHub Actions Security**: External contributor PRs don't get access to repository secrets (`N8N_API_URL`, `N8N_API_KEY`, etc.)
|
||||
2. **MSW Mock Server**: Mock Service Worker (MSW) is not properly intercepting HTTP requests in the CI environment
|
||||
3. **Test Configuration**: Integration tests expect `http://localhost:3001/mock-api` but the mock server isn't responding
|
||||
|
||||
### Evidence
|
||||
|
||||
From CI logs (PR #343):
|
||||
```
|
||||
[CI-DEBUG] Global setup complete, N8N_API_URL: http://localhost:3001/mock-api
|
||||
❌ No response from n8n server (repeated 60+ times across 20 tests)
|
||||
```
|
||||
|
||||
The tests ARE using the correct mock URL, but MSW isn't intercepting the requests.
|
||||
|
||||
### Why This Happens
|
||||
|
||||
**For External PRs:**
|
||||
- GitHub Actions doesn't expose repository secrets for security reasons
|
||||
- Prevents malicious PRs from exfiltrating secrets
|
||||
- MSW setup runs but requests don't get intercepted in CI
|
||||
|
||||
**Test Configuration:**
|
||||
- `.env.test` line 19: `N8N_API_URL=http://localhost:3001/mock-api`
|
||||
- `.env.test` line 67: `MSW_ENABLED=true`
|
||||
- CI workflow line 75-80: Secrets set but empty for external PRs
|
||||
|
||||
### Impact
|
||||
|
||||
- ✅ **Code Quality**: NOT affected - the actual code changes are correct
|
||||
- ✅ **Local Testing**: Works fine - MSW intercepts requests locally
|
||||
- ❌ **CI for External PRs**: Integration tests fail (infrastructure issue)
|
||||
- ✅ **CI for Internal PRs**: Works fine (has access to secrets)
|
||||
|
||||
### Current Workarounds
|
||||
|
||||
1. **For Maintainers**: Use `--admin` flag to merge despite failing tests when code is verified correct
|
||||
2. **For Contributors**: Run tests locally where MSW works properly
|
||||
3. **For CI**: Unit tests pass (don't require n8n API), integration tests fail
|
||||
|
||||
### Files Affected
|
||||
|
||||
- `tests/integration/setup/integration-setup.ts` - MSW server setup
|
||||
- `tests/setup/msw-setup.ts` - MSW configuration
|
||||
- `tests/mocks/n8n-api/handlers.ts` - Mock request handlers
|
||||
- `.github/workflows/test.yml` - CI configuration
|
||||
- `.env.test` - Test environment configuration
|
||||
|
||||
### Potential Solutions (Not Implemented)
|
||||
|
||||
1. **Separate Unit/Integration Runs**
|
||||
- Run integration tests only for internal PRs
|
||||
- Skip integration tests for external PRs
|
||||
- Rely on unit tests for external PR validation
|
||||
|
||||
2. **MSW CI Debugging**
|
||||
- Add extensive logging to MSW setup
|
||||
- Check if MSW server actually starts in CI
|
||||
- Verify request interception is working
|
||||
|
||||
3. **Mock Server Process**
|
||||
- Start actual HTTP server in CI instead of MSW
|
||||
- More reliable but adds complexity
|
||||
- Would require test infrastructure refactoring
|
||||
|
||||
4. **Public Test Instance**
|
||||
- Use publicly accessible test n8n instance
|
||||
- Exposes test data, security concerns
|
||||
- Would work for external PRs
|
||||
|
||||
### Decision
|
||||
|
||||
**Status**: Documented but not fixed
|
||||
|
||||
**Rationale**:
|
||||
- Integration test infrastructure refactoring is separate concern from code quality
|
||||
- External PRs are relatively rare compared to internal development
|
||||
- Unit tests provide sufficient coverage for most changes
|
||||
- Maintainers can verify integration tests locally before merging
|
||||
|
||||
### Testing Strategy
|
||||
|
||||
**For External Contributor PRs:**
|
||||
1. ✅ Unit tests must pass
|
||||
2. ✅ TypeScript compilation must pass
|
||||
3. ✅ Build must succeed
|
||||
4. ⚠️ Integration test failures are expected (infrastructure issue)
|
||||
5. ✅ Maintainer verifies locally before merge
|
||||
|
||||
**For Internal PRs:**
|
||||
1. ✅ All tests must pass (unit + integration)
|
||||
2. ✅ Full CI validation
|
||||
|
||||
### References
|
||||
|
||||
- PR #343: First occurrence of this issue
|
||||
- PR #345: Documented the infrastructure issue
|
||||
- Issue: External PRs don't get secrets (GitHub Actions security)
|
||||
|
||||
### Last Updated
|
||||
|
||||
2025-10-21 - Documented as part of PR #345 investigation
|
||||
@@ -80,6 +80,53 @@ Remove the server:
|
||||
claude mcp remove n8n-mcp
|
||||
```
|
||||
|
||||
## 🎓 Add Claude Skills (Optional)
|
||||
|
||||
Supercharge your n8n workflow building with specialized Claude Code skills! The [n8n-skills](https://github.com/czlonkowski/n8n-skills) repository provides 7 complementary skills that teach AI assistants how to build production-ready n8n workflows.
|
||||
|
||||
### What You Get
|
||||
|
||||
- ✅ **n8n Expression Syntax** - Correct {{}} patterns and common mistakes
|
||||
- ✅ **n8n MCP Tools Expert** - How to use n8n-mcp tools effectively
|
||||
- ✅ **n8n Workflow Patterns** - 5 proven architectural patterns
|
||||
- ✅ **n8n Validation Expert** - Interpret and fix validation errors
|
||||
- ✅ **n8n Node Configuration** - Operation-aware setup guidance
|
||||
- ✅ **n8n Code JavaScript** - Write effective JavaScript in Code nodes
|
||||
- ✅ **n8n Code Python** - Python patterns with limitation awareness
|
||||
|
||||
### Installation
|
||||
|
||||
**Method 1: Plugin Installation** (Recommended)
|
||||
```bash
|
||||
/plugin install czlonkowski/n8n-skills
|
||||
```
|
||||
|
||||
**Method 2: Via Marketplace**
|
||||
```bash
|
||||
# Add as marketplace, then browse and install
|
||||
/plugin marketplace add czlonkowski/n8n-skills
|
||||
|
||||
# Then browse available plugins
|
||||
/plugin install
|
||||
# Select "n8n-mcp-skills" from the list
|
||||
```
|
||||
|
||||
**Method 3: Manual Installation**
|
||||
```bash
|
||||
# 1. Clone the repository
|
||||
git clone https://github.com/czlonkowski/n8n-skills.git
|
||||
|
||||
# 2. Copy skills to your Claude Code skills directory
|
||||
cp -r n8n-skills/skills/* ~/.claude/skills/
|
||||
|
||||
# 3. Reload Claude Code
|
||||
# Skills will activate automatically
|
||||
```
|
||||
|
||||
For complete installation instructions, configuration options, and usage examples, see the [n8n-skills README](https://github.com/czlonkowski/n8n-skills#-installation).
|
||||
|
||||
Skills work seamlessly with n8n-mcp to provide expert guidance throughout the workflow building process!
|
||||
|
||||
## Project Instructions
|
||||
|
||||
For optimal results, create a `CLAUDE.md` file in your project root with the instructions from the [main README's Claude Project Setup section](../README.md#-claude-project-setup).
|
||||
|
||||
34
docs/CODEX_SETUP.md
Normal file
34
docs/CODEX_SETUP.md
Normal file
@@ -0,0 +1,34 @@
|
||||
# Codex Setup
|
||||
|
||||
Connect n8n-MCP to Codex for enhanced n8n workflow development.
|
||||
|
||||
## Update your Codex configuration
|
||||
|
||||
Go to your Codex settings at `~/.codex/config.toml` and add the following configuration:
|
||||
|
||||
### Basic configuration (documentation tools only):
|
||||
```toml
|
||||
[mcp_servers.n8n]
|
||||
command = "npx"
|
||||
args = ["n8n-mcp"]
|
||||
env = { "MCP_MODE" = "stdio", "LOG_LEVEL" = "error", "DISABLE_CONSOLE_OUTPUT" = "true" }
|
||||
```
|
||||
|
||||
### Full configuration (with n8n management tools):
|
||||
```toml
|
||||
[mcp_servers.n8n]
|
||||
command = "npx"
|
||||
args = ["n8n-mcp"]
|
||||
env = { "MCP_MODE" = "stdio", "LOG_LEVEL" = "error", "DISABLE_CONSOLE_OUTPUT" = "true", "N8N_API_URL" = "https://your-n8n-instance.com", "N8N_API_KEY" = "your-api-key" }
|
||||
```
|
||||
|
||||
Make sure to replace `https://your-n8n-instance.com` with your actual n8n URL and `your-api-key` with your n8n API key.
|
||||
|
||||
## Managing Your MCP Server
|
||||
Enter the Codex CLI and use the `/mcp` command to see server status and available tools.
|
||||
|
||||

|
||||
|
||||
## Project Instructions
|
||||
|
||||
For optimal results, create a `AGENTS.md` file in your project root with the instructions same with [main README's Claude Project Setup section](../README.md#-claude-project-setup).
|
||||
@@ -65,6 +65,9 @@ docker run -d \
|
||||
| `NODE_ENV` | Environment: `development` or `production` | `production` | No |
|
||||
| `LOG_LEVEL` | Logging level: `debug`, `info`, `warn`, `error` | `info` | No |
|
||||
| `NODE_DB_PATH` | Custom database path (v2.7.16+) | `/app/data/nodes.db` | No |
|
||||
| `AUTH_RATE_LIMIT_WINDOW` | Rate limit window in ms (v2.16.3+) | `900000` (15 min) | No |
|
||||
| `AUTH_RATE_LIMIT_MAX` | Max auth attempts per window (v2.16.3+) | `20` | No |
|
||||
| `WEBHOOK_SECURITY_MODE` | SSRF protection: `strict`/`moderate`/`permissive` (v2.16.3+) | `strict` | No |
|
||||
|
||||
*Either `AUTH_TOKEN` or `AUTH_TOKEN_FILE` must be set for HTTP mode. If both are set, `AUTH_TOKEN` takes precedence.
|
||||
|
||||
@@ -283,7 +286,36 @@ docker ps --format "table {{.Names}}\t{{.Status}}"
|
||||
docker inspect n8n-mcp | jq '.[0].State.Health'
|
||||
```
|
||||
|
||||
## 🔒 Security Considerations
|
||||
## 🔒 Security Features (v2.16.3+)
|
||||
|
||||
### Rate Limiting
|
||||
|
||||
Protects against brute force authentication attacks:
|
||||
|
||||
```bash
|
||||
# Configure in .env or docker-compose.yml
|
||||
AUTH_RATE_LIMIT_WINDOW=900000 # 15 minutes in milliseconds
|
||||
AUTH_RATE_LIMIT_MAX=20 # 20 attempts per IP per window
|
||||
```
|
||||
|
||||
### SSRF Protection
|
||||
|
||||
Prevents Server-Side Request Forgery when using webhook triggers:
|
||||
|
||||
```bash
|
||||
# For production (blocks localhost + private IPs + cloud metadata)
|
||||
WEBHOOK_SECURITY_MODE=strict
|
||||
|
||||
# For local development with local n8n instance
|
||||
WEBHOOK_SECURITY_MODE=moderate
|
||||
|
||||
# For internal testing only (allows private IPs)
|
||||
WEBHOOK_SECURITY_MODE=permissive
|
||||
```
|
||||
|
||||
**Note:** Cloud metadata endpoints (169.254.169.254, metadata.google.internal, etc.) are ALWAYS blocked in all modes.
|
||||
|
||||
## 🔒 Authentication
|
||||
|
||||
### Authentication
|
||||
|
||||
|
||||
@@ -196,6 +196,41 @@ docker ps -a | grep n8n-mcp | grep Exited | awk '{print $1}' | xargs -r docker r
|
||||
- Manually clean up containers periodically
|
||||
- Consider using HTTP mode instead
|
||||
|
||||
### Webhooks to Local n8n Fail (v2.16.3+)
|
||||
|
||||
**Symptoms:**
|
||||
- `n8n_trigger_webhook_workflow` fails with "SSRF protection" error
|
||||
- Error message: "SSRF protection: Localhost access is blocked"
|
||||
- Webhooks work from n8n UI but not from n8n-MCP
|
||||
|
||||
**Root Cause:** Default strict SSRF protection blocks localhost access to prevent attacks.
|
||||
|
||||
**Solution:** Use moderate security mode for local development
|
||||
|
||||
```bash
|
||||
# For Docker run
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
-e MCP_MODE=http \
|
||||
-e AUTH_TOKEN=your-token \
|
||||
-e WEBHOOK_SECURITY_MODE=moderate \
|
||||
-p 3000:3000 \
|
||||
ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
|
||||
# For Docker Compose - add to environment:
|
||||
services:
|
||||
n8n-mcp:
|
||||
environment:
|
||||
WEBHOOK_SECURITY_MODE: moderate
|
||||
```
|
||||
|
||||
**Security Modes Explained:**
|
||||
- `strict` (default): Blocks localhost + private IPs + cloud metadata (production)
|
||||
- `moderate`: Allows localhost, blocks private IPs + cloud metadata (local development)
|
||||
- `permissive`: Allows localhost + private IPs, blocks cloud metadata (testing only)
|
||||
|
||||
**Important:** Always use `strict` mode in production. Cloud metadata is blocked in all modes.
|
||||
|
||||
### n8n API Connection Issues
|
||||
|
||||
**Symptoms:**
|
||||
|
||||
3491
docs/FINAL_AI_VALIDATION_SPEC.md
Normal file
3491
docs/FINAL_AI_VALIDATION_SPEC.md
Normal file
File diff suppressed because it is too large
Load Diff
371
docs/FLEXIBLE_INSTANCE_CONFIGURATION.md
Normal file
371
docs/FLEXIBLE_INSTANCE_CONFIGURATION.md
Normal file
@@ -0,0 +1,371 @@
|
||||
# Flexible Instance Configuration
|
||||
|
||||
## Overview
|
||||
|
||||
The Flexible Instance Configuration feature enables n8n-mcp to serve multiple users with different n8n instances dynamically, without requiring separate deployments for each user. This feature is designed for scenarios where n8n-mcp is hosted centrally and needs to connect to different n8n instances based on runtime context.
|
||||
|
||||
## Architecture
|
||||
|
||||
### Core Components
|
||||
|
||||
1. **InstanceContext Interface** (`src/types/instance-context.ts`)
|
||||
- Runtime configuration container for instance-specific settings
|
||||
- Optional fields for backward compatibility
|
||||
- Comprehensive validation with security checks
|
||||
|
||||
2. **Dual-Mode API Client**
|
||||
- **Singleton Mode**: Uses environment variables (backward compatible)
|
||||
- **Instance Mode**: Uses runtime context for multi-instance support
|
||||
- Automatic fallback between modes
|
||||
|
||||
3. **LRU Cache with Security**
|
||||
- SHA-256 hashed cache keys for security
|
||||
- 30-minute TTL with automatic cleanup
|
||||
- Maximum 100 concurrent instances
|
||||
- Secure dispose callbacks without logging sensitive data
|
||||
|
||||
4. **Session Management**
|
||||
- HTTP server tracks session context
|
||||
- Each session can have different instance configuration
|
||||
- Automatic cleanup on session end
|
||||
|
||||
## Configuration
|
||||
|
||||
### Environment Variables
|
||||
|
||||
New environment variables for cache configuration:
|
||||
|
||||
- `INSTANCE_CACHE_MAX` - Maximum number of cached instances (default: 100, min: 1, max: 10000)
|
||||
- `INSTANCE_CACHE_TTL_MINUTES` - Cache TTL in minutes (default: 30, min: 1, max: 1440/24 hours)
|
||||
|
||||
Example:
|
||||
```bash
|
||||
# Increase cache size for high-volume deployments
|
||||
export INSTANCE_CACHE_MAX=500
|
||||
export INSTANCE_CACHE_TTL_MINUTES=60
|
||||
```
|
||||
|
||||
### InstanceContext Structure
|
||||
|
||||
```typescript
|
||||
interface InstanceContext {
|
||||
n8nApiUrl?: string; // n8n instance URL
|
||||
n8nApiKey?: string; // API key for authentication
|
||||
n8nApiTimeout?: number; // Request timeout in ms (default: 30000)
|
||||
n8nApiMaxRetries?: number; // Max retry attempts (default: 3)
|
||||
instanceId?: string; // Unique instance identifier
|
||||
sessionId?: string; // Session identifier
|
||||
metadata?: Record<string, any>; // Additional metadata
|
||||
}
|
||||
```
|
||||
|
||||
### Validation Rules
|
||||
|
||||
1. **URL Validation**:
|
||||
- Must be valid HTTP/HTTPS URL
|
||||
- No file://, javascript:, or other dangerous protocols
|
||||
- Proper URL format with protocol and host
|
||||
|
||||
2. **API Key Validation**:
|
||||
- Non-empty string required when provided
|
||||
- No placeholder values (e.g., "YOUR_API_KEY")
|
||||
- Case-insensitive placeholder detection
|
||||
|
||||
3. **Numeric Validation**:
|
||||
- Timeout must be positive number (>0)
|
||||
- Max retries must be non-negative (≥0)
|
||||
- No Infinity or NaN values
|
||||
|
||||
## Usage Examples
|
||||
|
||||
### Basic Usage
|
||||
|
||||
```typescript
|
||||
import { getN8nApiClient } from './mcp/handlers-n8n-manager';
|
||||
import { InstanceContext } from './types/instance-context';
|
||||
|
||||
// Create context for a specific instance
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: 'https://customer1.n8n.cloud',
|
||||
n8nApiKey: 'customer1-api-key',
|
||||
instanceId: 'customer1'
|
||||
};
|
||||
|
||||
// Get client for this instance
|
||||
const client = getN8nApiClient(context);
|
||||
if (client) {
|
||||
// Use client for API operations
|
||||
const workflows = await client.getWorkflows();
|
||||
}
|
||||
```
|
||||
|
||||
### HTTP Headers for Multi-Tenant Support
|
||||
|
||||
When using the HTTP server mode, clients can pass instance-specific configuration via HTTP headers:
|
||||
|
||||
```bash
|
||||
# Example curl request with instance headers
|
||||
curl -X POST http://localhost:3000/mcp \
|
||||
-H "Authorization: Bearer your-auth-token" \
|
||||
-H "Content-Type: application/json" \
|
||||
-H "X-N8n-Url: https://instance1.n8n.cloud" \
|
||||
-H "X-N8n-Key: instance1-api-key" \
|
||||
-H "X-Instance-Id: instance-1" \
|
||||
-H "X-Session-Id: session-123" \
|
||||
-d '{"method": "n8n_list_workflows", "params": {}, "id": 1}'
|
||||
```
|
||||
|
||||
#### Supported Headers
|
||||
|
||||
- **X-N8n-Url**: The n8n instance URL (e.g., `https://instance.n8n.cloud`)
|
||||
- **X-N8n-Key**: The API key for authentication with the n8n instance
|
||||
- **X-Instance-Id**: A unique identifier for the instance (optional, for tracking)
|
||||
- **X-Session-Id**: A session identifier (optional, for session tracking)
|
||||
|
||||
#### Header Extraction Logic
|
||||
|
||||
1. If either `X-N8n-Url` or `X-N8n-Key` header is present, an instance context is created
|
||||
2. All headers are extracted and passed to the MCP server
|
||||
3. The server uses the instance-specific configuration instead of environment variables
|
||||
4. If no headers are present, the server falls back to environment variables (backward compatible)
|
||||
|
||||
#### Example: JavaScript Client
|
||||
|
||||
```javascript
|
||||
const headers = {
|
||||
'Authorization': 'Bearer your-auth-token',
|
||||
'Content-Type': 'application/json',
|
||||
'X-N8n-Url': 'https://customer1.n8n.cloud',
|
||||
'X-N8n-Key': 'customer1-api-key',
|
||||
'X-Instance-Id': 'customer-1',
|
||||
'X-Session-Id': 'session-456'
|
||||
};
|
||||
|
||||
const response = await fetch('http://localhost:3000/mcp', {
|
||||
method: 'POST',
|
||||
headers: headers,
|
||||
body: JSON.stringify({
|
||||
method: 'n8n_list_workflows',
|
||||
params: {},
|
||||
id: 1
|
||||
})
|
||||
});
|
||||
|
||||
const result = await response.json();
|
||||
```
|
||||
|
||||
### HTTP Server Integration
|
||||
|
||||
```typescript
|
||||
// In HTTP request handler
|
||||
app.post('/mcp', (req, res) => {
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: req.headers['x-n8n-url'],
|
||||
n8nApiKey: req.headers['x-n8n-key'],
|
||||
sessionId: req.sessionID
|
||||
};
|
||||
|
||||
// Context passed to handlers
|
||||
const result = await handleRequest(req.body, context);
|
||||
res.json(result);
|
||||
});
|
||||
```
|
||||
|
||||
### Validation Example
|
||||
|
||||
```typescript
|
||||
import { validateInstanceContext } from './types/instance-context';
|
||||
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: 'https://api.n8n.cloud',
|
||||
n8nApiKey: 'valid-key'
|
||||
};
|
||||
|
||||
const validation = validateInstanceContext(context);
|
||||
if (!validation.valid) {
|
||||
console.error('Validation errors:', validation.errors);
|
||||
} else {
|
||||
// Context is valid, proceed
|
||||
const client = getN8nApiClient(context);
|
||||
}
|
||||
```
|
||||
|
||||
## Security Features
|
||||
|
||||
### 1. Cache Key Hashing
|
||||
- All cache keys use SHA-256 hashing with memoization
|
||||
- Prevents sensitive data exposure in logs
|
||||
- Example: `sha256(url:key:instance)` → 64-char hex string
|
||||
- Memoization cache limited to 1000 entries
|
||||
|
||||
### 2. Enhanced Input Validation
|
||||
- Field-specific error messages with detailed reasons
|
||||
- URL protocol restrictions (HTTP/HTTPS only)
|
||||
- API key placeholder detection (case-insensitive)
|
||||
- Numeric range validation with specific error messages
|
||||
- Example: "Invalid n8nApiUrl: ftp://example.com - URL must use HTTP or HTTPS protocol"
|
||||
|
||||
### 3. Secure Logging
|
||||
- Only first 8 characters of cache keys logged
|
||||
- No sensitive data in debug logs
|
||||
- URL sanitization (domain only, no paths)
|
||||
- Configuration fallback logging for debugging
|
||||
|
||||
### 4. Memory Management
|
||||
- Configurable LRU cache with automatic eviction
|
||||
- TTL-based expiration (configurable, default 30 minutes)
|
||||
- Dispose callbacks for cleanup
|
||||
- Maximum cache size limits with bounds checking
|
||||
|
||||
### 5. Concurrency Protection
|
||||
- Mutex-based locking for cache operations
|
||||
- Prevents duplicate client creation
|
||||
- Simple lock checking with timeout
|
||||
- Thread-safe cache operations
|
||||
|
||||
## Performance Optimization
|
||||
|
||||
### Cache Strategy
|
||||
- **Max Size**: Configurable via `INSTANCE_CACHE_MAX` (default: 100)
|
||||
- **TTL**: Configurable via `INSTANCE_CACHE_TTL_MINUTES` (default: 30)
|
||||
- **Update on Access**: Age refreshed on each use
|
||||
- **Eviction**: Least Recently Used (LRU) policy
|
||||
- **Memoization**: Hash creation uses memoization for frequently used keys
|
||||
|
||||
### Cache Metrics
|
||||
The system tracks comprehensive metrics:
|
||||
- Cache hits and misses
|
||||
- Hit rate percentage
|
||||
- Eviction count
|
||||
- Current size vs maximum size
|
||||
- Operation timing
|
||||
|
||||
Retrieve metrics using:
|
||||
```typescript
|
||||
import { getInstanceCacheStatistics } from './mcp/handlers-n8n-manager';
|
||||
console.log(getInstanceCacheStatistics());
|
||||
```
|
||||
|
||||
### Benefits
|
||||
- **Performance**: ~12ms average response time
|
||||
- **Memory Efficient**: Minimal footprint per instance
|
||||
- **Thread Safe**: Mutex protection for concurrent operations
|
||||
- **Auto Cleanup**: Unused instances automatically evicted
|
||||
- **No Memory Leaks**: Proper disposal callbacks
|
||||
|
||||
## Backward Compatibility
|
||||
|
||||
The feature maintains 100% backward compatibility:
|
||||
|
||||
1. **Environment Variables Still Work**:
|
||||
- If no context provided, falls back to env vars
|
||||
- Existing deployments continue working unchanged
|
||||
|
||||
2. **Optional Parameters**:
|
||||
- All context fields are optional
|
||||
- Missing fields use defaults or env vars
|
||||
|
||||
3. **API Unchanged**:
|
||||
- Same handler signatures with optional context
|
||||
- No breaking changes to existing code
|
||||
|
||||
## Testing
|
||||
|
||||
Comprehensive test coverage ensures reliability:
|
||||
|
||||
```bash
|
||||
# Run all flexible instance tests
|
||||
npm test -- tests/unit/flexible-instance-security-advanced.test.ts
|
||||
npm test -- tests/unit/mcp/lru-cache-behavior.test.ts
|
||||
npm test -- tests/unit/types/instance-context-coverage.test.ts
|
||||
npm test -- tests/unit/mcp/handlers-n8n-manager-simple.test.ts
|
||||
```
|
||||
|
||||
### Test Coverage Areas
|
||||
- Input validation edge cases
|
||||
- Cache behavior and eviction
|
||||
- Security (hashing, sanitization)
|
||||
- Session management
|
||||
- Memory leak prevention
|
||||
- Concurrent access patterns
|
||||
|
||||
## Migration Guide
|
||||
|
||||
### For Existing Deployments
|
||||
No changes required - environment variables continue to work.
|
||||
|
||||
### For Multi-Instance Support
|
||||
|
||||
1. **Update HTTP Server** (if using HTTP mode):
|
||||
```typescript
|
||||
// Add context extraction from headers
|
||||
const context = extractInstanceContext(req);
|
||||
```
|
||||
|
||||
2. **Pass Context to Handlers**:
|
||||
```typescript
|
||||
// Old way (still works)
|
||||
await handleListWorkflows(params);
|
||||
|
||||
// New way (with instance context)
|
||||
await handleListWorkflows(params, context);
|
||||
```
|
||||
|
||||
3. **Configure Clients** to send instance information:
|
||||
```typescript
|
||||
// Client sends instance info in headers
|
||||
headers: {
|
||||
'X-N8n-Url': 'https://instance.n8n.cloud',
|
||||
'X-N8n-Key': 'api-key',
|
||||
'X-Instance-Id': 'customer-123'
|
||||
}
|
||||
```
|
||||
|
||||
## Monitoring
|
||||
|
||||
### Metrics to Track
|
||||
- Cache hit/miss ratio
|
||||
- Instance count in cache
|
||||
- Average TTL utilization
|
||||
- Memory usage per instance
|
||||
- API client creation rate
|
||||
|
||||
### Debug Logging
|
||||
Enable debug logs to monitor cache behavior:
|
||||
```bash
|
||||
LOG_LEVEL=debug npm start
|
||||
```
|
||||
|
||||
## Limitations
|
||||
|
||||
1. **Maximum Instances**: 100 concurrent instances (configurable)
|
||||
2. **TTL**: 30-minute cache lifetime (configurable)
|
||||
3. **Memory**: ~1MB per cached instance (estimated)
|
||||
4. **Validation**: Strict validation may reject edge cases
|
||||
|
||||
## Security Considerations
|
||||
|
||||
1. **Never Log Sensitive Data**: API keys are never logged
|
||||
2. **Hash All Identifiers**: Use SHA-256 for cache keys
|
||||
3. **Validate All Input**: Comprehensive validation before use
|
||||
4. **Limit Resources**: Cache size and TTL limits
|
||||
5. **Clean Up Properly**: Dispose callbacks for resource cleanup
|
||||
|
||||
## Future Enhancements
|
||||
|
||||
Potential improvements for future versions:
|
||||
|
||||
1. **Configurable Cache Settings**: Runtime cache size/TTL configuration
|
||||
2. **Instance Metrics**: Per-instance usage tracking
|
||||
3. **Rate Limiting**: Per-instance rate limits
|
||||
4. **Instance Groups**: Logical grouping of instances
|
||||
5. **Persistent Cache**: Optional Redis/database backing
|
||||
6. **Instance Discovery**: Automatic instance detection
|
||||
|
||||
## Support
|
||||
|
||||
For issues or questions about flexible instance configuration:
|
||||
1. Check validation errors for specific problems
|
||||
2. Enable debug logging for detailed diagnostics
|
||||
3. Review test files for usage examples
|
||||
4. Open an issue on GitHub with details
|
||||
@@ -73,6 +73,13 @@ PORT=3000
|
||||
# Optional: Enable n8n management tools
|
||||
# N8N_API_URL=https://your-n8n-instance.com
|
||||
# N8N_API_KEY=your-api-key-here
|
||||
# Security Configuration (v2.16.3+)
|
||||
# Rate limiting (default: 20 attempts per 15 minutes)
|
||||
AUTH_RATE_LIMIT_WINDOW=900000
|
||||
AUTH_RATE_LIMIT_MAX=20
|
||||
# SSRF protection mode (default: strict)
|
||||
# Use 'moderate' for local n8n, 'strict' for production
|
||||
WEBHOOK_SECURITY_MODE=strict
|
||||
EOF
|
||||
|
||||
# 2. Deploy with Docker
|
||||
@@ -592,6 +599,67 @@ curl -H "Authorization: Bearer $AUTH_TOKEN" \
|
||||
}
|
||||
```
|
||||
|
||||
## 🔒 Security Features (v2.16.3+)
|
||||
|
||||
### Rate Limiting
|
||||
|
||||
Built-in rate limiting protects authentication endpoints from brute force attacks:
|
||||
|
||||
**Configuration:**
|
||||
```bash
|
||||
# Defaults (15 minutes window, 20 attempts per IP)
|
||||
AUTH_RATE_LIMIT_WINDOW=900000 # milliseconds
|
||||
AUTH_RATE_LIMIT_MAX=20
|
||||
```
|
||||
|
||||
**Features:**
|
||||
- Per-IP rate limiting with configurable window and max attempts
|
||||
- Standard rate limit headers (RateLimit-Limit, RateLimit-Remaining, RateLimit-Reset)
|
||||
- JSON-RPC formatted error responses
|
||||
- Automatic IP tracking behind reverse proxies (requires TRUST_PROXY=1)
|
||||
|
||||
**Behavior:**
|
||||
- First 20 attempts: Return 401 Unauthorized for invalid credentials
|
||||
- Attempts 21+: Return 429 Too Many Requests with Retry-After header
|
||||
- Counter resets after 15 minutes (configurable)
|
||||
|
||||
### SSRF Protection
|
||||
|
||||
Prevents Server-Side Request Forgery attacks when using webhook triggers:
|
||||
|
||||
**Three Security Modes:**
|
||||
|
||||
1. **Strict Mode (default)** - Production deployments
|
||||
```bash
|
||||
WEBHOOK_SECURITY_MODE=strict
|
||||
```
|
||||
- ✅ Block localhost (127.0.0.1, ::1)
|
||||
- ✅ Block private IPs (10.x, 192.168.x, 172.16-31.x)
|
||||
- ✅ Block cloud metadata (169.254.169.254, metadata.google.internal)
|
||||
- ✅ DNS rebinding prevention
|
||||
- 🎯 **Use for**: Cloud deployments, production environments
|
||||
|
||||
2. **Moderate Mode** - Local development with local n8n
|
||||
```bash
|
||||
WEBHOOK_SECURITY_MODE=moderate
|
||||
```
|
||||
- ✅ Allow localhost (for local n8n instances)
|
||||
- ✅ Block private IPs
|
||||
- ✅ Block cloud metadata
|
||||
- ✅ DNS rebinding prevention
|
||||
- 🎯 **Use for**: Development with n8n on localhost:5678
|
||||
|
||||
3. **Permissive Mode** - Internal networks only
|
||||
```bash
|
||||
WEBHOOK_SECURITY_MODE=permissive
|
||||
```
|
||||
- ✅ Allow localhost and private IPs
|
||||
- ✅ Block cloud metadata (always blocked)
|
||||
- ✅ DNS rebinding prevention
|
||||
- 🎯 **Use for**: Internal testing (NOT for production)
|
||||
|
||||
**Important:** Cloud metadata endpoints are ALWAYS blocked in all modes for security.
|
||||
|
||||
## 🔒 Security Best Practices
|
||||
|
||||
### 1. Token Management
|
||||
|
||||
724
docs/LIBRARY_USAGE.md
Normal file
724
docs/LIBRARY_USAGE.md
Normal file
@@ -0,0 +1,724 @@
|
||||
# Library Usage Guide - Multi-Tenant / Hosted Deployments
|
||||
|
||||
This guide covers using n8n-mcp as a library dependency for building multi-tenant hosted services.
|
||||
|
||||
## Overview
|
||||
|
||||
n8n-mcp can be used as a Node.js library to build multi-tenant backends that provide MCP services to multiple users or instances. The package exports all necessary components for integration into your existing services.
|
||||
|
||||
## Installation
|
||||
|
||||
```bash
|
||||
npm install n8n-mcp
|
||||
```
|
||||
|
||||
## Core Concepts
|
||||
|
||||
### Library Mode vs CLI Mode
|
||||
|
||||
- **CLI Mode** (default): Single-player usage via `npx n8n-mcp` or Docker
|
||||
- **Library Mode**: Multi-tenant usage by importing and using the `N8NMCPEngine` class
|
||||
|
||||
### Instance Context
|
||||
|
||||
The `InstanceContext` type allows you to pass per-request configuration to the MCP engine:
|
||||
|
||||
```typescript
|
||||
interface InstanceContext {
|
||||
// Instance-specific n8n API configuration
|
||||
n8nApiUrl?: string;
|
||||
n8nApiKey?: string;
|
||||
n8nApiTimeout?: number;
|
||||
n8nApiMaxRetries?: number;
|
||||
|
||||
// Instance identification
|
||||
instanceId?: string;
|
||||
sessionId?: string;
|
||||
|
||||
// Extensible metadata
|
||||
metadata?: Record<string, any>;
|
||||
}
|
||||
```
|
||||
|
||||
## Basic Example
|
||||
|
||||
```typescript
|
||||
import express from 'express';
|
||||
import { N8NMCPEngine } from 'n8n-mcp';
|
||||
|
||||
const app = express();
|
||||
const mcpEngine = new N8NMCPEngine({
|
||||
sessionTimeout: 3600000, // 1 hour
|
||||
logLevel: 'info'
|
||||
});
|
||||
|
||||
// Handle MCP requests with per-user context
|
||||
app.post('/mcp', async (req, res) => {
|
||||
const instanceContext = {
|
||||
n8nApiUrl: req.user.n8nUrl,
|
||||
n8nApiKey: req.user.n8nApiKey,
|
||||
instanceId: req.user.id
|
||||
};
|
||||
|
||||
await mcpEngine.processRequest(req, res, instanceContext);
|
||||
});
|
||||
|
||||
app.listen(3000);
|
||||
```
|
||||
|
||||
## Multi-Tenant Backend Example
|
||||
|
||||
This example shows a complete multi-tenant implementation with user authentication and instance management:
|
||||
|
||||
```typescript
|
||||
import express from 'express';
|
||||
import { N8NMCPEngine, InstanceContext, validateInstanceContext } from 'n8n-mcp';
|
||||
|
||||
const app = express();
|
||||
const mcpEngine = new N8NMCPEngine({
|
||||
sessionTimeout: 3600000, // 1 hour
|
||||
logLevel: 'info'
|
||||
});
|
||||
|
||||
// Start MCP engine
|
||||
await mcpEngine.start();
|
||||
|
||||
// Authentication middleware
|
||||
const authenticate = async (req, res, next) => {
|
||||
const token = req.headers.authorization?.replace('Bearer ', '');
|
||||
if (!token) {
|
||||
return res.status(401).json({ error: 'Unauthorized' });
|
||||
}
|
||||
|
||||
// Verify token and attach user to request
|
||||
req.user = await getUserFromToken(token);
|
||||
next();
|
||||
};
|
||||
|
||||
// Get instance configuration from database
|
||||
const getInstanceConfig = async (instanceId: string, userId: string) => {
|
||||
// Your database logic here
|
||||
const instance = await db.instances.findOne({
|
||||
where: { id: instanceId, userId }
|
||||
});
|
||||
|
||||
if (!instance) {
|
||||
throw new Error('Instance not found');
|
||||
}
|
||||
|
||||
return {
|
||||
n8nApiUrl: instance.n8nUrl,
|
||||
n8nApiKey: await decryptApiKey(instance.encryptedApiKey),
|
||||
instanceId: instance.id
|
||||
};
|
||||
};
|
||||
|
||||
// MCP endpoint with per-instance context
|
||||
app.post('/api/instances/:instanceId/mcp', authenticate, async (req, res) => {
|
||||
try {
|
||||
// Get instance configuration
|
||||
const instance = await getInstanceConfig(req.params.instanceId, req.user.id);
|
||||
|
||||
// Create instance context
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: instance.n8nApiUrl,
|
||||
n8nApiKey: instance.n8nApiKey,
|
||||
instanceId: instance.instanceId,
|
||||
metadata: {
|
||||
userId: req.user.id,
|
||||
userAgent: req.headers['user-agent'],
|
||||
ip: req.ip
|
||||
}
|
||||
};
|
||||
|
||||
// Validate context before processing
|
||||
const validation = validateInstanceContext(context);
|
||||
if (!validation.valid) {
|
||||
return res.status(400).json({
|
||||
error: 'Invalid instance configuration',
|
||||
details: validation.errors
|
||||
});
|
||||
}
|
||||
|
||||
// Process request with instance context
|
||||
await mcpEngine.processRequest(req, res, context);
|
||||
|
||||
} catch (error) {
|
||||
console.error('MCP request error:', error);
|
||||
res.status(500).json({ error: 'Internal server error' });
|
||||
}
|
||||
});
|
||||
|
||||
// Health endpoint
|
||||
app.get('/health', async (req, res) => {
|
||||
const health = await mcpEngine.healthCheck();
|
||||
res.status(health.status === 'healthy' ? 200 : 503).json(health);
|
||||
});
|
||||
|
||||
// Graceful shutdown
|
||||
process.on('SIGTERM', async () => {
|
||||
await mcpEngine.shutdown();
|
||||
process.exit(0);
|
||||
});
|
||||
|
||||
app.listen(3000);
|
||||
```
|
||||
|
||||
## API Reference
|
||||
|
||||
### N8NMCPEngine
|
||||
|
||||
#### Constructor
|
||||
|
||||
```typescript
|
||||
new N8NMCPEngine(options?: {
|
||||
sessionTimeout?: number; // Session TTL in ms (default: 1800000 = 30min)
|
||||
logLevel?: 'error' | 'warn' | 'info' | 'debug'; // Default: 'info'
|
||||
})
|
||||
```
|
||||
|
||||
#### Methods
|
||||
|
||||
##### `async processRequest(req, res, context?)`
|
||||
|
||||
Process a single MCP request with optional instance context.
|
||||
|
||||
**Parameters:**
|
||||
- `req`: Express request object
|
||||
- `res`: Express response object
|
||||
- `context` (optional): InstanceContext with per-instance configuration
|
||||
|
||||
**Example:**
|
||||
```typescript
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: 'https://instance1.n8n.cloud',
|
||||
n8nApiKey: 'instance1-key',
|
||||
instanceId: 'tenant-123'
|
||||
};
|
||||
|
||||
await engine.processRequest(req, res, context);
|
||||
```
|
||||
|
||||
##### `async healthCheck()`
|
||||
|
||||
Get engine health status for monitoring.
|
||||
|
||||
**Returns:** `EngineHealth`
|
||||
```typescript
|
||||
{
|
||||
status: 'healthy' | 'unhealthy';
|
||||
uptime: number; // seconds
|
||||
sessionActive: boolean;
|
||||
memoryUsage: {
|
||||
used: number;
|
||||
total: number;
|
||||
unit: string;
|
||||
};
|
||||
version: string;
|
||||
}
|
||||
```
|
||||
|
||||
**Example:**
|
||||
```typescript
|
||||
app.get('/health', async (req, res) => {
|
||||
const health = await engine.healthCheck();
|
||||
res.status(health.status === 'healthy' ? 200 : 503).json(health);
|
||||
});
|
||||
```
|
||||
|
||||
##### `getSessionInfo()`
|
||||
|
||||
Get current session information for debugging.
|
||||
|
||||
**Returns:**
|
||||
```typescript
|
||||
{
|
||||
active: boolean;
|
||||
sessionId?: string;
|
||||
age?: number; // milliseconds
|
||||
sessions?: {
|
||||
total: number;
|
||||
active: number;
|
||||
expired: number;
|
||||
max: number;
|
||||
sessionIds: string[];
|
||||
};
|
||||
}
|
||||
```
|
||||
|
||||
##### `async start()`
|
||||
|
||||
Start the engine (for standalone mode). Not needed when using `processRequest()` directly.
|
||||
|
||||
##### `async shutdown()`
|
||||
|
||||
Graceful shutdown for service lifecycle management.
|
||||
|
||||
**Example:**
|
||||
```typescript
|
||||
process.on('SIGTERM', async () => {
|
||||
await engine.shutdown();
|
||||
process.exit(0);
|
||||
});
|
||||
```
|
||||
|
||||
### Types
|
||||
|
||||
#### InstanceContext
|
||||
|
||||
Configuration for a specific user instance:
|
||||
|
||||
```typescript
|
||||
interface InstanceContext {
|
||||
n8nApiUrl?: string;
|
||||
n8nApiKey?: string;
|
||||
n8nApiTimeout?: number;
|
||||
n8nApiMaxRetries?: number;
|
||||
instanceId?: string;
|
||||
sessionId?: string;
|
||||
metadata?: Record<string, any>;
|
||||
}
|
||||
```
|
||||
|
||||
#### Validation Functions
|
||||
|
||||
##### `validateInstanceContext(context: InstanceContext)`
|
||||
|
||||
Validate and sanitize instance context.
|
||||
|
||||
**Returns:**
|
||||
```typescript
|
||||
{
|
||||
valid: boolean;
|
||||
errors?: string[];
|
||||
}
|
||||
```
|
||||
|
||||
**Example:**
|
||||
```typescript
|
||||
import { validateInstanceContext } from 'n8n-mcp';
|
||||
|
||||
const validation = validateInstanceContext(context);
|
||||
if (!validation.valid) {
|
||||
console.error('Invalid context:', validation.errors);
|
||||
}
|
||||
```
|
||||
|
||||
##### `isInstanceContext(obj: any)`
|
||||
|
||||
Type guard to check if an object is a valid InstanceContext.
|
||||
|
||||
**Example:**
|
||||
```typescript
|
||||
import { isInstanceContext } from 'n8n-mcp';
|
||||
|
||||
if (isInstanceContext(req.body.context)) {
|
||||
// TypeScript knows this is InstanceContext
|
||||
await engine.processRequest(req, res, req.body.context);
|
||||
}
|
||||
```
|
||||
|
||||
## Session Management
|
||||
|
||||
### Session Strategies
|
||||
|
||||
The MCP engine supports flexible session ID formats:
|
||||
|
||||
- **UUIDv4**: Internal n8n-mcp format (default)
|
||||
- **Instance-prefixed**: `instance-{userId}-{hash}-{uuid}` for multi-tenant isolation
|
||||
- **Custom formats**: Any non-empty string for mcp-remote and other proxies
|
||||
|
||||
Session validation happens via transport lookup, not format validation. This ensures compatibility with all MCP clients.
|
||||
|
||||
### Multi-Tenant Configuration
|
||||
|
||||
Set these environment variables for multi-tenant mode:
|
||||
|
||||
```bash
|
||||
# Enable multi-tenant mode
|
||||
ENABLE_MULTI_TENANT=true
|
||||
|
||||
# Session strategy: "instance" (default) or "shared"
|
||||
MULTI_TENANT_SESSION_STRATEGY=instance
|
||||
```
|
||||
|
||||
**Session Strategies:**
|
||||
|
||||
- **instance** (recommended): Each tenant gets isolated sessions
|
||||
- Session ID: `instance-{instanceId}-{configHash}-{uuid}`
|
||||
- Better isolation and security
|
||||
- Easier debugging per tenant
|
||||
|
||||
- **shared**: Multiple tenants share sessions with context switching
|
||||
- More efficient for high tenant count
|
||||
- Requires careful context management
|
||||
|
||||
## Security Considerations
|
||||
|
||||
### API Key Management
|
||||
|
||||
Always encrypt API keys server-side:
|
||||
|
||||
```typescript
|
||||
import { createCipheriv, createDecipheriv } from 'crypto';
|
||||
|
||||
// Encrypt before storing
|
||||
const encryptApiKey = (apiKey: string) => {
|
||||
const cipher = createCipheriv('aes-256-gcm', encryptionKey, iv);
|
||||
return cipher.update(apiKey, 'utf8', 'hex') + cipher.final('hex');
|
||||
};
|
||||
|
||||
// Decrypt before using
|
||||
const decryptApiKey = (encrypted: string) => {
|
||||
const decipher = createDecipheriv('aes-256-gcm', encryptionKey, iv);
|
||||
return decipher.update(encrypted, 'hex', 'utf8') + decipher.final('utf8');
|
||||
};
|
||||
|
||||
// Use decrypted key in context
|
||||
const context: InstanceContext = {
|
||||
n8nApiKey: await decryptApiKey(instance.encryptedApiKey),
|
||||
// ...
|
||||
};
|
||||
```
|
||||
|
||||
### Input Validation
|
||||
|
||||
Always validate instance context before processing:
|
||||
|
||||
```typescript
|
||||
import { validateInstanceContext } from 'n8n-mcp';
|
||||
|
||||
const validation = validateInstanceContext(context);
|
||||
if (!validation.valid) {
|
||||
throw new Error(`Invalid context: ${validation.errors?.join(', ')}`);
|
||||
}
|
||||
```
|
||||
|
||||
### Rate Limiting
|
||||
|
||||
Implement rate limiting per tenant:
|
||||
|
||||
```typescript
|
||||
import rateLimit from 'express-rate-limit';
|
||||
|
||||
const limiter = rateLimit({
|
||||
windowMs: 15 * 60 * 1000, // 15 minutes
|
||||
max: 100, // limit each IP to 100 requests per windowMs
|
||||
keyGenerator: (req) => req.user?.id || req.ip
|
||||
});
|
||||
|
||||
app.post('/api/instances/:instanceId/mcp', authenticate, limiter, async (req, res) => {
|
||||
// ...
|
||||
});
|
||||
```
|
||||
|
||||
## Error Handling
|
||||
|
||||
Always wrap MCP requests in try-catch blocks:
|
||||
|
||||
```typescript
|
||||
app.post('/api/instances/:instanceId/mcp', authenticate, async (req, res) => {
|
||||
try {
|
||||
const context = await getInstanceConfig(req.params.instanceId, req.user.id);
|
||||
await mcpEngine.processRequest(req, res, context);
|
||||
} catch (error) {
|
||||
console.error('MCP error:', error);
|
||||
|
||||
// Don't leak internal errors to clients
|
||||
if (error.message.includes('not found')) {
|
||||
return res.status(404).json({ error: 'Instance not found' });
|
||||
}
|
||||
|
||||
res.status(500).json({ error: 'Internal server error' });
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
## Monitoring
|
||||
|
||||
### Health Checks
|
||||
|
||||
Set up periodic health checks:
|
||||
|
||||
```typescript
|
||||
setInterval(async () => {
|
||||
const health = await mcpEngine.healthCheck();
|
||||
|
||||
if (health.status === 'unhealthy') {
|
||||
console.error('MCP engine unhealthy:', health);
|
||||
// Alert your monitoring system
|
||||
}
|
||||
|
||||
// Log metrics
|
||||
console.log('MCP engine metrics:', {
|
||||
uptime: health.uptime,
|
||||
memory: health.memoryUsage,
|
||||
sessionActive: health.sessionActive
|
||||
});
|
||||
}, 60000); // Every minute
|
||||
```
|
||||
|
||||
### Session Monitoring
|
||||
|
||||
Track active sessions:
|
||||
|
||||
```typescript
|
||||
app.get('/admin/sessions', authenticate, async (req, res) => {
|
||||
if (!req.user.isAdmin) {
|
||||
return res.status(403).json({ error: 'Forbidden' });
|
||||
}
|
||||
|
||||
const sessionInfo = mcpEngine.getSessionInfo();
|
||||
res.json(sessionInfo);
|
||||
});
|
||||
```
|
||||
|
||||
## Testing
|
||||
|
||||
### Unit Testing
|
||||
|
||||
```typescript
|
||||
import { N8NMCPEngine, InstanceContext } from 'n8n-mcp';
|
||||
|
||||
describe('MCP Engine', () => {
|
||||
let engine: N8NMCPEngine;
|
||||
|
||||
beforeEach(() => {
|
||||
engine = new N8NMCPEngine({ logLevel: 'error' });
|
||||
});
|
||||
|
||||
afterEach(async () => {
|
||||
await engine.shutdown();
|
||||
});
|
||||
|
||||
it('should process request with context', async () => {
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: 'https://test.n8n.io',
|
||||
n8nApiKey: 'test-key',
|
||||
instanceId: 'test-instance'
|
||||
};
|
||||
|
||||
const mockReq = createMockRequest();
|
||||
const mockRes = createMockResponse();
|
||||
|
||||
await engine.processRequest(mockReq, mockRes, context);
|
||||
|
||||
expect(mockRes.status).toBe(200);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### Integration Testing
|
||||
|
||||
```typescript
|
||||
import request from 'supertest';
|
||||
import { createApp } from './app';
|
||||
|
||||
describe('Multi-tenant MCP API', () => {
|
||||
let app;
|
||||
let authToken;
|
||||
|
||||
beforeAll(async () => {
|
||||
app = await createApp();
|
||||
authToken = await getTestAuthToken();
|
||||
});
|
||||
|
||||
it('should handle MCP request for instance', async () => {
|
||||
const response = await request(app)
|
||||
.post('/api/instances/test-instance/mcp')
|
||||
.set('Authorization', `Bearer ${authToken}`)
|
||||
.send({
|
||||
jsonrpc: '2.0',
|
||||
method: 'initialize',
|
||||
params: {
|
||||
protocolVersion: '2024-11-05',
|
||||
capabilities: {}
|
||||
},
|
||||
id: 1
|
||||
});
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.result).toBeDefined();
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Deployment Considerations
|
||||
|
||||
### Environment Variables
|
||||
|
||||
```bash
|
||||
# Required for multi-tenant mode
|
||||
ENABLE_MULTI_TENANT=true
|
||||
MULTI_TENANT_SESSION_STRATEGY=instance
|
||||
|
||||
# Optional: Logging
|
||||
LOG_LEVEL=info
|
||||
DISABLE_CONSOLE_OUTPUT=false
|
||||
|
||||
# Optional: Session configuration
|
||||
SESSION_TIMEOUT=1800000 # 30 minutes in milliseconds
|
||||
MAX_SESSIONS=100
|
||||
|
||||
# Optional: Performance
|
||||
NODE_ENV=production
|
||||
```
|
||||
|
||||
### Docker Deployment
|
||||
|
||||
```dockerfile
|
||||
FROM node:20-alpine
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
COPY package*.json ./
|
||||
RUN npm ci --only=production
|
||||
|
||||
COPY . .
|
||||
|
||||
ENV NODE_ENV=production
|
||||
ENV ENABLE_MULTI_TENANT=true
|
||||
ENV LOG_LEVEL=info
|
||||
|
||||
EXPOSE 3000
|
||||
|
||||
CMD ["node", "dist/server.js"]
|
||||
```
|
||||
|
||||
### Kubernetes Deployment
|
||||
|
||||
```yaml
|
||||
apiVersion: apps/v1
|
||||
kind: Deployment
|
||||
metadata:
|
||||
name: n8n-mcp-backend
|
||||
spec:
|
||||
replicas: 3
|
||||
selector:
|
||||
matchLabels:
|
||||
app: n8n-mcp-backend
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
app: n8n-mcp-backend
|
||||
spec:
|
||||
containers:
|
||||
- name: backend
|
||||
image: your-registry/n8n-mcp-backend:latest
|
||||
ports:
|
||||
- containerPort: 3000
|
||||
env:
|
||||
- name: ENABLE_MULTI_TENANT
|
||||
value: "true"
|
||||
- name: LOG_LEVEL
|
||||
value: "info"
|
||||
resources:
|
||||
requests:
|
||||
memory: "256Mi"
|
||||
cpu: "250m"
|
||||
limits:
|
||||
memory: "512Mi"
|
||||
cpu: "500m"
|
||||
livenessProbe:
|
||||
httpGet:
|
||||
path: /health
|
||||
port: 3000
|
||||
initialDelaySeconds: 10
|
||||
periodSeconds: 30
|
||||
readinessProbe:
|
||||
httpGet:
|
||||
path: /health
|
||||
port: 3000
|
||||
initialDelaySeconds: 5
|
||||
periodSeconds: 10
|
||||
```
|
||||
|
||||
## Examples
|
||||
|
||||
### Complete Multi-Tenant SaaS Example
|
||||
|
||||
For a complete implementation example, see:
|
||||
- [n8n-mcp-backend](https://github.com/czlonkowski/n8n-mcp-backend) - Full hosted service implementation
|
||||
|
||||
### Migration from Single-Player
|
||||
|
||||
If you're migrating from single-player (CLI/Docker) to multi-tenant:
|
||||
|
||||
1. **Keep backward compatibility** - Use environment fallback:
|
||||
```typescript
|
||||
const context: InstanceContext = {
|
||||
n8nApiUrl: instanceUrl || process.env.N8N_API_URL,
|
||||
n8nApiKey: instanceKey || process.env.N8N_API_KEY,
|
||||
instanceId: instanceId || 'default'
|
||||
};
|
||||
```
|
||||
|
||||
2. **Gradual rollout** - Start with a feature flag:
|
||||
```typescript
|
||||
const isMultiTenant = process.env.ENABLE_MULTI_TENANT === 'true';
|
||||
|
||||
if (isMultiTenant) {
|
||||
const context = await getInstanceConfig(req.params.instanceId);
|
||||
await engine.processRequest(req, res, context);
|
||||
} else {
|
||||
// Legacy single-player mode
|
||||
await engine.processRequest(req, res);
|
||||
}
|
||||
```
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
#### Module Resolution Errors
|
||||
|
||||
If you see `Cannot find module 'n8n-mcp'`:
|
||||
|
||||
```bash
|
||||
# Clear node_modules and reinstall
|
||||
rm -rf node_modules package-lock.json
|
||||
npm install
|
||||
|
||||
# Verify package has types field
|
||||
npm info n8n-mcp
|
||||
|
||||
# Check TypeScript can resolve it
|
||||
npx tsc --noEmit
|
||||
```
|
||||
|
||||
#### Session ID Validation Errors
|
||||
|
||||
If you see `Invalid session ID format` errors:
|
||||
|
||||
- Ensure you're using n8n-mcp v2.18.9 or later
|
||||
- Session IDs can be any non-empty string
|
||||
- No need to generate UUIDs - use your own format
|
||||
|
||||
#### Memory Leaks
|
||||
|
||||
If memory usage grows over time:
|
||||
|
||||
```typescript
|
||||
// Ensure proper cleanup
|
||||
process.on('SIGTERM', async () => {
|
||||
await engine.shutdown();
|
||||
process.exit(0);
|
||||
});
|
||||
|
||||
// Monitor session count
|
||||
const sessionInfo = engine.getSessionInfo();
|
||||
console.log('Active sessions:', sessionInfo.sessions?.active);
|
||||
```
|
||||
|
||||
## Further Reading
|
||||
|
||||
- [MCP Protocol Specification](https://modelcontextprotocol.io/docs)
|
||||
- [n8n API Documentation](https://docs.n8n.io/api/)
|
||||
- [Express.js Guide](https://expressjs.com/en/guide/routing.html)
|
||||
- [n8n-mcp Main README](../README.md)
|
||||
|
||||
## Support
|
||||
|
||||
- **Issues**: [GitHub Issues](https://github.com/czlonkowski/n8n-mcp/issues)
|
||||
- **Discussions**: [GitHub Discussions](https://github.com/czlonkowski/n8n-mcp/discussions)
|
||||
- **Security**: For security issues, see [SECURITY.md](../SECURITY.md)
|
||||
@@ -35,15 +35,15 @@ cd n8n-mcp
|
||||
npm install
|
||||
npm run build
|
||||
|
||||
# Run the test script
|
||||
./scripts/test-n8n-mode.sh
|
||||
# Run the integration test script
|
||||
./scripts/test-n8n-integration.sh
|
||||
```
|
||||
|
||||
This script will:
|
||||
1. Start n8n-MCP in n8n mode on port 3001
|
||||
2. Enable debug logging for troubleshooting
|
||||
3. Run comprehensive protocol tests
|
||||
4. Display results and any issues found
|
||||
1. Start a real n8n instance in Docker
|
||||
2. Start n8n-MCP server configured for n8n
|
||||
3. Guide you through API key setup for workflow management
|
||||
4. Test the complete integration between n8n and n8n-MCP
|
||||
|
||||
### Manual Local Setup
|
||||
|
||||
@@ -86,8 +86,8 @@ curl http://localhost:3001/mcp
|
||||
| `MCP_MODE` | Yes | Enables HTTP mode for n8n MCP Client | `http` |
|
||||
| `N8N_API_URL` | Yes* | URL of your n8n instance | `http://localhost:5678` |
|
||||
| `N8N_API_KEY` | Yes* | n8n API key for workflow management | `n8n_api_xxx...` |
|
||||
| `MCP_AUTH_TOKEN` | Yes | Authentication token for MCP requests | `secure-random-32-char-token` |
|
||||
| `AUTH_TOKEN` | Yes | Must match MCP_AUTH_TOKEN | `secure-random-32-char-token` |
|
||||
| `MCP_AUTH_TOKEN` | Yes | Authentication token for MCP requests (min 32 chars) | `secure-random-32-char-token` |
|
||||
| `AUTH_TOKEN` | Yes | **MUST match MCP_AUTH_TOKEN exactly** | `secure-random-32-char-token` |
|
||||
| `PORT` | No | Port for the HTTP server | `3000` (default) |
|
||||
| `LOG_LEVEL` | No | Logging verbosity | `info`, `debug`, `error` |
|
||||
|
||||
@@ -103,13 +103,48 @@ Starting with version 2.9.2, we use a single optimized Dockerfile for all deploy
|
||||
|
||||
## Production Deployment
|
||||
|
||||
> **⚠️ Critical**: Docker caches images locally. Always run `docker pull ghcr.io/czlonkowski/n8n-mcp:latest` before deploying to ensure you have the latest version. This simple step prevents most deployment issues.
|
||||
|
||||
### Same Server as n8n
|
||||
|
||||
If you're running n8n-MCP on the same server as your n8n instance:
|
||||
|
||||
### Building from Source (Recommended)
|
||||
### Using Pre-built Image (Recommended)
|
||||
|
||||
For the latest features and bug fixes, build from source:
|
||||
The pre-built images are automatically updated with each release and are the easiest way to get started.
|
||||
|
||||
**IMPORTANT**: Always pull the latest image to avoid using cached versions:
|
||||
|
||||
```bash
|
||||
# ALWAYS pull the latest image first
|
||||
docker pull ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
|
||||
# Generate a secure token (save this!)
|
||||
AUTH_TOKEN=$(openssl rand -hex 32)
|
||||
echo "Your AUTH_TOKEN: $AUTH_TOKEN"
|
||||
|
||||
# Create a Docker network if n8n uses one
|
||||
docker network create n8n-net
|
||||
|
||||
# Run n8n-MCP container
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
--network n8n-net \
|
||||
-p 3000:3000 \
|
||||
-e N8N_MODE=true \
|
||||
-e MCP_MODE=http \
|
||||
-e N8N_API_URL=http://n8n:5678 \
|
||||
-e N8N_API_KEY=your-n8n-api-key \
|
||||
-e MCP_AUTH_TOKEN=$AUTH_TOKEN \
|
||||
-e AUTH_TOKEN=$AUTH_TOKEN \
|
||||
-e LOG_LEVEL=info \
|
||||
--restart unless-stopped \
|
||||
ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
```
|
||||
|
||||
### Building from Source (Advanced Users)
|
||||
|
||||
Only build from source if you need custom modifications or are contributing to development:
|
||||
|
||||
```bash
|
||||
# Clone and build
|
||||
@@ -119,49 +154,18 @@ cd n8n-mcp
|
||||
# Build Docker image
|
||||
docker build -t n8n-mcp:latest .
|
||||
|
||||
# Create a Docker network if n8n uses one
|
||||
docker network create n8n-net
|
||||
|
||||
# Run n8n-MCP container
|
||||
# Run using your local image
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
--network n8n-net \
|
||||
-p 3000:3000 \
|
||||
-e N8N_MODE=true \
|
||||
-e MCP_MODE=http \
|
||||
-e N8N_API_URL=http://n8n:5678 \
|
||||
-e N8N_API_KEY=your-n8n-api-key \
|
||||
-e MCP_AUTH_TOKEN=$(openssl rand -hex 32) \
|
||||
-e AUTH_TOKEN=$(openssl rand -hex 32) \
|
||||
-e LOG_LEVEL=info \
|
||||
--restart unless-stopped \
|
||||
# ... other settings
|
||||
n8n-mcp:latest
|
||||
```
|
||||
|
||||
### Using Pre-built Image (May Be Outdated)
|
||||
|
||||
⚠️ **Warning**: Pre-built images may be outdated due to CI/CD synchronization issues. Always check the [GitHub releases](https://github.com/czlonkowski/n8n-mcp/releases) for the latest version.
|
||||
|
||||
```bash
|
||||
# Create a Docker network if n8n uses one
|
||||
docker network create n8n-net
|
||||
|
||||
# Run n8n-MCP container
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
--network n8n-net \
|
||||
-p 3000:3000 \
|
||||
-e N8N_MODE=true \
|
||||
-e MCP_MODE=http \
|
||||
-e N8N_API_URL=http://n8n:5678 \
|
||||
-e N8N_API_KEY=your-n8n-api-key \
|
||||
-e MCP_AUTH_TOKEN=$(openssl rand -hex 32) \
|
||||
-e AUTH_TOKEN=$(openssl rand -hex 32) \
|
||||
-e LOG_LEVEL=info \
|
||||
--restart unless-stopped \
|
||||
ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
```
|
||||
|
||||
### Using systemd (for native installation)
|
||||
|
||||
```bash
|
||||
@@ -198,43 +202,19 @@ sudo systemctl start n8n-mcp
|
||||
|
||||
Deploy n8n-MCP on a separate server from your n8n instance:
|
||||
|
||||
#### Quick Docker Deployment (Build from Source)
|
||||
#### Quick Docker Deployment (Recommended)
|
||||
|
||||
**Always pull the latest image to ensure you have the current version:**
|
||||
|
||||
```bash
|
||||
# On your cloud server (Hetzner, AWS, DigitalOcean, etc.)
|
||||
# First, clone and build
|
||||
git clone https://github.com/czlonkowski/n8n-mcp.git
|
||||
cd n8n-mcp
|
||||
docker build -t n8n-mcp:latest .
|
||||
# ALWAYS pull the latest image first
|
||||
docker pull ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
|
||||
# Generate auth tokens
|
||||
AUTH_TOKEN=$(openssl rand -hex 32)
|
||||
echo "Save this AUTH_TOKEN: $AUTH_TOKEN"
|
||||
|
||||
# Run the container
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
-p 3000:3000 \
|
||||
-e N8N_MODE=true \
|
||||
-e MCP_MODE=http \
|
||||
-e N8N_API_URL=https://your-n8n-instance.com \
|
||||
-e N8N_API_KEY=your-n8n-api-key \
|
||||
-e MCP_AUTH_TOKEN=$AUTH_TOKEN \
|
||||
-e AUTH_TOKEN=$AUTH_TOKEN \
|
||||
-e LOG_LEVEL=info \
|
||||
--restart unless-stopped \
|
||||
n8n-mcp:latest
|
||||
```
|
||||
|
||||
#### Quick Docker Deployment (Pre-built Image)
|
||||
|
||||
⚠️ **Warning**: May be outdated. Check [releases](https://github.com/czlonkowski/n8n-mcp/releases) first.
|
||||
|
||||
```bash
|
||||
# Generate auth tokens
|
||||
AUTH_TOKEN=$(openssl rand -hex 32)
|
||||
echo "Save this AUTH_TOKEN: $AUTH_TOKEN"
|
||||
|
||||
# Run the container
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
@@ -250,6 +230,24 @@ docker run -d \
|
||||
ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
```
|
||||
|
||||
#### Building from Source (Advanced)
|
||||
|
||||
Only needed if you're modifying the code:
|
||||
|
||||
```bash
|
||||
# Clone and build
|
||||
git clone https://github.com/czlonkowski/n8n-mcp.git
|
||||
cd n8n-mcp
|
||||
docker build -t n8n-mcp:latest .
|
||||
|
||||
# Run using local image
|
||||
docker run -d \
|
||||
--name n8n-mcp \
|
||||
-p 3000:3000 \
|
||||
# ... same environment variables as above
|
||||
n8n-mcp:latest
|
||||
```
|
||||
|
||||
#### Full Production Setup (Hetzner/AWS/DigitalOcean)
|
||||
|
||||
1. **Server Requirements**:
|
||||
@@ -269,61 +267,7 @@ curl -fsSL https://get.docker.com | sh
|
||||
|
||||
3. **Deploy n8n-MCP with SSL** (using Caddy for automatic HTTPS):
|
||||
|
||||
**Option A: Build from Source (Recommended)**
|
||||
```bash
|
||||
# Clone and prepare
|
||||
git clone https://github.com/czlonkowski/n8n-mcp.git
|
||||
cd n8n-mcp
|
||||
|
||||
# Build local image
|
||||
docker build -t n8n-mcp:latest .
|
||||
|
||||
# Create docker-compose.yml
|
||||
cat > docker-compose.yml << 'EOF'
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
n8n-mcp:
|
||||
image: n8n-mcp:latest # Using locally built image
|
||||
container_name: n8n-mcp
|
||||
restart: unless-stopped
|
||||
environment:
|
||||
- N8N_MODE=true
|
||||
- MCP_MODE=http
|
||||
- N8N_API_URL=${N8N_API_URL}
|
||||
- N8N_API_KEY=${N8N_API_KEY}
|
||||
- MCP_AUTH_TOKEN=${MCP_AUTH_TOKEN}
|
||||
- AUTH_TOKEN=${AUTH_TOKEN}
|
||||
- PORT=3000
|
||||
- LOG_LEVEL=info
|
||||
networks:
|
||||
- web
|
||||
|
||||
caddy:
|
||||
image: caddy:2-alpine
|
||||
container_name: caddy
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "80:80"
|
||||
- "443:443"
|
||||
volumes:
|
||||
- ./Caddyfile:/etc/caddy/Caddyfile
|
||||
- caddy_data:/data
|
||||
- caddy_config:/config
|
||||
networks:
|
||||
- web
|
||||
|
||||
networks:
|
||||
web:
|
||||
driver: bridge
|
||||
|
||||
volumes:
|
||||
caddy_data:
|
||||
caddy_config:
|
||||
EOF
|
||||
```
|
||||
|
||||
**Option B: Pre-built Image (May Be Outdated)**
|
||||
**Using Docker Compose (Recommended)**
|
||||
```bash
|
||||
# Create docker-compose.yml
|
||||
cat > docker-compose.yml << 'EOF'
|
||||
@@ -332,6 +276,7 @@ version: '3.8'
|
||||
services:
|
||||
n8n-mcp:
|
||||
image: ghcr.io/czlonkowski/n8n-mcp:latest
|
||||
pull_policy: always # Always pull latest image
|
||||
container_name: n8n-mcp
|
||||
restart: unless-stopped
|
||||
environment:
|
||||
@@ -370,7 +315,56 @@ volumes:
|
||||
EOF
|
||||
```
|
||||
|
||||
**Complete Setup (Both Options)**
|
||||
**Note**: The `pull_policy: always` ensures you always get the latest version.
|
||||
|
||||
**Building from Source (if needed)**
|
||||
```bash
|
||||
# Only if you need custom modifications
|
||||
git clone https://github.com/czlonkowski/n8n-mcp.git
|
||||
cd n8n-mcp
|
||||
docker build -t n8n-mcp:local .
|
||||
|
||||
# Then update docker-compose.yml to use:
|
||||
# image: n8n-mcp:local
|
||||
container_name: n8n-mcp
|
||||
restart: unless-stopped
|
||||
environment:
|
||||
- N8N_MODE=true
|
||||
- MCP_MODE=http
|
||||
- N8N_API_URL=${N8N_API_URL}
|
||||
- N8N_API_KEY=${N8N_API_KEY}
|
||||
- MCP_AUTH_TOKEN=${MCP_AUTH_TOKEN}
|
||||
- AUTH_TOKEN=${AUTH_TOKEN}
|
||||
- PORT=3000
|
||||
- LOG_LEVEL=info
|
||||
networks:
|
||||
- web
|
||||
|
||||
caddy:
|
||||
image: caddy:2-alpine
|
||||
container_name: caddy
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "80:80"
|
||||
- "443:443"
|
||||
volumes:
|
||||
- ./Caddyfile:/etc/caddy/Caddyfile
|
||||
- caddy_data:/data
|
||||
- caddy_config:/config
|
||||
networks:
|
||||
- web
|
||||
|
||||
networks:
|
||||
web:
|
||||
driver: bridge
|
||||
|
||||
volumes:
|
||||
caddy_data:
|
||||
caddy_config:
|
||||
EOF
|
||||
```
|
||||
|
||||
**Complete the Setup**
|
||||
```bash
|
||||
# Create Caddyfile
|
||||
cat > Caddyfile << 'EOF'
|
||||
@@ -481,12 +475,21 @@ You are an n8n workflow expert. Use the MCP tools to:
|
||||
- **IP Whitelisting**: Consider restricting access to known n8n instances
|
||||
|
||||
### Docker Security
|
||||
- **Always pull latest images**: Docker caches images locally, so run `docker pull` before deployment
|
||||
- Run containers with `--read-only` flag if possible
|
||||
- Use specific image versions instead of `:latest` in production
|
||||
- Regular updates: `docker pull ghcr.io/czlonkowski/n8n-mcp:latest`
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Docker Image Issues
|
||||
|
||||
**Using Outdated Cached Images**
|
||||
- **Symptom**: Missing features, old bugs reappearing, features not working as documented
|
||||
- **Cause**: Docker uses locally cached images instead of pulling the latest version
|
||||
- **Solution**: Always run `docker pull ghcr.io/czlonkowski/n8n-mcp:latest` before deployment
|
||||
- **Verification**: Check image age with `docker images | grep n8n-mcp`
|
||||
|
||||
### Common Configuration Issues
|
||||
|
||||
**Missing `MCP_MODE=http` Environment Variable**
|
||||
@@ -572,10 +575,10 @@ You are an n8n workflow expert. Use the MCP tools to:
|
||||
|
||||
### Version Compatibility Issues
|
||||
|
||||
**"Outdated Docker Image"**
|
||||
**"Features Not Working as Expected"**
|
||||
- **Symptom**: Missing features, old bugs, or compatibility issues
|
||||
- **Solution**: Build from source instead of using pre-built images
|
||||
- **Check**: Compare your image version with [GitHub releases](https://github.com/czlonkowski/n8n-mcp/releases)
|
||||
- **Solution**: Pull the latest image: `docker pull ghcr.io/czlonkowski/n8n-mcp:latest`
|
||||
- **Check**: Verify image date with `docker inspect ghcr.io/czlonkowski/n8n-mcp:latest | grep Created`
|
||||
|
||||
**"Protocol version mismatch"**
|
||||
- n8n-MCP automatically uses version 2024-11-05 for n8n compatibility
|
||||
@@ -752,4 +755,4 @@ curl http://localhost:3001/mcp
|
||||
|
||||
---
|
||||
|
||||
Need help? Open an issue on [GitHub](https://github.com/czlonkowski/n8n-mcp/issues) or check the [n8n forums](https://community.n8n.io).
|
||||
Need help? Open an issue on [GitHub](https://github.com/czlonkowski/n8n-mcp/issues) or check the [n8n forums](https://community.n8n.io)
|
||||
@@ -1,62 +0,0 @@
|
||||
# PR #104 Test Suite Improvements Summary
|
||||
|
||||
## Overview
|
||||
Based on comprehensive review feedback from PR #104, we've significantly improved the test suite quality, organization, and coverage.
|
||||
|
||||
## Test Results
|
||||
- **Before:** 78 failing tests
|
||||
- **After:** 0 failing tests (1,356 passed, 19 skipped)
|
||||
- **Coverage:** 85.34% statements, 85.3% branches
|
||||
|
||||
## Key Improvements
|
||||
|
||||
### 1. Fixed All Test Failures
|
||||
- Fixed logger test spy issues by properly handling DEBUG environment variable
|
||||
- Fixed MSW configuration test by restoring environment variables
|
||||
- Fixed workflow validator tests by adding proper node connections
|
||||
- Fixed mock setup issues in edge case tests
|
||||
|
||||
### 2. Improved Test Organization
|
||||
- Split large config-validator.test.ts (1,075 lines) into 4 focused files:
|
||||
- config-validator-basic.test.ts
|
||||
- config-validator-node-specific.test.ts
|
||||
- config-validator-security.test.ts
|
||||
- config-validator-edge-cases.test.ts
|
||||
|
||||
### 3. Enhanced Test Coverage
|
||||
- Added comprehensive edge case tests for all major validators
|
||||
- Added null/undefined handling tests
|
||||
- Added boundary value tests
|
||||
- Added performance tests with CI-aware timeouts
|
||||
- Added security validation tests
|
||||
|
||||
### 4. Improved Test Quality
|
||||
- Fixed test naming conventions (100% compliance with "should X when Y" pattern)
|
||||
- Added JSDoc comments to test utilities and factories
|
||||
- Created comprehensive test documentation (tests/README.md)
|
||||
- Improved test isolation to prevent cross-test pollution
|
||||
|
||||
### 5. New Features
|
||||
- Implemented validateBatch method for ConfigValidator
|
||||
- Added test factories for better test data management
|
||||
- Created test utilities for common scenarios
|
||||
|
||||
## Files Modified
|
||||
- 7 existing test files fixed
|
||||
- 8 new test files created
|
||||
- 1 source file enhanced (ConfigValidator)
|
||||
- 4 debug files removed before commit
|
||||
|
||||
## Skipped Tests
|
||||
19 tests remain skipped with documented reasons:
|
||||
- FTS5 search sync test (database corruption in CI)
|
||||
- Template clearing (not implemented)
|
||||
- Mock API configuration tests
|
||||
- Duplicate edge case tests with mocking issues (working versions exist)
|
||||
|
||||
## Next Steps
|
||||
The only remaining task from the improvement plan is:
|
||||
- Add performance regression tests and boundaries (low priority, future sprint)
|
||||
|
||||
## Conclusion
|
||||
The test suite is now robust, well-organized, and provides excellent coverage. All critical issues have been resolved, and the codebase is ready for merge.
|
||||
@@ -105,8 +105,30 @@ These are automatically set by the Railway template:
|
||||
| `CORS_ORIGIN` | `*` | Allow any origin |
|
||||
| `HOST` | `0.0.0.0` | Listen on all interfaces |
|
||||
| `PORT` | (Railway provides) | Don't set manually |
|
||||
| `AUTH_RATE_LIMIT_WINDOW` | `900000` (15 min) | Rate limit window (v2.16.3+) |
|
||||
| `AUTH_RATE_LIMIT_MAX` | `20` | Max auth attempts (v2.16.3+) |
|
||||
| `WEBHOOK_SECURITY_MODE` | `strict` | SSRF protection mode (v2.16.3+) |
|
||||
|
||||
### Optional: n8n API Integration
|
||||
### Optional Variables
|
||||
|
||||
| Variable | Default Value | Description |
|
||||
|----------|--------------|-------------|
|
||||
| `N8N_MODE` | `false` | Enable n8n integration mode for MCP Client Tool |
|
||||
| `N8N_API_URL` | - | URL of your n8n instance (for workflow management) |
|
||||
| `N8N_API_KEY` | - | API key from n8n Settings → API |
|
||||
|
||||
### Optional: n8n Integration
|
||||
|
||||
#### For n8n MCP Client Tool Integration
|
||||
|
||||
To use n8n-MCP with n8n's MCP Client Tool node:
|
||||
|
||||
1. **Go to Railway dashboard** → Your service → **Variables**
|
||||
2. **Add this variable**:
|
||||
- `N8N_MODE`: Set to `true` to enable n8n integration mode
|
||||
3. **Save changes** - Railway will redeploy automatically
|
||||
|
||||
#### For n8n API Integration (Workflow Management)
|
||||
|
||||
To enable workflow management features:
|
||||
|
||||
@@ -161,6 +183,46 @@ Claude Desktop → mcp-remote → Railway (HTTPS) → n8n-MCP Server
|
||||
- Ensure the URL is correct and includes `https://`
|
||||
- Check Railway logs for any errors
|
||||
|
||||
**Windows: "The filename, directory name, or volume label syntax is incorrect" or npx command not found:**
|
||||
|
||||
This is a common Windows issue with spaces in Node.js installation paths. The error occurs because Claude Desktop can't properly execute npx.
|
||||
|
||||
**Solution 1: Use node directly (Recommended)**
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"n8n-railway": {
|
||||
"command": "node",
|
||||
"args": [
|
||||
"C:\\Program Files\\nodejs\\node_modules\\npm\\bin\\npx-cli.js",
|
||||
"-y",
|
||||
"mcp-remote",
|
||||
"https://your-app-name.up.railway.app/mcp",
|
||||
"--header",
|
||||
"Authorization: Bearer YOUR_SECURE_TOKEN_HERE"
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**Solution 2: Use cmd wrapper**
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"n8n-railway": {
|
||||
"command": "cmd",
|
||||
"args": [
|
||||
"/C",
|
||||
"\"C:\\Program Files\\nodejs\\npx\" -y mcp-remote https://your-app-name.up.railway.app/mcp --header \"Authorization: Bearer YOUR_SECURE_TOKEN_HERE\""
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
To find your exact npx path, open Command Prompt and run: `where npx`
|
||||
|
||||
### Railway-Specific Issues
|
||||
|
||||
**Build failures:**
|
||||
@@ -225,6 +287,32 @@ Since the Railway template uses a specific Docker image tag, updates are manual:
|
||||
|
||||
You could use the `latest` tag, but this may cause unexpected breaking changes.
|
||||
|
||||
## 🔒 Security Features (v2.16.3+)
|
||||
|
||||
Railway deployments include enhanced security features:
|
||||
|
||||
### Rate Limiting
|
||||
- **Automatic brute force protection** - 20 attempts per 15 minutes per IP
|
||||
- **Configurable limits** via `AUTH_RATE_LIMIT_WINDOW` and `AUTH_RATE_LIMIT_MAX`
|
||||
- **Standard rate limit headers** for client awareness
|
||||
|
||||
### SSRF Protection
|
||||
- **Default strict mode** blocks localhost, private IPs, and cloud metadata
|
||||
- **Cloud metadata always blocked** (169.254.169.254, metadata.google.internal, etc.)
|
||||
- **Use `moderate` mode only if** connecting to local n8n instance
|
||||
|
||||
**Security Configuration:**
|
||||
```bash
|
||||
# In Railway Variables tab:
|
||||
WEBHOOK_SECURITY_MODE=strict # Production (recommended)
|
||||
# or
|
||||
WEBHOOK_SECURITY_MODE=moderate # If using local n8n with port forwarding
|
||||
|
||||
# Rate limiting (defaults are good for most use cases)
|
||||
AUTH_RATE_LIMIT_WINDOW=900000 # 15 minutes
|
||||
AUTH_RATE_LIMIT_MAX=20 # 20 attempts per IP
|
||||
```
|
||||
|
||||
## 📝 Best Practices
|
||||
|
||||
1. **Always change the default AUTH_TOKEN immediately**
|
||||
|
||||
BIN
docs/img/codex_connected.png
Normal file
BIN
docs/img/codex_connected.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 125 KiB |
BIN
docs/img/skills.png
Normal file
BIN
docs/img/skills.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 430 KiB |
@@ -1,162 +0,0 @@
|
||||
# Issue #90: "propertyValues[itemName] is not iterable" Error - Research Findings
|
||||
|
||||
## Executive Summary
|
||||
|
||||
The error "propertyValues[itemName] is not iterable" occurs when AI agents create workflows with incorrect data structures for n8n nodes that use `fixedCollection` properties. This primarily affects Switch Node v2, If Node, and Filter Node. The error prevents workflows from loading in the n8n UI, resulting in empty canvases.
|
||||
|
||||
## Root Cause Analysis
|
||||
|
||||
### 1. Data Structure Mismatch
|
||||
|
||||
The error occurs when n8n's validation engine expects an iterable array but encounters a non-iterable object. This happens with nodes using `fixedCollection` type properties.
|
||||
|
||||
**Incorrect Structure (causes error):**
|
||||
```json
|
||||
{
|
||||
"rules": {
|
||||
"conditions": {
|
||||
"values": [
|
||||
{
|
||||
"value1": "={{$json.status}}",
|
||||
"operation": "equals",
|
||||
"value2": "active"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**Correct Structure:**
|
||||
```json
|
||||
{
|
||||
"rules": {
|
||||
"conditions": [
|
||||
{
|
||||
"value1": "={{$json.status}}",
|
||||
"operation": "equals",
|
||||
"value2": "active"
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 2. Affected Nodes
|
||||
|
||||
Based on the research and issue comments, the following nodes are affected:
|
||||
|
||||
1. **Switch Node v2** (`n8n-nodes-base.switch` with typeVersion: 2)
|
||||
- Uses `rules` parameter with `conditions` fixedCollection
|
||||
- v3 doesn't have this issue due to restructured schema
|
||||
|
||||
2. **If Node** (`n8n-nodes-base.if` with typeVersion: 1)
|
||||
- Uses `conditions` parameter with nested conditions array
|
||||
- Similar structure to Switch v2
|
||||
|
||||
3. **Filter Node** (`n8n-nodes-base.filter`)
|
||||
- Uses `conditions` parameter
|
||||
- Same fixedCollection pattern
|
||||
|
||||
### 3. Why AI Agents Create Incorrect Structures
|
||||
|
||||
1. **Training Data Issues**: AI models may have been trained on outdated or incorrect n8n workflow examples
|
||||
2. **Nested Object Inference**: AI tends to create unnecessarily nested structures when it sees collection-type parameters
|
||||
3. **Legacy Format Confusion**: Mixing v2 and v3 Switch node formats
|
||||
4. **Schema Misinterpretation**: The term "fixedCollection" may lead AI to create object wrappers
|
||||
|
||||
## Current Impact
|
||||
|
||||
From issue #90 comments:
|
||||
- Multiple users experiencing the issue
|
||||
- Workflows fail to load completely (empty canvas)
|
||||
- Users resort to using Switch Node v3 or direct API calls
|
||||
- The issue appears in "most MCPs" according to user feedback
|
||||
|
||||
## Recommended Actions
|
||||
|
||||
### 1. Immediate Validation Enhancement
|
||||
|
||||
Add specific validation for fixedCollection properties in the workflow validator:
|
||||
|
||||
```typescript
|
||||
// In workflow-validator.ts or enhanced-config-validator.ts
|
||||
function validateFixedCollectionParameters(node, result) {
|
||||
const problematicNodes = {
|
||||
'n8n-nodes-base.switch': { version: 2, fields: ['rules'] },
|
||||
'n8n-nodes-base.if': { version: 1, fields: ['conditions'] },
|
||||
'n8n-nodes-base.filter': { version: 1, fields: ['conditions'] }
|
||||
};
|
||||
|
||||
const nodeConfig = problematicNodes[node.type];
|
||||
if (nodeConfig && node.typeVersion === nodeConfig.version) {
|
||||
// Validate structure
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 2. Enhanced MCP Tool Validation
|
||||
|
||||
Update the validation tools to detect and prevent this specific error pattern:
|
||||
|
||||
1. **In `validate_node_operation` tool**: Add checks for fixedCollection structures
|
||||
2. **In `validate_workflow` tool**: Include specific validation for Switch/If nodes
|
||||
3. **In `n8n_create_workflow` tool**: Pre-validate parameters before submission
|
||||
|
||||
### 3. AI-Friendly Examples
|
||||
|
||||
Update workflow examples to show correct structures:
|
||||
|
||||
```typescript
|
||||
// In workflow-examples.ts
|
||||
export const SWITCH_NODE_EXAMPLE = {
|
||||
name: "Switch",
|
||||
type: "n8n-nodes-base.switch",
|
||||
typeVersion: 3, // Prefer v3 over v2
|
||||
parameters: {
|
||||
// Correct v3 structure
|
||||
}
|
||||
};
|
||||
```
|
||||
|
||||
### 4. Migration Strategy
|
||||
|
||||
For existing workflows with Switch v2:
|
||||
1. Detect Switch v2 nodes in validation
|
||||
2. Suggest migration to v3
|
||||
3. Provide automatic conversion utility
|
||||
|
||||
### 5. Documentation Updates
|
||||
|
||||
1. Add warnings about fixedCollection structures in tool documentation
|
||||
2. Include specific examples of correct vs incorrect structures
|
||||
3. Document the Switch v2 to v3 migration path
|
||||
|
||||
## Proposed Implementation Priority
|
||||
|
||||
1. **High Priority**: Add validation to prevent creation of invalid structures
|
||||
2. **High Priority**: Update existing validation tools to catch this error
|
||||
3. **Medium Priority**: Add auto-fix capabilities to correct structures
|
||||
4. **Medium Priority**: Update examples and documentation
|
||||
5. **Low Priority**: Create migration utilities for v2 to v3
|
||||
|
||||
## Testing Strategy
|
||||
|
||||
1. Create test cases for each affected node type
|
||||
2. Test both correct and incorrect structures
|
||||
3. Verify validation catches all variants of the error
|
||||
4. Test auto-fix suggestions work correctly
|
||||
|
||||
## Success Metrics
|
||||
|
||||
- Zero instances of "propertyValues[itemName] is not iterable" in newly created workflows
|
||||
- Clear error messages that guide users to correct structures
|
||||
- Successful validation of all Switch/If node configurations before workflow creation
|
||||
|
||||
## Next Steps
|
||||
|
||||
1. Implement validation enhancements in the workflow validator
|
||||
2. Update MCP tools to include these validations
|
||||
3. Add comprehensive tests
|
||||
4. Update documentation with clear examples
|
||||
5. Consider adding a migration tool for existing workflows
|
||||
1213
docs/local/DEEP_DIVE_ANALYSIS_2025-10-02.md
Normal file
1213
docs/local/DEEP_DIVE_ANALYSIS_2025-10-02.md
Normal file
File diff suppressed because it is too large
Load Diff
225
docs/local/DEEP_DIVE_ANALYSIS_README.md
Normal file
225
docs/local/DEEP_DIVE_ANALYSIS_README.md
Normal file
@@ -0,0 +1,225 @@
|
||||
# N8N-MCP Deep Dive Analysis - October 2, 2025
|
||||
|
||||
## Overview
|
||||
|
||||
This directory contains a comprehensive deep-dive analysis of n8n-mcp usage data from September 26 - October 2, 2025.
|
||||
|
||||
**Data Volume Analyzed:**
|
||||
- 212,375 telemetry events
|
||||
- 5,751 workflow creations
|
||||
- 2,119 unique users
|
||||
- 6 days of usage data
|
||||
|
||||
## Report Structure
|
||||
|
||||
|
||||
###: `DEEP_DIVE_ANALYSIS_2025-10-02.md` (Main Report)
|
||||
|
||||
**Sections Covered:**
|
||||
1. **Executive Summary** - Key findings and recommendations
|
||||
2. **Tool Performance Analysis** - Success rates, performance metrics, critical findings
|
||||
3. **Validation Catastrophe** - The node type prefix disaster analysis
|
||||
4. **Usage Patterns & User Segmentation** - User distribution, daily trends
|
||||
5. **Tool Sequence Analysis** - How AI agents use tools together
|
||||
6. **Workflow Creation Patterns** - Complexity distribution, popular nodes
|
||||
7. **Platform & Version Distribution** - OS, architecture, version adoption
|
||||
8. **Error Patterns & Root Causes** - TypeErrors, validation errors, discovery failures
|
||||
9. **P0-P1 Refactoring Recommendations** - Detailed implementation guides
|
||||
|
||||
**Sections Covered:**
|
||||
- Remaining P1 and P2 recommendations
|
||||
- Architectural refactoring suggestions
|
||||
- Telemetry enhancements
|
||||
- CHANGELOG integration
|
||||
- Final recommendations summary
|
||||
|
||||
## Key Findings Summary
|
||||
|
||||
### Critical Issues (P0 - Fix Immediately)
|
||||
|
||||
1. **Node Type Prefix Validation Catastrophe**
|
||||
- 5,000+ validation errors from single root cause
|
||||
- `nodes-base.X` vs `n8n-nodes-base.X` confusion
|
||||
- **Solution**: Auto-normalize prefixes (2-4 hours effort)
|
||||
|
||||
2. **TypeError in Node Information Tools**
|
||||
- 10-18% failure rate in get_node_essentials/info
|
||||
- 1,000+ failures affecting hundreds of users
|
||||
- **Solution**: Complete null-safety audit (1 day effort)
|
||||
|
||||
3. **Task Discovery Failures**
|
||||
- `get_node_for_task` failing 28% of the time
|
||||
- Worst-performing tool in entire system
|
||||
- **Solution**: Expand task library + fuzzy matching (3 days effort)
|
||||
|
||||
### Performance Metrics
|
||||
|
||||
**Excellent Reliability (96-100% success):**
|
||||
- n8n_update_partial_workflow: 98.7%
|
||||
- search_nodes: 99.8%
|
||||
- n8n_create_workflow: 96.1%
|
||||
- All workflow management tools: 100%
|
||||
|
||||
**User Distribution:**
|
||||
- Power Users (12): 2,112 events/user, 33 workflows
|
||||
- Heavy Users (47): 673 events/user, 18 workflows
|
||||
- Regular Users (516): 199 events/user, 7 workflows (CORE AUDIENCE)
|
||||
- Active Users (919): 52 events/user, 2 workflows
|
||||
- Casual Users (625): 8 events/user, 1 workflow
|
||||
|
||||
### Usage Insights
|
||||
|
||||
**Most Used Tools:**
|
||||
1. n8n_update_partial_workflow: 10,177 calls (iterative refinement)
|
||||
2. search_nodes: 8,839 calls (node discovery)
|
||||
3. n8n_create_workflow: 6,046 calls (workflow creation)
|
||||
|
||||
**Most Common Tool Sequences:**
|
||||
1. update → update → update (549x) - Iterative refinement pattern
|
||||
2. create → update (297x) - Create then refine
|
||||
3. update → get_workflow (265x) - Update then verify
|
||||
|
||||
**Most Popular Nodes:**
|
||||
1. code (53% of workflows) - AI agents love programmatic control
|
||||
2. httpRequest (47%) - Integration-heavy usage
|
||||
3. webhook (32%) - Event-driven automation
|
||||
|
||||
## SQL Analytical Views Created
|
||||
|
||||
15 comprehensive views were created in Supabase for ongoing analysis:
|
||||
|
||||
1. `vw_tool_performance` - Performance metrics per tool
|
||||
2. `vw_error_analysis` - Error patterns and frequencies
|
||||
3. `vw_validation_analysis` - Validation failure details
|
||||
4. `vw_tool_sequences` - Tool-to-tool transition patterns
|
||||
5. `vw_workflow_creation_patterns` - Workflow characteristics
|
||||
6. `vw_node_usage_analysis` - Node popularity and complexity
|
||||
7. `vw_node_cooccurrence` - Which nodes are used together
|
||||
8. `vw_user_activity` - Per-user activity metrics
|
||||
9. `vw_session_analysis` - Platform/version distribution
|
||||
10. `vw_workflow_validation_failures` - Workflow validation issues
|
||||
11. `vw_temporal_patterns` - Time-based usage patterns
|
||||
12. `vw_tool_funnel` - User progression through tools
|
||||
13. `vw_search_analysis` - Search behavior
|
||||
14. `vw_tool_success_summary` - Success/failure rates
|
||||
15. `vw_user_journeys` - Complete user session reconstruction
|
||||
|
||||
## Priority Recommendations
|
||||
|
||||
### Immediate Actions (This Week)
|
||||
|
||||
✅ **P0-R1**: Auto-normalize node type prefixes → Eliminate 4,800 errors
|
||||
✅ **P0-R2**: Complete null-safety audit → Fix 10-18% TypeError failures
|
||||
✅ **P0-R3**: Expand get_node_for_task library → 72% → 95% success rate
|
||||
|
||||
**Expected Impact**: Reduce error rate from 5-10% to <2% overall
|
||||
|
||||
### Next Release (2-3 Weeks)
|
||||
|
||||
✅ **P1-R4**: Batch workflow operations → Save 30-50% tokens
|
||||
✅ **P1-R5**: Proactive node suggestions → Reduce search iterations
|
||||
✅ **P1-R6**: Auto-fix suggestions in errors → Self-service recovery
|
||||
|
||||
**Expected Impact**: 40% faster workflow creation, better UX
|
||||
|
||||
### Future Roadmap (1-3 Months)
|
||||
|
||||
✅ **A1**: Service layer consolidation → Cleaner architecture
|
||||
✅ **A2**: Repository caching → 50% faster node operations
|
||||
✅ **R10**: Workflow template library from usage → 80% coverage
|
||||
✅ **T1-T3**: Enhanced telemetry → Better observability
|
||||
|
||||
**Expected Impact**: Scalable foundation for 10x growth
|
||||
|
||||
## Methodology
|
||||
|
||||
### Data Sources
|
||||
|
||||
1. **Supabase Telemetry Database**
|
||||
- `telemetry_events` table: 212,375 rows
|
||||
- `telemetry_workflows` table: 5,751 rows
|
||||
|
||||
2. **Analytical Views**
|
||||
- Created 15 SQL views for multi-dimensional analysis
|
||||
- Enabled complex queries and pattern recognition
|
||||
|
||||
3. **CHANGELOG Review**
|
||||
- Analyzed recent changes (v2.14.0 - v2.14.6)
|
||||
- Correlated fixes with error patterns
|
||||
|
||||
### Analysis Approach
|
||||
|
||||
1. **Quantitative Analysis**
|
||||
- Success/failure rates per tool
|
||||
- Performance metrics (avg, median, p95, p99)
|
||||
- User segmentation and cohort analysis
|
||||
- Temporal trends and growth patterns
|
||||
|
||||
2. **Pattern Recognition**
|
||||
- Tool sequence analysis (Markov chains)
|
||||
- Node co-occurrence patterns
|
||||
- Workflow complexity distribution
|
||||
- Error clustering and root cause analysis
|
||||
|
||||
3. **Qualitative Insights**
|
||||
- CHANGELOG integration
|
||||
- Error message analysis
|
||||
- User journey reconstruction
|
||||
- Best practice identification
|
||||
|
||||
## How to Use This Analysis
|
||||
|
||||
### For Development Priorities
|
||||
|
||||
1. Review **P0 Critical Recommendations** (Section 8)
|
||||
2. Check estimated effort and impact
|
||||
3. Prioritize based on ROI (impact/effort ratio)
|
||||
4. Follow implementation guides with code examples
|
||||
|
||||
### For Architecture Decisions
|
||||
|
||||
1. Review **Architectural Recommendations** (Section 9)
|
||||
2. Consider service layer consolidation
|
||||
3. Evaluate repository caching opportunities
|
||||
4. Plan for 10x scale
|
||||
|
||||
### For Product Strategy
|
||||
|
||||
1. Review **Usage Patterns** (Section 3 & 5)
|
||||
2. Understand user segments (power vs casual)
|
||||
3. Identify high-value features (most-used tools)
|
||||
4. Focus on reliability over features (96% success rate target)
|
||||
|
||||
### For Telemetry Enhancement
|
||||
|
||||
1. Review **Telemetry Enhancements** (Section 10)
|
||||
2. Add fine-grained timing metrics
|
||||
3. Track workflow creation funnels
|
||||
4. Monitor node-level analytics
|
||||
|
||||
## Contact & Feedback
|
||||
|
||||
For questions about this analysis or to request additional insights:
|
||||
- Data Analyst: Claude Code with Supabase MCP
|
||||
- Analysis Date: October 2, 2025
|
||||
- Data Period: September 26 - October 2, 2025
|
||||
|
||||
## Change Log
|
||||
|
||||
- **2025-10-02**: Initial comprehensive analysis completed
|
||||
- 15 SQL analytical views created
|
||||
- 13 sections of detailed findings
|
||||
- P0/P1/P2 recommendations with implementation guides
|
||||
- Code examples and effort estimates provided
|
||||
|
||||
## Next Steps
|
||||
|
||||
1. ✅ Review findings with development team
|
||||
2. ✅ Prioritize P0 recommendations for immediate implementation
|
||||
3. ✅ Plan P1 features for next release cycle
|
||||
4. ✅ Set up monitoring for key metrics
|
||||
5. ✅ Schedule follow-up analysis (weekly recommended)
|
||||
|
||||
---
|
||||
|
||||
*This analysis represents a snapshot of n8n-mcp usage during early adoption phase. Patterns may evolve as the user base grows and matures.*
|
||||
1328
docs/local/Deep_dive_p1_p2.md
Normal file
1328
docs/local/Deep_dive_p1_p2.md
Normal file
File diff suppressed because it is too large
Load Diff
3396
docs/local/N8N_AI_WORKFLOW_BUILDER_ANALYSIS.md
Normal file
3396
docs/local/N8N_AI_WORKFLOW_BUILDER_ANALYSIS.md
Normal file
File diff suppressed because it is too large
Load Diff
1489
docs/local/P0_IMPLEMENTATION_PLAN.md
Normal file
1489
docs/local/P0_IMPLEMENTATION_PLAN.md
Normal file
File diff suppressed because it is too large
Load Diff
369
docs/local/TEMPLATE_MINING_ANALYSIS.md
Normal file
369
docs/local/TEMPLATE_MINING_ANALYSIS.md
Normal file
@@ -0,0 +1,369 @@
|
||||
# Template Mining Analysis - Alternative to P0-R3
|
||||
|
||||
**Date**: 2025-10-02
|
||||
**Context**: Analyzing whether to fix `get_node_for_task` (28% failure rate) or replace it with template-based configuration extraction
|
||||
|
||||
## Executive Summary
|
||||
|
||||
**RECOMMENDATION**: Replace `get_node_for_task` with template-based configuration extraction. The template database contains 2,646 real-world workflows with rich node configurations that far exceed the 31 hardcoded task templates.
|
||||
|
||||
## Key Findings
|
||||
|
||||
### 1. Template Database Coverage
|
||||
|
||||
- **Total Templates**: 2,646 production workflows from n8n.io
|
||||
- **Unique Node Types**: 543 (covers 103% of our 525 core nodes)
|
||||
- **Metadata Coverage**: 100% (AI-generated structured metadata)
|
||||
|
||||
### 2. Node Type Coverage in Templates
|
||||
|
||||
Top node types by template usage:
|
||||
```
|
||||
3,820 templates: n8n-nodes-base.httpRequest (144% of total templates!)
|
||||
3,678 templates: n8n-nodes-base.set
|
||||
2,445 templates: n8n-nodes-base.code
|
||||
1,700 templates: n8n-nodes-base.googleSheets
|
||||
1,471 templates: @n8n/n8n-nodes-langchain.agent
|
||||
1,269 templates: @n8n/n8n-nodes-langchain.lmChatOpenAi
|
||||
792 templates: n8n-nodes-base.telegram
|
||||
702 templates: n8n-nodes-base.httpRequestTool
|
||||
596 templates: n8n-nodes-base.gmail
|
||||
466 templates: n8n-nodes-base.webhook
|
||||
```
|
||||
|
||||
**Comparison**:
|
||||
- Hardcoded task templates: 31 tasks covering 5.9% of nodes
|
||||
- Real templates: 2,646 templates with 2-3k examples for common nodes
|
||||
|
||||
### 3. Database Structure
|
||||
|
||||
```sql
|
||||
CREATE TABLE templates (
|
||||
id INTEGER PRIMARY KEY,
|
||||
workflow_id INTEGER UNIQUE NOT NULL,
|
||||
name TEXT NOT NULL,
|
||||
description TEXT,
|
||||
-- Node information
|
||||
nodes_used TEXT, -- JSON array: ["n8n-nodes-base.httpRequest", ...]
|
||||
workflow_json_compressed TEXT, -- Base64 encoded gzip of full workflow
|
||||
-- Metadata (100% coverage)
|
||||
metadata_json TEXT, -- AI-generated structured metadata
|
||||
-- Stats
|
||||
views INTEGER DEFAULT 0,
|
||||
created_at DATETIME,
|
||||
-- ...
|
||||
);
|
||||
```
|
||||
|
||||
### 4. Real Configuration Examples
|
||||
|
||||
#### HTTP Request Node Configurations
|
||||
|
||||
**Simple URL fetch**:
|
||||
```json
|
||||
{
|
||||
"url": "https://api.example.com/data",
|
||||
"options": {}
|
||||
}
|
||||
```
|
||||
|
||||
**With authentication**:
|
||||
```json
|
||||
{
|
||||
"url": "=https://api.wavespeed.ai/api/v3/predictions/{{ $json.data.id }}/result",
|
||||
"options": {},
|
||||
"authentication": "genericCredentialType",
|
||||
"genericAuthType": "httpHeaderAuth"
|
||||
}
|
||||
```
|
||||
|
||||
**Complex expressions**:
|
||||
```json
|
||||
{
|
||||
"url": "=https://image.pollinations.ai/prompt/{{$('Social Media Content Factory').item.json.output.description.replaceAll(' ','-').replaceAll(',','').replaceAll('.','') }}",
|
||||
"options": {}
|
||||
}
|
||||
```
|
||||
|
||||
#### Webhook Node Configurations
|
||||
|
||||
**Basic webhook**:
|
||||
```json
|
||||
{
|
||||
"path": "ytube",
|
||||
"options": {},
|
||||
"httpMethod": "POST",
|
||||
"responseMode": "responseNode"
|
||||
}
|
||||
```
|
||||
|
||||
**With binary data**:
|
||||
```json
|
||||
{
|
||||
"path": "your-endpoint",
|
||||
"options": {
|
||||
"binaryPropertyName": "data"
|
||||
},
|
||||
"httpMethod": "POST"
|
||||
}
|
||||
```
|
||||
|
||||
### 5. AI-Generated Metadata
|
||||
|
||||
Each template has structured metadata including:
|
||||
|
||||
```json
|
||||
{
|
||||
"categories": ["automation", "integration", "data processing"],
|
||||
"complexity": "medium",
|
||||
"use_cases": [
|
||||
"Extract transaction data from Gmail",
|
||||
"Automate bookkeeping",
|
||||
"Expense tracking"
|
||||
],
|
||||
"estimated_setup_minutes": 30,
|
||||
"required_services": ["Gmail", "Google Sheets", "Google Gemini"],
|
||||
"key_features": [
|
||||
"Fetch emails by label",
|
||||
"Extract transaction data",
|
||||
"Use LLM for structured output"
|
||||
],
|
||||
"target_audience": ["Accountants", "Small business owners"]
|
||||
}
|
||||
```
|
||||
|
||||
## Comparison: Task Templates vs Real Templates
|
||||
|
||||
### Current Approach (get_node_for_task)
|
||||
|
||||
**Pros**:
|
||||
- Curated configurations with best practices
|
||||
- Predictable, stable responses
|
||||
- Fast lookup (no decompression needed)
|
||||
|
||||
**Cons**:
|
||||
- Only 31 tasks (5.9% node coverage)
|
||||
- 28% failure rate (users can't find what they need)
|
||||
- Requires manual maintenance
|
||||
- Static configurations without real-world context
|
||||
- Usage ratio 22.5:1 (search_nodes is preferred)
|
||||
|
||||
### Template-Based Approach
|
||||
|
||||
**Pros**:
|
||||
- 2,646 real workflows with 2-3k examples for common nodes
|
||||
- 100% metadata coverage for semantic matching
|
||||
- Real-world patterns and best practices
|
||||
- Covers 543 node types (103% coverage)
|
||||
- Self-updating (templates fetched from n8n.io)
|
||||
- Rich context (use cases, complexity, setup time)
|
||||
|
||||
**Cons**:
|
||||
- Requires decompression for full workflow access
|
||||
- May contain template-specific context (but can be filtered)
|
||||
- Need ranking/filtering logic for best matches
|
||||
|
||||
## Proposed Implementation Strategy
|
||||
|
||||
### Phase 1: Extract Node Configurations from Templates
|
||||
|
||||
Create a new service: `TemplateConfigExtractor`
|
||||
|
||||
```typescript
|
||||
interface ExtractedNodeConfig {
|
||||
nodeType: string;
|
||||
configuration: Record<string, any>;
|
||||
source: {
|
||||
templateId: number;
|
||||
templateName: string;
|
||||
templateViews: number;
|
||||
useCases: string[];
|
||||
complexity: 'simple' | 'medium' | 'complex';
|
||||
};
|
||||
patterns: {
|
||||
hasAuthentication: boolean;
|
||||
hasExpressions: boolean;
|
||||
hasOptionalFields: boolean;
|
||||
};
|
||||
}
|
||||
|
||||
class TemplateConfigExtractor {
|
||||
async extractConfigsForNode(
|
||||
nodeType: string,
|
||||
options?: {
|
||||
complexity?: 'simple' | 'medium' | 'complex';
|
||||
requiresAuth?: boolean;
|
||||
limit?: number;
|
||||
}
|
||||
): Promise<ExtractedNodeConfig[]> {
|
||||
// 1. Query templates containing nodeType
|
||||
// 2. Decompress workflow_json_compressed
|
||||
// 3. Extract node configurations
|
||||
// 4. Rank by popularity + complexity match
|
||||
// 5. Return top N configurations
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Phase 2: Integrate with Existing Tools
|
||||
|
||||
**Option A**: Enhance `get_node_essentials`
|
||||
- Add `includeExamples: boolean` parameter
|
||||
- Return 2-3 real configurations from templates
|
||||
- Preserve existing compact format
|
||||
|
||||
**Option B**: Enhance `get_node_info`
|
||||
- Add `examples` section with template-sourced configs
|
||||
- Include source attribution (template name, views)
|
||||
|
||||
**Option C**: New tool `get_node_examples`
|
||||
- Dedicated tool for retrieving configuration examples
|
||||
- Query by node type, complexity, use case
|
||||
- Returns ranked list of real configurations
|
||||
|
||||
### Phase 3: Deprecate get_node_for_task
|
||||
|
||||
- Mark as deprecated in tool documentation
|
||||
- Redirect to enhanced tools
|
||||
- Remove after 2-3 version cycles
|
||||
|
||||
## Performance Considerations
|
||||
|
||||
### Decompression Cost
|
||||
|
||||
- Average compressed size: 6-12 KB
|
||||
- Decompression time: ~5-10ms per template
|
||||
- Caching strategy needed for frequently accessed templates
|
||||
|
||||
### Query Strategy
|
||||
|
||||
```sql
|
||||
-- Fast: Get templates for a node type (no decompression)
|
||||
SELECT id, name, views, metadata_json
|
||||
FROM templates
|
||||
WHERE nodes_used LIKE '%n8n-nodes-base.httpRequest%'
|
||||
ORDER BY views DESC
|
||||
LIMIT 10;
|
||||
|
||||
-- Then decompress only top matches
|
||||
```
|
||||
|
||||
### Caching
|
||||
|
||||
- Cache decompressed workflows for popular templates (top 100)
|
||||
- TTL: 1 hour
|
||||
- Estimated memory: 100 * 50KB = 5MB
|
||||
|
||||
## Impact on P0-R3
|
||||
|
||||
**Original P0-R3 Plan**: Expand task library from 31 to 100+ tasks using fuzzy matching
|
||||
|
||||
**New Approach**: Mine 2,646 templates for real configurations
|
||||
|
||||
**Impact Assessment**:
|
||||
|
||||
| Metric | Original Plan | Template Mining |
|
||||
|--------|--------------|-----------------|
|
||||
| Configuration examples | 100 (estimated) | 2,646+ actual |
|
||||
| Node coverage | ~20% | 103% |
|
||||
| Maintenance | High (manual) | Low (auto-fetch) |
|
||||
| Accuracy | Curated | Production-tested |
|
||||
| Context richness | Limited | Rich metadata |
|
||||
| Development time | 2-3 weeks | 1 week |
|
||||
|
||||
**Recommendation**: PIVOT to template mining approach for P0-R3
|
||||
|
||||
## Implementation Estimate
|
||||
|
||||
### Week 1: Core Infrastructure
|
||||
- Day 1-2: Create `TemplateConfigExtractor` service
|
||||
- Day 3: Implement caching layer
|
||||
- Day 4-5: Testing and optimization
|
||||
|
||||
### Week 2: Integration
|
||||
- Day 1-2: Enhance `get_node_essentials` with examples
|
||||
- Day 3: Update tool documentation
|
||||
- Day 4-5: Integration testing
|
||||
|
||||
**Total**: 2 weeks vs 3 weeks for original plan
|
||||
|
||||
## Validation Tests
|
||||
|
||||
```typescript
|
||||
// Test: Extract HTTP Request configs
|
||||
const configs = await extractor.extractConfigsForNode(
|
||||
'n8n-nodes-base.httpRequest',
|
||||
{ complexity: 'simple', limit: 5 }
|
||||
);
|
||||
|
||||
// Expected: 5 configs from top templates
|
||||
// - Simple URL fetch
|
||||
// - With authentication
|
||||
// - With custom headers
|
||||
// - With expressions
|
||||
// - With error handling
|
||||
|
||||
// Test: Extract webhook configs
|
||||
const webhookConfigs = await extractor.extractConfigsForNode(
|
||||
'n8n-nodes-base.webhook',
|
||||
{ limit: 3 }
|
||||
);
|
||||
|
||||
// Expected: 3 configs showing different patterns
|
||||
// - Basic POST webhook
|
||||
// - With response node
|
||||
// - With binary data handling
|
||||
```
|
||||
|
||||
## Risks and Mitigation
|
||||
|
||||
### Risk 1: Template Quality Varies
|
||||
- **Mitigation**: Filter by views (popularity) and metadata complexity rating
|
||||
- Only use templates with >1000 views for examples
|
||||
|
||||
### Risk 2: Decompression Performance
|
||||
- **Mitigation**: Cache decompressed popular templates
|
||||
- Implement lazy loading (decompress on demand)
|
||||
|
||||
### Risk 3: Template-Specific Context
|
||||
- **Mitigation**: Extract only node configuration, strip workflow-specific context
|
||||
- Provide source attribution for context
|
||||
|
||||
### Risk 4: Breaking Changes in Template Structure
|
||||
- **Mitigation**: Robust error handling in decompression
|
||||
- Fallback to cached configs if template fetch fails
|
||||
|
||||
## Success Metrics
|
||||
|
||||
**Before** (get_node_for_task):
|
||||
- 392 calls, 72% success rate
|
||||
- 28% failure rate
|
||||
- 31 task templates
|
||||
- 5.9% node coverage
|
||||
|
||||
**Target** (template-based):
|
||||
- 90%+ success rate for configuration discovery
|
||||
- 100%+ node coverage
|
||||
- 2,646+ real-world examples
|
||||
- Self-updating from n8n.io
|
||||
|
||||
## Next Steps
|
||||
|
||||
1. ✅ Complete template database analysis
|
||||
2. ⏳ Create `TemplateConfigExtractor` service
|
||||
3. ⏳ Implement caching layer
|
||||
4. ⏳ Enhance `get_node_essentials` with examples
|
||||
5. ⏳ Update P0 implementation plan
|
||||
6. ⏳ Begin implementation
|
||||
|
||||
## Conclusion
|
||||
|
||||
The template database provides a vastly superior alternative to hardcoded task templates:
|
||||
|
||||
- **2,646 templates** vs 31 tasks (85x more examples)
|
||||
- **103% node coverage** vs 5.9% coverage (17x improvement)
|
||||
- **Real-world configurations** vs synthetic examples
|
||||
- **Self-updating** vs manual maintenance
|
||||
- **Rich metadata** for semantic matching
|
||||
|
||||
**Recommendation**: Pivot P0-R3 from "expand task library" to "mine template configurations"
|
||||
1306
docs/local/integration-testing-plan.md
Normal file
1306
docs/local/integration-testing-plan.md
Normal file
File diff suppressed because it is too large
Load Diff
260
docs/local/integration-tests-phase1-summary.md
Normal file
260
docs/local/integration-tests-phase1-summary.md
Normal file
@@ -0,0 +1,260 @@
|
||||
# Integration Tests Phase 1: Foundation - COMPLETED
|
||||
|
||||
## Overview
|
||||
Phase 1 establishes the foundation for n8n API integration testing. All core utilities, fixtures, and infrastructure are now in place.
|
||||
|
||||
## Branch
|
||||
`feat/integration-tests-foundation`
|
||||
|
||||
## Completed Tasks
|
||||
|
||||
### 1. Environment Configuration
|
||||
- ✅ Updated `.env.example` with integration testing configuration
|
||||
- ✅ Added environment variables for:
|
||||
- n8n API credentials (`N8N_API_URL`, `N8N_API_KEY`)
|
||||
- Webhook workflow IDs (4 workflows for GET/POST/PUT/DELETE)
|
||||
- Test configuration (cleanup, tags, naming)
|
||||
- ✅ Included detailed setup instructions in comments
|
||||
|
||||
### 2. Directory Structure
|
||||
```
|
||||
tests/integration/n8n-api/
|
||||
├── workflows/ (empty - for Phase 2+)
|
||||
├── executions/ (empty - for Phase 2+)
|
||||
├── system/ (empty - for Phase 2+)
|
||||
├── scripts/
|
||||
│ └── cleanup-orphans.ts
|
||||
└── utils/
|
||||
├── credentials.ts
|
||||
├── n8n-client.ts
|
||||
├── test-context.ts
|
||||
├── cleanup-helpers.ts
|
||||
├── fixtures.ts
|
||||
├── factories.ts
|
||||
└── webhook-workflows.ts
|
||||
```
|
||||
|
||||
### 3. Core Utilities
|
||||
|
||||
#### `credentials.ts` (200 lines)
|
||||
- Environment-aware credential loading
|
||||
- Detects CI vs local environment automatically
|
||||
- Validation functions with helpful error messages
|
||||
- Non-throwing credential check functions
|
||||
|
||||
**Key Functions:**
|
||||
- `getN8nCredentials()` - Load credentials from .env or GitHub secrets
|
||||
- `validateCredentials()` - Ensure required credentials are present
|
||||
- `validateWebhookWorkflows()` - Check webhook workflow IDs with setup instructions
|
||||
- `hasCredentials()` - Non-throwing credential check
|
||||
- `hasWebhookWorkflows()` - Non-throwing webhook check
|
||||
|
||||
#### `n8n-client.ts` (45 lines)
|
||||
- Singleton n8n API client wrapper
|
||||
- Pre-configured with test credentials
|
||||
- Health check functionality
|
||||
|
||||
**Key Functions:**
|
||||
- `getTestN8nClient()` - Get/create configured API client
|
||||
- `resetTestN8nClient()` - Reset client instance
|
||||
- `isN8nApiAccessible()` - Check API connectivity
|
||||
|
||||
#### `test-context.ts` (120 lines)
|
||||
- Resource tracking for automatic cleanup
|
||||
- Test workflow naming utilities
|
||||
- Tag management
|
||||
|
||||
**Key Functions:**
|
||||
- `createTestContext()` - Create context for tracking resources
|
||||
- `TestContext.trackWorkflow()` - Track workflow for cleanup
|
||||
- `TestContext.trackExecution()` - Track execution for cleanup
|
||||
- `TestContext.cleanup()` - Delete all tracked resources
|
||||
- `createTestWorkflowName()` - Generate unique workflow names
|
||||
- `getTestTag()` - Get configured test tag
|
||||
|
||||
#### `cleanup-helpers.ts` (275 lines)
|
||||
- Multi-level cleanup strategies
|
||||
- Orphaned resource detection
|
||||
- Age-based execution cleanup
|
||||
- Tag-based workflow cleanup
|
||||
|
||||
**Key Functions:**
|
||||
- `cleanupOrphanedWorkflows()` - Find and delete test workflows
|
||||
- `cleanupOldExecutions()` - Delete executions older than X hours
|
||||
- `cleanupAllTestResources()` - Comprehensive cleanup
|
||||
- `cleanupWorkflowsByTag()` - Delete workflows by tag
|
||||
- `cleanupExecutionsByWorkflow()` - Delete workflow's executions
|
||||
|
||||
#### `fixtures.ts` (310 lines)
|
||||
- Pre-built workflow templates
|
||||
- All using FULL node type format (n8n-nodes-base.*)
|
||||
|
||||
**Available Fixtures:**
|
||||
- `SIMPLE_WEBHOOK_WORKFLOW` - Single webhook node
|
||||
- `SIMPLE_HTTP_WORKFLOW` - Webhook + HTTP Request
|
||||
- `MULTI_NODE_WORKFLOW` - Complex branching workflow
|
||||
- `ERROR_HANDLING_WORKFLOW` - Error output configuration
|
||||
- `AI_AGENT_WORKFLOW` - Langchain agent node
|
||||
- `EXPRESSION_WORKFLOW` - n8n expressions testing
|
||||
|
||||
**Helper Functions:**
|
||||
- `getFixture()` - Get fixture by name (with deep clone)
|
||||
- `createCustomWorkflow()` - Build custom workflow from nodes
|
||||
|
||||
#### `factories.ts` (315 lines)
|
||||
- Dynamic test data generation
|
||||
- Node builders with sensible defaults
|
||||
- Workflow composition helpers
|
||||
|
||||
**Node Factories:**
|
||||
- `createWebhookNode()` - Webhook node with customization
|
||||
- `createHttpRequestNode()` - HTTP Request node
|
||||
- `createSetNode()` - Set node with assignments
|
||||
- `createManualTriggerNode()` - Manual trigger node
|
||||
|
||||
**Connection Factories:**
|
||||
- `createConnection()` - Simple node connection
|
||||
- `createSequentialWorkflow()` - Auto-connected sequential nodes
|
||||
- `createParallelWorkflow()` - Trigger with parallel branches
|
||||
- `createErrorHandlingWorkflow()` - Workflow with error handling
|
||||
|
||||
**Utilities:**
|
||||
- `randomString()` - Generate random test data
|
||||
- `uniqueId()` - Unique IDs for testing
|
||||
- `createTestTags()` - Test workflow tags
|
||||
- `createWorkflowSettings()` - Common settings
|
||||
|
||||
#### `webhook-workflows.ts` (215 lines)
|
||||
- Webhook workflow configuration templates
|
||||
- Setup instructions generator
|
||||
- URL generation utilities
|
||||
|
||||
**Key Features:**
|
||||
- `WEBHOOK_WORKFLOW_CONFIGS` - Configurations for all 4 HTTP methods
|
||||
- `printSetupInstructions()` - Print detailed setup guide
|
||||
- `generateWebhookWorkflowJson()` - Generate workflow JSON
|
||||
- `exportAllWebhookWorkflows()` - Export all 4 configs
|
||||
- `getWebhookUrl()` - Get webhook URL for testing
|
||||
- `isValidWebhookWorkflow()` - Validate workflow structure
|
||||
|
||||
### 4. Scripts
|
||||
|
||||
#### `cleanup-orphans.ts` (40 lines)
|
||||
- Standalone cleanup script
|
||||
- Can be run manually or in CI
|
||||
- Comprehensive output logging
|
||||
|
||||
**Usage:**
|
||||
```bash
|
||||
npm run test:cleanup:orphans
|
||||
```
|
||||
|
||||
### 5. npm Scripts
|
||||
Added to `package.json`:
|
||||
```json
|
||||
{
|
||||
"test:integration:n8n": "vitest run tests/integration/n8n-api",
|
||||
"test:cleanup:orphans": "tsx tests/integration/n8n-api/scripts/cleanup-orphans.ts"
|
||||
}
|
||||
```
|
||||
|
||||
## Code Quality
|
||||
|
||||
### TypeScript
|
||||
- ✅ All code passes `npm run typecheck`
|
||||
- ✅ All code compiles with `npm run build`
|
||||
- ✅ No TypeScript errors
|
||||
- ✅ Proper type annotations throughout
|
||||
|
||||
### Error Handling
|
||||
- ✅ Comprehensive error messages
|
||||
- ✅ Helpful setup instructions in error messages
|
||||
- ✅ Non-throwing validation functions where appropriate
|
||||
- ✅ Graceful handling of missing credentials
|
||||
|
||||
### Documentation
|
||||
- ✅ All functions have JSDoc comments
|
||||
- ✅ Usage examples in comments
|
||||
- ✅ Clear parameter descriptions
|
||||
- ✅ Return type documentation
|
||||
|
||||
## Files Created
|
||||
|
||||
### Documentation
|
||||
1. `docs/local/integration-testing-plan.md` (550 lines)
|
||||
2. `docs/local/integration-tests-phase1-summary.md` (this file)
|
||||
|
||||
### Code
|
||||
1. `.env.example` - Updated with test configuration (32 new lines)
|
||||
2. `package.json` - Added 2 npm scripts
|
||||
3. `tests/integration/n8n-api/utils/credentials.ts` (200 lines)
|
||||
4. `tests/integration/n8n-api/utils/n8n-client.ts` (45 lines)
|
||||
5. `tests/integration/n8n-api/utils/test-context.ts` (120 lines)
|
||||
6. `tests/integration/n8n-api/utils/cleanup-helpers.ts` (275 lines)
|
||||
7. `tests/integration/n8n-api/utils/fixtures.ts` (310 lines)
|
||||
8. `tests/integration/n8n-api/utils/factories.ts` (315 lines)
|
||||
9. `tests/integration/n8n-api/utils/webhook-workflows.ts` (215 lines)
|
||||
10. `tests/integration/n8n-api/scripts/cleanup-orphans.ts` (40 lines)
|
||||
|
||||
**Total New Code:** ~1,520 lines of production-ready TypeScript
|
||||
|
||||
## Next Steps (Phase 2)
|
||||
|
||||
Phase 2 will implement the first actual integration tests:
|
||||
- Create workflow creation tests (10+ scenarios)
|
||||
- Test P0 bug fix (SHORT vs FULL node types)
|
||||
- Test workflow retrieval
|
||||
- Test workflow deletion
|
||||
|
||||
**Branch:** `feat/integration-tests-workflow-creation`
|
||||
|
||||
## Prerequisites for Running Tests
|
||||
|
||||
Before running integration tests, you need to:
|
||||
|
||||
1. **Set up n8n instance:**
|
||||
- Local: `npx n8n start`
|
||||
- Or use cloud/self-hosted n8n
|
||||
|
||||
2. **Configure credentials in `.env`:**
|
||||
```bash
|
||||
N8N_API_URL=http://localhost:5678
|
||||
N8N_API_KEY=<your-api-key>
|
||||
```
|
||||
|
||||
3. **Create 4 webhook workflows manually:**
|
||||
- One for each HTTP method (GET, POST, PUT, DELETE)
|
||||
- Activate each workflow in n8n UI
|
||||
- Set workflow IDs in `.env`:
|
||||
```bash
|
||||
N8N_TEST_WEBHOOK_GET_ID=<workflow-id>
|
||||
N8N_TEST_WEBHOOK_POST_ID=<workflow-id>
|
||||
N8N_TEST_WEBHOOK_PUT_ID=<workflow-id>
|
||||
N8N_TEST_WEBHOOK_DELETE_ID=<workflow-id>
|
||||
```
|
||||
|
||||
See `docs/local/integration-testing-plan.md` for detailed setup instructions.
|
||||
|
||||
## Success Metrics
|
||||
|
||||
Phase 1 Success Criteria - ALL MET:
|
||||
- ✅ All utilities implemented and tested
|
||||
- ✅ TypeScript compiles without errors
|
||||
- ✅ Code follows project conventions
|
||||
- ✅ Comprehensive documentation
|
||||
- ✅ Environment configuration complete
|
||||
- ✅ Cleanup infrastructure in place
|
||||
- ✅ Ready for Phase 2 test implementation
|
||||
|
||||
## Lessons Learned
|
||||
|
||||
1. **N8nApiClient Constructor:** Uses config object, not separate parameters
|
||||
2. **Cursor Handling:** n8n API returns `null` for no more pages, need to convert to `undefined`
|
||||
3. **Workflow ID Validation:** Some workflows might have undefined IDs, need null checks
|
||||
4. **Connection Types:** Error connections need explicit typing to avoid TypeScript errors
|
||||
5. **Webhook Activation:** Cannot be done via API, must be manual - hence pre-activated workflow requirement
|
||||
|
||||
## Time Invested
|
||||
|
||||
Phase 1 actual time: ~2 hours (estimated 2-3 days in plan)
|
||||
- Faster than expected due to clear architecture and reusable patterns
|
||||
@@ -1,712 +0,0 @@
|
||||
# MCP Tools Documentation for LLMs
|
||||
|
||||
This document provides comprehensive documentation for the most commonly used MCP tools in the n8n-mcp server. Each tool includes parameters, return formats, examples, and best practices.
|
||||
|
||||
## Table of Contents
|
||||
1. [search_nodes](#search_nodes)
|
||||
2. [get_node_essentials](#get_node_essentials)
|
||||
3. [list_nodes](#list_nodes)
|
||||
4. [validate_node_minimal](#validate_node_minimal)
|
||||
5. [validate_node_operation](#validate_node_operation)
|
||||
6. [get_node_for_task](#get_node_for_task)
|
||||
7. [n8n_create_workflow](#n8n_create_workflow)
|
||||
8. [n8n_update_partial_workflow](#n8n_update_partial_workflow)
|
||||
|
||||
---
|
||||
|
||||
## search_nodes
|
||||
|
||||
**Brief Description**: Search for n8n nodes by keywords in names and descriptions.
|
||||
|
||||
### Parameters
|
||||
- `query` (string, required): Search term - single word recommended for best results
|
||||
- `limit` (number, optional): Maximum results to return (default: 20)
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"nodes": [
|
||||
{
|
||||
"nodeType": "nodes-base.slack",
|
||||
"displayName": "Slack",
|
||||
"description": "Send messages to Slack channels"
|
||||
}
|
||||
],
|
||||
"totalFound": 5
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Finding integration nodes**: `search_nodes("slack")` to find Slack integration
|
||||
2. **Finding HTTP nodes**: `search_nodes("http")` for HTTP/webhook nodes
|
||||
3. **Finding database nodes**: `search_nodes("postgres")` for PostgreSQL nodes
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Search for Slack-related nodes
|
||||
{
|
||||
"query": "slack",
|
||||
"limit": 10
|
||||
}
|
||||
|
||||
// Search for webhook nodes
|
||||
{
|
||||
"query": "webhook",
|
||||
"limit": 20
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- Fast operation (cached results)
|
||||
- Single-word queries are more precise
|
||||
- Returns results with OR logic (any word matches)
|
||||
|
||||
### Best Practices
|
||||
- Use single words for precise results: "slack" not "send slack message"
|
||||
- Try shorter terms if no results: "sheet" instead of "spreadsheet"
|
||||
- Search is case-insensitive
|
||||
- Common searches: "http", "webhook", "email", "database", "slack"
|
||||
|
||||
### Common Pitfalls
|
||||
- Multi-word searches return too many results (OR logic)
|
||||
- Searching for exact phrases doesn't work
|
||||
- Node types aren't searchable here (use exact type with get_node_info)
|
||||
|
||||
### Related Tools
|
||||
- `list_nodes` - Browse nodes by category
|
||||
- `get_node_essentials` - Get node configuration after finding it
|
||||
- `list_ai_tools` - Find AI-capable nodes specifically
|
||||
|
||||
---
|
||||
|
||||
## get_node_essentials
|
||||
|
||||
**Brief Description**: Get only the 10-20 most important properties for a node with working examples.
|
||||
|
||||
### Parameters
|
||||
- `nodeType` (string, required): Full node type with prefix (e.g., "nodes-base.httpRequest")
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"nodeType": "nodes-base.httpRequest",
|
||||
"displayName": "HTTP Request",
|
||||
"essentialProperties": [
|
||||
{
|
||||
"name": "method",
|
||||
"type": "options",
|
||||
"default": "GET",
|
||||
"options": ["GET", "POST", "PUT", "DELETE"],
|
||||
"required": true
|
||||
},
|
||||
{
|
||||
"name": "url",
|
||||
"type": "string",
|
||||
"required": true,
|
||||
"placeholder": "https://api.example.com/endpoint"
|
||||
}
|
||||
],
|
||||
"examples": [
|
||||
{
|
||||
"name": "Simple GET Request",
|
||||
"configuration": {
|
||||
"method": "GET",
|
||||
"url": "https://api.example.com/users"
|
||||
}
|
||||
}
|
||||
],
|
||||
"tips": [
|
||||
"Use expressions like {{$json.url}} to make URLs dynamic",
|
||||
"Enable 'Split Into Items' for array responses"
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Quick node configuration**: Get just what you need without parsing 100KB+ of data
|
||||
2. **Learning node basics**: Understand essential properties with examples
|
||||
3. **Building workflows efficiently**: 95% smaller responses than get_node_info
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Get essentials for HTTP Request node
|
||||
{
|
||||
"nodeType": "nodes-base.httpRequest"
|
||||
}
|
||||
|
||||
// Get essentials for Slack node
|
||||
{
|
||||
"nodeType": "nodes-base.slack"
|
||||
}
|
||||
|
||||
// Get essentials for OpenAI node
|
||||
{
|
||||
"nodeType": "nodes-langchain.openAi"
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- Very fast (<5KB responses vs 100KB+ for full info)
|
||||
- Curated for 20+ common nodes
|
||||
- Automatic fallback for unconfigured nodes
|
||||
|
||||
### Best Practices
|
||||
- Always use this before get_node_info
|
||||
- Node type must include prefix: "nodes-base.slack" not "slack"
|
||||
- Check examples section for working configurations
|
||||
- Use tips section for common patterns
|
||||
|
||||
### Common Pitfalls
|
||||
- Forgetting the prefix in node type
|
||||
- Using wrong package name (n8n-nodes-base vs @n8n/n8n-nodes-langchain)
|
||||
- Case sensitivity in node types
|
||||
|
||||
### Related Tools
|
||||
- `get_node_info` - Full schema when essentials aren't enough
|
||||
- `search_node_properties` - Find specific properties
|
||||
- `get_node_for_task` - Pre-configured for common tasks
|
||||
|
||||
---
|
||||
|
||||
## list_nodes
|
||||
|
||||
**Brief Description**: List available n8n nodes with optional filtering by package, category, or capabilities.
|
||||
|
||||
### Parameters
|
||||
- `package` (string, optional): Filter by exact package name
|
||||
- `category` (string, optional): Filter by category (trigger, transform, output, input)
|
||||
- `developmentStyle` (string, optional): Filter by implementation style
|
||||
- `isAITool` (boolean, optional): Filter for AI-capable nodes
|
||||
- `limit` (number, optional): Maximum results (default: 50, max: 500)
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"nodes": [
|
||||
{
|
||||
"nodeType": "nodes-base.webhook",
|
||||
"displayName": "Webhook",
|
||||
"description": "Receive HTTP requests",
|
||||
"categories": ["trigger"],
|
||||
"version": 2
|
||||
}
|
||||
],
|
||||
"total": 104,
|
||||
"hasMore": false
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Browse all triggers**: `list_nodes({category: "trigger", limit: 200})`
|
||||
2. **List all nodes**: `list_nodes({limit: 500})`
|
||||
3. **Find AI nodes**: `list_nodes({isAITool: true})`
|
||||
4. **Browse core nodes**: `list_nodes({package: "n8n-nodes-base"})`
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// List all trigger nodes
|
||||
{
|
||||
"category": "trigger",
|
||||
"limit": 200
|
||||
}
|
||||
|
||||
// List all AI-capable nodes
|
||||
{
|
||||
"isAITool": true,
|
||||
"limit": 100
|
||||
}
|
||||
|
||||
// List nodes from core package
|
||||
{
|
||||
"package": "n8n-nodes-base",
|
||||
"limit": 200
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- Fast operation (cached results)
|
||||
- Default limit of 50 may miss nodes - use 200+
|
||||
- Returns metadata only, not full schemas
|
||||
|
||||
### Best Practices
|
||||
- Always set limit to 200+ for complete results
|
||||
- Use exact package names: "n8n-nodes-base" not "@n8n/n8n-nodes-base"
|
||||
- Categories are singular: "trigger" not "triggers"
|
||||
- Common categories: trigger (104), transform, output, input
|
||||
|
||||
### Common Pitfalls
|
||||
- Default limit (50) misses many nodes
|
||||
- Using wrong package name format
|
||||
- Multiple filters may return empty results
|
||||
|
||||
### Related Tools
|
||||
- `search_nodes` - Search by keywords
|
||||
- `list_ai_tools` - Specifically for AI nodes
|
||||
- `get_database_statistics` - Overview of all nodes
|
||||
|
||||
---
|
||||
|
||||
## validate_node_minimal
|
||||
|
||||
**Brief Description**: Quick validation checking only for missing required fields.
|
||||
|
||||
### Parameters
|
||||
- `nodeType` (string, required): Node type to validate (e.g., "nodes-base.slack")
|
||||
- `config` (object, required): Node configuration to check
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"valid": false,
|
||||
"missingRequired": ["channel", "messageType"],
|
||||
"message": "Missing 2 required fields"
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Quick validation**: Check if all required fields are present
|
||||
2. **Pre-flight check**: Validate before creating workflow
|
||||
3. **Minimal overhead**: Fastest validation option
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Validate Slack message configuration
|
||||
{
|
||||
"nodeType": "nodes-base.slack",
|
||||
"config": {
|
||||
"resource": "message",
|
||||
"operation": "send",
|
||||
"text": "Hello World"
|
||||
// Missing: channel
|
||||
}
|
||||
}
|
||||
|
||||
// Validate HTTP Request
|
||||
{
|
||||
"nodeType": "nodes-base.httpRequest",
|
||||
"config": {
|
||||
"method": "POST"
|
||||
// Missing: url
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- Fastest validation option
|
||||
- No schema loading overhead
|
||||
- Returns only missing fields
|
||||
|
||||
### Best Practices
|
||||
- Use for quick checks during workflow building
|
||||
- Follow up with validate_node_operation for complex nodes
|
||||
- Check operation-specific requirements
|
||||
|
||||
### Common Pitfalls
|
||||
- Doesn't validate field values or types
|
||||
- Doesn't check operation-specific requirements
|
||||
- Won't catch configuration errors beyond missing fields
|
||||
|
||||
### Related Tools
|
||||
- `validate_node_operation` - Comprehensive validation
|
||||
- `validate_workflow` - Full workflow validation
|
||||
|
||||
---
|
||||
|
||||
## validate_node_operation
|
||||
|
||||
**Brief Description**: Comprehensive node configuration validation with operation awareness and helpful error messages.
|
||||
|
||||
### Parameters
|
||||
- `nodeType` (string, required): Node type to validate
|
||||
- `config` (object, required): Complete node configuration including operation fields
|
||||
- `profile` (string, optional): Validation profile (minimal, runtime, ai-friendly, strict)
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"valid": false,
|
||||
"errors": [
|
||||
{
|
||||
"field": "channel",
|
||||
"message": "Channel is required to send Slack message",
|
||||
"suggestion": "Add channel: '#general' or '@username'"
|
||||
}
|
||||
],
|
||||
"warnings": [
|
||||
{
|
||||
"field": "unfurl_links",
|
||||
"message": "Consider setting unfurl_links: false for better performance"
|
||||
}
|
||||
],
|
||||
"examples": {
|
||||
"minimal": {
|
||||
"resource": "message",
|
||||
"operation": "send",
|
||||
"channel": "#general",
|
||||
"text": "Hello World"
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Complex node validation**: Slack, Google Sheets, databases
|
||||
2. **Operation-specific checks**: Different rules per operation
|
||||
3. **Getting fix suggestions**: Helpful error messages with solutions
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Validate Slack configuration
|
||||
{
|
||||
"nodeType": "nodes-base.slack",
|
||||
"config": {
|
||||
"resource": "message",
|
||||
"operation": "send",
|
||||
"text": "Hello team!"
|
||||
},
|
||||
"profile": "ai-friendly"
|
||||
}
|
||||
|
||||
// Validate Google Sheets operation
|
||||
{
|
||||
"nodeType": "nodes-base.googleSheets",
|
||||
"config": {
|
||||
"operation": "append",
|
||||
"sheetId": "1234567890",
|
||||
"range": "Sheet1!A:Z"
|
||||
},
|
||||
"profile": "runtime"
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- Slower than minimal validation
|
||||
- Loads full node schema
|
||||
- Operation-aware validation rules
|
||||
|
||||
### Best Practices
|
||||
- Use "ai-friendly" profile for balanced validation
|
||||
- Check examples in response for working configurations
|
||||
- Follow suggestions to fix errors
|
||||
- Essential for complex nodes (Slack, databases, APIs)
|
||||
|
||||
### Common Pitfalls
|
||||
- Forgetting operation fields (resource, operation, action)
|
||||
- Using wrong profile (too strict or too lenient)
|
||||
- Ignoring warnings that could cause runtime issues
|
||||
|
||||
### Related Tools
|
||||
- `validate_node_minimal` - Quick required field check
|
||||
- `get_property_dependencies` - Understand field relationships
|
||||
- `validate_workflow` - Validate entire workflow
|
||||
|
||||
---
|
||||
|
||||
## get_node_for_task
|
||||
|
||||
**Brief Description**: Get pre-configured node settings for common automation tasks.
|
||||
|
||||
### Parameters
|
||||
- `task` (string, required): Task identifier (e.g., "post_json_request", "receive_webhook")
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"task": "post_json_request",
|
||||
"nodeType": "nodes-base.httpRequest",
|
||||
"displayName": "HTTP Request",
|
||||
"configuration": {
|
||||
"method": "POST",
|
||||
"url": "={{ $json.api_endpoint }}",
|
||||
"responseFormat": "json",
|
||||
"options": {
|
||||
"bodyContentType": "json"
|
||||
},
|
||||
"bodyParametersJson": "={{ JSON.stringify($json) }}"
|
||||
},
|
||||
"userMustProvide": [
|
||||
"url - The API endpoint URL",
|
||||
"bodyParametersJson - The JSON data to send"
|
||||
],
|
||||
"tips": [
|
||||
"Use expressions to make values dynamic",
|
||||
"Enable 'Split Into Items' for batch processing"
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Quick task setup**: Configure nodes for specific tasks instantly
|
||||
2. **Learning patterns**: See how to configure nodes properly
|
||||
3. **Common workflows**: Standard patterns like webhooks, API calls, database queries
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Get configuration for JSON POST request
|
||||
{
|
||||
"task": "post_json_request"
|
||||
}
|
||||
|
||||
// Get webhook receiver configuration
|
||||
{
|
||||
"task": "receive_webhook"
|
||||
}
|
||||
|
||||
// Get AI chat configuration
|
||||
{
|
||||
"task": "chat_with_ai"
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- Instant response (pre-configured templates)
|
||||
- No database lookups required
|
||||
- Includes working examples
|
||||
|
||||
### Best Practices
|
||||
- Use list_tasks first to see available options
|
||||
- Check userMustProvide section
|
||||
- Follow tips for best results
|
||||
- Common tasks: API calls, webhooks, database queries, AI chat
|
||||
|
||||
### Common Pitfalls
|
||||
- Not all tasks available (use list_tasks)
|
||||
- Configuration needs customization
|
||||
- Some fields still need user input
|
||||
|
||||
### Related Tools
|
||||
- `list_tasks` - See all available tasks
|
||||
- `get_node_essentials` - Alternative approach
|
||||
- `search_templates` - Find complete workflow templates
|
||||
|
||||
---
|
||||
|
||||
## n8n_create_workflow
|
||||
|
||||
**Brief Description**: Create a new workflow in n8n with nodes and connections.
|
||||
|
||||
### Parameters
|
||||
- `name` (string, required): Workflow name
|
||||
- `nodes` (array, required): Array of node definitions
|
||||
- `connections` (object, required): Node connections mapping
|
||||
- `settings` (object, optional): Workflow settings
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"id": "workflow-uuid",
|
||||
"name": "My Workflow",
|
||||
"active": false,
|
||||
"createdAt": "2024-01-15T10:30:00Z",
|
||||
"updatedAt": "2024-01-15T10:30:00Z",
|
||||
"nodes": [...],
|
||||
"connections": {...}
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Automated workflow creation**: Build workflows programmatically
|
||||
2. **Template deployment**: Deploy pre-built workflow patterns
|
||||
3. **Multi-workflow systems**: Create interconnected workflows
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Create simple webhook → HTTP request workflow
|
||||
{
|
||||
"name": "Webhook to API",
|
||||
"nodes": [
|
||||
{
|
||||
"id": "webhook-1",
|
||||
"name": "Webhook",
|
||||
"type": "n8n-nodes-base.webhook",
|
||||
"typeVersion": 2,
|
||||
"position": [250, 300],
|
||||
"parameters": {
|
||||
"path": "/my-webhook",
|
||||
"httpMethod": "POST"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "http-1",
|
||||
"name": "HTTP Request",
|
||||
"type": "n8n-nodes-base.httpRequest",
|
||||
"typeVersion": 4.2,
|
||||
"position": [450, 300],
|
||||
"parameters": {
|
||||
"method": "POST",
|
||||
"url": "https://api.example.com/process",
|
||||
"responseFormat": "json"
|
||||
}
|
||||
}
|
||||
],
|
||||
"connections": {
|
||||
"Webhook": {
|
||||
"main": [[{"node": "HTTP Request", "type": "main", "index": 0}]]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- API call to n8n instance required
|
||||
- Workflow created in inactive state
|
||||
- Must be manually activated in UI
|
||||
|
||||
### Best Practices
|
||||
- Always include typeVersion for nodes
|
||||
- Use node names (not IDs) in connections
|
||||
- Position nodes logically ([x, y] coordinates)
|
||||
- Test with validate_workflow first
|
||||
- Start simple, add complexity gradually
|
||||
|
||||
### Common Pitfalls
|
||||
- Missing typeVersion causes errors
|
||||
- Using node IDs instead of names in connections
|
||||
- Forgetting required node properties
|
||||
- Creating cycles in connections
|
||||
- Workflow can't be activated via API
|
||||
|
||||
### Related Tools
|
||||
- `validate_workflow` - Validate before creating
|
||||
- `n8n_update_partial_workflow` - Modify existing workflows
|
||||
- `n8n_trigger_webhook_workflow` - Execute workflows
|
||||
|
||||
---
|
||||
|
||||
## n8n_update_partial_workflow
|
||||
|
||||
**Brief Description**: Update workflows using diff operations for precise, incremental changes without sending the entire workflow.
|
||||
|
||||
### Parameters
|
||||
- `id` (string, required): Workflow ID to update
|
||||
- `operations` (array, required): Array of diff operations (max 5)
|
||||
- `validateOnly` (boolean, optional): Test without applying changes
|
||||
|
||||
### Return Format
|
||||
```json
|
||||
{
|
||||
"success": true,
|
||||
"workflow": {
|
||||
"id": "workflow-uuid",
|
||||
"name": "Updated Workflow",
|
||||
"nodes": [...],
|
||||
"connections": {...}
|
||||
},
|
||||
"appliedOperations": 3
|
||||
}
|
||||
```
|
||||
|
||||
### Common Use Cases
|
||||
1. **Add nodes to existing workflows**: Insert new functionality
|
||||
2. **Update node configurations**: Change parameters without full replacement
|
||||
3. **Manage connections**: Add/remove node connections
|
||||
4. **Quick edits**: Rename, enable/disable nodes, update settings
|
||||
|
||||
### Examples
|
||||
```json
|
||||
// Add a new node and connect it
|
||||
{
|
||||
"id": "workflow-123",
|
||||
"operations": [
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"id": "set-1",
|
||||
"name": "Set Data",
|
||||
"type": "n8n-nodes-base.set",
|
||||
"typeVersion": 3,
|
||||
"position": [600, 300],
|
||||
"parameters": {
|
||||
"values": {
|
||||
"string": [{
|
||||
"name": "status",
|
||||
"value": "processed"
|
||||
}]
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "HTTP Request",
|
||||
"target": "Set Data"
|
||||
}
|
||||
]
|
||||
}
|
||||
|
||||
// Update multiple properties
|
||||
{
|
||||
"id": "workflow-123",
|
||||
"operations": [
|
||||
{
|
||||
"type": "updateName",
|
||||
"name": "Production Workflow v2"
|
||||
},
|
||||
{
|
||||
"type": "updateNode",
|
||||
"nodeName": "Webhook",
|
||||
"changes": {
|
||||
"parameters.path": "/v2/webhook"
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addTag",
|
||||
"tag": "production"
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
### Performance Notes
|
||||
- 80-90% token savings vs full updates
|
||||
- Maximum 5 operations per request
|
||||
- Two-pass processing handles dependencies
|
||||
- Transactional: all or nothing
|
||||
|
||||
### Best Practices
|
||||
- Use validateOnly: true to test first
|
||||
- Keep operations under 5 for reliability
|
||||
- Operations can be in any order (v2.7.0+)
|
||||
- Use node names, not IDs in operations
|
||||
- For updateNode, use dot notation for nested paths
|
||||
|
||||
### Common Pitfalls
|
||||
- Exceeding 5 operations limit
|
||||
- Using node IDs instead of names
|
||||
- Forgetting required node properties in addNode
|
||||
- Not testing with validateOnly first
|
||||
|
||||
### Related Tools
|
||||
- `n8n_update_full_workflow` - Complete workflow replacement
|
||||
- `n8n_get_workflow` - Fetch current workflow state
|
||||
- `validate_workflow` - Validate changes before applying
|
||||
|
||||
---
|
||||
|
||||
## Quick Reference
|
||||
|
||||
### Workflow Building Process
|
||||
1. **Discovery**: `search_nodes` → `list_nodes`
|
||||
2. **Configuration**: `get_node_essentials` → `get_node_for_task`
|
||||
3. **Validation**: `validate_node_minimal` → `validate_node_operation`
|
||||
4. **Creation**: `validate_workflow` → `n8n_create_workflow`
|
||||
5. **Updates**: `n8n_update_partial_workflow`
|
||||
|
||||
### Performance Tips
|
||||
- Use `get_node_essentials` instead of `get_node_info` (95% smaller)
|
||||
- Set high limits on `list_nodes` (200+)
|
||||
- Use single words in `search_nodes`
|
||||
- Validate incrementally while building
|
||||
|
||||
### Common Node Types
|
||||
- **Triggers**: webhook, schedule, emailReadImap, slackTrigger
|
||||
- **Core**: httpRequest, code, set, if, merge, splitInBatches
|
||||
- **Integrations**: slack, gmail, googleSheets, postgres, mongodb
|
||||
- **AI**: agent, openAi, chainLlm, documentLoader
|
||||
|
||||
### Error Prevention
|
||||
- Always include node type prefixes: "nodes-base.slack"
|
||||
- Use node names (not IDs) in connections
|
||||
- Include typeVersion in all nodes
|
||||
- Test with validateOnly before applying changes
|
||||
- Check userMustProvide sections in templates
|
||||
@@ -1,514 +0,0 @@
|
||||
# n8n MCP Client Tool Integration - Implementation Plan (Simplified)
|
||||
|
||||
## Overview
|
||||
|
||||
This document provides a **simplified** implementation plan for making n8n-mcp compatible with n8n's MCP Client Tool (v1.1). Based on expert review, we're taking a minimal approach that extends the existing single-session server rather than creating new architecture.
|
||||
|
||||
## Key Design Principles
|
||||
|
||||
1. **Minimal Changes**: Extend existing single-session server with n8n compatibility mode
|
||||
2. **No Overengineering**: No complex session management or multi-session architecture
|
||||
3. **Docker-Native**: Separate Docker image for n8n deployment
|
||||
4. **Remote Deployment**: Designed to run alongside n8n in production
|
||||
5. **Backward Compatible**: Existing functionality remains unchanged
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- Docker and Docker Compose
|
||||
- n8n version 1.104.2 or higher (with MCP Client Tool v1.1)
|
||||
- Basic understanding of Docker networking
|
||||
|
||||
## Implementation Approach
|
||||
|
||||
Instead of creating new multi-session architecture, we'll extend the existing single-session server with an n8n compatibility mode. This approach was recommended by all three expert reviewers as simpler and more maintainable.
|
||||
|
||||
## Architecture Changes
|
||||
|
||||
```
|
||||
src/
|
||||
├── http-server-single-session.ts # MODIFY: Add n8n mode flag
|
||||
└── mcp/
|
||||
└── server.ts # NO CHANGES NEEDED
|
||||
|
||||
Docker/
|
||||
├── Dockerfile.n8n # NEW: n8n-specific image
|
||||
├── docker-compose.n8n.yml # NEW: Simplified stack
|
||||
└── .github/workflows/
|
||||
└── docker-build-n8n.yml # NEW: Build workflow
|
||||
```
|
||||
|
||||
## Implementation Steps
|
||||
|
||||
### Step 1: Modify Existing Single-Session Server
|
||||
|
||||
#### 1.1 Update `src/http-server-single-session.ts`
|
||||
|
||||
Add n8n compatibility mode to the existing server with minimal changes:
|
||||
|
||||
```typescript
|
||||
// Add these constants at the top (after imports)
|
||||
const PROTOCOL_VERSION = "2024-11-05";
|
||||
const N8N_MODE = process.env.N8N_MODE === 'true';
|
||||
|
||||
// In the constructor or start method, add logging
|
||||
if (N8N_MODE) {
|
||||
logger.info('Running in n8n compatibility mode');
|
||||
}
|
||||
|
||||
// In setupRoutes method, add the protocol version endpoint
|
||||
if (N8N_MODE) {
|
||||
app.get('/mcp', (req, res) => {
|
||||
res.json({
|
||||
protocolVersion: PROTOCOL_VERSION,
|
||||
serverInfo: {
|
||||
name: "n8n-mcp",
|
||||
version: PROJECT_VERSION,
|
||||
capabilities: {
|
||||
tools: true,
|
||||
resources: false,
|
||||
prompts: false,
|
||||
},
|
||||
},
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
// In handleMCPRequest method, add session header
|
||||
if (N8N_MODE && this.session) {
|
||||
res.setHeader('Mcp-Session-Id', this.session.sessionId);
|
||||
}
|
||||
|
||||
// Update error handling to use JSON-RPC format
|
||||
catch (error) {
|
||||
logger.error('MCP request error:', error);
|
||||
|
||||
if (N8N_MODE) {
|
||||
res.status(500).json({
|
||||
jsonrpc: '2.0',
|
||||
error: {
|
||||
code: -32603,
|
||||
message: 'Internal error',
|
||||
data: error instanceof Error ? error.message : 'Unknown error',
|
||||
},
|
||||
id: null,
|
||||
});
|
||||
} else {
|
||||
// Keep existing error handling for backward compatibility
|
||||
res.status(500).json({
|
||||
error: 'Internal server error',
|
||||
details: error instanceof Error ? error.message : 'Unknown error'
|
||||
});
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
That's it! No new files, no complex session management. Just a few lines of code.
|
||||
|
||||
### Step 2: Update Package Scripts
|
||||
|
||||
#### 2.1 Update `package.json`
|
||||
|
||||
Add a simple script for n8n mode:
|
||||
|
||||
```json
|
||||
{
|
||||
"scripts": {
|
||||
"start:n8n": "N8N_MODE=true MCP_MODE=http node dist/mcp/index.js"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Step 3: Create Docker Infrastructure for n8n
|
||||
|
||||
#### 3.1 Create `Dockerfile.n8n`
|
||||
|
||||
```dockerfile
|
||||
# Dockerfile.n8n - Optimized for n8n integration
|
||||
FROM node:22-alpine AS builder
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
# Install build dependencies
|
||||
RUN apk add --no-cache python3 make g++
|
||||
|
||||
# Copy package files
|
||||
COPY package*.json tsconfig*.json ./
|
||||
|
||||
# Install ALL dependencies
|
||||
RUN npm ci --no-audit --no-fund
|
||||
|
||||
# Copy source and build
|
||||
COPY src ./src
|
||||
RUN npm run build && npm run rebuild
|
||||
|
||||
# Runtime stage
|
||||
FROM node:22-alpine
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
# Install runtime dependencies
|
||||
RUN apk add --no-cache curl dumb-init
|
||||
|
||||
# Create non-root user
|
||||
RUN addgroup -g 1001 -S nodejs && adduser -S nodejs -u 1001
|
||||
|
||||
# Copy application from builder
|
||||
COPY --from=builder --chown=nodejs:nodejs /app/dist ./dist
|
||||
COPY --from=builder --chown=nodejs:nodejs /app/data ./data
|
||||
COPY --from=builder --chown=nodejs:nodejs /app/node_modules ./node_modules
|
||||
COPY --chown=nodejs:nodejs package.json ./
|
||||
|
||||
USER nodejs
|
||||
|
||||
EXPOSE 3001
|
||||
|
||||
HEALTHCHECK CMD curl -f http://localhost:3001/health || exit 1
|
||||
|
||||
ENTRYPOINT ["dumb-init", "--"]
|
||||
CMD ["node", "dist/mcp/index.js"]
|
||||
```
|
||||
|
||||
#### 3.2 Create `docker-compose.n8n.yml`
|
||||
|
||||
```yaml
|
||||
# docker-compose.n8n.yml - Simple stack for n8n + n8n-mcp
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
n8n:
|
||||
image: n8nio/n8n:latest
|
||||
container_name: n8n
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "5678:5678"
|
||||
environment:
|
||||
- N8N_BASIC_AUTH_ACTIVE=${N8N_BASIC_AUTH_ACTIVE:-true}
|
||||
- N8N_BASIC_AUTH_USER=${N8N_USER:-admin}
|
||||
- N8N_BASIC_AUTH_PASSWORD=${N8N_PASSWORD:-changeme}
|
||||
- N8N_COMMUNITY_PACKAGES_ALLOW_TOOL_USAGE=true
|
||||
volumes:
|
||||
- n8n_data:/home/node/.n8n
|
||||
networks:
|
||||
- n8n-net
|
||||
depends_on:
|
||||
n8n-mcp:
|
||||
condition: service_healthy
|
||||
|
||||
n8n-mcp:
|
||||
image: ghcr.io/${GITHUB_USER:-czlonkowski}/n8n-mcp-n8n:latest
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile.n8n
|
||||
container_name: n8n-mcp
|
||||
restart: unless-stopped
|
||||
environment:
|
||||
- MCP_MODE=http
|
||||
- N8N_MODE=true
|
||||
- AUTH_TOKEN=${MCP_AUTH_TOKEN}
|
||||
- NODE_ENV=production
|
||||
- HTTP_PORT=3001
|
||||
networks:
|
||||
- n8n-net
|
||||
healthcheck:
|
||||
test: ["CMD", "curl", "-f", "http://localhost:3001/health"]
|
||||
interval: 30s
|
||||
timeout: 10s
|
||||
retries: 3
|
||||
|
||||
networks:
|
||||
n8n-net:
|
||||
driver: bridge
|
||||
|
||||
volumes:
|
||||
n8n_data:
|
||||
```
|
||||
|
||||
#### 3.3 Create `.env.n8n.example`
|
||||
|
||||
```bash
|
||||
# .env.n8n.example - Copy to .env and configure
|
||||
|
||||
# n8n Configuration
|
||||
N8N_USER=admin
|
||||
N8N_PASSWORD=changeme
|
||||
N8N_BASIC_AUTH_ACTIVE=true
|
||||
|
||||
# MCP Configuration
|
||||
# Generate with: openssl rand -base64 32
|
||||
MCP_AUTH_TOKEN=your-secure-token-minimum-32-characters
|
||||
|
||||
# GitHub username for image registry
|
||||
GITHUB_USER=czlonkowski
|
||||
```
|
||||
|
||||
### Step 4: Create GitHub Actions Workflow
|
||||
|
||||
#### 4.1 Create `.github/workflows/docker-build-n8n.yml`
|
||||
|
||||
```yaml
|
||||
name: Build n8n Docker Image
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
tags: ['v*']
|
||||
paths:
|
||||
- 'src/**'
|
||||
- 'package*.json'
|
||||
- 'Dockerfile.n8n'
|
||||
workflow_dispatch:
|
||||
|
||||
env:
|
||||
REGISTRY: ghcr.io
|
||||
IMAGE_NAME: ${{ github.repository }}-n8n
|
||||
|
||||
jobs:
|
||||
build:
|
||||
runs-on: ubuntu-latest
|
||||
permissions:
|
||||
contents: read
|
||||
packages: write
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
|
||||
- uses: docker/setup-buildx-action@v3
|
||||
|
||||
- uses: docker/login-action@v3
|
||||
with:
|
||||
registry: ${{ env.REGISTRY }}
|
||||
username: ${{ github.actor }}
|
||||
password: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
- uses: docker/metadata-action@v5
|
||||
id: meta
|
||||
with:
|
||||
images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
||||
tags: |
|
||||
type=ref,event=branch
|
||||
type=semver,pattern={{version}}
|
||||
type=raw,value=latest,enable={{is_default_branch}}
|
||||
|
||||
- uses: docker/build-push-action@v5
|
||||
with:
|
||||
context: .
|
||||
file: ./Dockerfile.n8n
|
||||
push: true
|
||||
tags: ${{ steps.meta.outputs.tags }}
|
||||
cache-from: type=gha
|
||||
cache-to: type=gha,mode=max
|
||||
```
|
||||
|
||||
### Step 5: Testing
|
||||
|
||||
#### 5.1 Unit Tests for n8n Mode
|
||||
|
||||
Create `tests/unit/http-server-n8n-mode.test.ts`:
|
||||
|
||||
```typescript
|
||||
import { describe, it, expect, vi } from 'vitest';
|
||||
import request from 'supertest';
|
||||
|
||||
describe('n8n Mode', () => {
|
||||
it('should return protocol version on GET /mcp', async () => {
|
||||
process.env.N8N_MODE = 'true';
|
||||
const app = await createTestApp();
|
||||
|
||||
const response = await request(app)
|
||||
.get('/mcp')
|
||||
.expect(200);
|
||||
|
||||
expect(response.body.protocolVersion).toBe('2024-11-05');
|
||||
expect(response.body.serverInfo.capabilities.tools).toBe(true);
|
||||
});
|
||||
|
||||
it('should include session ID in response headers', async () => {
|
||||
process.env.N8N_MODE = 'true';
|
||||
const app = await createTestApp();
|
||||
|
||||
const response = await request(app)
|
||||
.post('/mcp')
|
||||
.set('Authorization', 'Bearer test-token')
|
||||
.send({ jsonrpc: '2.0', method: 'initialize', id: 1 });
|
||||
|
||||
expect(response.headers['mcp-session-id']).toBeDefined();
|
||||
});
|
||||
|
||||
it('should format errors as JSON-RPC', async () => {
|
||||
process.env.N8N_MODE = 'true';
|
||||
const app = await createTestApp();
|
||||
|
||||
const response = await request(app)
|
||||
.post('/mcp')
|
||||
.send({ invalid: 'request' })
|
||||
.expect(500);
|
||||
|
||||
expect(response.body.jsonrpc).toBe('2.0');
|
||||
expect(response.body.error.code).toBe(-32603);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
#### 5.2 Quick Deployment Script
|
||||
|
||||
Create `deploy/quick-deploy-n8n.sh`:
|
||||
|
||||
```bash
|
||||
#!/bin/bash
|
||||
set -e
|
||||
|
||||
echo "🚀 Quick Deploy n8n + n8n-mcp"
|
||||
|
||||
# Check prerequisites
|
||||
command -v docker >/dev/null 2>&1 || { echo "Docker required"; exit 1; }
|
||||
command -v docker-compose >/dev/null 2>&1 || { echo "Docker Compose required"; exit 1; }
|
||||
|
||||
# Generate auth token if not exists
|
||||
if [ ! -f .env ]; then
|
||||
cp .env.n8n.example .env
|
||||
TOKEN=$(openssl rand -base64 32)
|
||||
sed -i "s/your-secure-token-minimum-32-characters/$TOKEN/" .env
|
||||
echo "Generated MCP_AUTH_TOKEN: $TOKEN"
|
||||
fi
|
||||
|
||||
# Deploy
|
||||
docker-compose -f docker-compose.n8n.yml up -d
|
||||
|
||||
echo ""
|
||||
echo "✅ Deployment complete!"
|
||||
echo ""
|
||||
echo "📋 Next steps:"
|
||||
echo "1. Access n8n at http://localhost:5678"
|
||||
echo " Username: admin (or check .env)"
|
||||
echo " Password: changeme (or check .env)"
|
||||
echo ""
|
||||
echo "2. Create a workflow with MCP Client Tool:"
|
||||
echo " - Server URL: http://n8n-mcp:3001/mcp"
|
||||
echo " - Authentication: Bearer Token"
|
||||
echo " - Token: Check .env file for MCP_AUTH_TOKEN"
|
||||
echo ""
|
||||
echo "📊 View logs: docker-compose -f docker-compose.n8n.yml logs -f"
|
||||
echo "🛑 Stop: docker-compose -f docker-compose.n8n.yml down"
|
||||
```
|
||||
|
||||
## Implementation Checklist (Simplified)
|
||||
|
||||
### Code Changes
|
||||
- [ ] Add N8N_MODE flag to `http-server-single-session.ts`
|
||||
- [ ] Add protocol version endpoint (GET /mcp) when N8N_MODE=true
|
||||
- [ ] Add Mcp-Session-Id header to responses
|
||||
- [ ] Update error responses to JSON-RPC format when N8N_MODE=true
|
||||
- [ ] Add npm script `start:n8n` to package.json
|
||||
|
||||
### Docker Infrastructure
|
||||
- [ ] Create `Dockerfile.n8n` for n8n-specific image
|
||||
- [ ] Create `docker-compose.n8n.yml` for simple deployment
|
||||
- [ ] Create `.env.n8n.example` template
|
||||
- [ ] Create GitHub Actions workflow `docker-build-n8n.yml`
|
||||
- [ ] Create `deploy/quick-deploy-n8n.sh` script
|
||||
|
||||
### Testing
|
||||
- [ ] Write unit tests for n8n mode functionality
|
||||
- [ ] Test with actual n8n MCP Client Tool
|
||||
- [ ] Verify protocol version endpoint
|
||||
- [ ] Test authentication flow
|
||||
- [ ] Validate error formatting
|
||||
|
||||
### Documentation
|
||||
- [ ] Update README with n8n deployment section
|
||||
- [ ] Document N8N_MODE environment variable
|
||||
- [ ] Add troubleshooting guide for common issues
|
||||
|
||||
## Quick Start Guide
|
||||
|
||||
### 1. One-Command Deployment
|
||||
|
||||
```bash
|
||||
# Clone and deploy
|
||||
git clone https://github.com/czlonkowski/n8n-mcp.git
|
||||
cd n8n-mcp
|
||||
./deploy/quick-deploy-n8n.sh
|
||||
```
|
||||
|
||||
### 2. Manual Configuration in n8n
|
||||
|
||||
After deployment, configure the MCP Client Tool in n8n:
|
||||
|
||||
1. Open n8n at `http://localhost:5678`
|
||||
2. Create a new workflow
|
||||
3. Add "MCP Client Tool" node (under AI category)
|
||||
4. Configure:
|
||||
- **Server URL**: `http://n8n-mcp:3001/mcp`
|
||||
- **Authentication**: Bearer Token
|
||||
- **Token**: Check your `.env` file for MCP_AUTH_TOKEN
|
||||
5. Select a tool (e.g., `list_nodes`)
|
||||
6. Execute the workflow
|
||||
|
||||
### 3. Production Deployment
|
||||
|
||||
For production with SSL, use a reverse proxy:
|
||||
|
||||
```nginx
|
||||
# nginx configuration
|
||||
server {
|
||||
listen 443 ssl;
|
||||
server_name n8n.yourdomain.com;
|
||||
|
||||
location / {
|
||||
proxy_pass http://localhost:5678;
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection "upgrade";
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
The MCP server should remain internal only - n8n connects via Docker network.
|
||||
|
||||
## Success Criteria
|
||||
|
||||
The implementation is successful when:
|
||||
|
||||
1. **Minimal Code Changes**: Only ~20 lines added to existing server
|
||||
2. **Protocol Compliance**: GET /mcp returns correct protocol version
|
||||
3. **n8n Connection**: MCP Client Tool connects successfully
|
||||
4. **Tool Execution**: Tools work without modification
|
||||
5. **Backward Compatible**: Existing Claude Desktop usage unaffected
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
1. **"Protocol version mismatch"**
|
||||
- Ensure N8N_MODE=true is set
|
||||
- Check GET /mcp returns "2024-11-05"
|
||||
|
||||
2. **"Authentication failed"**
|
||||
- Verify AUTH_TOKEN matches in .env and n8n
|
||||
- Token must be 32+ characters
|
||||
- Use "Bearer Token" auth type in n8n
|
||||
|
||||
3. **"Connection refused"**
|
||||
- Check containers are on same network
|
||||
- Use internal hostname: `http://n8n-mcp:3001/mcp`
|
||||
- Verify health check passes
|
||||
|
||||
4. **Testing the Setup**
|
||||
```bash
|
||||
# Check protocol version
|
||||
docker exec n8n-mcp curl http://localhost:3001/mcp
|
||||
|
||||
# View logs
|
||||
docker-compose -f docker-compose.n8n.yml logs -f n8n-mcp
|
||||
```
|
||||
|
||||
## Summary
|
||||
|
||||
This simplified approach:
|
||||
- **Extends existing code** rather than creating new architecture
|
||||
- **Adds n8n compatibility** with minimal changes
|
||||
- **Uses separate Docker image** for clean deployment
|
||||
- **Maintains backward compatibility** for existing users
|
||||
- **Avoids overengineering** with simple, practical solutions
|
||||
|
||||
Total implementation effort: ~2-3 hours (vs. 2-3 days for multi-session approach)
|
||||
@@ -1,146 +0,0 @@
|
||||
# Test Artifacts Documentation
|
||||
|
||||
This document describes the comprehensive test result artifact storage system implemented in the n8n-mcp project.
|
||||
|
||||
## Overview
|
||||
|
||||
The test artifact system captures, stores, and presents test results in multiple formats to facilitate debugging, analysis, and historical tracking of test performance.
|
||||
|
||||
## Artifact Types
|
||||
|
||||
### 1. Test Results
|
||||
- **JUnit XML** (`test-results/junit.xml`): Standard format for CI integration
|
||||
- **JSON Results** (`test-results/results.json`): Detailed test data for analysis
|
||||
- **HTML Report** (`test-results/html/index.html`): Interactive test report
|
||||
- **Test Summary** (`test-summary.md`): Markdown summary for PR comments
|
||||
|
||||
### 2. Coverage Reports
|
||||
- **LCOV** (`coverage/lcov.info`): Standard coverage format
|
||||
- **HTML Coverage** (`coverage/html/index.html`): Interactive coverage browser
|
||||
- **Coverage Summary** (`coverage/coverage-summary.json`): JSON coverage data
|
||||
|
||||
### 3. Benchmark Results
|
||||
- **Benchmark JSON** (`benchmark-results.json`): Raw benchmark data
|
||||
- **Comparison Reports** (`benchmark-comparison.md`): PR benchmark comparisons
|
||||
|
||||
### 4. Detailed Reports
|
||||
- **HTML Report** (`test-reports/report.html`): Comprehensive styled report
|
||||
- **Markdown Report** (`test-reports/report.md`): Full markdown report
|
||||
- **JSON Report** (`test-reports/report.json`): Complete test data
|
||||
|
||||
## GitHub Actions Integration
|
||||
|
||||
### Test Workflow (`test.yml`)
|
||||
|
||||
The main test workflow:
|
||||
1. Runs tests with coverage using multiple reporters
|
||||
2. Generates test summaries and detailed reports
|
||||
3. Uploads artifacts with metadata
|
||||
4. Posts summaries to PRs
|
||||
5. Creates a combined artifact index
|
||||
|
||||
### Benchmark PR Workflow (`benchmark-pr.yml`)
|
||||
|
||||
For pull requests:
|
||||
1. Runs benchmarks on PR branch
|
||||
2. Runs benchmarks on base branch
|
||||
3. Compares results
|
||||
4. Posts comparison to PR
|
||||
5. Sets status checks for regressions
|
||||
|
||||
## Artifact Retention
|
||||
|
||||
- **Test Results**: 30 days
|
||||
- **Coverage Reports**: 30 days
|
||||
- **Benchmark Results**: 30 days
|
||||
- **Combined Results**: 90 days
|
||||
- **Test Metadata**: 30 days
|
||||
|
||||
## PR Comment Integration
|
||||
|
||||
The system automatically:
|
||||
- Posts test summaries to PR comments
|
||||
- Updates existing comments instead of creating duplicates
|
||||
- Includes links to full artifacts
|
||||
- Shows coverage and benchmark changes
|
||||
|
||||
## Job Summary
|
||||
|
||||
Each workflow run includes a job summary with:
|
||||
- Test results overview
|
||||
- Coverage summary
|
||||
- Benchmark results
|
||||
- Direct links to download artifacts
|
||||
|
||||
## Local Development
|
||||
|
||||
### Running Tests with Reports
|
||||
|
||||
```bash
|
||||
# Run tests with all reporters
|
||||
CI=true npm run test:coverage
|
||||
|
||||
# Generate detailed reports
|
||||
node scripts/generate-detailed-reports.js
|
||||
|
||||
# Generate test summary
|
||||
node scripts/generate-test-summary.js
|
||||
|
||||
# Compare benchmarks
|
||||
node scripts/compare-benchmarks.js benchmark-results.json benchmark-baseline.json
|
||||
```
|
||||
|
||||
### Report Locations
|
||||
|
||||
When running locally, reports are generated in:
|
||||
- `test-results/` - Vitest outputs
|
||||
- `test-reports/` - Detailed reports
|
||||
- `coverage/` - Coverage reports
|
||||
- Root directory - Summary files
|
||||
|
||||
## Report Formats
|
||||
|
||||
### HTML Report Features
|
||||
- Responsive design
|
||||
- Test suite breakdown
|
||||
- Failed test details with error messages
|
||||
- Coverage visualization with progress bars
|
||||
- Benchmark performance metrics
|
||||
- Sortable tables
|
||||
|
||||
### Markdown Report Features
|
||||
- GitHub-compatible formatting
|
||||
- Summary statistics
|
||||
- Failed test listings
|
||||
- Coverage breakdown
|
||||
- Benchmark comparisons
|
||||
|
||||
### JSON Report Features
|
||||
- Complete test data
|
||||
- Programmatic access
|
||||
- Historical comparison
|
||||
- CI/CD integration
|
||||
|
||||
## Best Practices
|
||||
|
||||
1. **Always Check Artifacts**: When tests fail in CI, download and review the HTML report
|
||||
2. **Monitor Coverage**: Use the coverage reports to identify untested code
|
||||
3. **Track Benchmarks**: Review benchmark comparisons on performance-critical PRs
|
||||
4. **Archive Important Runs**: Download artifacts from significant releases
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Missing Artifacts
|
||||
- Check if tests ran to completion
|
||||
- Verify artifact upload steps executed
|
||||
- Check retention period hasn't expired
|
||||
|
||||
### Report Generation Failures
|
||||
- Ensure all dependencies are installed
|
||||
- Check for valid test/coverage output files
|
||||
- Review workflow logs for errors
|
||||
|
||||
### PR Comment Issues
|
||||
- Verify GitHub Actions permissions
|
||||
- Check bot authentication
|
||||
- Review comment posting logs
|
||||
@@ -1,802 +0,0 @@
|
||||
# n8n-MCP Testing Architecture
|
||||
|
||||
## Overview
|
||||
|
||||
This document describes the comprehensive testing infrastructure implemented for the n8n-MCP project. The testing suite includes over 1,100 tests split between unit and integration tests, benchmarks, and a complete CI/CD pipeline ensuring code quality and reliability.
|
||||
|
||||
### Test Suite Statistics (from CI Run #41)
|
||||
|
||||
- **Total Tests**: 1,182 tests
|
||||
- **Unit Tests**: 933 tests (932 passed, 1 skipped)
|
||||
- **Integration Tests**: 249 tests (245 passed, 4 skipped)
|
||||
- **Test Files**:
|
||||
- 30 unit test files
|
||||
- 14 integration test files
|
||||
- **Test Execution Time**:
|
||||
- Unit tests: ~2 minutes with coverage
|
||||
- Integration tests: ~23 seconds
|
||||
- Total CI time: ~2.5 minutes
|
||||
- **Success Rate**: 99.5% (only 5 tests skipped, 0 failures)
|
||||
- **CI/CD Pipeline**: Fully automated with GitHub Actions
|
||||
- **Test Artifacts**: JUnit XML, coverage reports, benchmark results
|
||||
- **Parallel Execution**: Configurable with thread pool
|
||||
|
||||
## Testing Framework: Vitest
|
||||
|
||||
We use **Vitest** as our primary testing framework, chosen for its:
|
||||
- **Speed**: Native ESM support and fast execution
|
||||
- **TypeScript Integration**: First-class TypeScript support
|
||||
- **Watch Mode**: Instant feedback during development
|
||||
- **Jest Compatibility**: Easy migration from Jest
|
||||
- **Built-in Mocking**: Powerful mocking capabilities
|
||||
- **Coverage**: Integrated code coverage with v8
|
||||
|
||||
### Configuration
|
||||
|
||||
```typescript
|
||||
// vitest.config.ts
|
||||
export default defineConfig({
|
||||
test: {
|
||||
globals: true,
|
||||
environment: 'node',
|
||||
setupFiles: ['./tests/setup/global-setup.ts'],
|
||||
pool: 'threads',
|
||||
poolOptions: {
|
||||
threads: {
|
||||
singleThread: process.env.TEST_PARALLEL !== 'true',
|
||||
maxThreads: parseInt(process.env.TEST_MAX_WORKERS || '4', 10)
|
||||
}
|
||||
},
|
||||
coverage: {
|
||||
provider: 'v8',
|
||||
reporter: ['lcov', 'html', 'text-summary'],
|
||||
exclude: ['node_modules/', 'tests/', '**/*.test.ts', 'scripts/']
|
||||
}
|
||||
},
|
||||
resolve: {
|
||||
alias: {
|
||||
'@': path.resolve(__dirname, './src'),
|
||||
'@tests': path.resolve(__dirname, './tests')
|
||||
}
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
## Directory Structure
|
||||
|
||||
```
|
||||
tests/
|
||||
├── unit/ # Unit tests with mocks (933 tests, 30 files)
|
||||
│ ├── __mocks__/ # Mock implementations
|
||||
│ │ └── n8n-nodes-base.test.ts
|
||||
│ ├── database/ # Database layer tests
|
||||
│ │ ├── database-adapter-unit.test.ts
|
||||
│ │ ├── node-repository-core.test.ts
|
||||
│ │ └── template-repository-core.test.ts
|
||||
│ ├── loaders/ # Node loader tests
|
||||
│ │ └── node-loader.test.ts
|
||||
│ ├── mappers/ # Data mapper tests
|
||||
│ │ └── docs-mapper.test.ts
|
||||
│ ├── mcp/ # MCP server and tools tests
|
||||
│ │ ├── handlers-n8n-manager.test.ts
|
||||
│ │ ├── handlers-workflow-diff.test.ts
|
||||
│ │ ├── tools-documentation.test.ts
|
||||
│ │ └── tools.test.ts
|
||||
│ ├── parsers/ # Parser tests
|
||||
│ │ ├── node-parser.test.ts
|
||||
│ │ ├── property-extractor.test.ts
|
||||
│ │ └── simple-parser.test.ts
|
||||
│ ├── services/ # Service layer tests (largest test suite)
|
||||
│ │ ├── config-validator.test.ts
|
||||
│ │ ├── enhanced-config-validator.test.ts
|
||||
│ │ ├── example-generator.test.ts
|
||||
│ │ ├── expression-validator.test.ts
|
||||
│ │ ├── n8n-api-client.test.ts
|
||||
│ │ ├── n8n-validation.test.ts
|
||||
│ │ ├── node-specific-validators.test.ts
|
||||
│ │ ├── property-dependencies.test.ts
|
||||
│ │ ├── property-filter.test.ts
|
||||
│ │ ├── task-templates.test.ts
|
||||
│ │ ├── workflow-diff-engine.test.ts
|
||||
│ │ ├── workflow-validator-comprehensive.test.ts
|
||||
│ │ └── workflow-validator.test.ts
|
||||
│ └── utils/ # Utility function tests
|
||||
│ └── database-utils.test.ts
|
||||
├── integration/ # Integration tests (249 tests, 14 files)
|
||||
│ ├── database/ # Database integration tests
|
||||
│ │ ├── connection-management.test.ts
|
||||
│ │ ├── fts5-search.test.ts
|
||||
│ │ ├── node-repository.test.ts
|
||||
│ │ ├── performance.test.ts
|
||||
│ │ └── transactions.test.ts
|
||||
│ ├── mcp-protocol/ # MCP protocol tests
|
||||
│ │ ├── basic-connection.test.ts
|
||||
│ │ ├── error-handling.test.ts
|
||||
│ │ ├── performance.test.ts
|
||||
│ │ ├── protocol-compliance.test.ts
|
||||
│ │ ├── session-management.test.ts
|
||||
│ │ └── tool-invocation.test.ts
|
||||
│ └── setup/ # Integration test setup
|
||||
│ ├── integration-setup.ts
|
||||
│ └── msw-test-server.ts
|
||||
├── benchmarks/ # Performance benchmarks
|
||||
│ ├── database-queries.bench.ts
|
||||
│ └── sample.bench.ts
|
||||
├── setup/ # Global test configuration
|
||||
│ ├── global-setup.ts # Global test setup
|
||||
│ ├── msw-setup.ts # Mock Service Worker setup
|
||||
│ └── test-env.ts # Test environment configuration
|
||||
├── utils/ # Test utilities
|
||||
│ ├── assertions.ts # Custom assertions
|
||||
│ ├── builders/ # Test data builders
|
||||
│ │ └── workflow.builder.ts
|
||||
│ ├── data-generators.ts # Test data generators
|
||||
│ ├── database-utils.ts # Database test utilities
|
||||
│ └── test-helpers.ts # General test helpers
|
||||
├── mocks/ # Mock implementations
|
||||
│ └── n8n-api/ # n8n API mocks
|
||||
│ ├── handlers.ts # MSW request handlers
|
||||
│ └── data/ # Mock data
|
||||
└── fixtures/ # Test fixtures
|
||||
├── database/ # Database fixtures
|
||||
├── factories/ # Data factories
|
||||
└── workflows/ # Workflow fixtures
|
||||
```
|
||||
|
||||
## Mock Strategy
|
||||
|
||||
### 1. Mock Service Worker (MSW) for API Mocking
|
||||
|
||||
We use MSW for intercepting and mocking HTTP requests:
|
||||
|
||||
```typescript
|
||||
// tests/mocks/n8n-api/handlers.ts
|
||||
import { http, HttpResponse } from 'msw';
|
||||
|
||||
export const handlers = [
|
||||
// Workflow endpoints
|
||||
http.get('*/workflows/:id', ({ params }) => {
|
||||
const workflow = mockWorkflows.find(w => w.id === params.id);
|
||||
if (!workflow) {
|
||||
return new HttpResponse(null, { status: 404 });
|
||||
}
|
||||
return HttpResponse.json(workflow);
|
||||
}),
|
||||
|
||||
// Execution endpoints
|
||||
http.post('*/workflows/:id/run', async ({ params, request }) => {
|
||||
const body = await request.json();
|
||||
return HttpResponse.json({
|
||||
executionId: generateExecutionId(),
|
||||
status: 'running'
|
||||
});
|
||||
})
|
||||
];
|
||||
```
|
||||
|
||||
### 2. Database Mocking
|
||||
|
||||
For unit tests, we mock the database layer:
|
||||
|
||||
```typescript
|
||||
// tests/unit/__mocks__/better-sqlite3.ts
|
||||
import { vi } from 'vitest';
|
||||
|
||||
export default vi.fn(() => ({
|
||||
prepare: vi.fn(() => ({
|
||||
all: vi.fn().mockReturnValue([]),
|
||||
get: vi.fn().mockReturnValue(undefined),
|
||||
run: vi.fn().mockReturnValue({ changes: 1 }),
|
||||
finalize: vi.fn()
|
||||
})),
|
||||
exec: vi.fn(),
|
||||
close: vi.fn(),
|
||||
pragma: vi.fn()
|
||||
}));
|
||||
```
|
||||
|
||||
### 3. MCP SDK Mocking
|
||||
|
||||
For testing MCP protocol interactions:
|
||||
|
||||
```typescript
|
||||
// tests/integration/mcp-protocol/test-helpers.ts
|
||||
export class TestableN8NMCPServer extends N8NMCPServer {
|
||||
private transports = new Set<Transport>();
|
||||
|
||||
async connectToTransport(transport: Transport): Promise<void> {
|
||||
this.transports.add(transport);
|
||||
await this.connect(transport);
|
||||
}
|
||||
|
||||
async close(): Promise<void> {
|
||||
for (const transport of this.transports) {
|
||||
await transport.close();
|
||||
}
|
||||
this.transports.clear();
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Test Patterns and Utilities
|
||||
|
||||
### 1. Database Test Utilities
|
||||
|
||||
```typescript
|
||||
// tests/utils/database-utils.ts
|
||||
export class TestDatabase {
|
||||
constructor(options: TestDatabaseOptions = {}) {
|
||||
this.options = {
|
||||
mode: 'memory',
|
||||
enableFTS5: true,
|
||||
...options
|
||||
};
|
||||
}
|
||||
|
||||
async initialize(): Promise<Database.Database> {
|
||||
const db = this.options.mode === 'memory'
|
||||
? new Database(':memory:')
|
||||
: new Database(this.dbPath);
|
||||
|
||||
if (this.options.enableFTS5) {
|
||||
await this.enableFTS5(db);
|
||||
}
|
||||
|
||||
return db;
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 2. Data Generators
|
||||
|
||||
```typescript
|
||||
// tests/utils/data-generators.ts
|
||||
export class TestDataGenerator {
|
||||
static generateNode(overrides: Partial<ParsedNode> = {}): ParsedNode {
|
||||
return {
|
||||
nodeType: `test.node${faker.number.int()}`,
|
||||
displayName: faker.commerce.productName(),
|
||||
description: faker.lorem.sentence(),
|
||||
properties: this.generateProperties(5),
|
||||
...overrides
|
||||
};
|
||||
}
|
||||
|
||||
static generateWorkflow(nodeCount = 3): any {
|
||||
const nodes = Array.from({ length: nodeCount }, (_, i) => ({
|
||||
id: `node_${i}`,
|
||||
type: 'test.node',
|
||||
position: [i * 100, 0],
|
||||
parameters: {}
|
||||
}));
|
||||
|
||||
return { nodes, connections: {} };
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 3. Custom Assertions
|
||||
|
||||
```typescript
|
||||
// tests/utils/assertions.ts
|
||||
export function expectValidMCPResponse(response: any): void {
|
||||
expect(response).toBeDefined();
|
||||
expect(response.content).toBeDefined();
|
||||
expect(Array.isArray(response.content)).toBe(true);
|
||||
expect(response.content[0]).toHaveProperty('type', 'text');
|
||||
expect(response.content[0]).toHaveProperty('text');
|
||||
}
|
||||
|
||||
export function expectNodeStructure(node: any): void {
|
||||
expect(node).toHaveProperty('nodeType');
|
||||
expect(node).toHaveProperty('displayName');
|
||||
expect(node).toHaveProperty('properties');
|
||||
expect(Array.isArray(node.properties)).toBe(true);
|
||||
}
|
||||
```
|
||||
|
||||
## Unit Testing
|
||||
|
||||
Our unit tests focus on testing individual components in isolation with mocked dependencies:
|
||||
|
||||
### Service Layer Tests
|
||||
|
||||
The bulk of our unit tests (400+ tests) are in the services layer:
|
||||
|
||||
```typescript
|
||||
// tests/unit/services/workflow-validator-comprehensive.test.ts
|
||||
describe('WorkflowValidator Comprehensive Tests', () => {
|
||||
it('should validate complex workflow with AI nodes', () => {
|
||||
const workflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: 'ai_agent',
|
||||
type: '@n8n/n8n-nodes-langchain.agent',
|
||||
parameters: { prompt: 'Analyze data' }
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result = validator.validateWorkflow(workflow);
|
||||
expect(result.valid).toBe(true);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### Parser Tests
|
||||
|
||||
Testing the node parsing logic:
|
||||
|
||||
```typescript
|
||||
// tests/unit/parsers/property-extractor.test.ts
|
||||
describe('PropertyExtractor', () => {
|
||||
it('should extract nested properties correctly', () => {
|
||||
const node = {
|
||||
properties: [
|
||||
{
|
||||
displayName: 'Options',
|
||||
name: 'options',
|
||||
type: 'collection',
|
||||
options: [
|
||||
{ name: 'timeout', type: 'number' }
|
||||
]
|
||||
}
|
||||
]
|
||||
};
|
||||
|
||||
const extracted = extractor.extractProperties(node);
|
||||
expect(extracted).toHaveProperty('options.timeout');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### Mock Testing
|
||||
|
||||
Testing our mock implementations:
|
||||
|
||||
```typescript
|
||||
// tests/unit/__mocks__/n8n-nodes-base.test.ts
|
||||
describe('n8n-nodes-base mock', () => {
|
||||
it('should provide mocked node definitions', () => {
|
||||
const httpNode = mockNodes['n8n-nodes-base.httpRequest'];
|
||||
expect(httpNode).toBeDefined();
|
||||
expect(httpNode.description.displayName).toBe('HTTP Request');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Integration Testing
|
||||
|
||||
Our integration tests verify the complete system behavior:
|
||||
|
||||
### MCP Protocol Testing
|
||||
|
||||
```typescript
|
||||
// tests/integration/mcp-protocol/tool-invocation.test.ts
|
||||
describe('MCP Tool Invocation', () => {
|
||||
let mcpServer: TestableN8NMCPServer;
|
||||
let client: Client;
|
||||
|
||||
beforeEach(async () => {
|
||||
mcpServer = new TestableN8NMCPServer();
|
||||
await mcpServer.initialize();
|
||||
|
||||
const [serverTransport, clientTransport] = InMemoryTransport.createLinkedPair();
|
||||
await mcpServer.connectToTransport(serverTransport);
|
||||
|
||||
client = new Client({ name: 'test-client', version: '1.0.0' }, {});
|
||||
await client.connect(clientTransport);
|
||||
});
|
||||
|
||||
it('should list nodes with filtering', async () => {
|
||||
const response = await client.callTool({
|
||||
name: 'list_nodes',
|
||||
arguments: { category: 'trigger', limit: 10 }
|
||||
});
|
||||
|
||||
expectValidMCPResponse(response);
|
||||
const result = JSON.parse(response.content[0].text);
|
||||
expect(result.nodes).toHaveLength(10);
|
||||
expect(result.nodes.every(n => n.category === 'trigger')).toBe(true);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### Database Integration Testing
|
||||
|
||||
```typescript
|
||||
// tests/integration/database/fts5-search.test.ts
|
||||
describe('FTS5 Search Integration', () => {
|
||||
it('should perform fuzzy search', async () => {
|
||||
const results = await nodeRepo.searchNodes('HTT', 'FUZZY');
|
||||
|
||||
expect(results.some(n => n.nodeType.includes('httpRequest'))).toBe(true);
|
||||
expect(results.some(n => n.displayName.includes('HTTP'))).toBe(true);
|
||||
});
|
||||
|
||||
it('should handle complex boolean queries', async () => {
|
||||
const results = await nodeRepo.searchNodes('webhook OR http', 'OR');
|
||||
|
||||
expect(results.length).toBeGreaterThan(0);
|
||||
expect(results.some(n =>
|
||||
n.description?.includes('webhook') ||
|
||||
n.description?.includes('http')
|
||||
)).toBe(true);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Test Distribution and Coverage
|
||||
|
||||
### Test Distribution by Component
|
||||
|
||||
Based on our 1,182 tests:
|
||||
|
||||
1. **Services Layer** (~450 tests)
|
||||
- `workflow-validator-comprehensive.test.ts`: 150+ tests
|
||||
- `node-specific-validators.test.ts`: 120+ tests
|
||||
- `n8n-validation.test.ts`: 80+ tests
|
||||
- `n8n-api-client.test.ts`: 60+ tests
|
||||
|
||||
2. **Parsers** (~200 tests)
|
||||
- `simple-parser.test.ts`: 80+ tests
|
||||
- `property-extractor.test.ts`: 70+ tests
|
||||
- `node-parser.test.ts`: 50+ tests
|
||||
|
||||
3. **MCP Integration** (~150 tests)
|
||||
- `tool-invocation.test.ts`: 50+ tests
|
||||
- `error-handling.test.ts`: 40+ tests
|
||||
- `session-management.test.ts`: 30+ tests
|
||||
|
||||
4. **Database** (~300 tests)
|
||||
- Unit tests for repositories: 100+ tests
|
||||
- Integration tests for FTS5 search: 80+ tests
|
||||
- Transaction tests: 60+ tests
|
||||
- Performance tests: 60+ tests
|
||||
|
||||
### Test Execution Performance
|
||||
|
||||
From our CI runs:
|
||||
- **Fastest tests**: Unit tests with mocks (<1ms each)
|
||||
- **Slowest tests**: Integration tests with real database (100-5000ms)
|
||||
- **Average test time**: ~20ms per test
|
||||
- **Total suite execution**: Under 3 minutes in CI
|
||||
|
||||
## CI/CD Pipeline
|
||||
|
||||
Our GitHub Actions workflow runs all tests automatically:
|
||||
|
||||
```yaml
|
||||
# .github/workflows/test.yml
|
||||
name: Test Suite
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
pull_request:
|
||||
branches: [main]
|
||||
|
||||
jobs:
|
||||
test:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: 20
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Run unit tests with coverage
|
||||
run: npm run test:unit -- --coverage
|
||||
|
||||
- name: Run integration tests
|
||||
run: npm run test:integration
|
||||
|
||||
- name: Upload coverage to Codecov
|
||||
uses: codecov/codecov-action@v4
|
||||
```
|
||||
|
||||
### Test Execution Scripts
|
||||
|
||||
```json
|
||||
// package.json
|
||||
{
|
||||
"scripts": {
|
||||
"test": "vitest",
|
||||
"test:unit": "vitest run tests/unit",
|
||||
"test:integration": "vitest run tests/integration --config vitest.config.integration.ts",
|
||||
"test:coverage": "vitest run --coverage",
|
||||
"test:watch": "vitest watch",
|
||||
"test:bench": "vitest bench --config vitest.config.benchmark.ts",
|
||||
"benchmark:ci": "CI=true node scripts/run-benchmarks-ci.js"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### CI Test Results Summary
|
||||
|
||||
From our latest CI run (#41):
|
||||
|
||||
```
|
||||
UNIT TESTS:
|
||||
Test Files 30 passed (30)
|
||||
Tests 932 passed | 1 skipped (933)
|
||||
|
||||
INTEGRATION TESTS:
|
||||
Test Files 14 passed (14)
|
||||
Tests 245 passed | 4 skipped (249)
|
||||
|
||||
TOTAL: 1,177 passed | 5 skipped | 0 failed
|
||||
```
|
||||
|
||||
## Performance Testing
|
||||
|
||||
We use Vitest's built-in benchmark functionality:
|
||||
|
||||
```typescript
|
||||
// tests/benchmarks/database-queries.bench.ts
|
||||
import { bench, describe } from 'vitest';
|
||||
|
||||
describe('Database Query Performance', () => {
|
||||
bench('search nodes by category', async () => {
|
||||
await nodeRepo.getNodesByCategory('trigger');
|
||||
});
|
||||
|
||||
bench('FTS5 search performance', async () => {
|
||||
await nodeRepo.searchNodes('webhook http request', 'AND');
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Environment Configuration
|
||||
|
||||
Test environment is configured via `.env.test`:
|
||||
|
||||
```bash
|
||||
# Test Environment Configuration
|
||||
NODE_ENV=test
|
||||
TEST_DB_PATH=:memory:
|
||||
TEST_PARALLEL=false
|
||||
TEST_MAX_WORKERS=4
|
||||
FEATURE_TEST_COVERAGE=true
|
||||
MSW_ENABLED=true
|
||||
```
|
||||
|
||||
## Key Patterns and Lessons Learned
|
||||
|
||||
### 1. Response Structure Consistency
|
||||
|
||||
All MCP responses follow a specific structure that must be handled correctly:
|
||||
|
||||
```typescript
|
||||
// Common pattern for handling MCP responses
|
||||
const response = await client.callTool({ name: 'list_nodes', arguments: {} });
|
||||
|
||||
// MCP responses have content array with text objects
|
||||
expect(response.content).toBeDefined();
|
||||
expect(response.content[0].type).toBe('text');
|
||||
|
||||
// Parse the actual data
|
||||
const data = JSON.parse(response.content[0].text);
|
||||
```
|
||||
|
||||
### 2. MSW Integration Setup
|
||||
|
||||
Proper MSW setup is crucial for integration tests:
|
||||
|
||||
```typescript
|
||||
// tests/integration/setup/integration-setup.ts
|
||||
import { setupServer } from 'msw/node';
|
||||
import { handlers } from '@tests/mocks/n8n-api/handlers';
|
||||
|
||||
// Create server but don't start it globally
|
||||
const server = setupServer(...handlers);
|
||||
|
||||
beforeAll(async () => {
|
||||
// Only start MSW for integration tests
|
||||
if (process.env.MSW_ENABLED === 'true') {
|
||||
server.listen({ onUnhandledRequest: 'bypass' });
|
||||
}
|
||||
});
|
||||
|
||||
afterAll(async () => {
|
||||
server.close();
|
||||
});
|
||||
```
|
||||
|
||||
### 3. Database Isolation for Parallel Tests
|
||||
|
||||
Each test gets its own database to enable parallel execution:
|
||||
|
||||
```typescript
|
||||
// tests/utils/database-utils.ts
|
||||
export function createTestDatabaseAdapter(
|
||||
db?: Database.Database,
|
||||
options: TestDatabaseOptions = {}
|
||||
): DatabaseAdapter {
|
||||
const database = db || new Database(':memory:');
|
||||
|
||||
// Enable FTS5 if needed
|
||||
if (options.enableFTS5) {
|
||||
database.exec('PRAGMA main.compile_options;');
|
||||
}
|
||||
|
||||
return new DatabaseAdapter(database);
|
||||
}
|
||||
```
|
||||
|
||||
### 4. Environment-Aware Performance Thresholds
|
||||
|
||||
CI environments are slower, so we adjust expectations:
|
||||
|
||||
```typescript
|
||||
// Environment-aware thresholds
|
||||
const getThreshold = (local: number, ci: number) =>
|
||||
process.env.CI ? ci : local;
|
||||
|
||||
it('should respond quickly', async () => {
|
||||
const start = performance.now();
|
||||
await someOperation();
|
||||
const duration = performance.now() - start;
|
||||
|
||||
expect(duration).toBeLessThan(getThreshold(50, 200));
|
||||
});
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
### 1. Test Isolation
|
||||
- Each test creates its own database instance
|
||||
- Tests clean up after themselves
|
||||
- No shared state between tests
|
||||
|
||||
### 2. Proper Cleanup Order
|
||||
```typescript
|
||||
afterEach(async () => {
|
||||
// Close client first to ensure no pending requests
|
||||
await client.close();
|
||||
|
||||
// Give time for client to fully close
|
||||
await new Promise(resolve => setTimeout(resolve, 50));
|
||||
|
||||
// Then close server
|
||||
await mcpServer.close();
|
||||
|
||||
// Finally cleanup database
|
||||
await testDb.cleanup();
|
||||
});
|
||||
```
|
||||
|
||||
### 3. Handle Async Operations Carefully
|
||||
```typescript
|
||||
// Avoid race conditions in cleanup
|
||||
it('should handle disconnection', async () => {
|
||||
// ... test code ...
|
||||
|
||||
// Ensure operations complete before cleanup
|
||||
await transport.close();
|
||||
await new Promise(resolve => setTimeout(resolve, 100));
|
||||
});
|
||||
```
|
||||
|
||||
### 4. Meaningful Test Organization
|
||||
- Group related tests using `describe` blocks
|
||||
- Use descriptive test names that explain the behavior
|
||||
- Follow AAA pattern: Arrange, Act, Assert
|
||||
- Keep tests focused on single behaviors
|
||||
|
||||
## Debugging Tests
|
||||
|
||||
### Running Specific Tests
|
||||
```bash
|
||||
# Run a single test file
|
||||
npm test tests/integration/mcp-protocol/tool-invocation.test.ts
|
||||
|
||||
# Run tests matching a pattern
|
||||
npm test -- --grep "should list nodes"
|
||||
|
||||
# Run with debugging output
|
||||
DEBUG=* npm test
|
||||
```
|
||||
|
||||
### VSCode Integration
|
||||
```json
|
||||
// .vscode/launch.json
|
||||
{
|
||||
"configurations": [
|
||||
{
|
||||
"type": "node",
|
||||
"request": "launch",
|
||||
"name": "Debug Tests",
|
||||
"program": "${workspaceFolder}/node_modules/vitest/vitest.mjs",
|
||||
"args": ["run", "${file}"],
|
||||
"console": "integratedTerminal"
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
## Test Coverage
|
||||
|
||||
While we don't enforce strict coverage thresholds yet, the infrastructure is in place:
|
||||
- Coverage reports generated in `lcov`, `html`, and `text` formats
|
||||
- Integration with Codecov for tracking coverage over time
|
||||
- Per-file coverage visible in VSCode with extensions
|
||||
|
||||
## Future Improvements
|
||||
|
||||
1. **E2E Testing**: Add Playwright for testing the full MCP server interaction
|
||||
2. **Load Testing**: Implement k6 or Artillery for stress testing
|
||||
3. **Contract Testing**: Add Pact for ensuring API compatibility
|
||||
4. **Visual Regression**: For any UI components that may be added
|
||||
5. **Mutation Testing**: Use Stryker to ensure test quality
|
||||
|
||||
## Common Issues and Solutions
|
||||
|
||||
### 1. Tests Hanging in CI
|
||||
|
||||
**Problem**: Tests would hang indefinitely in CI due to `process.exit()` calls.
|
||||
|
||||
**Solution**: Remove all `process.exit()` calls from test code and use proper cleanup:
|
||||
```typescript
|
||||
// Bad
|
||||
afterAll(() => {
|
||||
process.exit(0); // This causes Vitest to hang
|
||||
});
|
||||
|
||||
// Good
|
||||
afterAll(async () => {
|
||||
await cleanup();
|
||||
// Let Vitest handle process termination
|
||||
});
|
||||
```
|
||||
|
||||
### 2. MCP Response Structure
|
||||
|
||||
**Problem**: Tests expecting wrong response format from MCP tools.
|
||||
|
||||
**Solution**: Always access responses through `content[0].text`:
|
||||
```typescript
|
||||
// Wrong
|
||||
const data = response[0].text;
|
||||
|
||||
// Correct
|
||||
const data = JSON.parse(response.content[0].text);
|
||||
```
|
||||
|
||||
### 3. Database Not Found Errors
|
||||
|
||||
**Problem**: Tests failing with "node not found" when database is empty.
|
||||
|
||||
**Solution**: Check for empty databases before assertions:
|
||||
```typescript
|
||||
const stats = await server.executeTool('get_database_statistics', {});
|
||||
if (stats.totalNodes > 0) {
|
||||
expect(result.nodes.length).toBeGreaterThan(0);
|
||||
} else {
|
||||
expect(result.nodes).toHaveLength(0);
|
||||
}
|
||||
```
|
||||
|
||||
### 4. MSW Loading Globally
|
||||
|
||||
**Problem**: MSW interfering with unit tests when loaded globally.
|
||||
|
||||
**Solution**: Only load MSW in integration test setup:
|
||||
```typescript
|
||||
// vitest.config.integration.ts
|
||||
setupFiles: [
|
||||
'./tests/setup/global-setup.ts',
|
||||
'./tests/integration/setup/integration-setup.ts' // MSW only here
|
||||
]
|
||||
```
|
||||
|
||||
## Resources
|
||||
|
||||
- [Vitest Documentation](https://vitest.dev/)
|
||||
- [MSW Documentation](https://mswjs.io/)
|
||||
- [Testing Best Practices](https://github.com/goldbergyoni/javascript-testing-best-practices)
|
||||
- [MCP SDK Documentation](https://modelcontextprotocol.io/)
|
||||
@@ -1,276 +0,0 @@
|
||||
# n8n-MCP Testing Implementation Checklist
|
||||
|
||||
## Test Suite Development Status
|
||||
|
||||
### Context
|
||||
- **Situation**: Building comprehensive test suite from scratch
|
||||
- **Branch**: feat/comprehensive-testing-suite (separate from main)
|
||||
- **Main Branch Status**: Working in production without tests
|
||||
- **Goal**: Add test coverage without disrupting development
|
||||
|
||||
## Immediate Actions (Day 1)
|
||||
|
||||
- [x] ~~Fix failing tests (Phase 0)~~ ✅ COMPLETED
|
||||
- [x] ~~Create GitHub Actions workflow file~~ ✅ COMPLETED
|
||||
- [x] ~~Install Vitest and remove Jest~~ ✅ COMPLETED
|
||||
- [x] ~~Create vitest.config.ts~~ ✅ COMPLETED
|
||||
- [x] ~~Setup global test configuration~~ ✅ COMPLETED
|
||||
- [x] ~~Migrate existing tests to Vitest syntax~~ ✅ COMPLETED
|
||||
- [x] ~~Setup coverage reporting with Codecov~~ ✅ COMPLETED
|
||||
|
||||
## Phase 1: Vitest Migration ✅ COMPLETED
|
||||
|
||||
All tests have been successfully migrated from Jest to Vitest:
|
||||
- ✅ Removed Jest and installed Vitest
|
||||
- ✅ Created vitest.config.ts with path aliases
|
||||
- ✅ Set up global test configuration
|
||||
- ✅ Migrated all 6 test files (68 tests passing)
|
||||
- ✅ Updated TypeScript configuration
|
||||
- ✅ Cleaned up Jest configuration files
|
||||
|
||||
## Week 1: Foundation
|
||||
|
||||
### Testing Infrastructure ✅ COMPLETED (Phase 2)
|
||||
- [x] ~~Create test directory structure~~ ✅ COMPLETED
|
||||
- [x] ~~Setup mock infrastructure for better-sqlite3~~ ✅ COMPLETED
|
||||
- [x] ~~Create mock for n8n-nodes-base package~~ ✅ COMPLETED
|
||||
- [x] ~~Setup test database utilities~~ ✅ COMPLETED
|
||||
- [x] ~~Create factory pattern for nodes~~ ✅ COMPLETED
|
||||
- [x] ~~Create builder pattern for workflows~~ ✅ COMPLETED
|
||||
- [x] ~~Setup global test utilities~~ ✅ COMPLETED
|
||||
- [x] ~~Configure test environment variables~~ ✅ COMPLETED
|
||||
|
||||
### CI/CD Pipeline ✅ COMPLETED (Phase 3.8)
|
||||
- [x] ~~GitHub Actions for test execution~~ ✅ COMPLETED & VERIFIED
|
||||
- Successfully running with Vitest
|
||||
- 1021 tests passing in CI
|
||||
- Build time: ~2 minutes
|
||||
- [x] ~~Coverage reporting integration~~ ✅ COMPLETED (Codecov setup)
|
||||
- [x] ~~Performance benchmark tracking~~ ✅ COMPLETED
|
||||
- [x] ~~Test result artifacts~~ ✅ COMPLETED
|
||||
- [ ] Branch protection rules
|
||||
- [ ] Required status checks
|
||||
|
||||
## Week 2: Mock Infrastructure
|
||||
|
||||
### Database Mocking
|
||||
- [ ] Complete better-sqlite3 mock implementation
|
||||
- [ ] Mock prepared statements
|
||||
- [ ] Mock transactions
|
||||
- [ ] Mock FTS5 search functionality
|
||||
- [ ] Test data seeding utilities
|
||||
|
||||
### External Dependencies
|
||||
- [ ] Mock axios for API calls
|
||||
- [ ] Mock file system operations
|
||||
- [ ] Mock MCP SDK
|
||||
- [ ] Mock Express server
|
||||
- [ ] Mock WebSocket connections
|
||||
|
||||
## Week 3-4: Unit Tests ✅ COMPLETED (Phase 3)
|
||||
|
||||
### Core Services (Priority 1) ✅ COMPLETED
|
||||
- [x] ~~`config-validator.ts` - 95% coverage~~ ✅ 96.9%
|
||||
- [x] ~~`enhanced-config-validator.ts` - 95% coverage~~ ✅ 94.55%
|
||||
- [x] ~~`workflow-validator.ts` - 90% coverage~~ ✅ 97.59%
|
||||
- [x] ~~`expression-validator.ts` - 90% coverage~~ ✅ 97.22%
|
||||
- [x] ~~`property-filter.ts` - 90% coverage~~ ✅ 95.25%
|
||||
- [x] ~~`example-generator.ts` - 85% coverage~~ ✅ 94.34%
|
||||
|
||||
### Parsers (Priority 2) ✅ COMPLETED
|
||||
- [x] ~~`node-parser.ts` - 90% coverage~~ ✅ 97.42%
|
||||
- [x] ~~`property-extractor.ts` - 90% coverage~~ ✅ 95.49%
|
||||
|
||||
### MCP Layer (Priority 3) ✅ COMPLETED
|
||||
- [x] ~~`tools.ts` - 90% coverage~~ ✅ 94.11%
|
||||
- [x] ~~`handlers-n8n-manager.ts` - 85% coverage~~ ✅ 92.71%
|
||||
- [x] ~~`handlers-workflow-diff.ts` - 85% coverage~~ ✅ 96.34%
|
||||
- [x] ~~`tools-documentation.ts` - 80% coverage~~ ✅ 94.12%
|
||||
|
||||
### Database Layer (Priority 4) ✅ COMPLETED
|
||||
- [x] ~~`node-repository.ts` - 85% coverage~~ ✅ 91.48%
|
||||
- [x] ~~`database-adapter.ts` - 85% coverage~~ ✅ 89.29%
|
||||
- [x] ~~`template-repository.ts` - 80% coverage~~ ✅ 86.78%
|
||||
|
||||
### Loaders and Mappers (Priority 5) ✅ COMPLETED
|
||||
- [x] ~~`node-loader.ts` - 85% coverage~~ ✅ 91.89%
|
||||
- [x] ~~`docs-mapper.ts` - 80% coverage~~ ✅ 95.45%
|
||||
|
||||
### Additional Critical Services Tested ✅ COMPLETED (Phase 3.5)
|
||||
- [x] ~~`n8n-api-client.ts`~~ ✅ 83.87%
|
||||
- [x] ~~`workflow-diff-engine.ts`~~ ✅ 90.06%
|
||||
- [x] ~~`n8n-validation.ts`~~ ✅ 97.14%
|
||||
- [x] ~~`node-specific-validators.ts`~~ ✅ 98.7%
|
||||
|
||||
## Week 5-6: Integration Tests 🚧 IN PROGRESS
|
||||
|
||||
### Real Status (July 29, 2025)
|
||||
**Context**: Building test suite from scratch on testing branch. Main branch has no tests.
|
||||
|
||||
**Overall Status**: 187/246 tests passing (76% pass rate)
|
||||
**Critical Issue**: CI shows green despite 58 failing tests due to `|| true` in workflow
|
||||
|
||||
### MCP Protocol Tests 🔄 MIXED STATUS
|
||||
- [x] ~~Full MCP server initialization~~ ✅ COMPLETED
|
||||
- [x] ~~Tool invocation flow~~ ✅ FIXED (30 tests in tool-invocation.test.ts)
|
||||
- [ ] Error handling and recovery ⚠️ 16 FAILING (error-handling.test.ts)
|
||||
- [x] ~~Concurrent request handling~~ ✅ COMPLETED
|
||||
- [ ] Session management ⚠️ 5 FAILING (timeout issues)
|
||||
|
||||
### n8n API Integration 🔄 PENDING
|
||||
- [ ] Workflow CRUD operations (MSW mocks ready)
|
||||
- [ ] Webhook triggering
|
||||
- [ ] Execution monitoring
|
||||
- [ ] Authentication handling
|
||||
- [ ] Error scenarios
|
||||
|
||||
### Database Integration ⚠️ ISSUES FOUND
|
||||
- [x] ~~SQLite operations with real DB~~ ✅ BASIC TESTS PASS
|
||||
- [ ] FTS5 search functionality ⚠️ 7 FAILING (syntax errors)
|
||||
- [ ] Transaction handling ⚠️ 1 FAILING (isolation issues)
|
||||
- [ ] Migration testing 🔄 NOT STARTED
|
||||
- [ ] Performance under load ⚠️ 4 FAILING (slower than thresholds)
|
||||
|
||||
## Week 7-8: E2E & Performance
|
||||
|
||||
### End-to-End Scenarios
|
||||
- [ ] Complete workflow creation flow
|
||||
- [ ] AI agent workflow setup
|
||||
- [ ] Template import and validation
|
||||
- [ ] Workflow execution monitoring
|
||||
- [ ] Error recovery scenarios
|
||||
|
||||
### Performance Benchmarks
|
||||
- [ ] Node loading speed (< 50ms per node)
|
||||
- [ ] Search performance (< 100ms for 1000 nodes)
|
||||
- [ ] Validation speed (< 10ms simple, < 100ms complex)
|
||||
- [ ] Database query performance
|
||||
- [ ] Memory usage profiling
|
||||
- [ ] Concurrent request handling
|
||||
|
||||
### Load Testing
|
||||
- [ ] 100 concurrent MCP requests
|
||||
- [ ] 10,000 nodes in database
|
||||
- [ ] 1,000 workflow validations/minute
|
||||
- [ ] Memory leak detection
|
||||
- [ ] Resource cleanup verification
|
||||
|
||||
## Testing Quality Gates
|
||||
|
||||
### Coverage Requirements
|
||||
- [ ] Overall: 80%+ (Currently: 62.67%)
|
||||
- [x] ~~Core services: 90%+~~ ✅ COMPLETED
|
||||
- [x] ~~MCP tools: 90%+~~ ✅ COMPLETED
|
||||
- [x] ~~Critical paths: 95%+~~ ✅ COMPLETED
|
||||
- [x] ~~New code: 90%+~~ ✅ COMPLETED
|
||||
|
||||
### Performance Requirements
|
||||
- [x] ~~All unit tests < 10ms~~ ✅ COMPLETED
|
||||
- [ ] Integration tests < 1s
|
||||
- [ ] E2E tests < 10s
|
||||
- [x] ~~Full suite < 5 minutes~~ ✅ COMPLETED (~2 minutes)
|
||||
- [x] ~~No memory leaks~~ ✅ COMPLETED
|
||||
|
||||
### Code Quality
|
||||
- [x] ~~No ESLint errors~~ ✅ COMPLETED
|
||||
- [x] ~~No TypeScript errors~~ ✅ COMPLETED
|
||||
- [x] ~~No console.log in tests~~ ✅ COMPLETED
|
||||
- [x] ~~All tests have descriptions~~ ✅ COMPLETED
|
||||
- [x] ~~No hardcoded values~~ ✅ COMPLETED
|
||||
|
||||
## Monitoring & Maintenance
|
||||
|
||||
### Daily
|
||||
- [ ] Check CI pipeline status
|
||||
- [ ] Review failed tests
|
||||
- [ ] Monitor flaky tests
|
||||
|
||||
### Weekly
|
||||
- [ ] Review coverage reports
|
||||
- [ ] Update test documentation
|
||||
- [ ] Performance benchmark review
|
||||
- [ ] Team sync on testing progress
|
||||
|
||||
### Monthly
|
||||
- [ ] Update baseline benchmarks
|
||||
- [ ] Review and refactor tests
|
||||
- [ ] Update testing strategy
|
||||
- [ ] Training/knowledge sharing
|
||||
|
||||
## Risk Mitigation
|
||||
|
||||
### Technical Risks
|
||||
- [ ] Mock complexity - Use simple, maintainable mocks
|
||||
- [ ] Test brittleness - Focus on behavior, not implementation
|
||||
- [ ] Performance impact - Run heavy tests in parallel
|
||||
- [ ] Flaky tests - Proper async handling and isolation
|
||||
|
||||
### Process Risks
|
||||
- [ ] Slow adoption - Provide training and examples
|
||||
- [ ] Coverage gaming - Review test quality, not just numbers
|
||||
- [ ] Maintenance burden - Automate what's possible
|
||||
- [ ] Integration complexity - Use test containers
|
||||
|
||||
## Success Criteria
|
||||
|
||||
### Current Reality Check
|
||||
- **Unit Tests**: ✅ SOLID (932 passing, 87.8% coverage)
|
||||
- **Integration Tests**: ⚠️ NEEDS WORK (58 failing, 76% pass rate)
|
||||
- **E2E Tests**: 🔄 NOT STARTED
|
||||
- **CI/CD**: ⚠️ BROKEN (hiding failures with || true)
|
||||
|
||||
### Revised Technical Metrics
|
||||
- Coverage: Currently 87.8% for unit tests ✅
|
||||
- Integration test pass rate: Target 100% (currently 76%)
|
||||
- Performance: Adjust thresholds based on reality
|
||||
- Reliability: Fix flaky tests during repair
|
||||
- Speed: CI pipeline < 5 minutes ✅ (~2 minutes)
|
||||
|
||||
### Team Metrics
|
||||
- All developers writing tests ✅
|
||||
- Tests reviewed in PRs ✅
|
||||
- No production bugs from tested code
|
||||
- Improved development velocity ✅
|
||||
|
||||
## Phases Completed
|
||||
|
||||
- **Phase 0**: Immediate Fixes ✅ COMPLETED
|
||||
- **Phase 1**: Vitest Migration ✅ COMPLETED
|
||||
- **Phase 2**: Test Infrastructure ✅ COMPLETED
|
||||
- **Phase 3**: Unit Tests (All 943 tests) ✅ COMPLETED
|
||||
- **Phase 3.5**: Critical Service Testing ✅ COMPLETED
|
||||
- **Phase 3.8**: CI/CD & Infrastructure ✅ COMPLETED
|
||||
- **Phase 4**: Integration Tests 🚧 IN PROGRESS
|
||||
- **Status**: 58 out of 246 tests failing (23.6% failure rate)
|
||||
- **CI Issue**: Tests appear green due to `|| true` error suppression
|
||||
- **Categories of Failures**:
|
||||
- Database: 9 tests (state isolation, FTS5 syntax)
|
||||
- MCP Protocol: 16 tests (response structure in error-handling.test.ts)
|
||||
- MSW: 6 tests (not initialized properly)
|
||||
- FTS5 Search: 7 tests (query syntax issues)
|
||||
- Session Management: 5 tests (async cleanup)
|
||||
- Performance: 15 tests (threshold mismatches)
|
||||
- **Next Steps**:
|
||||
1. Get team buy-in for "red" CI
|
||||
2. Remove `|| true` from workflow
|
||||
3. Fix tests systematically by category
|
||||
- **Phase 5**: E2E Tests 🔄 PENDING
|
||||
|
||||
## Resources & Tools
|
||||
|
||||
### Documentation
|
||||
- Vitest: https://vitest.dev/
|
||||
- Testing Library: https://testing-library.com/
|
||||
- MSW: https://mswjs.io/
|
||||
- Testcontainers: https://www.testcontainers.com/
|
||||
|
||||
### Monitoring
|
||||
- Codecov: https://codecov.io/
|
||||
- GitHub Actions: https://github.com/features/actions
|
||||
- Benchmark Action: https://github.com/benchmark-action/github-action-benchmark
|
||||
|
||||
### Team Resources
|
||||
- Testing best practices guide
|
||||
- Example test implementations
|
||||
- Mock usage patterns
|
||||
- Performance optimization tips
|
||||
@@ -1,472 +0,0 @@
|
||||
# n8n-MCP Testing Implementation Guide
|
||||
|
||||
## Phase 1: Foundation Setup (Week 1-2)
|
||||
|
||||
### 1.1 Install Vitest and Dependencies
|
||||
|
||||
```bash
|
||||
# Remove Jest
|
||||
npm uninstall jest ts-jest @types/jest
|
||||
|
||||
# Install Vitest and related packages
|
||||
npm install -D vitest @vitest/ui @vitest/coverage-v8
|
||||
npm install -D @testing-library/jest-dom
|
||||
npm install -D msw # For API mocking
|
||||
npm install -D @faker-js/faker # For test data
|
||||
npm install -D fishery # For factories
|
||||
```
|
||||
|
||||
### 1.2 Update package.json Scripts
|
||||
|
||||
```json
|
||||
{
|
||||
"scripts": {
|
||||
// Testing
|
||||
"test": "vitest",
|
||||
"test:ui": "vitest --ui",
|
||||
"test:unit": "vitest run tests/unit",
|
||||
"test:integration": "vitest run tests/integration",
|
||||
"test:e2e": "vitest run tests/e2e",
|
||||
"test:watch": "vitest watch",
|
||||
"test:coverage": "vitest run --coverage",
|
||||
"test:coverage:check": "vitest run --coverage --coverage.thresholdAutoUpdate=false",
|
||||
|
||||
// Benchmarks
|
||||
"bench": "vitest bench",
|
||||
"bench:compare": "vitest bench --compare",
|
||||
|
||||
// CI specific
|
||||
"test:ci": "vitest run --reporter=junit --reporter=default",
|
||||
"test:ci:coverage": "vitest run --coverage --reporter=junit --reporter=default"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 1.3 Migrate Existing Tests
|
||||
|
||||
```typescript
|
||||
// Before (Jest)
|
||||
import { describe, test, expect } from '@jest/globals';
|
||||
|
||||
// After (Vitest)
|
||||
import { describe, it, expect, vi } from 'vitest';
|
||||
|
||||
// Update mock syntax
|
||||
// Jest: jest.mock('module')
|
||||
// Vitest: vi.mock('module')
|
||||
|
||||
// Update timer mocks
|
||||
// Jest: jest.useFakeTimers()
|
||||
// Vitest: vi.useFakeTimers()
|
||||
```
|
||||
|
||||
### 1.4 Create Test Database Setup
|
||||
|
||||
```typescript
|
||||
// tests/setup/test-database.ts
|
||||
import Database from 'better-sqlite3';
|
||||
import { readFileSync } from 'fs';
|
||||
import { join } from 'path';
|
||||
|
||||
export class TestDatabase {
|
||||
private db: Database.Database;
|
||||
|
||||
constructor() {
|
||||
this.db = new Database(':memory:');
|
||||
this.initialize();
|
||||
}
|
||||
|
||||
private initialize() {
|
||||
const schema = readFileSync(
|
||||
join(__dirname, '../../src/database/schema.sql'),
|
||||
'utf8'
|
||||
);
|
||||
this.db.exec(schema);
|
||||
}
|
||||
|
||||
seedNodes(nodes: any[]) {
|
||||
const stmt = this.db.prepare(`
|
||||
INSERT INTO nodes (type, displayName, name, group, version, description, properties)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?)
|
||||
`);
|
||||
|
||||
const insertMany = this.db.transaction((nodes) => {
|
||||
for (const node of nodes) {
|
||||
stmt.run(
|
||||
node.type,
|
||||
node.displayName,
|
||||
node.name,
|
||||
node.group,
|
||||
node.version,
|
||||
node.description,
|
||||
JSON.stringify(node.properties)
|
||||
);
|
||||
}
|
||||
});
|
||||
|
||||
insertMany(nodes);
|
||||
}
|
||||
|
||||
close() {
|
||||
this.db.close();
|
||||
}
|
||||
|
||||
getDb() {
|
||||
return this.db;
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Phase 2: Core Unit Tests (Week 3-4)
|
||||
|
||||
### 2.1 Test Organization Template
|
||||
|
||||
```typescript
|
||||
// tests/unit/services/[service-name].test.ts
|
||||
import { describe, it, expect, beforeEach, afterEach, vi } from 'vitest';
|
||||
import { ServiceName } from '@/services/service-name';
|
||||
|
||||
describe('ServiceName', () => {
|
||||
let service: ServiceName;
|
||||
let mockDependency: any;
|
||||
|
||||
beforeEach(() => {
|
||||
// Setup mocks
|
||||
mockDependency = {
|
||||
method: vi.fn()
|
||||
};
|
||||
|
||||
// Create service instance
|
||||
service = new ServiceName(mockDependency);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
describe('methodName', () => {
|
||||
it('should handle happy path', async () => {
|
||||
// Arrange
|
||||
const input = { /* test data */ };
|
||||
mockDependency.method.mockResolvedValue({ /* mock response */ });
|
||||
|
||||
// Act
|
||||
const result = await service.methodName(input);
|
||||
|
||||
// Assert
|
||||
expect(result).toEqual(/* expected output */);
|
||||
expect(mockDependency.method).toHaveBeenCalledWith(/* expected args */);
|
||||
});
|
||||
|
||||
it('should handle errors gracefully', async () => {
|
||||
// Arrange
|
||||
mockDependency.method.mockRejectedValue(new Error('Test error'));
|
||||
|
||||
// Act & Assert
|
||||
await expect(service.methodName({})).rejects.toThrow('Expected error message');
|
||||
});
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### 2.2 Mock Strategies by Layer
|
||||
|
||||
#### Database Layer
|
||||
```typescript
|
||||
// tests/unit/database/node-repository.test.ts
|
||||
import { vi } from 'vitest';
|
||||
|
||||
vi.mock('better-sqlite3', () => ({
|
||||
default: vi.fn(() => ({
|
||||
prepare: vi.fn(() => ({
|
||||
all: vi.fn(() => mockData),
|
||||
get: vi.fn((id) => mockData.find(d => d.id === id)),
|
||||
run: vi.fn(() => ({ changes: 1 }))
|
||||
})),
|
||||
exec: vi.fn(),
|
||||
close: vi.fn()
|
||||
}))
|
||||
}));
|
||||
```
|
||||
|
||||
#### External APIs
|
||||
```typescript
|
||||
// tests/unit/services/__mocks__/axios.ts
|
||||
export default {
|
||||
create: vi.fn(() => ({
|
||||
get: vi.fn(() => Promise.resolve({ data: {} })),
|
||||
post: vi.fn(() => Promise.resolve({ data: { id: '123' } })),
|
||||
put: vi.fn(() => Promise.resolve({ data: {} })),
|
||||
delete: vi.fn(() => Promise.resolve({ data: {} }))
|
||||
}))
|
||||
};
|
||||
```
|
||||
|
||||
#### File System
|
||||
```typescript
|
||||
// Use memfs for file system mocking
|
||||
import { vol } from 'memfs';
|
||||
|
||||
vi.mock('fs', () => vol);
|
||||
|
||||
beforeEach(() => {
|
||||
vol.reset();
|
||||
vol.fromJSON({
|
||||
'/test/file.json': JSON.stringify({ test: 'data' })
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### 2.3 Critical Path Tests
|
||||
|
||||
```typescript
|
||||
// Priority 1: Node Loading and Parsing
|
||||
// tests/unit/loaders/node-loader.test.ts
|
||||
|
||||
// Priority 2: Configuration Validation
|
||||
// tests/unit/services/config-validator.test.ts
|
||||
|
||||
// Priority 3: MCP Tools
|
||||
// tests/unit/mcp/tools.test.ts
|
||||
|
||||
// Priority 4: Database Operations
|
||||
// tests/unit/database/node-repository.test.ts
|
||||
|
||||
// Priority 5: Workflow Validation
|
||||
// tests/unit/services/workflow-validator.test.ts
|
||||
```
|
||||
|
||||
## Phase 3: Integration Tests (Week 5-6)
|
||||
|
||||
### 3.1 Test Container Setup
|
||||
|
||||
```typescript
|
||||
// tests/setup/test-containers.ts
|
||||
import { GenericContainer, StartedTestContainer } from 'testcontainers';
|
||||
|
||||
export class N8nTestContainer {
|
||||
private container: StartedTestContainer;
|
||||
|
||||
async start() {
|
||||
this.container = await new GenericContainer('n8nio/n8n:latest')
|
||||
.withExposedPorts(5678)
|
||||
.withEnv('N8N_BASIC_AUTH_ACTIVE', 'false')
|
||||
.withEnv('N8N_ENCRYPTION_KEY', 'test-key')
|
||||
.start();
|
||||
|
||||
return {
|
||||
url: `http://localhost:${this.container.getMappedPort(5678)}`,
|
||||
stop: () => this.container.stop()
|
||||
};
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### 3.2 Integration Test Pattern
|
||||
|
||||
```typescript
|
||||
// tests/integration/n8n-api/workflow-crud.test.ts
|
||||
import { N8nTestContainer } from '@tests/setup/test-containers';
|
||||
import { N8nAPIClient } from '@/services/n8n-api-client';
|
||||
|
||||
describe('n8n API Integration', () => {
|
||||
let container: any;
|
||||
let apiClient: N8nAPIClient;
|
||||
|
||||
beforeAll(async () => {
|
||||
container = await new N8nTestContainer().start();
|
||||
apiClient = new N8nAPIClient(container.url);
|
||||
}, 30000);
|
||||
|
||||
afterAll(async () => {
|
||||
await container.stop();
|
||||
});
|
||||
|
||||
it('should create and retrieve workflow', async () => {
|
||||
// Create workflow
|
||||
const workflow = createTestWorkflow();
|
||||
const created = await apiClient.createWorkflow(workflow);
|
||||
|
||||
expect(created.id).toBeDefined();
|
||||
|
||||
// Retrieve workflow
|
||||
const retrieved = await apiClient.getWorkflow(created.id);
|
||||
expect(retrieved.name).toBe(workflow.name);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Phase 4: E2E & Performance (Week 7-8)
|
||||
|
||||
### 4.1 E2E Test Setup
|
||||
|
||||
```typescript
|
||||
// tests/e2e/workflows/complete-workflow.test.ts
|
||||
import { MCPClient } from '@tests/utils/mcp-client';
|
||||
import { N8nTestContainer } from '@tests/setup/test-containers';
|
||||
|
||||
describe('Complete Workflow E2E', () => {
|
||||
let mcpServer: any;
|
||||
let n8nContainer: any;
|
||||
let mcpClient: MCPClient;
|
||||
|
||||
beforeAll(async () => {
|
||||
// Start n8n
|
||||
n8nContainer = await new N8nTestContainer().start();
|
||||
|
||||
// Start MCP server
|
||||
mcpServer = await startMCPServer({
|
||||
n8nUrl: n8nContainer.url
|
||||
});
|
||||
|
||||
// Create MCP client
|
||||
mcpClient = new MCPClient(mcpServer.url);
|
||||
}, 60000);
|
||||
|
||||
it('should execute complete workflow creation flow', async () => {
|
||||
// 1. Search for nodes
|
||||
const searchResult = await mcpClient.call('search_nodes', {
|
||||
query: 'webhook http slack'
|
||||
});
|
||||
|
||||
// 2. Get node details
|
||||
const webhookInfo = await mcpClient.call('get_node_info', {
|
||||
nodeType: 'nodes-base.webhook'
|
||||
});
|
||||
|
||||
// 3. Create workflow
|
||||
const workflow = new WorkflowBuilder('E2E Test')
|
||||
.addWebhookNode()
|
||||
.addHttpRequestNode()
|
||||
.addSlackNode()
|
||||
.connectSequentially()
|
||||
.build();
|
||||
|
||||
// 4. Validate workflow
|
||||
const validation = await mcpClient.call('validate_workflow', {
|
||||
workflow
|
||||
});
|
||||
|
||||
expect(validation.isValid).toBe(true);
|
||||
|
||||
// 5. Deploy to n8n
|
||||
const deployed = await mcpClient.call('n8n_create_workflow', {
|
||||
...workflow
|
||||
});
|
||||
|
||||
expect(deployed.id).toBeDefined();
|
||||
expect(deployed.active).toBe(false);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### 4.2 Performance Benchmarks
|
||||
|
||||
```typescript
|
||||
// vitest.benchmark.config.ts
|
||||
export default {
|
||||
test: {
|
||||
benchmark: {
|
||||
// Output benchmark results
|
||||
outputFile: './benchmark-results.json',
|
||||
|
||||
// Compare with baseline
|
||||
compare: './benchmark-baseline.json',
|
||||
|
||||
// Fail if performance degrades by more than 10%
|
||||
threshold: {
|
||||
p95: 1.1, // 110% of baseline
|
||||
p99: 1.2 // 120% of baseline
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
```
|
||||
|
||||
## Testing Best Practices
|
||||
|
||||
### 1. Test Naming Convention
|
||||
```typescript
|
||||
// Format: should [expected behavior] when [condition]
|
||||
it('should return user data when valid ID is provided')
|
||||
it('should throw ValidationError when email is invalid')
|
||||
it('should retry 3 times when network fails')
|
||||
```
|
||||
|
||||
### 2. Test Data Builders
|
||||
```typescript
|
||||
// Use builders for complex test data
|
||||
const user = new UserBuilder()
|
||||
.withEmail('test@example.com')
|
||||
.withRole('admin')
|
||||
.build();
|
||||
```
|
||||
|
||||
### 3. Custom Matchers
|
||||
```typescript
|
||||
// tests/utils/matchers.ts
|
||||
export const toBeValidNode = (received: any) => {
|
||||
const pass =
|
||||
received.type &&
|
||||
received.displayName &&
|
||||
received.properties &&
|
||||
Array.isArray(received.properties);
|
||||
|
||||
return {
|
||||
pass,
|
||||
message: () => `expected ${received} to be a valid node`
|
||||
};
|
||||
};
|
||||
|
||||
// Usage
|
||||
expect(node).toBeValidNode();
|
||||
```
|
||||
|
||||
### 4. Snapshot Testing
|
||||
```typescript
|
||||
// For complex structures
|
||||
it('should generate correct node schema', () => {
|
||||
const schema = generateNodeSchema(node);
|
||||
expect(schema).toMatchSnapshot();
|
||||
});
|
||||
```
|
||||
|
||||
### 5. Test Isolation
|
||||
```typescript
|
||||
// Always clean up after tests
|
||||
afterEach(async () => {
|
||||
await cleanup();
|
||||
vi.clearAllMocks();
|
||||
vi.restoreAllMocks();
|
||||
});
|
||||
```
|
||||
|
||||
## Coverage Goals by Module
|
||||
|
||||
| Module | Target | Priority | Notes |
|
||||
|--------|--------|----------|-------|
|
||||
| services/config-validator | 95% | High | Critical for reliability |
|
||||
| services/workflow-validator | 90% | High | Core functionality |
|
||||
| mcp/tools | 90% | High | User-facing API |
|
||||
| database/node-repository | 85% | Medium | Well-tested DB layer |
|
||||
| loaders/node-loader | 85% | Medium | External dependencies |
|
||||
| parsers/* | 90% | High | Data transformation |
|
||||
| utils/* | 80% | Low | Helper functions |
|
||||
| scripts/* | 50% | Low | One-time scripts |
|
||||
|
||||
## Continuous Improvement
|
||||
|
||||
1. **Weekly Reviews**: Review test coverage and identify gaps
|
||||
2. **Performance Baselines**: Update benchmarks monthly
|
||||
3. **Flaky Test Detection**: Monitor and fix within 48 hours
|
||||
4. **Test Documentation**: Keep examples updated
|
||||
5. **Developer Training**: Pair programming on tests
|
||||
|
||||
## Success Metrics
|
||||
|
||||
- [ ] All tests pass in CI (0 failures)
|
||||
- [ ] Coverage > 80% overall
|
||||
- [ ] No flaky tests
|
||||
- [ ] CI runs < 5 minutes
|
||||
- [ ] Performance benchmarks stable
|
||||
- [ ] Zero production bugs from tested code
|
||||
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
@@ -1,66 +0,0 @@
|
||||
# Token Efficiency Improvements Summary
|
||||
|
||||
## Overview
|
||||
Made all MCP tool descriptions concise and token-efficient while preserving essential information.
|
||||
|
||||
## Key Improvements
|
||||
|
||||
### Before vs After Examples
|
||||
|
||||
1. **search_nodes**
|
||||
- Before: ~350 chars with verbose explanation
|
||||
- After: 165 chars
|
||||
- `Search nodes by keywords. Modes: OR (any word), AND (all words), FUZZY (typos OK). Primary nodes ranked first. Examples: "webhook"→Webhook, "http call"→HTTP Request.`
|
||||
|
||||
2. **get_node_info**
|
||||
- Before: ~450 chars with warnings about size
|
||||
- After: 174 chars
|
||||
- `Get FULL node schema (100KB+). TIP: Use get_node_essentials first! Returns all properties/operations/credentials. Prefix required: "nodes-base.httpRequest" not "httpRequest".`
|
||||
|
||||
3. **validate_node_minimal**
|
||||
- Before: ~350 chars explaining what it doesn't do
|
||||
- After: 102 chars
|
||||
- `Fast check for missing required fields only. No warnings/suggestions. Returns: list of missing fields.`
|
||||
|
||||
4. **get_property_dependencies**
|
||||
- Before: ~400 chars with full example
|
||||
- After: 131 chars
|
||||
- `Shows property dependencies and visibility rules. Example: sendBody=true reveals body fields. Test visibility with optional config.`
|
||||
|
||||
## Statistics
|
||||
|
||||
### Documentation Tools (22 tools)
|
||||
- Average description length: **129 characters**
|
||||
- Total characters: 2,836
|
||||
- Tools over 200 chars: 1 (list_nodes at 204)
|
||||
|
||||
### Management Tools (17 tools)
|
||||
- Average description length: **93 characters**
|
||||
- Total characters: 1,578
|
||||
- Tools over 200 chars: 1 (n8n_update_partial_workflow at 284)
|
||||
|
||||
## Strategy Used
|
||||
|
||||
1. **Remove redundancy**: Eliminated repeated information available in parameter descriptions
|
||||
2. **Use abbreviations**: "vs" instead of "versus", "&" instead of "and" where appropriate
|
||||
3. **Compact examples**: `"webhook"→Webhook` instead of verbose explanations
|
||||
4. **Direct language**: "Fast check" instead of "Quick validation that only checks"
|
||||
5. **Move details to documentation**: Complex tools reference `tools_documentation()` for full details
|
||||
6. **Essential info only**: Focus on what the tool does, not how it works internally
|
||||
|
||||
## Special Cases
|
||||
|
||||
### n8n_update_partial_workflow
|
||||
This tool's description is necessarily longer (284 chars) because:
|
||||
- Lists all 13 operation types
|
||||
- Critical for users to know available operations
|
||||
- Directs to full documentation for details
|
||||
|
||||
### Complex Documentation Preserved
|
||||
For tools like `n8n_update_partial_workflow`, detailed documentation was moved to `tools-documentation.ts` rather than deleted, ensuring users can still access comprehensive information when needed.
|
||||
|
||||
## Impact
|
||||
- **Token savings**: ~65-70% reduction in description tokens
|
||||
- **Faster AI responses**: Less context used for tool descriptions
|
||||
- **Better UX**: Clearer, more scannable tool list
|
||||
- **Maintained functionality**: All essential information preserved
|
||||
@@ -1,118 +0,0 @@
|
||||
# Transactional Updates Example
|
||||
|
||||
This example demonstrates the new transactional update capabilities in v2.7.0.
|
||||
|
||||
## Before (v2.6.x and earlier)
|
||||
|
||||
Previously, you had to carefully order operations to ensure nodes existed before connecting them:
|
||||
|
||||
```json
|
||||
{
|
||||
"id": "workflow-123",
|
||||
"operations": [
|
||||
// 1. First add all nodes
|
||||
{ "type": "addNode", "node": { "name": "Process", "type": "n8n-nodes-base.set", ... }},
|
||||
{ "type": "addNode", "node": { "name": "Notify", "type": "n8n-nodes-base.slack", ... }},
|
||||
|
||||
// 2. Then add connections (would fail if done before nodes)
|
||||
{ "type": "addConnection", "source": "Webhook", "target": "Process" },
|
||||
{ "type": "addConnection", "source": "Process", "target": "Notify" }
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
## After (v2.7.0+)
|
||||
|
||||
Now you can write operations in any order - the engine automatically handles dependencies:
|
||||
|
||||
```json
|
||||
{
|
||||
"id": "workflow-123",
|
||||
"operations": [
|
||||
// Connections can come first!
|
||||
{ "type": "addConnection", "source": "Webhook", "target": "Process" },
|
||||
{ "type": "addConnection", "source": "Process", "target": "Notify" },
|
||||
|
||||
// Nodes added later - still works!
|
||||
{ "type": "addNode", "node": { "name": "Process", "type": "n8n-nodes-base.set", "position": [400, 300] }},
|
||||
{ "type": "addNode", "node": { "name": "Notify", "type": "n8n-nodes-base.slack", "position": [600, 300] }}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
## How It Works
|
||||
|
||||
1. **Two-Pass Processing**:
|
||||
- Pass 1: All node operations (add, remove, update, move, enable, disable)
|
||||
- Pass 2: All other operations (connections, settings, metadata)
|
||||
|
||||
2. **Operation Limit**: Maximum 5 operations per request keeps complexity manageable
|
||||
|
||||
3. **Atomic Updates**: All operations succeed or all fail - no partial updates
|
||||
|
||||
## Benefits for AI Agents
|
||||
|
||||
- **Intuitive**: Write operations in the order that makes sense logically
|
||||
- **Reliable**: No need to track dependencies manually
|
||||
- **Simple**: Focus on what to change, not how to order changes
|
||||
- **Safe**: Built-in limits prevent overly complex operations
|
||||
|
||||
## Complete Example
|
||||
|
||||
Here's a real-world example of adding error handling to a workflow:
|
||||
|
||||
```json
|
||||
{
|
||||
"id": "workflow-123",
|
||||
"operations": [
|
||||
// Define the flow first (makes logical sense)
|
||||
{
|
||||
"type": "removeConnection",
|
||||
"source": "HTTP Request",
|
||||
"target": "Save to DB"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "HTTP Request",
|
||||
"target": "Error Handler"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Error Handler",
|
||||
"target": "Send Alert"
|
||||
},
|
||||
|
||||
// Then add the nodes
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Error Handler",
|
||||
"type": "n8n-nodes-base.if",
|
||||
"position": [500, 400],
|
||||
"parameters": {
|
||||
"conditions": {
|
||||
"boolean": [{
|
||||
"value1": "={{$json.error}}",
|
||||
"value2": true
|
||||
}]
|
||||
}
|
||||
}
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Send Alert",
|
||||
"type": "n8n-nodes-base.emailSend",
|
||||
"position": [700, 400],
|
||||
"parameters": {
|
||||
"to": "alerts@company.com",
|
||||
"subject": "Workflow Error Alert"
|
||||
}
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
All operations will be processed correctly, even though connections reference nodes that don't exist yet!
|
||||
@@ -1,92 +0,0 @@
|
||||
# Validation Improvements v2.4.2
|
||||
|
||||
Based on AI agent feedback, we've implemented several improvements to the `validate_node_operation` tool:
|
||||
|
||||
## 🎯 Issues Addressed
|
||||
|
||||
### 1. **@version Warnings** ✅ FIXED
|
||||
- **Issue**: Showed confusing warnings about `@version` property not being used
|
||||
- **Fix**: Filter out internal properties starting with `@` or `_`
|
||||
- **Result**: No more false warnings about internal n8n properties
|
||||
|
||||
### 2. **Duplicate Errors** ✅ FIXED
|
||||
- **Issue**: Same error shown multiple times (e.g., missing `ts` field)
|
||||
- **Fix**: Implemented deduplication that keeps the most specific error message
|
||||
- **Result**: Each error shown only once with the best description
|
||||
|
||||
### 3. **Basic Code Validation** ✅ ADDED
|
||||
- **Issue**: No syntax validation for Code node
|
||||
- **Fix**: Added basic syntax checks for JavaScript and Python
|
||||
- **Features**:
|
||||
- Unbalanced braces/parentheses detection
|
||||
- Python indentation consistency check
|
||||
- n8n-specific patterns (return statement, input access)
|
||||
- Security warnings (eval/exec usage)
|
||||
|
||||
## 📊 Before & After
|
||||
|
||||
### Before (v2.4.1):
|
||||
```json
|
||||
{
|
||||
"errors": [
|
||||
{ "property": "ts", "message": "Required property 'Message Timestamp' is missing" },
|
||||
{ "property": "ts", "message": "Message timestamp (ts) is required to update a message" }
|
||||
],
|
||||
"warnings": [
|
||||
{ "property": "@version", "message": "Property '@version' is configured but won't be used" }
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
### After (v2.4.2):
|
||||
```json
|
||||
{
|
||||
"errors": [
|
||||
{ "property": "ts", "message": "Message timestamp (ts) is required to update a message",
|
||||
"fix": "Provide the timestamp of the message to update" }
|
||||
],
|
||||
"warnings": [] // No @version warning
|
||||
}
|
||||
```
|
||||
|
||||
## 🆕 Code Validation Examples
|
||||
|
||||
### JavaScript Syntax Check:
|
||||
```javascript
|
||||
// Missing closing brace
|
||||
if (true) {
|
||||
return items;
|
||||
// Error: "Unbalanced braces detected"
|
||||
```
|
||||
|
||||
### Python Indentation Check:
|
||||
```python
|
||||
def process():
|
||||
if True: # Tab
|
||||
return items # Spaces
|
||||
# Error: "Mixed tabs and spaces in indentation"
|
||||
```
|
||||
|
||||
### n8n Pattern Check:
|
||||
```javascript
|
||||
const result = items.map(item => item.json);
|
||||
// Warning: "No return statement found"
|
||||
// Suggestion: "Add: return items;"
|
||||
```
|
||||
|
||||
## 🚀 Impact
|
||||
|
||||
- **Cleaner validation results** - No more noise from internal properties
|
||||
- **Clearer error messages** - Each issue reported once with best description
|
||||
- **Better code quality** - Basic syntax validation catches common mistakes
|
||||
- **n8n best practices** - Warns about missing return statements and input handling
|
||||
|
||||
## 📝 Summary
|
||||
|
||||
The `validate_node_operation` tool is now even more helpful for AI agents and developers:
|
||||
- 95% reduction in false positives (operation-aware)
|
||||
- No duplicate or confusing warnings
|
||||
- Basic code validation for common syntax errors
|
||||
- n8n-specific pattern checking
|
||||
|
||||
**Rating improved from 9/10 to 9.5/10!** 🎉
|
||||
@@ -116,17 +116,46 @@ The `n8n_update_partial_workflow` tool allows you to make targeted changes to wo
|
||||
}
|
||||
```
|
||||
|
||||
#### Update Connection (Change routing)
|
||||
#### Rewire Connection
|
||||
```json
|
||||
{
|
||||
"type": "updateConnection",
|
||||
"type": "rewireConnection",
|
||||
"source": "Webhook",
|
||||
"from": "Old Handler",
|
||||
"to": "New Handler",
|
||||
"description": "Rewire connection to new handler"
|
||||
}
|
||||
```
|
||||
|
||||
#### Smart Parameters for IF Nodes
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "IF",
|
||||
"target": "Send Email",
|
||||
"changes": {
|
||||
"sourceOutput": "false", // Change from 'true' to 'false' output
|
||||
"targetInput": "main"
|
||||
},
|
||||
"description": "Route failed conditions to email"
|
||||
"target": "Success Handler",
|
||||
"branch": "true", // Semantic parameter instead of sourceIndex
|
||||
"description": "Route true branch to success handler"
|
||||
}
|
||||
```
|
||||
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "IF",
|
||||
"target": "Error Handler",
|
||||
"branch": "false", // Routes to false branch (sourceIndex=1)
|
||||
"description": "Route false branch to error handler"
|
||||
}
|
||||
```
|
||||
|
||||
#### Smart Parameters for Switch Nodes
|
||||
```json
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Switch",
|
||||
"target": "Handler A",
|
||||
"case": 0, // First output
|
||||
"description": "Route case 0 to Handler A"
|
||||
}
|
||||
```
|
||||
|
||||
@@ -296,6 +325,193 @@ The `n8n_update_partial_workflow` tool allows you to make targeted changes to wo
|
||||
}
|
||||
```
|
||||
|
||||
### Example 5: Large Batch Workflow Refactoring
|
||||
Demonstrates handling many operations in a single request - no longer limited to 5 operations!
|
||||
|
||||
```json
|
||||
{
|
||||
"id": "workflow-batch",
|
||||
"operations": [
|
||||
// Add 10 processing nodes
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Filter Active Users",
|
||||
"type": "n8n-nodes-base.filter",
|
||||
"position": [400, 200],
|
||||
"parameters": { "conditions": { "boolean": [{ "value1": "={{$json.active}}", "value2": true }] } }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Transform User Data",
|
||||
"type": "n8n-nodes-base.set",
|
||||
"position": [600, 200],
|
||||
"parameters": { "values": { "string": [{ "name": "formatted_name", "value": "={{$json.firstName}} {{$json.lastName}}" }] } }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Validate Email",
|
||||
"type": "n8n-nodes-base.if",
|
||||
"position": [800, 200],
|
||||
"parameters": { "conditions": { "string": [{ "value1": "={{$json.email}}", "operation": "contains", "value2": "@" }] } }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Enrich with API",
|
||||
"type": "n8n-nodes-base.httpRequest",
|
||||
"position": [1000, 150],
|
||||
"parameters": { "url": "https://api.example.com/enrich", "method": "POST" }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Log Invalid Emails",
|
||||
"type": "n8n-nodes-base.code",
|
||||
"position": [1000, 350],
|
||||
"parameters": { "jsCode": "console.log('Invalid email:', $json.email);\nreturn $json;" }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Merge Results",
|
||||
"type": "n8n-nodes-base.merge",
|
||||
"position": [1200, 250]
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Deduplicate",
|
||||
"type": "n8n-nodes-base.removeDuplicates",
|
||||
"position": [1400, 250],
|
||||
"parameters": { "propertyName": "id" }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Sort by Date",
|
||||
"type": "n8n-nodes-base.sort",
|
||||
"position": [1600, 250],
|
||||
"parameters": { "sortFieldsUi": { "sortField": [{ "fieldName": "created_at", "order": "descending" }] } }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Batch for DB",
|
||||
"type": "n8n-nodes-base.splitInBatches",
|
||||
"position": [1800, 250],
|
||||
"parameters": { "batchSize": 100 }
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addNode",
|
||||
"node": {
|
||||
"name": "Save to Database",
|
||||
"type": "n8n-nodes-base.postgres",
|
||||
"position": [2000, 250],
|
||||
"parameters": { "operation": "insert", "table": "processed_users" }
|
||||
}
|
||||
},
|
||||
// Connect all the nodes
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Get Users",
|
||||
"target": "Filter Active Users"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Filter Active Users",
|
||||
"target": "Transform User Data"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Transform User Data",
|
||||
"target": "Validate Email"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Validate Email",
|
||||
"sourceOutput": "true",
|
||||
"target": "Enrich with API"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Validate Email",
|
||||
"sourceOutput": "false",
|
||||
"target": "Log Invalid Emails"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Enrich with API",
|
||||
"target": "Merge Results"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Log Invalid Emails",
|
||||
"target": "Merge Results",
|
||||
"targetInput": "input2"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Merge Results",
|
||||
"target": "Deduplicate"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Deduplicate",
|
||||
"target": "Sort by Date"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Sort by Date",
|
||||
"target": "Batch for DB"
|
||||
},
|
||||
{
|
||||
"type": "addConnection",
|
||||
"source": "Batch for DB",
|
||||
"target": "Save to Database"
|
||||
},
|
||||
// Update workflow metadata
|
||||
{
|
||||
"type": "updateName",
|
||||
"name": "User Processing Pipeline v2"
|
||||
},
|
||||
{
|
||||
"type": "updateSettings",
|
||||
"settings": {
|
||||
"executionOrder": "v1",
|
||||
"timezone": "UTC",
|
||||
"saveDataSuccessExecution": "all"
|
||||
}
|
||||
},
|
||||
{
|
||||
"type": "addTag",
|
||||
"tag": "production"
|
||||
},
|
||||
{
|
||||
"type": "addTag",
|
||||
"tag": "user-processing"
|
||||
},
|
||||
{
|
||||
"type": "addTag",
|
||||
"tag": "v2"
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
This example shows 26 operations in a single request, creating a complete data processing pipeline with proper error handling, validation, and batch processing.
|
||||
|
||||
## Best Practices
|
||||
|
||||
1. **Use Descriptive Names**: Always provide clear node names and descriptions for operations
|
||||
@@ -390,13 +606,13 @@ The tool validates all operations before applying any changes. Common errors inc
|
||||
|
||||
Always check the response for validation errors and adjust your operations accordingly.
|
||||
|
||||
## Transactional Updates (v2.7.0+)
|
||||
## Transactional Updates
|
||||
|
||||
The diff engine now supports transactional updates using a **two-pass processing** approach:
|
||||
|
||||
### How It Works
|
||||
|
||||
1. **Operation Limit**: Maximum 5 operations per request to ensure reliability
|
||||
1. **No Operation Limit**: Process unlimited operations in a single request
|
||||
2. **Two-Pass Processing**:
|
||||
- **Pass 1**: All node operations (add, remove, update, move, enable, disable)
|
||||
- **Pass 2**: All other operations (connections, settings, metadata)
|
||||
@@ -446,9 +662,9 @@ This allows you to add nodes and connect them in the same request:
|
||||
### Benefits
|
||||
|
||||
- **Order Independence**: You don't need to worry about operation order
|
||||
- **Atomic Updates**: All operations succeed or all fail
|
||||
- **Atomic Updates**: All operations succeed or all fail (unless continueOnError is enabled)
|
||||
- **Intuitive Usage**: Add complex workflow structures in one call
|
||||
- **Clear Limits**: 5 operations max keeps things simple and reliable
|
||||
- **No Hard Limits**: Process unlimited operations efficiently
|
||||
|
||||
### Example: Complete Workflow Addition
|
||||
|
||||
@@ -507,4 +723,4 @@ This allows you to add nodes and connect them in the same request:
|
||||
}
|
||||
```
|
||||
|
||||
All 5 operations will be processed correctly regardless of order!
|
||||
All operations will be processed correctly regardless of order!
|
||||
15435
fetch_log.txt
Normal file
15435
fetch_log.txt
Normal file
File diff suppressed because one or more lines are too long
32
monitor_fetch.sh
Normal file
32
monitor_fetch.sh
Normal file
@@ -0,0 +1,32 @@
|
||||
#!/bin/bash
|
||||
|
||||
echo "Monitoring template fetch progress..."
|
||||
echo "=================================="
|
||||
|
||||
while true; do
|
||||
# Check if process is still running
|
||||
if ! pgrep -f "fetch-templates" > /dev/null; then
|
||||
echo "Fetch process completed!"
|
||||
break
|
||||
fi
|
||||
|
||||
# Get database size
|
||||
DB_SIZE=$(ls -lh data/nodes.db 2>/dev/null | awk '{print $5}')
|
||||
|
||||
# Get template count
|
||||
TEMPLATE_COUNT=$(sqlite3 data/nodes.db "SELECT COUNT(*) FROM templates" 2>/dev/null || echo "0")
|
||||
|
||||
# Get last log entry
|
||||
LAST_LOG=$(tail -n 1 fetch_log.txt 2>/dev/null | grep "Fetching template details" | tail -1)
|
||||
|
||||
# Display status
|
||||
echo -ne "\rDB Size: $DB_SIZE | Templates: $TEMPLATE_COUNT | $LAST_LOG"
|
||||
|
||||
sleep 5
|
||||
done
|
||||
|
||||
echo ""
|
||||
echo "Final statistics:"
|
||||
echo "-----------------"
|
||||
ls -lh data/nodes.db
|
||||
sqlite3 data/nodes.db "SELECT COUNT(*) as count, printf('%.1f MB', SUM(LENGTH(workflow_json_compressed))/1024.0/1024.0) as compressed_size FROM templates"
|
||||
0
n8n-nodes.db
Normal file
0
n8n-nodes.db
Normal file
22143
package-lock.json
generated
22143
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
36
package.json
36
package.json
@@ -1,8 +1,16 @@
|
||||
{
|
||||
"name": "n8n-mcp",
|
||||
"version": "2.10.0",
|
||||
"version": "2.20.5",
|
||||
"description": "Integration between n8n workflow automation and Model Context Protocol (MCP)",
|
||||
"main": "dist/index.js",
|
||||
"types": "dist/index.d.ts",
|
||||
"exports": {
|
||||
".": {
|
||||
"types": "./dist/index.d.ts",
|
||||
"require": "./dist/index.js",
|
||||
"import": "./dist/index.js"
|
||||
}
|
||||
},
|
||||
"bin": {
|
||||
"n8n-mcp": "./dist/mcp/index.js"
|
||||
},
|
||||
@@ -31,12 +39,16 @@
|
||||
"test:watch": "vitest watch",
|
||||
"test:unit": "vitest run tests/unit",
|
||||
"test:integration": "vitest run --config vitest.config.integration.ts",
|
||||
"test:integration:n8n": "vitest run tests/integration/n8n-api",
|
||||
"test:cleanup:orphans": "tsx tests/integration/n8n-api/scripts/cleanup-orphans.ts",
|
||||
"test:e2e": "vitest run tests/e2e",
|
||||
"lint": "tsc --noEmit",
|
||||
"typecheck": "tsc --noEmit",
|
||||
"update:n8n": "node scripts/update-n8n-deps.js",
|
||||
"update:n8n:check": "node scripts/update-n8n-deps.js --dry-run",
|
||||
"fetch:templates": "node dist/scripts/fetch-templates.js",
|
||||
"fetch:templates:update": "node dist/scripts/fetch-templates.js --update",
|
||||
"fetch:templates:extract": "node dist/scripts/fetch-templates.js --extract-only",
|
||||
"fetch:templates:robust": "node dist/scripts/fetch-templates-robust.js",
|
||||
"prebuild:fts5": "npx tsx scripts/prebuild-fts5.ts",
|
||||
"test:templates": "node dist/scripts/test-templates.js",
|
||||
@@ -127,17 +139,27 @@
|
||||
"vitest": "^3.2.4"
|
||||
},
|
||||
"dependencies": {
|
||||
"@modelcontextprotocol/sdk": "^1.13.2",
|
||||
"@n8n/n8n-nodes-langchain": "^1.103.1",
|
||||
"@modelcontextprotocol/sdk": "^1.20.1",
|
||||
"@n8n/n8n-nodes-langchain": "^1.114.1",
|
||||
"@supabase/supabase-js": "^2.57.4",
|
||||
"dotenv": "^16.5.0",
|
||||
"express": "^5.1.0",
|
||||
"n8n": "^1.104.1",
|
||||
"n8n-core": "^1.103.1",
|
||||
"n8n-workflow": "^1.101.0",
|
||||
"express-rate-limit": "^7.1.5",
|
||||
"lru-cache": "^11.2.1",
|
||||
"n8n": "^1.115.2",
|
||||
"n8n-core": "^1.114.0",
|
||||
"n8n-workflow": "^1.112.0",
|
||||
"openai": "^4.77.0",
|
||||
"sql.js": "^1.13.0",
|
||||
"uuid": "^10.0.0"
|
||||
"uuid": "^10.0.0",
|
||||
"zod": "^3.24.1"
|
||||
},
|
||||
"optionalDependencies": {
|
||||
"@rollup/rollup-darwin-arm64": "^4.50.0",
|
||||
"@rollup/rollup-linux-x64-gnu": "^4.50.0",
|
||||
"better-sqlite3": "^11.10.0"
|
||||
},
|
||||
"overrides": {
|
||||
"pyodide": "0.26.4"
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,12 +1,15 @@
|
||||
{
|
||||
"name": "n8n-mcp-runtime",
|
||||
"version": "2.10.0",
|
||||
"version": "2.20.2",
|
||||
"description": "n8n MCP Server Runtime Dependencies Only",
|
||||
"private": true,
|
||||
"dependencies": {
|
||||
"@modelcontextprotocol/sdk": "^1.13.2",
|
||||
"@supabase/supabase-js": "^2.57.4",
|
||||
"express": "^5.1.0",
|
||||
"express-rate-limit": "^7.1.5",
|
||||
"dotenv": "^16.5.0",
|
||||
"lru-cache": "^11.2.1",
|
||||
"sql.js": "^1.13.0",
|
||||
"uuid": "^10.0.0",
|
||||
"axios": "^1.7.7"
|
||||
|
||||
@@ -1,60 +0,0 @@
|
||||
# n8n-MCP v2.7.0 Release Notes
|
||||
|
||||
## 🎉 What's New
|
||||
|
||||
### 🔧 File Refactoring & Version Management
|
||||
- **Renamed core MCP files** to remove unnecessary suffixes for cleaner codebase:
|
||||
- `tools-update.ts` → `tools.ts`
|
||||
- `server-update.ts` → `server.ts`
|
||||
- `http-server-fixed.ts` → `http-server.ts`
|
||||
- **Fixed version management** - Now reads from package.json as single source of truth (fixes #5)
|
||||
- **Updated imports** across 21+ files to use the new file names
|
||||
|
||||
### 🔍 New Diagnostic Tool
|
||||
- **Added `n8n_diagnostic` tool** - Helps troubleshoot why n8n management tools might not be appearing
|
||||
- Shows environment variable status, API connectivity, and tool availability
|
||||
- Provides step-by-step troubleshooting guidance
|
||||
- Includes verbose mode for additional debug information
|
||||
|
||||
### 🧹 Code Cleanup
|
||||
- Removed legacy HTTP server implementation with known issues
|
||||
- Removed unused legacy API client
|
||||
- Added version utility for consistent version handling
|
||||
- Added script to sync runtime package version
|
||||
|
||||
## 📦 Installation
|
||||
|
||||
### Docker (Recommended)
|
||||
```bash
|
||||
docker pull ghcr.io/czlonkowski/n8n-mcp:2.7.0
|
||||
```
|
||||
|
||||
### Claude Desktop
|
||||
Update your configuration to use the latest version:
|
||||
```json
|
||||
{
|
||||
"mcpServers": {
|
||||
"n8n-mcp": {
|
||||
"command": "docker",
|
||||
"args": ["run", "-i", "--rm", "ghcr.io/czlonkowski/n8n-mcp:2.7.0"]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## 🐛 Bug Fixes
|
||||
- Fixed version mismatch where version was hardcoded as 2.4.1 instead of reading from package.json
|
||||
- Improved error messages for better debugging
|
||||
|
||||
## 📚 Documentation Updates
|
||||
- Condensed version history in CLAUDE.md
|
||||
- Updated documentation structure in README.md
|
||||
- Removed outdated documentation files
|
||||
- Added n8n_diagnostic tool to documentation
|
||||
|
||||
## 🙏 Acknowledgments
|
||||
Thanks to all contributors and users who reported issues!
|
||||
|
||||
---
|
||||
|
||||
**Full Changelog**: https://github.com/czlonkowski/n8n-mcp/blob/main/CHANGELOG.md
|
||||
78
scripts/audit-schema-coverage.ts
Normal file
78
scripts/audit-schema-coverage.ts
Normal file
@@ -0,0 +1,78 @@
|
||||
/**
|
||||
* Database Schema Coverage Audit Script
|
||||
*
|
||||
* Audits the database to determine how many nodes have complete schema information
|
||||
* for resourceLocator mode validation. This helps assess the coverage of our
|
||||
* schema-driven validation approach.
|
||||
*/
|
||||
|
||||
import Database from 'better-sqlite3';
|
||||
import path from 'path';
|
||||
|
||||
const dbPath = path.join(__dirname, '../data/nodes.db');
|
||||
const db = new Database(dbPath, { readonly: true });
|
||||
|
||||
console.log('=== Schema Coverage Audit ===\n');
|
||||
|
||||
// Query 1: How many nodes have resourceLocator properties?
|
||||
const totalResourceLocator = db.prepare(`
|
||||
SELECT COUNT(*) as count FROM nodes
|
||||
WHERE properties_schema LIKE '%resourceLocator%'
|
||||
`).get() as { count: number };
|
||||
|
||||
console.log(`Nodes with resourceLocator properties: ${totalResourceLocator.count}`);
|
||||
|
||||
// Query 2: Of those, how many have modes defined?
|
||||
const withModes = db.prepare(`
|
||||
SELECT COUNT(*) as count FROM nodes
|
||||
WHERE properties_schema LIKE '%resourceLocator%'
|
||||
AND properties_schema LIKE '%modes%'
|
||||
`).get() as { count: number };
|
||||
|
||||
console.log(`Nodes with modes defined: ${withModes.count}`);
|
||||
|
||||
// Query 3: Which nodes have resourceLocator but NO modes?
|
||||
const withoutModes = db.prepare(`
|
||||
SELECT node_type, display_name
|
||||
FROM nodes
|
||||
WHERE properties_schema LIKE '%resourceLocator%'
|
||||
AND properties_schema NOT LIKE '%modes%'
|
||||
LIMIT 10
|
||||
`).all() as Array<{ node_type: string; display_name: string }>;
|
||||
|
||||
console.log(`\nSample nodes WITHOUT modes (showing 10):`);
|
||||
withoutModes.forEach(node => {
|
||||
console.log(` - ${node.display_name} (${node.node_type})`);
|
||||
});
|
||||
|
||||
// Calculate coverage percentage
|
||||
const coverage = totalResourceLocator.count > 0
|
||||
? (withModes.count / totalResourceLocator.count) * 100
|
||||
: 0;
|
||||
|
||||
console.log(`\nSchema coverage: ${coverage.toFixed(1)}% of resourceLocator nodes have modes defined`);
|
||||
|
||||
// Query 4: Get some examples of nodes WITH modes for verification
|
||||
console.log('\nSample nodes WITH modes (showing 5):');
|
||||
const withModesExamples = db.prepare(`
|
||||
SELECT node_type, display_name
|
||||
FROM nodes
|
||||
WHERE properties_schema LIKE '%resourceLocator%'
|
||||
AND properties_schema LIKE '%modes%'
|
||||
LIMIT 5
|
||||
`).all() as Array<{ node_type: string; display_name: string }>;
|
||||
|
||||
withModesExamples.forEach(node => {
|
||||
console.log(` - ${node.display_name} (${node.node_type})`);
|
||||
});
|
||||
|
||||
// Summary
|
||||
console.log('\n=== Summary ===');
|
||||
console.log(`Total nodes in database: ${db.prepare('SELECT COUNT(*) as count FROM nodes').get() as any as { count: number }.count}`);
|
||||
console.log(`Nodes with resourceLocator: ${totalResourceLocator.count}`);
|
||||
console.log(`Nodes with complete mode schemas: ${withModes.count}`);
|
||||
console.log(`Nodes without mode schemas: ${totalResourceLocator.count - withModes.count}`);
|
||||
console.log(`\nImplication: Schema-driven validation will apply to ${withModes.count} nodes.`);
|
||||
console.log(`For the remaining ${totalResourceLocator.count - withModes.count} nodes, validation will be skipped (graceful degradation).`);
|
||||
|
||||
db.close();
|
||||
@@ -1,327 +0,0 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Debug script for n8n integration issues
|
||||
* Tests MCP protocol compliance and identifies schema validation problems
|
||||
*/
|
||||
|
||||
const http = require('http');
|
||||
const crypto = require('crypto');
|
||||
|
||||
const MCP_PORT = process.env.MCP_PORT || 3001;
|
||||
const AUTH_TOKEN = process.env.AUTH_TOKEN || 'test-token-for-n8n-testing-minimum-32-chars';
|
||||
|
||||
console.log('🔍 Debugging n8n MCP Integration Issues');
|
||||
console.log('=====================================\n');
|
||||
|
||||
// Test data for different MCP protocol calls
|
||||
const testCases = [
|
||||
{
|
||||
name: 'MCP Initialize',
|
||||
path: '/mcp',
|
||||
method: 'POST',
|
||||
data: {
|
||||
jsonrpc: '2.0',
|
||||
method: 'initialize',
|
||||
params: {
|
||||
protocolVersion: '2025-03-26',
|
||||
capabilities: {
|
||||
tools: {}
|
||||
},
|
||||
clientInfo: {
|
||||
name: 'n8n-debug-test',
|
||||
version: '1.0.0'
|
||||
}
|
||||
},
|
||||
id: 1
|
||||
}
|
||||
},
|
||||
{
|
||||
name: 'Tools List',
|
||||
path: '/mcp',
|
||||
method: 'POST',
|
||||
sessionId: null, // Will be set after initialize
|
||||
data: {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools/list',
|
||||
params: {},
|
||||
id: 2
|
||||
}
|
||||
},
|
||||
{
|
||||
name: 'Tools Call - tools_documentation',
|
||||
path: '/mcp',
|
||||
method: 'POST',
|
||||
sessionId: null, // Will be set after initialize
|
||||
data: {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools/call',
|
||||
params: {
|
||||
name: 'tools_documentation',
|
||||
arguments: {}
|
||||
},
|
||||
id: 3
|
||||
}
|
||||
},
|
||||
{
|
||||
name: 'Tools Call - get_node_essentials',
|
||||
path: '/mcp',
|
||||
method: 'POST',
|
||||
sessionId: null, // Will be set after initialize
|
||||
data: {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools/call',
|
||||
params: {
|
||||
name: 'get_node_essentials',
|
||||
arguments: {
|
||||
nodeType: 'nodes-base.httpRequest'
|
||||
}
|
||||
},
|
||||
id: 4
|
||||
}
|
||||
}
|
||||
];
|
||||
|
||||
async function makeRequest(testCase) {
|
||||
return new Promise((resolve, reject) => {
|
||||
const data = JSON.stringify(testCase.data);
|
||||
|
||||
const options = {
|
||||
hostname: 'localhost',
|
||||
port: MCP_PORT,
|
||||
path: testCase.path,
|
||||
method: testCase.method,
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
'Content-Length': Buffer.byteLength(data),
|
||||
'Authorization': `Bearer ${AUTH_TOKEN}`,
|
||||
'Accept': 'application/json, text/event-stream' // Fix for StreamableHTTPServerTransport
|
||||
}
|
||||
};
|
||||
|
||||
// Add session ID header if available
|
||||
if (testCase.sessionId) {
|
||||
options.headers['Mcp-Session-Id'] = testCase.sessionId;
|
||||
}
|
||||
|
||||
console.log(`📤 Making request: ${testCase.name}`);
|
||||
console.log(` Method: ${testCase.method} ${testCase.path}`);
|
||||
if (testCase.sessionId) {
|
||||
console.log(` Session-ID: ${testCase.sessionId}`);
|
||||
}
|
||||
console.log(` Data: ${data}`);
|
||||
|
||||
const req = http.request(options, (res) => {
|
||||
let responseData = '';
|
||||
|
||||
console.log(`📥 Response Status: ${res.statusCode}`);
|
||||
console.log(` Headers:`, res.headers);
|
||||
|
||||
res.on('data', (chunk) => {
|
||||
responseData += chunk;
|
||||
});
|
||||
|
||||
res.on('end', () => {
|
||||
try {
|
||||
let parsed;
|
||||
|
||||
// Handle SSE format response
|
||||
if (responseData.startsWith('event: message\ndata: ')) {
|
||||
const dataLine = responseData.split('\n').find(line => line.startsWith('data: '));
|
||||
if (dataLine) {
|
||||
const jsonData = dataLine.substring(6); // Remove 'data: '
|
||||
parsed = JSON.parse(jsonData);
|
||||
} else {
|
||||
throw new Error('Could not extract JSON from SSE response');
|
||||
}
|
||||
} else {
|
||||
parsed = JSON.parse(responseData);
|
||||
}
|
||||
|
||||
resolve({
|
||||
statusCode: res.statusCode,
|
||||
headers: res.headers,
|
||||
data: parsed,
|
||||
raw: responseData
|
||||
});
|
||||
} catch (e) {
|
||||
resolve({
|
||||
statusCode: res.statusCode,
|
||||
headers: res.headers,
|
||||
data: null,
|
||||
raw: responseData,
|
||||
parseError: e.message
|
||||
});
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
req.on('error', (err) => {
|
||||
reject(err);
|
||||
});
|
||||
|
||||
req.write(data);
|
||||
req.end();
|
||||
});
|
||||
}
|
||||
|
||||
async function validateMCPResponse(testCase, response) {
|
||||
console.log(`✅ Validating response for: ${testCase.name}`);
|
||||
|
||||
const issues = [];
|
||||
|
||||
// Check HTTP status
|
||||
if (response.statusCode !== 200) {
|
||||
issues.push(`❌ Expected HTTP 200, got ${response.statusCode}`);
|
||||
}
|
||||
|
||||
// Check JSON-RPC structure
|
||||
if (!response.data) {
|
||||
issues.push(`❌ Response is not valid JSON: ${response.parseError}`);
|
||||
return issues;
|
||||
}
|
||||
|
||||
if (response.data.jsonrpc !== '2.0') {
|
||||
issues.push(`❌ Missing or invalid jsonrpc field: ${response.data.jsonrpc}`);
|
||||
}
|
||||
|
||||
if (response.data.id !== testCase.data.id) {
|
||||
issues.push(`❌ ID mismatch: expected ${testCase.data.id}, got ${response.data.id}`);
|
||||
}
|
||||
|
||||
// Method-specific validation
|
||||
if (testCase.data.method === 'initialize') {
|
||||
if (!response.data.result) {
|
||||
issues.push(`❌ Initialize response missing result field`);
|
||||
} else {
|
||||
if (!response.data.result.protocolVersion) {
|
||||
issues.push(`❌ Initialize response missing protocolVersion`);
|
||||
} else if (response.data.result.protocolVersion !== '2025-03-26') {
|
||||
issues.push(`❌ Protocol version mismatch: expected 2025-03-26, got ${response.data.result.protocolVersion}`);
|
||||
}
|
||||
|
||||
if (!response.data.result.capabilities) {
|
||||
issues.push(`❌ Initialize response missing capabilities`);
|
||||
}
|
||||
|
||||
if (!response.data.result.serverInfo) {
|
||||
issues.push(`❌ Initialize response missing serverInfo`);
|
||||
}
|
||||
}
|
||||
|
||||
// Extract session ID for subsequent requests
|
||||
if (response.headers['mcp-session-id']) {
|
||||
console.log(`📋 Session ID: ${response.headers['mcp-session-id']}`);
|
||||
return { issues, sessionId: response.headers['mcp-session-id'] };
|
||||
} else {
|
||||
issues.push(`❌ Initialize response missing Mcp-Session-Id header`);
|
||||
}
|
||||
}
|
||||
|
||||
if (testCase.data.method === 'tools/list') {
|
||||
if (!response.data.result || !response.data.result.tools) {
|
||||
issues.push(`❌ Tools list response missing tools array`);
|
||||
} else {
|
||||
console.log(`📋 Found ${response.data.result.tools.length} tools`);
|
||||
}
|
||||
}
|
||||
|
||||
if (testCase.data.method === 'tools/call') {
|
||||
if (!response.data.result) {
|
||||
issues.push(`❌ Tool call response missing result field`);
|
||||
} else if (!response.data.result.content) {
|
||||
issues.push(`❌ Tool call response missing content array`);
|
||||
} else if (!Array.isArray(response.data.result.content)) {
|
||||
issues.push(`❌ Tool call response content is not an array`);
|
||||
} else {
|
||||
// Validate content structure
|
||||
for (let i = 0; i < response.data.result.content.length; i++) {
|
||||
const content = response.data.result.content[i];
|
||||
if (!content.type) {
|
||||
issues.push(`❌ Content item ${i} missing type field`);
|
||||
}
|
||||
if (content.type === 'text' && !content.text) {
|
||||
issues.push(`❌ Text content item ${i} missing text field`);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (issues.length === 0) {
|
||||
console.log(`✅ ${testCase.name} validation passed`);
|
||||
} else {
|
||||
console.log(`❌ ${testCase.name} validation failed:`);
|
||||
issues.forEach(issue => console.log(` ${issue}`));
|
||||
}
|
||||
|
||||
return { issues };
|
||||
}
|
||||
|
||||
async function runTests() {
|
||||
console.log('Starting MCP protocol compliance tests...\n');
|
||||
|
||||
let sessionId = null;
|
||||
let allIssues = [];
|
||||
|
||||
for (const testCase of testCases) {
|
||||
try {
|
||||
// Set session ID from previous test
|
||||
if (sessionId && testCase.name !== 'MCP Initialize') {
|
||||
testCase.sessionId = sessionId;
|
||||
}
|
||||
|
||||
const response = await makeRequest(testCase);
|
||||
console.log(`📄 Raw Response: ${response.raw}\n`);
|
||||
|
||||
const validation = await validateMCPResponse(testCase, response);
|
||||
|
||||
if (validation.sessionId) {
|
||||
sessionId = validation.sessionId;
|
||||
}
|
||||
|
||||
allIssues.push(...validation.issues);
|
||||
|
||||
console.log('─'.repeat(50));
|
||||
|
||||
} catch (error) {
|
||||
console.error(`❌ Request failed for ${testCase.name}:`, error.message);
|
||||
allIssues.push(`Request failed for ${testCase.name}: ${error.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
// Summary
|
||||
console.log('\n📊 SUMMARY');
|
||||
console.log('==========');
|
||||
|
||||
if (allIssues.length === 0) {
|
||||
console.log('🎉 All tests passed! MCP protocol compliance looks good.');
|
||||
} else {
|
||||
console.log(`❌ Found ${allIssues.length} issues:`);
|
||||
allIssues.forEach((issue, i) => {
|
||||
console.log(` ${i + 1}. ${issue}`);
|
||||
});
|
||||
}
|
||||
|
||||
console.log('\n🔍 Recommendations:');
|
||||
console.log('1. Check MCP server logs at /tmp/mcp-server.log');
|
||||
console.log('2. Verify protocol version consistency (should be 2025-03-26)');
|
||||
console.log('3. Ensure tool schemas match MCP specification exactly');
|
||||
console.log('4. Test with actual n8n MCP Client Tool node');
|
||||
}
|
||||
|
||||
// Check if MCP server is running
|
||||
console.log(`Checking if MCP server is running at localhost:${MCP_PORT}...`);
|
||||
|
||||
const healthCheck = http.get(`http://localhost:${MCP_PORT}/health`, (res) => {
|
||||
if (res.statusCode === 200) {
|
||||
console.log('✅ MCP server is running\n');
|
||||
runTests().catch(console.error);
|
||||
} else {
|
||||
console.error('❌ MCP server health check failed:', res.statusCode);
|
||||
process.exit(1);
|
||||
}
|
||||
}).on('error', (err) => {
|
||||
console.error('❌ MCP server is not running. Please start it first:', err.message);
|
||||
console.error('Use: npm run start:n8n');
|
||||
process.exit(1);
|
||||
});
|
||||
41
scripts/export-webhook-workflows.ts
Normal file
41
scripts/export-webhook-workflows.ts
Normal file
@@ -0,0 +1,41 @@
|
||||
#!/usr/bin/env tsx
|
||||
|
||||
/**
|
||||
* Export Webhook Workflow JSONs
|
||||
*
|
||||
* Generates the 4 webhook workflow JSON files needed for integration testing.
|
||||
* These workflows must be imported into n8n and activated manually.
|
||||
*/
|
||||
|
||||
import { writeFileSync, mkdirSync } from 'fs';
|
||||
import { join } from 'path';
|
||||
import { exportAllWebhookWorkflows } from '../tests/integration/n8n-api/utils/webhook-workflows';
|
||||
|
||||
const OUTPUT_DIR = join(process.cwd(), 'workflows-for-import');
|
||||
|
||||
// Create output directory
|
||||
mkdirSync(OUTPUT_DIR, { recursive: true });
|
||||
|
||||
// Generate all workflow JSONs
|
||||
const workflows = exportAllWebhookWorkflows();
|
||||
|
||||
// Write each workflow to a separate file
|
||||
Object.entries(workflows).forEach(([method, workflow]) => {
|
||||
const filename = `webhook-${method.toLowerCase()}.json`;
|
||||
const filepath = join(OUTPUT_DIR, filename);
|
||||
|
||||
writeFileSync(filepath, JSON.stringify(workflow, null, 2), 'utf-8');
|
||||
|
||||
console.log(`✓ Generated: ${filename}`);
|
||||
});
|
||||
|
||||
console.log(`\n✓ All workflow JSONs written to: ${OUTPUT_DIR}`);
|
||||
console.log('\nNext steps:');
|
||||
console.log('1. Import each JSON file into your n8n instance');
|
||||
console.log('2. Activate each workflow in the n8n UI');
|
||||
console.log('3. Copy the webhook URLs from each workflow (open workflow → Webhook node → copy URL)');
|
||||
console.log('4. Add them to your .env file:');
|
||||
console.log(' N8N_TEST_WEBHOOK_GET_URL=https://your-n8n.com/webhook/mcp-test-get');
|
||||
console.log(' N8N_TEST_WEBHOOK_POST_URL=https://your-n8n.com/webhook/mcp-test-post');
|
||||
console.log(' N8N_TEST_WEBHOOK_PUT_URL=https://your-n8n.com/webhook/mcp-test-put');
|
||||
console.log(' N8N_TEST_WEBHOOK_DELETE_URL=https://your-n8n.com/webhook/mcp-test-delete');
|
||||
84
scripts/extract-changelog.js
Executable file
84
scripts/extract-changelog.js
Executable file
@@ -0,0 +1,84 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Extract changelog content for a specific version
|
||||
* Used by GitHub Actions to extract release notes
|
||||
*/
|
||||
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
|
||||
function extractChangelog(version, changelogPath) {
|
||||
try {
|
||||
if (!fs.existsSync(changelogPath)) {
|
||||
console.error(`Changelog file not found at ${changelogPath}`);
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
const content = fs.readFileSync(changelogPath, 'utf8');
|
||||
const lines = content.split('\n');
|
||||
|
||||
// Find the start of this version's section
|
||||
const versionHeaderRegex = new RegExp(`^## \\[${version.replace(/[.*+?^${}()|[\]\\]/g, '\\$&')}\\]`);
|
||||
let startIndex = -1;
|
||||
let endIndex = -1;
|
||||
|
||||
for (let i = 0; i < lines.length; i++) {
|
||||
if (versionHeaderRegex.test(lines[i])) {
|
||||
startIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (startIndex === -1) {
|
||||
console.error(`No changelog entries found for version ${version}`);
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
// Find the end of this version's section (next version or end of file)
|
||||
for (let i = startIndex + 1; i < lines.length; i++) {
|
||||
if (lines[i].startsWith('## [') && !lines[i].includes('Unreleased')) {
|
||||
endIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (endIndex === -1) {
|
||||
endIndex = lines.length;
|
||||
}
|
||||
|
||||
// Extract the section content
|
||||
const sectionLines = lines.slice(startIndex, endIndex);
|
||||
|
||||
// Remove the version header and any trailing empty lines
|
||||
let contentLines = sectionLines.slice(1);
|
||||
while (contentLines.length > 0 && contentLines[contentLines.length - 1].trim() === '') {
|
||||
contentLines.pop();
|
||||
}
|
||||
|
||||
if (contentLines.length === 0) {
|
||||
console.error(`No content found for version ${version}`);
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
const releaseNotes = contentLines.join('\n').trim();
|
||||
|
||||
// Write to stdout for GitHub Actions
|
||||
console.log(releaseNotes);
|
||||
|
||||
} catch (error) {
|
||||
console.error(`Error extracting changelog: ${error.message}`);
|
||||
process.exit(1);
|
||||
}
|
||||
}
|
||||
|
||||
// Parse command line arguments
|
||||
const version = process.argv[2];
|
||||
const changelogPath = process.argv[3];
|
||||
|
||||
if (!version || !changelogPath) {
|
||||
console.error('Usage: extract-changelog.js <version> <changelog-path>');
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
extractChangelog(version, changelogPath);
|
||||
400
scripts/prepare-release.js
Executable file
400
scripts/prepare-release.js
Executable file
@@ -0,0 +1,400 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Pre-release preparation script
|
||||
* Validates and prepares everything needed for a successful release
|
||||
*/
|
||||
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
const { execSync, spawnSync } = require('child_process');
|
||||
const readline = require('readline');
|
||||
|
||||
// Color codes
|
||||
const colors = {
|
||||
reset: '\x1b[0m',
|
||||
red: '\x1b[31m',
|
||||
green: '\x1b[32m',
|
||||
yellow: '\x1b[33m',
|
||||
blue: '\x1b[34m',
|
||||
magenta: '\x1b[35m',
|
||||
cyan: '\x1b[36m'
|
||||
};
|
||||
|
||||
function log(message, color = 'reset') {
|
||||
console.log(`${colors[color]}${message}${colors.reset}`);
|
||||
}
|
||||
|
||||
function success(message) {
|
||||
log(`✅ ${message}`, 'green');
|
||||
}
|
||||
|
||||
function warning(message) {
|
||||
log(`⚠️ ${message}`, 'yellow');
|
||||
}
|
||||
|
||||
function error(message) {
|
||||
log(`❌ ${message}`, 'red');
|
||||
}
|
||||
|
||||
function info(message) {
|
||||
log(`ℹ️ ${message}`, 'blue');
|
||||
}
|
||||
|
||||
function header(title) {
|
||||
log(`\n${'='.repeat(60)}`, 'cyan');
|
||||
log(`🚀 ${title}`, 'cyan');
|
||||
log(`${'='.repeat(60)}`, 'cyan');
|
||||
}
|
||||
|
||||
class ReleasePreparation {
|
||||
constructor() {
|
||||
this.rootDir = path.resolve(__dirname, '..');
|
||||
this.rl = readline.createInterface({
|
||||
input: process.stdin,
|
||||
output: process.stdout
|
||||
});
|
||||
}
|
||||
|
||||
async askQuestion(question) {
|
||||
return new Promise((resolve) => {
|
||||
this.rl.question(question, resolve);
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Get current version and ask for new version
|
||||
*/
|
||||
async getVersionInfo() {
|
||||
const packageJson = require(path.join(this.rootDir, 'package.json'));
|
||||
const currentVersion = packageJson.version;
|
||||
|
||||
log(`\nCurrent version: ${currentVersion}`, 'blue');
|
||||
|
||||
const newVersion = await this.askQuestion('\nEnter new version (e.g., 2.10.0): ');
|
||||
|
||||
if (!newVersion || !this.isValidSemver(newVersion)) {
|
||||
error('Invalid semantic version format');
|
||||
throw new Error('Invalid version');
|
||||
}
|
||||
|
||||
if (this.compareVersions(newVersion, currentVersion) <= 0) {
|
||||
error('New version must be greater than current version');
|
||||
throw new Error('Version not incremented');
|
||||
}
|
||||
|
||||
return { currentVersion, newVersion };
|
||||
}
|
||||
|
||||
/**
|
||||
* Validate semantic version format (strict semver compliance)
|
||||
*/
|
||||
isValidSemver(version) {
|
||||
// Strict semantic versioning regex
|
||||
const semverRegex = /^(0|[1-9]\d*)\.(0|[1-9]\d*)\.(0|[1-9]\d*)(?:-((?:0|[1-9]\d*|\d*[a-zA-Z-][0-9a-zA-Z-]*)(?:\.(?:0|[1-9]\d*|\d*[a-zA-Z-][0-9a-zA-Z-]*))*))?(?:\+([0-9a-zA-Z-]+(?:\.[0-9a-zA-Z-]+)*))?$/;
|
||||
return semverRegex.test(version);
|
||||
}
|
||||
|
||||
/**
|
||||
* Compare two semantic versions
|
||||
*/
|
||||
compareVersions(v1, v2) {
|
||||
const parseVersion = (v) => v.split('-')[0].split('.').map(Number);
|
||||
const [v1Parts, v2Parts] = [parseVersion(v1), parseVersion(v2)];
|
||||
|
||||
for (let i = 0; i < 3; i++) {
|
||||
if (v1Parts[i] > v2Parts[i]) return 1;
|
||||
if (v1Parts[i] < v2Parts[i]) return -1;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* Update version in package files
|
||||
*/
|
||||
updateVersions(newVersion) {
|
||||
log('\n📝 Updating version in package files...', 'blue');
|
||||
|
||||
// Update package.json
|
||||
const packageJsonPath = path.join(this.rootDir, 'package.json');
|
||||
const packageJson = require(packageJsonPath);
|
||||
packageJson.version = newVersion;
|
||||
fs.writeFileSync(packageJsonPath, JSON.stringify(packageJson, null, 2) + '\n');
|
||||
success('Updated package.json');
|
||||
|
||||
// Sync to runtime package
|
||||
try {
|
||||
execSync('npm run sync:runtime-version', { cwd: this.rootDir, stdio: 'pipe' });
|
||||
success('Synced package.runtime.json');
|
||||
} catch (err) {
|
||||
warning('Could not sync runtime version automatically');
|
||||
|
||||
// Manual sync
|
||||
const runtimeJsonPath = path.join(this.rootDir, 'package.runtime.json');
|
||||
if (fs.existsSync(runtimeJsonPath)) {
|
||||
const runtimeJson = require(runtimeJsonPath);
|
||||
runtimeJson.version = newVersion;
|
||||
fs.writeFileSync(runtimeJsonPath, JSON.stringify(runtimeJson, null, 2) + '\n');
|
||||
success('Manually synced package.runtime.json');
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Update changelog
|
||||
*/
|
||||
async updateChangelog(newVersion) {
|
||||
const changelogPath = path.join(this.rootDir, 'docs/CHANGELOG.md');
|
||||
|
||||
if (!fs.existsSync(changelogPath)) {
|
||||
warning('Changelog file not found, skipping update');
|
||||
return;
|
||||
}
|
||||
|
||||
log('\n📋 Updating changelog...', 'blue');
|
||||
|
||||
const content = fs.readFileSync(changelogPath, 'utf8');
|
||||
const today = new Date().toISOString().split('T')[0];
|
||||
|
||||
// Check if version already exists in changelog
|
||||
const versionRegex = new RegExp(`^## \\[${newVersion.replace(/[.*+?^${}()|[\]\\]/g, '\\$&')}\\]`, 'm');
|
||||
if (versionRegex.test(content)) {
|
||||
info(`Version ${newVersion} already exists in changelog`);
|
||||
return;
|
||||
}
|
||||
|
||||
// Find the Unreleased section
|
||||
const unreleasedMatch = content.match(/^## \[Unreleased\]\s*\n([\s\S]*?)(?=\n## \[|$)/m);
|
||||
|
||||
if (unreleasedMatch) {
|
||||
const unreleasedContent = unreleasedMatch[1].trim();
|
||||
|
||||
if (unreleasedContent) {
|
||||
log('\nFound content in Unreleased section:', 'blue');
|
||||
log(unreleasedContent.substring(0, 200) + '...', 'yellow');
|
||||
|
||||
const moveContent = await this.askQuestion('\nMove this content to the new version? (y/n): ');
|
||||
|
||||
if (moveContent.toLowerCase() === 'y') {
|
||||
// Move unreleased content to new version
|
||||
const newVersionSection = `## [${newVersion}] - ${today}\n\n${unreleasedContent}\n\n`;
|
||||
const updatedContent = content.replace(
|
||||
/^## \[Unreleased\]\s*\n[\s\S]*?(?=\n## \[)/m,
|
||||
`## [Unreleased]\n\n${newVersionSection}## [`
|
||||
);
|
||||
|
||||
fs.writeFileSync(changelogPath, updatedContent);
|
||||
success(`Moved unreleased content to version ${newVersion}`);
|
||||
} else {
|
||||
// Just add empty version section
|
||||
const newVersionSection = `## [${newVersion}] - ${today}\n\n### Added\n- \n\n### Changed\n- \n\n### Fixed\n- \n\n`;
|
||||
const updatedContent = content.replace(
|
||||
/^## \[Unreleased\]\s*\n/m,
|
||||
`## [Unreleased]\n\n${newVersionSection}`
|
||||
);
|
||||
|
||||
fs.writeFileSync(changelogPath, updatedContent);
|
||||
warning(`Added empty version section for ${newVersion} - please fill in the changes`);
|
||||
}
|
||||
} else {
|
||||
// Add empty version section
|
||||
const newVersionSection = `## [${newVersion}] - ${today}\n\n### Added\n- \n\n### Changed\n- \n\n### Fixed\n- \n\n`;
|
||||
const updatedContent = content.replace(
|
||||
/^## \[Unreleased\]\s*\n/m,
|
||||
`## [Unreleased]\n\n${newVersionSection}`
|
||||
);
|
||||
|
||||
fs.writeFileSync(changelogPath, updatedContent);
|
||||
warning(`Added empty version section for ${newVersion} - please fill in the changes`);
|
||||
}
|
||||
} else {
|
||||
warning('Could not find Unreleased section in changelog');
|
||||
}
|
||||
|
||||
info('Please review and edit the changelog before committing');
|
||||
}
|
||||
|
||||
/**
|
||||
* Run tests and build
|
||||
*/
|
||||
async runChecks() {
|
||||
log('\n🧪 Running pre-release checks...', 'blue');
|
||||
|
||||
try {
|
||||
// Run tests
|
||||
log('Running tests...', 'blue');
|
||||
execSync('npm test', { cwd: this.rootDir, stdio: 'inherit' });
|
||||
success('All tests passed');
|
||||
|
||||
// Run build
|
||||
log('Building project...', 'blue');
|
||||
execSync('npm run build', { cwd: this.rootDir, stdio: 'inherit' });
|
||||
success('Build completed');
|
||||
|
||||
// Rebuild database
|
||||
log('Rebuilding database...', 'blue');
|
||||
execSync('npm run rebuild', { cwd: this.rootDir, stdio: 'inherit' });
|
||||
success('Database rebuilt');
|
||||
|
||||
// Run type checking
|
||||
log('Type checking...', 'blue');
|
||||
execSync('npm run typecheck', { cwd: this.rootDir, stdio: 'inherit' });
|
||||
success('Type checking passed');
|
||||
|
||||
} catch (err) {
|
||||
error('Pre-release checks failed');
|
||||
throw err;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Create git commit
|
||||
*/
|
||||
async createCommit(newVersion) {
|
||||
log('\n📝 Creating git commit...', 'blue');
|
||||
|
||||
try {
|
||||
// Check git status
|
||||
const status = execSync('git status --porcelain', {
|
||||
cwd: this.rootDir,
|
||||
encoding: 'utf8'
|
||||
});
|
||||
|
||||
if (!status.trim()) {
|
||||
info('No changes to commit');
|
||||
return;
|
||||
}
|
||||
|
||||
// Show what will be committed
|
||||
log('\nFiles to be committed:', 'blue');
|
||||
execSync('git diff --name-only', { cwd: this.rootDir, stdio: 'inherit' });
|
||||
|
||||
const commit = await this.askQuestion('\nCreate commit for release? (y/n): ');
|
||||
|
||||
if (commit.toLowerCase() === 'y') {
|
||||
// Add files
|
||||
execSync('git add package.json package.runtime.json docs/CHANGELOG.md', {
|
||||
cwd: this.rootDir,
|
||||
stdio: 'pipe'
|
||||
});
|
||||
|
||||
// Create commit
|
||||
const commitMessage = `chore: release v${newVersion}
|
||||
|
||||
🤖 Generated with [Claude Code](https://claude.ai/code)
|
||||
|
||||
Co-Authored-By: Claude <noreply@anthropic.com>`;
|
||||
|
||||
const result = spawnSync('git', ['commit', '-m', commitMessage], {
|
||||
cwd: this.rootDir,
|
||||
stdio: 'pipe',
|
||||
encoding: 'utf8'
|
||||
});
|
||||
|
||||
if (result.error || result.status !== 0) {
|
||||
throw new Error(`Git commit failed: ${result.stderr || result.error?.message}`);
|
||||
}
|
||||
|
||||
success(`Created commit for v${newVersion}`);
|
||||
|
||||
const push = await this.askQuestion('\nPush to trigger release workflow? (y/n): ');
|
||||
|
||||
if (push.toLowerCase() === 'y') {
|
||||
// Add confirmation for destructive operation
|
||||
warning('\n⚠️ DESTRUCTIVE OPERATION WARNING ⚠️');
|
||||
warning('This will trigger a PUBLIC RELEASE that cannot be undone!');
|
||||
warning('The following will happen automatically:');
|
||||
warning('• Create GitHub release with tag');
|
||||
warning('• Publish package to NPM registry');
|
||||
warning('• Build and push Docker images');
|
||||
warning('• Update documentation');
|
||||
|
||||
const confirmation = await this.askQuestion('\nType "RELEASE" (all caps) to confirm: ');
|
||||
|
||||
if (confirmation === 'RELEASE') {
|
||||
execSync('git push', { cwd: this.rootDir, stdio: 'inherit' });
|
||||
success('Pushed to remote repository');
|
||||
log('\n🎉 Release workflow will be triggered automatically!', 'green');
|
||||
log('Monitor progress at: https://github.com/czlonkowski/n8n-mcp/actions', 'blue');
|
||||
} else {
|
||||
warning('Release cancelled. Commit created but not pushed.');
|
||||
info('You can push manually later to trigger the release.');
|
||||
}
|
||||
} else {
|
||||
info('Commit created but not pushed. Push manually to trigger release.');
|
||||
}
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Git operations failed: ${err.message}`);
|
||||
throw err;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Display final instructions
|
||||
*/
|
||||
displayInstructions(newVersion) {
|
||||
header('Release Preparation Complete');
|
||||
|
||||
log('📋 What happens next:', 'blue');
|
||||
log(`1. The GitHub Actions workflow will detect the version change to v${newVersion}`, 'green');
|
||||
log('2. It will automatically:', 'green');
|
||||
log(' • Create a GitHub release with changelog content', 'green');
|
||||
log(' • Publish the npm package', 'green');
|
||||
log(' • Build and push Docker images', 'green');
|
||||
log(' • Update documentation badges', 'green');
|
||||
log('\n🔍 Monitor the release at:', 'blue');
|
||||
log(' • GitHub Actions: https://github.com/czlonkowski/n8n-mcp/actions', 'blue');
|
||||
log(' • NPM Package: https://www.npmjs.com/package/n8n-mcp', 'blue');
|
||||
log(' • Docker Images: https://github.com/czlonkowski/n8n-mcp/pkgs/container/n8n-mcp', 'blue');
|
||||
|
||||
log('\n✅ Release preparation completed successfully!', 'green');
|
||||
}
|
||||
|
||||
/**
|
||||
* Main execution flow
|
||||
*/
|
||||
async run() {
|
||||
try {
|
||||
header('n8n-MCP Release Preparation');
|
||||
|
||||
// Get version information
|
||||
const { currentVersion, newVersion } = await this.getVersionInfo();
|
||||
|
||||
log(`\n🔄 Preparing release: ${currentVersion} → ${newVersion}`, 'magenta');
|
||||
|
||||
// Update versions
|
||||
this.updateVersions(newVersion);
|
||||
|
||||
// Update changelog
|
||||
await this.updateChangelog(newVersion);
|
||||
|
||||
// Run pre-release checks
|
||||
await this.runChecks();
|
||||
|
||||
// Create git commit
|
||||
await this.createCommit(newVersion);
|
||||
|
||||
// Display final instructions
|
||||
this.displayInstructions(newVersion);
|
||||
|
||||
} catch (err) {
|
||||
error(`Release preparation failed: ${err.message}`);
|
||||
process.exit(1);
|
||||
} finally {
|
||||
this.rl.close();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Run the script
|
||||
if (require.main === module) {
|
||||
const preparation = new ReleasePreparation();
|
||||
preparation.run().catch(err => {
|
||||
console.error('Release preparation failed:', err);
|
||||
process.exit(1);
|
||||
});
|
||||
}
|
||||
|
||||
module.exports = ReleasePreparation;
|
||||
62
scripts/publish-npm-quick.sh
Executable file
62
scripts/publish-npm-quick.sh
Executable file
@@ -0,0 +1,62 @@
|
||||
#!/bin/bash
|
||||
# Quick publish script that skips tests
|
||||
set -e
|
||||
|
||||
# Color codes
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
NC='\033[0m'
|
||||
|
||||
echo "🚀 Preparing n8n-mcp for npm publish (quick mode)..."
|
||||
|
||||
# Sync version
|
||||
echo "🔄 Syncing version to package.runtime.json..."
|
||||
npm run sync:runtime-version
|
||||
|
||||
VERSION=$(node -e "console.log(require('./package.json').version)")
|
||||
echo -e "${GREEN}📌 Version: $VERSION${NC}"
|
||||
|
||||
# Prepare publish directory
|
||||
PUBLISH_DIR="npm-publish-temp"
|
||||
rm -rf $PUBLISH_DIR
|
||||
mkdir -p $PUBLISH_DIR
|
||||
|
||||
echo "📦 Copying files..."
|
||||
cp -r dist $PUBLISH_DIR/
|
||||
cp -r data $PUBLISH_DIR/
|
||||
cp README.md LICENSE .env.example $PUBLISH_DIR/
|
||||
cp .npmignore $PUBLISH_DIR/ 2>/dev/null || true
|
||||
cp package.runtime.json $PUBLISH_DIR/package.json
|
||||
|
||||
cd $PUBLISH_DIR
|
||||
|
||||
# Configure package.json
|
||||
node -e "
|
||||
const pkg = require('./package.json');
|
||||
pkg.name = 'n8n-mcp';
|
||||
pkg.description = 'Integration between n8n workflow automation and Model Context Protocol (MCP)';
|
||||
pkg.bin = { 'n8n-mcp': './dist/mcp/index.js' };
|
||||
pkg.repository = { type: 'git', url: 'git+https://github.com/czlonkowski/n8n-mcp.git' };
|
||||
pkg.keywords = ['n8n', 'mcp', 'model-context-protocol', 'ai', 'workflow', 'automation'];
|
||||
pkg.author = 'Romuald Czlonkowski @ www.aiadvisors.pl/en';
|
||||
pkg.license = 'MIT';
|
||||
pkg.bugs = { url: 'https://github.com/czlonkowski/n8n-mcp/issues' };
|
||||
pkg.homepage = 'https://github.com/czlonkowski/n8n-mcp#readme';
|
||||
pkg.files = ['dist/**/*', 'data/nodes.db', '.env.example', 'README.md', 'LICENSE'];
|
||||
delete pkg.private;
|
||||
require('fs').writeFileSync('./package.json', JSON.stringify(pkg, null, 2));
|
||||
"
|
||||
|
||||
echo ""
|
||||
echo "📋 Package details:"
|
||||
echo -e "${GREEN}Name:${NC} $(node -e "console.log(require('./package.json').name)")"
|
||||
echo -e "${GREEN}Version:${NC} $(node -e "console.log(require('./package.json').version)")"
|
||||
echo -e "${GREEN}Size:${NC} ~50MB"
|
||||
echo ""
|
||||
echo "✅ Ready to publish!"
|
||||
echo ""
|
||||
echo -e "${YELLOW}⚠️ Note: Tests were skipped in quick mode${NC}"
|
||||
echo ""
|
||||
echo "To publish, run:"
|
||||
echo -e " ${GREEN}cd $PUBLISH_DIR${NC}"
|
||||
echo -e " ${GREEN}npm publish --otp=YOUR_OTP_CODE${NC}"
|
||||
@@ -11,14 +11,8 @@ NC='\033[0m' # No Color
|
||||
|
||||
echo "🚀 Preparing n8n-mcp for npm publish..."
|
||||
|
||||
# Run tests first to ensure quality
|
||||
echo "🧪 Running tests..."
|
||||
npm test
|
||||
if [ $? -ne 0 ]; then
|
||||
echo -e "${RED}❌ Tests failed. Aborting publish.${NC}"
|
||||
exit 1
|
||||
fi
|
||||
echo -e "${GREEN}✅ All tests passed!${NC}"
|
||||
# Skip tests - they already run in CI before merge/publish
|
||||
echo "⏭️ Skipping tests (already verified in CI)"
|
||||
|
||||
# Sync version to runtime package first
|
||||
echo "🔄 Syncing version to package.runtime.json..."
|
||||
@@ -65,6 +59,15 @@ node -e "
|
||||
const pkg = require('./package.json');
|
||||
pkg.name = 'n8n-mcp';
|
||||
pkg.description = 'Integration between n8n workflow automation and Model Context Protocol (MCP)';
|
||||
pkg.main = 'dist/index.js';
|
||||
pkg.types = 'dist/index.d.ts';
|
||||
pkg.exports = {
|
||||
'.': {
|
||||
types: './dist/index.d.ts',
|
||||
require: './dist/index.js',
|
||||
import: './dist/index.js'
|
||||
}
|
||||
};
|
||||
pkg.bin = { 'n8n-mcp': './dist/mcp/index.js' };
|
||||
pkg.repository = { type: 'git', url: 'git+https://github.com/czlonkowski/n8n-mcp.git' };
|
||||
pkg.keywords = ['n8n', 'mcp', 'model-context-protocol', 'ai', 'workflow', 'automation'];
|
||||
|
||||
189
scripts/test-ai-validation-debug.ts
Normal file
189
scripts/test-ai-validation-debug.ts
Normal file
@@ -0,0 +1,189 @@
|
||||
#!/usr/bin/env node
|
||||
/**
|
||||
* Debug test for AI validation issues
|
||||
* Reproduces the bugs found by n8n-mcp-tester
|
||||
*/
|
||||
|
||||
import { validateAISpecificNodes, buildReverseConnectionMap } from '../src/services/ai-node-validator';
|
||||
import type { WorkflowJson } from '../src/services/ai-tool-validators';
|
||||
import { NodeTypeNormalizer } from '../src/utils/node-type-normalizer';
|
||||
|
||||
console.log('=== AI Validation Debug Tests ===\n');
|
||||
|
||||
// Test 1: AI Agent with NO language model connection
|
||||
console.log('Test 1: Missing Language Model Detection');
|
||||
const workflow1: WorkflowJson = {
|
||||
name: 'Test Missing LM',
|
||||
nodes: [
|
||||
{
|
||||
id: 'ai-agent-1',
|
||||
name: 'AI Agent',
|
||||
type: '@n8n/n8n-nodes-langchain.agent',
|
||||
position: [500, 300],
|
||||
parameters: {
|
||||
promptType: 'define',
|
||||
text: 'You are a helpful assistant'
|
||||
},
|
||||
typeVersion: 1.7
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
// NO connections - AI Agent is isolated
|
||||
}
|
||||
};
|
||||
|
||||
console.log('Workflow:', JSON.stringify(workflow1, null, 2));
|
||||
|
||||
const reverseMap1 = buildReverseConnectionMap(workflow1);
|
||||
console.log('\nReverse connection map for AI Agent:');
|
||||
console.log('Entries:', Array.from(reverseMap1.entries()));
|
||||
console.log('AI Agent connections:', reverseMap1.get('AI Agent'));
|
||||
|
||||
// Check node normalization
|
||||
const normalizedType1 = NodeTypeNormalizer.normalizeToFullForm(workflow1.nodes[0].type);
|
||||
console.log(`\nNode type: ${workflow1.nodes[0].type}`);
|
||||
console.log(`Normalized type: ${normalizedType1}`);
|
||||
console.log(`Match check: ${normalizedType1 === '@n8n/n8n-nodes-langchain.agent'}`);
|
||||
|
||||
const issues1 = validateAISpecificNodes(workflow1);
|
||||
console.log('\nValidation issues:');
|
||||
console.log(JSON.stringify(issues1, null, 2));
|
||||
|
||||
const hasMissingLMError = issues1.some(
|
||||
i => i.severity === 'error' && i.code === 'MISSING_LANGUAGE_MODEL'
|
||||
);
|
||||
console.log(`\n✓ Has MISSING_LANGUAGE_MODEL error: ${hasMissingLMError}`);
|
||||
console.log(`✗ Expected: true, Got: ${hasMissingLMError}`);
|
||||
|
||||
// Test 2: AI Agent WITH language model connection
|
||||
console.log('\n\n' + '='.repeat(60));
|
||||
console.log('Test 2: AI Agent WITH Language Model (Should be valid)');
|
||||
const workflow2: WorkflowJson = {
|
||||
name: 'Test With LM',
|
||||
nodes: [
|
||||
{
|
||||
id: 'openai-1',
|
||||
name: 'OpenAI Chat Model',
|
||||
type: '@n8n/n8n-nodes-langchain.lmChatOpenAi',
|
||||
position: [200, 300],
|
||||
parameters: {
|
||||
modelName: 'gpt-4'
|
||||
},
|
||||
typeVersion: 1
|
||||
},
|
||||
{
|
||||
id: 'ai-agent-1',
|
||||
name: 'AI Agent',
|
||||
type: '@n8n/n8n-nodes-langchain.agent',
|
||||
position: [500, 300],
|
||||
parameters: {
|
||||
promptType: 'define',
|
||||
text: 'You are a helpful assistant'
|
||||
},
|
||||
typeVersion: 1.7
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'OpenAI Chat Model': {
|
||||
ai_languageModel: [
|
||||
[
|
||||
{
|
||||
node: 'AI Agent',
|
||||
type: 'ai_languageModel',
|
||||
index: 0
|
||||
}
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
console.log('\nConnections:', JSON.stringify(workflow2.connections, null, 2));
|
||||
|
||||
const reverseMap2 = buildReverseConnectionMap(workflow2);
|
||||
console.log('\nReverse connection map for AI Agent:');
|
||||
console.log('AI Agent connections:', reverseMap2.get('AI Agent'));
|
||||
|
||||
const issues2 = validateAISpecificNodes(workflow2);
|
||||
console.log('\nValidation issues:');
|
||||
console.log(JSON.stringify(issues2, null, 2));
|
||||
|
||||
const hasMissingLMError2 = issues2.some(
|
||||
i => i.severity === 'error' && i.code === 'MISSING_LANGUAGE_MODEL'
|
||||
);
|
||||
console.log(`\n✓ Should NOT have MISSING_LANGUAGE_MODEL error: ${!hasMissingLMError2}`);
|
||||
console.log(`Expected: false, Got: ${hasMissingLMError2}`);
|
||||
|
||||
// Test 3: AI Agent with tools but no language model
|
||||
console.log('\n\n' + '='.repeat(60));
|
||||
console.log('Test 3: AI Agent with Tools but NO Language Model');
|
||||
const workflow3: WorkflowJson = {
|
||||
name: 'Test Tools No LM',
|
||||
nodes: [
|
||||
{
|
||||
id: 'http-tool-1',
|
||||
name: 'HTTP Request Tool',
|
||||
type: '@n8n/n8n-nodes-langchain.toolHttpRequest',
|
||||
position: [200, 300],
|
||||
parameters: {
|
||||
toolDescription: 'Calls an API',
|
||||
url: 'https://api.example.com'
|
||||
},
|
||||
typeVersion: 1.1
|
||||
},
|
||||
{
|
||||
id: 'ai-agent-1',
|
||||
name: 'AI Agent',
|
||||
type: '@n8n/n8n-nodes-langchain.agent',
|
||||
position: [500, 300],
|
||||
parameters: {
|
||||
promptType: 'define',
|
||||
text: 'You are a helpful assistant'
|
||||
},
|
||||
typeVersion: 1.7
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'HTTP Request Tool': {
|
||||
ai_tool: [
|
||||
[
|
||||
{
|
||||
node: 'AI Agent',
|
||||
type: 'ai_tool',
|
||||
index: 0
|
||||
}
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
console.log('\nConnections:', JSON.stringify(workflow3.connections, null, 2));
|
||||
|
||||
const reverseMap3 = buildReverseConnectionMap(workflow3);
|
||||
console.log('\nReverse connection map for AI Agent:');
|
||||
const aiAgentConns = reverseMap3.get('AI Agent');
|
||||
console.log('AI Agent connections:', aiAgentConns);
|
||||
console.log('Connection types:', aiAgentConns?.map(c => c.type));
|
||||
|
||||
const issues3 = validateAISpecificNodes(workflow3);
|
||||
console.log('\nValidation issues:');
|
||||
console.log(JSON.stringify(issues3, null, 2));
|
||||
|
||||
const hasMissingLMError3 = issues3.some(
|
||||
i => i.severity === 'error' && i.code === 'MISSING_LANGUAGE_MODEL'
|
||||
);
|
||||
const hasNoToolsInfo3 = issues3.some(
|
||||
i => i.severity === 'info' && i.message.includes('no ai_tool connections')
|
||||
);
|
||||
|
||||
console.log(`\n✓ Should have MISSING_LANGUAGE_MODEL error: ${hasMissingLMError3}`);
|
||||
console.log(`Expected: true, Got: ${hasMissingLMError3}`);
|
||||
console.log(`✗ Should NOT have "no tools" info: ${!hasNoToolsInfo3}`);
|
||||
console.log(`Expected: false, Got: ${hasNoToolsInfo3}`);
|
||||
|
||||
console.log('\n' + '='.repeat(60));
|
||||
console.log('Summary:');
|
||||
console.log(`Test 1 (No LM): ${hasMissingLMError ? 'PASS ✓' : 'FAIL ✗'}`);
|
||||
console.log(`Test 2 (With LM): ${!hasMissingLMError2 ? 'PASS ✓' : 'FAIL ✗'}`);
|
||||
console.log(`Test 3 (Tools, No LM): ${hasMissingLMError3 && !hasNoToolsInfo3 ? 'PASS ✓' : 'FAIL ✗'}`);
|
||||
@@ -10,7 +10,7 @@ import { getToolDocumentation } from '../src/mcp/tools-documentation';
|
||||
import { ExampleGenerator } from '../src/services/example-generator';
|
||||
import { EnhancedConfigValidator } from '../src/services/enhanced-config-validator';
|
||||
|
||||
const dbPath = process.env.NODE_DB_PATH || './nodes.db';
|
||||
const dbPath = process.env.NODE_DB_PATH || './data/nodes.db';
|
||||
|
||||
async function main() {
|
||||
console.log('🧪 Testing Code Node Documentation Fixes\n');
|
||||
|
||||
163
scripts/test-docker-fingerprint.ts
Normal file
163
scripts/test-docker-fingerprint.ts
Normal file
@@ -0,0 +1,163 @@
|
||||
/**
|
||||
* Test Docker Host Fingerprinting
|
||||
* Verifies that host machine characteristics are stable across container recreations
|
||||
*/
|
||||
|
||||
import { existsSync, readFileSync } from 'fs';
|
||||
import { platform, arch } from 'os';
|
||||
import { createHash } from 'crypto';
|
||||
|
||||
console.log('=== Docker Host Fingerprinting Test ===\n');
|
||||
|
||||
function generateHostFingerprint(): string {
|
||||
try {
|
||||
const signals: string[] = [];
|
||||
|
||||
console.log('Collecting host signals...\n');
|
||||
|
||||
// CPU info (stable across container recreations)
|
||||
if (existsSync('/proc/cpuinfo')) {
|
||||
const cpuinfo = readFileSync('/proc/cpuinfo', 'utf-8');
|
||||
const modelMatch = cpuinfo.match(/model name\s*:\s*(.+)/);
|
||||
const coresMatch = cpuinfo.match(/processor\s*:/g);
|
||||
|
||||
if (modelMatch) {
|
||||
const cpuModel = modelMatch[1].trim();
|
||||
signals.push(cpuModel);
|
||||
console.log('✓ CPU Model:', cpuModel);
|
||||
}
|
||||
|
||||
if (coresMatch) {
|
||||
const cores = `cores:${coresMatch.length}`;
|
||||
signals.push(cores);
|
||||
console.log('✓ CPU Cores:', coresMatch.length);
|
||||
}
|
||||
} else {
|
||||
console.log('✗ /proc/cpuinfo not available (Windows/Mac Docker)');
|
||||
}
|
||||
|
||||
// Memory (stable)
|
||||
if (existsSync('/proc/meminfo')) {
|
||||
const meminfo = readFileSync('/proc/meminfo', 'utf-8');
|
||||
const totalMatch = meminfo.match(/MemTotal:\s+(\d+)/);
|
||||
|
||||
if (totalMatch) {
|
||||
const memory = `mem:${totalMatch[1]}`;
|
||||
signals.push(memory);
|
||||
console.log('✓ Total Memory:', totalMatch[1], 'kB');
|
||||
}
|
||||
} else {
|
||||
console.log('✗ /proc/meminfo not available (Windows/Mac Docker)');
|
||||
}
|
||||
|
||||
// Docker network subnet
|
||||
const networkInfo = getDockerNetworkInfo();
|
||||
if (networkInfo) {
|
||||
signals.push(networkInfo);
|
||||
console.log('✓ Network Info:', networkInfo);
|
||||
} else {
|
||||
console.log('✗ Network info not available');
|
||||
}
|
||||
|
||||
// Platform basics (stable)
|
||||
signals.push(platform(), arch());
|
||||
console.log('✓ Platform:', platform());
|
||||
console.log('✓ Architecture:', arch());
|
||||
|
||||
// Generate stable ID from all signals
|
||||
console.log('\nCombined signals:', signals.join(' | '));
|
||||
const fingerprint = signals.join('-');
|
||||
const userId = createHash('sha256').update(fingerprint).digest('hex').substring(0, 16);
|
||||
|
||||
return userId;
|
||||
|
||||
} catch (error) {
|
||||
console.error('Error generating fingerprint:', error);
|
||||
// Fallback
|
||||
return createHash('sha256')
|
||||
.update(`${platform()}-${arch()}-docker`)
|
||||
.digest('hex')
|
||||
.substring(0, 16);
|
||||
}
|
||||
}
|
||||
|
||||
function getDockerNetworkInfo(): string | null {
|
||||
try {
|
||||
// Read routing table to get bridge network
|
||||
if (existsSync('/proc/net/route')) {
|
||||
const routes = readFileSync('/proc/net/route', 'utf-8');
|
||||
const lines = routes.split('\n');
|
||||
|
||||
for (const line of lines) {
|
||||
if (line.includes('eth0')) {
|
||||
const parts = line.split(/\s+/);
|
||||
if (parts[2]) {
|
||||
const gateway = parseInt(parts[2], 16).toString(16);
|
||||
return `net:${gateway}`;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch {
|
||||
// Ignore errors
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
// Test environment detection
|
||||
console.log('\n=== Environment Detection ===\n');
|
||||
|
||||
const isDocker = process.env.IS_DOCKER === 'true';
|
||||
const isCloudEnvironment = !!(
|
||||
process.env.RAILWAY_ENVIRONMENT ||
|
||||
process.env.RENDER ||
|
||||
process.env.FLY_APP_NAME ||
|
||||
process.env.HEROKU_APP_NAME ||
|
||||
process.env.AWS_EXECUTION_ENV ||
|
||||
process.env.KUBERNETES_SERVICE_HOST
|
||||
);
|
||||
|
||||
console.log('IS_DOCKER env:', process.env.IS_DOCKER);
|
||||
console.log('Docker detected:', isDocker);
|
||||
console.log('Cloud environment:', isCloudEnvironment);
|
||||
|
||||
// Generate fingerprints
|
||||
console.log('\n=== Fingerprint Generation ===\n');
|
||||
|
||||
const fingerprint1 = generateHostFingerprint();
|
||||
const fingerprint2 = generateHostFingerprint();
|
||||
const fingerprint3 = generateHostFingerprint();
|
||||
|
||||
console.log('\nFingerprint 1:', fingerprint1);
|
||||
console.log('Fingerprint 2:', fingerprint2);
|
||||
console.log('Fingerprint 3:', fingerprint3);
|
||||
|
||||
const consistent = fingerprint1 === fingerprint2 && fingerprint2 === fingerprint3;
|
||||
console.log('\nConsistent:', consistent ? '✓ YES' : '✗ NO');
|
||||
|
||||
// Test explicit ID override
|
||||
console.log('\n=== Environment Variable Override Test ===\n');
|
||||
|
||||
if (process.env.N8N_MCP_USER_ID) {
|
||||
console.log('Explicit user ID:', process.env.N8N_MCP_USER_ID);
|
||||
console.log('This would override the fingerprint');
|
||||
} else {
|
||||
console.log('No explicit user ID set');
|
||||
console.log('To test: N8N_MCP_USER_ID=my-custom-id npx tsx ' + process.argv[1]);
|
||||
}
|
||||
|
||||
// Stability estimate
|
||||
console.log('\n=== Stability Analysis ===\n');
|
||||
|
||||
const hasStableSignals = existsSync('/proc/cpuinfo') || existsSync('/proc/meminfo');
|
||||
if (hasStableSignals) {
|
||||
console.log('✓ Host-based signals available');
|
||||
console.log('✓ Fingerprint should be stable across container recreations');
|
||||
console.log('✓ Different fingerprints on different physical hosts');
|
||||
} else {
|
||||
console.log('⚠️ Limited host signals (Windows/Mac Docker Desktop)');
|
||||
console.log('⚠️ Fingerprint may not be fully stable');
|
||||
console.log('💡 Recommendation: Use N8N_MCP_USER_ID env var for stability');
|
||||
}
|
||||
|
||||
console.log('\n');
|
||||
58
scripts/test-error-message-tracking.ts
Normal file
58
scripts/test-error-message-tracking.ts
Normal file
@@ -0,0 +1,58 @@
|
||||
/**
|
||||
* Test script to verify error message tracking is working
|
||||
*/
|
||||
|
||||
import { telemetry } from '../src/telemetry';
|
||||
|
||||
async function testErrorTracking() {
|
||||
console.log('=== Testing Error Message Tracking ===\n');
|
||||
|
||||
// Track session first
|
||||
console.log('1. Starting session...');
|
||||
telemetry.trackSessionStart();
|
||||
|
||||
// Track an error WITH a message
|
||||
console.log('\n2. Tracking error WITH message:');
|
||||
const testErrorMessage = 'This is a test error message with sensitive data: password=secret123 and test@example.com';
|
||||
telemetry.trackError(
|
||||
'TypeError',
|
||||
'tool_execution',
|
||||
'test_tool',
|
||||
testErrorMessage
|
||||
);
|
||||
console.log(` Original message: "${testErrorMessage}"`);
|
||||
|
||||
// Track an error WITHOUT a message
|
||||
console.log('\n3. Tracking error WITHOUT message:');
|
||||
telemetry.trackError(
|
||||
'Error',
|
||||
'tool_execution',
|
||||
'test_tool2'
|
||||
);
|
||||
|
||||
// Check the event queue
|
||||
const metrics = telemetry.getMetrics();
|
||||
console.log('\n4. Telemetry metrics:');
|
||||
console.log(' Status:', metrics.status);
|
||||
console.log(' Events queued:', metrics.tracking.eventsQueued);
|
||||
|
||||
// Get raw event queue to inspect
|
||||
const eventTracker = (telemetry as any).eventTracker;
|
||||
const queue = eventTracker.getEventQueue();
|
||||
|
||||
console.log('\n5. Event queue contents:');
|
||||
queue.forEach((event, i) => {
|
||||
console.log(`\n Event ${i + 1}:`);
|
||||
console.log(` - Type: ${event.event}`);
|
||||
console.log(` - Properties:`, JSON.stringify(event.properties, null, 6));
|
||||
});
|
||||
|
||||
// Flush to database
|
||||
console.log('\n6. Flushing to database...');
|
||||
await telemetry.flush();
|
||||
|
||||
console.log('\n7. Done! Check Supabase for error events with "error" field.');
|
||||
console.log(' Query: SELECT * FROM telemetry_events WHERE event = \'error_occurred\' ORDER BY created_at DESC LIMIT 5;');
|
||||
}
|
||||
|
||||
testErrorTracking().catch(console.error);
|
||||
274
scripts/test-error-output-validation.ts
Normal file
274
scripts/test-error-output-validation.ts
Normal file
@@ -0,0 +1,274 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
|
||||
/**
|
||||
* Test script for error output validation improvements
|
||||
* Tests both incorrect and correct error output configurations
|
||||
*/
|
||||
|
||||
import { WorkflowValidator } from '../dist/services/workflow-validator.js';
|
||||
import { NodeRepository } from '../dist/database/node-repository.js';
|
||||
import { EnhancedConfigValidator } from '../dist/services/enhanced-config-validator.js';
|
||||
import { DatabaseAdapter } from '../dist/database/database-adapter.js';
|
||||
import { Logger } from '../dist/utils/logger.js';
|
||||
import path from 'path';
|
||||
import { fileURLToPath } from 'url';
|
||||
|
||||
const __filename = fileURLToPath(import.meta.url);
|
||||
const __dirname = path.dirname(__filename);
|
||||
|
||||
const logger = new Logger({ prefix: '[TestErrorValidation]' });
|
||||
|
||||
async function runTests() {
|
||||
// Initialize database
|
||||
const dbPath = path.join(__dirname, '..', 'data', 'n8n-nodes.db');
|
||||
const adapter = new DatabaseAdapter();
|
||||
adapter.initialize({
|
||||
type: 'better-sqlite3',
|
||||
filename: dbPath
|
||||
});
|
||||
const db = adapter.getDatabase();
|
||||
|
||||
const nodeRepository = new NodeRepository(db);
|
||||
const validator = new WorkflowValidator(nodeRepository, EnhancedConfigValidator);
|
||||
|
||||
console.log('\n🧪 Testing Error Output Validation Improvements\n');
|
||||
console.log('=' .repeat(60));
|
||||
|
||||
// Test 1: Incorrect configuration - multiple nodes in same array
|
||||
console.log('\n📝 Test 1: INCORRECT - Multiple nodes in main[0]');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const incorrectWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '132ef0dc-87af-41de-a95d-cabe3a0a5342',
|
||||
name: 'Validate Input',
|
||||
type: 'n8n-nodes-base.set',
|
||||
typeVersion: 3.4,
|
||||
position: [-400, 64] as [number, number],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '5dedf217-63f9-409f-b34e-7780b22e199a',
|
||||
name: 'Filter URLs',
|
||||
type: 'n8n-nodes-base.filter',
|
||||
typeVersion: 2.2,
|
||||
position: [-176, 64] as [number, number],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '9d5407cc-ca5a-4966-b4b7-0e5dfbf54ad3',
|
||||
name: 'Error Response1',
|
||||
type: 'n8n-nodes-base.respondToWebhook',
|
||||
typeVersion: 1.5,
|
||||
position: [-160, 240] as [number, number],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Validate Input': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Filter URLs', type: 'main', index: 0 },
|
||||
{ node: 'Error Response1', type: 'main', index: 0 } // WRONG!
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result1 = await validator.validateWorkflow(incorrectWorkflow);
|
||||
|
||||
if (result1.errors.length > 0) {
|
||||
console.log('❌ ERROR DETECTED (as expected):');
|
||||
const errorMessage = result1.errors.find(e =>
|
||||
e.message.includes('Incorrect error output configuration')
|
||||
);
|
||||
if (errorMessage) {
|
||||
console.log('\n' + errorMessage.message);
|
||||
}
|
||||
} else {
|
||||
console.log('✅ No errors found (but should have detected the issue!)');
|
||||
}
|
||||
|
||||
// Test 2: Correct configuration - separate arrays
|
||||
console.log('\n📝 Test 2: CORRECT - Separate main[0] and main[1]');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const correctWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '132ef0dc-87af-41de-a95d-cabe3a0a5342',
|
||||
name: 'Validate Input',
|
||||
type: 'n8n-nodes-base.set',
|
||||
typeVersion: 3.4,
|
||||
position: [-400, 64] as [number, number],
|
||||
parameters: {},
|
||||
onError: 'continueErrorOutput' as const
|
||||
},
|
||||
{
|
||||
id: '5dedf217-63f9-409f-b34e-7780b22e199a',
|
||||
name: 'Filter URLs',
|
||||
type: 'n8n-nodes-base.filter',
|
||||
typeVersion: 2.2,
|
||||
position: [-176, 64] as [number, number],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '9d5407cc-ca5a-4966-b4b7-0e5dfbf54ad3',
|
||||
name: 'Error Response1',
|
||||
type: 'n8n-nodes-base.respondToWebhook',
|
||||
typeVersion: 1.5,
|
||||
position: [-160, 240] as [number, number],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Validate Input': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Filter URLs', type: 'main', index: 0 }
|
||||
],
|
||||
[
|
||||
{ node: 'Error Response1', type: 'main', index: 0 } // CORRECT!
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result2 = await validator.validateWorkflow(correctWorkflow);
|
||||
|
||||
const hasIncorrectError = result2.errors.some(e =>
|
||||
e.message.includes('Incorrect error output configuration')
|
||||
);
|
||||
|
||||
if (!hasIncorrectError) {
|
||||
console.log('✅ No error output configuration issues (correct!)');
|
||||
} else {
|
||||
console.log('❌ Unexpected error found');
|
||||
}
|
||||
|
||||
// Test 3: onError without error connections
|
||||
console.log('\n📝 Test 3: onError without error connections');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const mismatchWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'HTTP Request',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
typeVersion: 4,
|
||||
position: [100, 100] as [number, number],
|
||||
parameters: {},
|
||||
onError: 'continueErrorOutput' as const
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Process Data',
|
||||
type: 'n8n-nodes-base.set',
|
||||
typeVersion: 2,
|
||||
position: [300, 100] as [number, number],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'HTTP Request': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Process Data', type: 'main', index: 0 }
|
||||
]
|
||||
// No main[1] for error output
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result3 = await validator.validateWorkflow(mismatchWorkflow);
|
||||
|
||||
const mismatchError = result3.errors.find(e =>
|
||||
e.message.includes("has onError: 'continueErrorOutput' but no error output connections")
|
||||
);
|
||||
|
||||
if (mismatchError) {
|
||||
console.log('❌ ERROR DETECTED (as expected):');
|
||||
console.log(`Node: ${mismatchError.nodeName}`);
|
||||
console.log(`Message: ${mismatchError.message}`);
|
||||
} else {
|
||||
console.log('✅ No mismatch detected (but should have!)');
|
||||
}
|
||||
|
||||
// Test 4: Error connections without onError
|
||||
console.log('\n📝 Test 4: Error connections without onError property');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const missingOnErrorWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'HTTP Request',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
typeVersion: 4,
|
||||
position: [100, 100] as [number, number],
|
||||
parameters: {}
|
||||
// Missing onError property
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Process Data',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100] as [number, number],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'Error Handler',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 300] as [number, number],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'HTTP Request': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Process Data', type: 'main', index: 0 }
|
||||
],
|
||||
[
|
||||
{ node: 'Error Handler', type: 'main', index: 0 }
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result4 = await validator.validateWorkflow(missingOnErrorWorkflow);
|
||||
|
||||
const missingOnErrorWarning = result4.warnings.find(w =>
|
||||
w.message.includes('error output connections in main[1] but missing onError')
|
||||
);
|
||||
|
||||
if (missingOnErrorWarning) {
|
||||
console.log('⚠️ WARNING DETECTED (as expected):');
|
||||
console.log(`Node: ${missingOnErrorWarning.nodeName}`);
|
||||
console.log(`Message: ${missingOnErrorWarning.message}`);
|
||||
} else {
|
||||
console.log('✅ No warning (but should have warned!)');
|
||||
}
|
||||
|
||||
console.log('\n' + '='.repeat(60));
|
||||
console.log('\n📊 Summary:');
|
||||
console.log('- Error output validation is working correctly');
|
||||
console.log('- Detects incorrect configurations (multiple nodes in main[0])');
|
||||
console.log('- Validates onError property matches connections');
|
||||
console.log('- Provides clear error messages with fix examples');
|
||||
|
||||
// Close database
|
||||
adapter.close();
|
||||
}
|
||||
|
||||
runTests().catch(error => {
|
||||
console.error('Test failed:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
158
scripts/test-error-validation.js
Normal file
158
scripts/test-error-validation.js
Normal file
@@ -0,0 +1,158 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Test script for error output validation improvements
|
||||
*/
|
||||
|
||||
const { WorkflowValidator } = require('../dist/services/workflow-validator.js');
|
||||
const { NodeRepository } = require('../dist/database/node-repository.js');
|
||||
const { EnhancedConfigValidator } = require('../dist/services/enhanced-config-validator.js');
|
||||
const Database = require('better-sqlite3');
|
||||
const path = require('path');
|
||||
|
||||
async function runTests() {
|
||||
// Initialize database
|
||||
const dbPath = path.join(__dirname, '..', 'data', 'nodes.db');
|
||||
const db = new Database(dbPath, { readonly: true });
|
||||
|
||||
const nodeRepository = new NodeRepository(db);
|
||||
const validator = new WorkflowValidator(nodeRepository, EnhancedConfigValidator);
|
||||
|
||||
console.log('\n🧪 Testing Error Output Validation Improvements\n');
|
||||
console.log('=' .repeat(60));
|
||||
|
||||
// Test 1: Incorrect configuration - multiple nodes in same array
|
||||
console.log('\n📝 Test 1: INCORRECT - Multiple nodes in main[0]');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const incorrectWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '132ef0dc-87af-41de-a95d-cabe3a0a5342',
|
||||
name: 'Validate Input',
|
||||
type: 'n8n-nodes-base.set',
|
||||
typeVersion: 3.4,
|
||||
position: [-400, 64],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '5dedf217-63f9-409f-b34e-7780b22e199a',
|
||||
name: 'Filter URLs',
|
||||
type: 'n8n-nodes-base.filter',
|
||||
typeVersion: 2.2,
|
||||
position: [-176, 64],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '9d5407cc-ca5a-4966-b4b7-0e5dfbf54ad3',
|
||||
name: 'Error Response1',
|
||||
type: 'n8n-nodes-base.respondToWebhook',
|
||||
typeVersion: 1.5,
|
||||
position: [-160, 240],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Validate Input': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Filter URLs', type: 'main', index: 0 },
|
||||
{ node: 'Error Response1', type: 'main', index: 0 } // WRONG!
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result1 = await validator.validateWorkflow(incorrectWorkflow);
|
||||
|
||||
if (result1.errors.length > 0) {
|
||||
console.log('❌ ERROR DETECTED (as expected):');
|
||||
const errorMessage = result1.errors.find(e =>
|
||||
e.message.includes('Incorrect error output configuration')
|
||||
);
|
||||
if (errorMessage) {
|
||||
console.log('\nError Summary:');
|
||||
console.log(`Node: ${errorMessage.nodeName || 'Validate Input'}`);
|
||||
console.log('\nFull Error Message:');
|
||||
console.log(errorMessage.message);
|
||||
} else {
|
||||
console.log('Other errors found:', result1.errors.map(e => e.message));
|
||||
}
|
||||
} else {
|
||||
console.log('⚠️ No errors found - validation may not be working correctly');
|
||||
}
|
||||
|
||||
// Test 2: Correct configuration - separate arrays
|
||||
console.log('\n📝 Test 2: CORRECT - Separate main[0] and main[1]');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const correctWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '132ef0dc-87af-41de-a95d-cabe3a0a5342',
|
||||
name: 'Validate Input',
|
||||
type: 'n8n-nodes-base.set',
|
||||
typeVersion: 3.4,
|
||||
position: [-400, 64],
|
||||
parameters: {},
|
||||
onError: 'continueErrorOutput'
|
||||
},
|
||||
{
|
||||
id: '5dedf217-63f9-409f-b34e-7780b22e199a',
|
||||
name: 'Filter URLs',
|
||||
type: 'n8n-nodes-base.filter',
|
||||
typeVersion: 2.2,
|
||||
position: [-176, 64],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '9d5407cc-ca5a-4966-b4b7-0e5dfbf54ad3',
|
||||
name: 'Error Response1',
|
||||
type: 'n8n-nodes-base.respondToWebhook',
|
||||
typeVersion: 1.5,
|
||||
position: [-160, 240],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Validate Input': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Filter URLs', type: 'main', index: 0 }
|
||||
],
|
||||
[
|
||||
{ node: 'Error Response1', type: 'main', index: 0 } // CORRECT!
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result2 = await validator.validateWorkflow(correctWorkflow);
|
||||
|
||||
const hasIncorrectError = result2.errors.some(e =>
|
||||
e.message.includes('Incorrect error output configuration')
|
||||
);
|
||||
|
||||
if (!hasIncorrectError) {
|
||||
console.log('✅ No error output configuration issues (correct!)');
|
||||
} else {
|
||||
console.log('❌ Unexpected error found');
|
||||
}
|
||||
|
||||
console.log('\n' + '='.repeat(60));
|
||||
console.log('\n✨ Error output validation is working correctly!');
|
||||
console.log('The validator now properly detects:');
|
||||
console.log(' 1. Multiple nodes incorrectly placed in main[0]');
|
||||
console.log(' 2. Provides clear JSON examples for fixing issues');
|
||||
console.log(' 3. Validates onError property matches connections');
|
||||
|
||||
// Close database
|
||||
db.close();
|
||||
}
|
||||
|
||||
runTests().catch(error => {
|
||||
console.error('Test failed:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
230
scripts/test-expression-format-validation.js
Normal file
230
scripts/test-expression-format-validation.js
Normal file
@@ -0,0 +1,230 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Test script for expression format validation
|
||||
* Tests the validation of expression prefixes and resource locator formats
|
||||
*/
|
||||
|
||||
const { WorkflowValidator } = require('../dist/services/workflow-validator.js');
|
||||
const { NodeRepository } = require('../dist/database/node-repository.js');
|
||||
const { EnhancedConfigValidator } = require('../dist/services/enhanced-config-validator.js');
|
||||
const { createDatabaseAdapter } = require('../dist/database/database-adapter.js');
|
||||
const path = require('path');
|
||||
|
||||
async function runTests() {
|
||||
// Initialize database
|
||||
const dbPath = path.join(__dirname, '..', 'data', 'nodes.db');
|
||||
const adapter = await createDatabaseAdapter(dbPath);
|
||||
const db = adapter;
|
||||
|
||||
const nodeRepository = new NodeRepository(db);
|
||||
const validator = new WorkflowValidator(nodeRepository, EnhancedConfigValidator);
|
||||
|
||||
console.log('\n🧪 Testing Expression Format Validation\n');
|
||||
console.log('=' .repeat(60));
|
||||
|
||||
// Test 1: Email node with missing = prefix
|
||||
console.log('\n📝 Test 1: Email Send node - Missing = prefix');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const emailWorkflowIncorrect = {
|
||||
nodes: [
|
||||
{
|
||||
id: 'b9dd1cfd-ee66-4049-97e7-1af6d976a4e0',
|
||||
name: 'Error Handler',
|
||||
type: 'n8n-nodes-base.emailSend',
|
||||
typeVersion: 2.1,
|
||||
position: [-128, 400],
|
||||
parameters: {
|
||||
fromEmail: '{{ $env.ADMIN_EMAIL }}', // INCORRECT - missing =
|
||||
toEmail: 'admin@company.com',
|
||||
subject: 'GitHub Issue Workflow Error - HIGH PRIORITY',
|
||||
options: {}
|
||||
},
|
||||
credentials: {
|
||||
smtp: {
|
||||
id: '7AQ08VMFHubmfvzR',
|
||||
name: 'romuald@aiadvisors.pl'
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result1 = await validator.validateWorkflow(emailWorkflowIncorrect);
|
||||
|
||||
if (result1.errors.some(e => e.message.includes('Expression format'))) {
|
||||
console.log('✅ ERROR DETECTED (correct behavior):');
|
||||
const formatError = result1.errors.find(e => e.message.includes('Expression format'));
|
||||
console.log('\n' + formatError.message);
|
||||
} else {
|
||||
console.log('❌ No expression format error detected (should have detected missing prefix)');
|
||||
}
|
||||
|
||||
// Test 2: Email node with correct = prefix
|
||||
console.log('\n📝 Test 2: Email Send node - Correct = prefix');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const emailWorkflowCorrect = {
|
||||
nodes: [
|
||||
{
|
||||
id: 'b9dd1cfd-ee66-4049-97e7-1af6d976a4e0',
|
||||
name: 'Error Handler',
|
||||
type: 'n8n-nodes-base.emailSend',
|
||||
typeVersion: 2.1,
|
||||
position: [-128, 400],
|
||||
parameters: {
|
||||
fromEmail: '={{ $env.ADMIN_EMAIL }}', // CORRECT - has =
|
||||
toEmail: 'admin@company.com',
|
||||
subject: 'GitHub Issue Workflow Error - HIGH PRIORITY',
|
||||
options: {}
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result2 = await validator.validateWorkflow(emailWorkflowCorrect);
|
||||
|
||||
if (result2.errors.some(e => e.message.includes('Expression format'))) {
|
||||
console.log('❌ Unexpected expression format error (should accept = prefix)');
|
||||
} else {
|
||||
console.log('✅ No expression format errors (correct!)');
|
||||
}
|
||||
|
||||
// Test 3: GitHub node without resource locator format
|
||||
console.log('\n📝 Test 3: GitHub node - Missing resource locator format');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const githubWorkflowIncorrect = {
|
||||
nodes: [
|
||||
{
|
||||
id: '3c742ca1-af8f-4d80-a47e-e68fb1ced491',
|
||||
name: 'Send Welcome Comment',
|
||||
type: 'n8n-nodes-base.github',
|
||||
typeVersion: 1.1,
|
||||
position: [-240, 96],
|
||||
parameters: {
|
||||
operation: 'createComment',
|
||||
owner: '{{ $vars.GITHUB_OWNER }}', // INCORRECT - needs RL format
|
||||
repository: '{{ $vars.GITHUB_REPO }}', // INCORRECT - needs RL format
|
||||
issueNumber: null,
|
||||
body: '👋 Hi @{{ $(\'Extract Issue Data\').first().json.author }}!' // INCORRECT - missing =
|
||||
},
|
||||
credentials: {
|
||||
githubApi: {
|
||||
id: 'edgpwh6ldYN07MXx',
|
||||
name: 'GitHub account'
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result3 = await validator.validateWorkflow(githubWorkflowIncorrect);
|
||||
|
||||
const formatErrors = result3.errors.filter(e => e.message.includes('Expression format'));
|
||||
console.log(`\nFound ${formatErrors.length} expression format errors:`);
|
||||
|
||||
if (formatErrors.length >= 3) {
|
||||
console.log('✅ All format issues detected:');
|
||||
formatErrors.forEach((error, index) => {
|
||||
const field = error.message.match(/Field '([^']+)'/)?.[1] || 'unknown';
|
||||
console.log(` ${index + 1}. Field '${field}' - ${error.message.includes('resource locator') ? 'Needs RL format' : 'Missing = prefix'}`);
|
||||
});
|
||||
} else {
|
||||
console.log('❌ Not all format issues detected');
|
||||
}
|
||||
|
||||
// Test 4: GitHub node with correct resource locator format
|
||||
console.log('\n📝 Test 4: GitHub node - Correct resource locator format');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const githubWorkflowCorrect = {
|
||||
nodes: [
|
||||
{
|
||||
id: '3c742ca1-af8f-4d80-a47e-e68fb1ced491',
|
||||
name: 'Send Welcome Comment',
|
||||
type: 'n8n-nodes-base.github',
|
||||
typeVersion: 1.1,
|
||||
position: [-240, 96],
|
||||
parameters: {
|
||||
operation: 'createComment',
|
||||
owner: {
|
||||
__rl: true,
|
||||
value: '={{ $vars.GITHUB_OWNER }}', // CORRECT - RL format with =
|
||||
mode: 'expression'
|
||||
},
|
||||
repository: {
|
||||
__rl: true,
|
||||
value: '={{ $vars.GITHUB_REPO }}', // CORRECT - RL format with =
|
||||
mode: 'expression'
|
||||
},
|
||||
issueNumber: 123,
|
||||
body: '=👋 Hi @{{ $(\'Extract Issue Data\').first().json.author }}!' // CORRECT - has =
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result4 = await validator.validateWorkflow(githubWorkflowCorrect);
|
||||
|
||||
const formatErrors4 = result4.errors.filter(e => e.message.includes('Expression format'));
|
||||
if (formatErrors4.length === 0) {
|
||||
console.log('✅ No expression format errors (correct!)');
|
||||
} else {
|
||||
console.log(`❌ Unexpected expression format errors: ${formatErrors4.length}`);
|
||||
formatErrors4.forEach(e => console.log(' - ' + e.message.split('\n')[0]));
|
||||
}
|
||||
|
||||
// Test 5: Mixed content expressions
|
||||
console.log('\n📝 Test 5: Mixed content with expressions');
|
||||
console.log('-'.repeat(40));
|
||||
|
||||
const mixedContentWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'HTTP Request',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
typeVersion: 4,
|
||||
position: [0, 0],
|
||||
parameters: {
|
||||
url: 'https://api.example.com/users/{{ $json.userId }}', // INCORRECT
|
||||
headers: {
|
||||
'Authorization': '=Bearer {{ $env.API_TOKEN }}' // CORRECT
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result5 = await validator.validateWorkflow(mixedContentWorkflow);
|
||||
|
||||
const urlError = result5.errors.find(e => e.message.includes('url') && e.message.includes('Expression format'));
|
||||
if (urlError) {
|
||||
console.log('✅ Mixed content error detected for URL field');
|
||||
console.log(' Should be: "=https://api.example.com/users/{{ $json.userId }}"');
|
||||
} else {
|
||||
console.log('❌ Mixed content error not detected');
|
||||
}
|
||||
|
||||
console.log('\n' + '='.repeat(60));
|
||||
console.log('\n✨ Expression Format Validation Summary:');
|
||||
console.log(' - Detects missing = prefix in expressions');
|
||||
console.log(' - Identifies fields needing resource locator format');
|
||||
console.log(' - Provides clear correction examples');
|
||||
console.log(' - Handles mixed literal and expression content');
|
||||
|
||||
// Close database
|
||||
db.close();
|
||||
}
|
||||
|
||||
runTests().catch(error => {
|
||||
console.error('Test failed:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
126
scripts/test-multi-tenant-simple.ts
Normal file
126
scripts/test-multi-tenant-simple.ts
Normal file
@@ -0,0 +1,126 @@
|
||||
#!/usr/bin/env ts-node
|
||||
|
||||
/**
|
||||
* Simple test for multi-tenant functionality
|
||||
* Tests that tools are registered correctly based on configuration
|
||||
*/
|
||||
|
||||
import { isN8nApiConfigured } from '../src/config/n8n-api';
|
||||
import { InstanceContext } from '../src/types/instance-context';
|
||||
import dotenv from 'dotenv';
|
||||
|
||||
dotenv.config();
|
||||
|
||||
async function testMultiTenant() {
|
||||
console.log('🧪 Testing Multi-Tenant Tool Registration\n');
|
||||
console.log('=' .repeat(60));
|
||||
|
||||
// Save original environment
|
||||
const originalEnv = {
|
||||
ENABLE_MULTI_TENANT: process.env.ENABLE_MULTI_TENANT,
|
||||
N8N_API_URL: process.env.N8N_API_URL,
|
||||
N8N_API_KEY: process.env.N8N_API_KEY
|
||||
};
|
||||
|
||||
try {
|
||||
// Test 1: Default - no API config
|
||||
console.log('\n✅ Test 1: No API configuration');
|
||||
delete process.env.N8N_API_URL;
|
||||
delete process.env.N8N_API_KEY;
|
||||
delete process.env.ENABLE_MULTI_TENANT;
|
||||
|
||||
const hasConfig1 = isN8nApiConfigured();
|
||||
console.log(` Environment API configured: ${hasConfig1}`);
|
||||
console.log(` Multi-tenant enabled: ${process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
console.log(` Should show tools: ${hasConfig1 || process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
|
||||
// Test 2: Multi-tenant enabled
|
||||
console.log('\n✅ Test 2: Multi-tenant enabled (no env API)');
|
||||
process.env.ENABLE_MULTI_TENANT = 'true';
|
||||
|
||||
const hasConfig2 = isN8nApiConfigured();
|
||||
console.log(` Environment API configured: ${hasConfig2}`);
|
||||
console.log(` Multi-tenant enabled: ${process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
console.log(` Should show tools: ${hasConfig2 || process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
|
||||
// Test 3: Environment variables set
|
||||
console.log('\n✅ Test 3: Environment variables set');
|
||||
process.env.ENABLE_MULTI_TENANT = 'false';
|
||||
process.env.N8N_API_URL = 'https://test.n8n.cloud';
|
||||
process.env.N8N_API_KEY = 'test-key';
|
||||
|
||||
const hasConfig3 = isN8nApiConfigured();
|
||||
console.log(` Environment API configured: ${hasConfig3}`);
|
||||
console.log(` Multi-tenant enabled: ${process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
console.log(` Should show tools: ${hasConfig3 || process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
|
||||
// Test 4: Instance context simulation
|
||||
console.log('\n✅ Test 4: Instance context (simulated)');
|
||||
const instanceContext: InstanceContext = {
|
||||
n8nApiUrl: 'https://instance.n8n.cloud',
|
||||
n8nApiKey: 'instance-key',
|
||||
instanceId: 'test-instance'
|
||||
};
|
||||
|
||||
const hasInstanceConfig = !!(instanceContext.n8nApiUrl && instanceContext.n8nApiKey);
|
||||
console.log(` Instance has API config: ${hasInstanceConfig}`);
|
||||
console.log(` Environment API configured: ${hasConfig3}`);
|
||||
console.log(` Multi-tenant enabled: ${process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
console.log(` Should show tools: ${hasConfig3 || hasInstanceConfig || process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
|
||||
// Test 5: Multi-tenant with instance strategy
|
||||
console.log('\n✅ Test 5: Multi-tenant with instance strategy');
|
||||
process.env.ENABLE_MULTI_TENANT = 'true';
|
||||
process.env.MULTI_TENANT_SESSION_STRATEGY = 'instance';
|
||||
delete process.env.N8N_API_URL;
|
||||
delete process.env.N8N_API_KEY;
|
||||
|
||||
const hasConfig5 = isN8nApiConfigured();
|
||||
const sessionStrategy = process.env.MULTI_TENANT_SESSION_STRATEGY || 'instance';
|
||||
console.log(` Environment API configured: ${hasConfig5}`);
|
||||
console.log(` Multi-tenant enabled: ${process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
console.log(` Session strategy: ${sessionStrategy}`);
|
||||
console.log(` Should show tools: ${hasConfig5 || process.env.ENABLE_MULTI_TENANT === 'true'}`);
|
||||
|
||||
if (instanceContext.instanceId) {
|
||||
const sessionId = `instance-${instanceContext.instanceId}-uuid`;
|
||||
console.log(` Session ID format: ${sessionId}`);
|
||||
}
|
||||
|
||||
console.log('\n' + '=' .repeat(60));
|
||||
console.log('✅ All configuration tests passed!');
|
||||
console.log('\n📝 Summary:');
|
||||
console.log(' - Tools are shown when: env API configured OR multi-tenant enabled OR instance context provided');
|
||||
console.log(' - Session isolation works with instance-based session IDs in multi-tenant mode');
|
||||
console.log(' - Backward compatibility maintained for env-based configuration');
|
||||
|
||||
} catch (error) {
|
||||
console.error('\n❌ Test failed:', error);
|
||||
process.exit(1);
|
||||
} finally {
|
||||
// Restore original environment
|
||||
if (originalEnv.ENABLE_MULTI_TENANT !== undefined) {
|
||||
process.env.ENABLE_MULTI_TENANT = originalEnv.ENABLE_MULTI_TENANT;
|
||||
} else {
|
||||
delete process.env.ENABLE_MULTI_TENANT;
|
||||
}
|
||||
|
||||
if (originalEnv.N8N_API_URL !== undefined) {
|
||||
process.env.N8N_API_URL = originalEnv.N8N_API_URL;
|
||||
} else {
|
||||
delete process.env.N8N_API_URL;
|
||||
}
|
||||
|
||||
if (originalEnv.N8N_API_KEY !== undefined) {
|
||||
process.env.N8N_API_KEY = originalEnv.N8N_API_KEY;
|
||||
} else {
|
||||
delete process.env.N8N_API_KEY;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Run tests
|
||||
testMultiTenant().catch(error => {
|
||||
console.error('Test execution failed:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
136
scripts/test-multi-tenant.ts
Normal file
136
scripts/test-multi-tenant.ts
Normal file
@@ -0,0 +1,136 @@
|
||||
#!/usr/bin/env ts-node
|
||||
|
||||
/**
|
||||
* Test script for multi-tenant functionality
|
||||
* Verifies that instance context from headers enables n8n API tools
|
||||
*/
|
||||
|
||||
import { N8NDocumentationMCPServer } from '../src/mcp/server';
|
||||
import { InstanceContext } from '../src/types/instance-context';
|
||||
import { logger } from '../src/utils/logger';
|
||||
import dotenv from 'dotenv';
|
||||
|
||||
dotenv.config();
|
||||
|
||||
async function testMultiTenant() {
|
||||
console.log('🧪 Testing Multi-Tenant Functionality\n');
|
||||
console.log('=' .repeat(60));
|
||||
|
||||
// Save original environment
|
||||
const originalEnv = {
|
||||
ENABLE_MULTI_TENANT: process.env.ENABLE_MULTI_TENANT,
|
||||
N8N_API_URL: process.env.N8N_API_URL,
|
||||
N8N_API_KEY: process.env.N8N_API_KEY
|
||||
};
|
||||
|
||||
// Wait a moment for database initialization
|
||||
await new Promise(resolve => setTimeout(resolve, 100));
|
||||
|
||||
try {
|
||||
// Test 1: Without multi-tenant mode (default)
|
||||
console.log('\n📌 Test 1: Without multi-tenant mode (no env vars)');
|
||||
delete process.env.N8N_API_URL;
|
||||
delete process.env.N8N_API_KEY;
|
||||
process.env.ENABLE_MULTI_TENANT = 'false';
|
||||
|
||||
const server1 = new N8NDocumentationMCPServer();
|
||||
const tools1 = await getToolsFromServer(server1);
|
||||
const hasManagementTools1 = tools1.some(t => t.name.startsWith('n8n_'));
|
||||
console.log(` Tools available: ${tools1.length}`);
|
||||
console.log(` Has management tools: ${hasManagementTools1}`);
|
||||
console.log(` ✅ Expected: No management tools (correct: ${!hasManagementTools1})`);
|
||||
|
||||
// Test 2: With instance context but multi-tenant disabled
|
||||
console.log('\n📌 Test 2: With instance context but multi-tenant disabled');
|
||||
const instanceContext: InstanceContext = {
|
||||
n8nApiUrl: 'https://instance1.n8n.cloud',
|
||||
n8nApiKey: 'test-api-key',
|
||||
instanceId: 'instance-1'
|
||||
};
|
||||
|
||||
const server2 = new N8NDocumentationMCPServer(instanceContext);
|
||||
const tools2 = await getToolsFromServer(server2);
|
||||
const hasManagementTools2 = tools2.some(t => t.name.startsWith('n8n_'));
|
||||
console.log(` Tools available: ${tools2.length}`);
|
||||
console.log(` Has management tools: ${hasManagementTools2}`);
|
||||
console.log(` ✅ Expected: Has management tools (correct: ${hasManagementTools2})`);
|
||||
|
||||
// Test 3: With multi-tenant mode enabled
|
||||
console.log('\n📌 Test 3: With multi-tenant mode enabled');
|
||||
process.env.ENABLE_MULTI_TENANT = 'true';
|
||||
|
||||
const server3 = new N8NDocumentationMCPServer();
|
||||
const tools3 = await getToolsFromServer(server3);
|
||||
const hasManagementTools3 = tools3.some(t => t.name.startsWith('n8n_'));
|
||||
console.log(` Tools available: ${tools3.length}`);
|
||||
console.log(` Has management tools: ${hasManagementTools3}`);
|
||||
console.log(` ✅ Expected: Has management tools (correct: ${hasManagementTools3})`);
|
||||
|
||||
// Test 4: Multi-tenant with instance context
|
||||
console.log('\n📌 Test 4: Multi-tenant with instance context');
|
||||
const server4 = new N8NDocumentationMCPServer(instanceContext);
|
||||
const tools4 = await getToolsFromServer(server4);
|
||||
const hasManagementTools4 = tools4.some(t => t.name.startsWith('n8n_'));
|
||||
console.log(` Tools available: ${tools4.length}`);
|
||||
console.log(` Has management tools: ${hasManagementTools4}`);
|
||||
console.log(` ✅ Expected: Has management tools (correct: ${hasManagementTools4})`);
|
||||
|
||||
// Test 5: Environment variables (backward compatibility)
|
||||
console.log('\n📌 Test 5: Environment variables (backward compatibility)');
|
||||
process.env.ENABLE_MULTI_TENANT = 'false';
|
||||
process.env.N8N_API_URL = 'https://env.n8n.cloud';
|
||||
process.env.N8N_API_KEY = 'env-api-key';
|
||||
|
||||
const server5 = new N8NDocumentationMCPServer();
|
||||
const tools5 = await getToolsFromServer(server5);
|
||||
const hasManagementTools5 = tools5.some(t => t.name.startsWith('n8n_'));
|
||||
console.log(` Tools available: ${tools5.length}`);
|
||||
console.log(` Has management tools: ${hasManagementTools5}`);
|
||||
console.log(` ✅ Expected: Has management tools (correct: ${hasManagementTools5})`);
|
||||
|
||||
console.log('\n' + '=' .repeat(60));
|
||||
console.log('✅ All multi-tenant tests passed!');
|
||||
|
||||
} catch (error) {
|
||||
console.error('\n❌ Test failed:', error);
|
||||
process.exit(1);
|
||||
} finally {
|
||||
// Restore original environment
|
||||
Object.assign(process.env, originalEnv);
|
||||
}
|
||||
}
|
||||
|
||||
// Helper function to get tools from server
|
||||
async function getToolsFromServer(server: N8NDocumentationMCPServer): Promise<any[]> {
|
||||
// Access the private server instance to simulate tool listing
|
||||
const serverInstance = (server as any).server;
|
||||
const handlers = (serverInstance as any)._requestHandlers;
|
||||
|
||||
// Find and call the ListToolsRequestSchema handler
|
||||
if (handlers && handlers.size > 0) {
|
||||
for (const [schema, handler] of handlers) {
|
||||
// Check for the tools/list schema
|
||||
if (schema && schema.method === 'tools/list') {
|
||||
const result = await handler({ params: {} });
|
||||
return result.tools || [];
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Fallback: directly check the handlers map
|
||||
const ListToolsRequestSchema = { method: 'tools/list' };
|
||||
const handler = handlers?.get(ListToolsRequestSchema);
|
||||
if (handler) {
|
||||
const result = await handler({ params: {} });
|
||||
return result.tools || [];
|
||||
}
|
||||
|
||||
console.log(' ⚠️ Warning: Could not find tools/list handler');
|
||||
return [];
|
||||
}
|
||||
|
||||
// Run tests
|
||||
testMultiTenant().catch(error => {
|
||||
console.error('Test execution failed:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
@@ -1,95 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Test script for n8n MCP integration fixes
|
||||
set -e
|
||||
|
||||
echo "🔧 Testing n8n MCP Integration Fixes"
|
||||
echo "===================================="
|
||||
|
||||
# Configuration
|
||||
MCP_PORT=${MCP_PORT:-3001}
|
||||
AUTH_TOKEN=${AUTH_TOKEN:-"test-token-for-n8n-testing-minimum-32-chars"}
|
||||
|
||||
# Colors
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
RED='\033[0;31m'
|
||||
BLUE='\033[0;34m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
# Cleanup function
|
||||
cleanup() {
|
||||
echo -e "\n${YELLOW}🧹 Cleaning up...${NC}"
|
||||
if [ -n "$MCP_PID" ] && kill -0 $MCP_PID 2>/dev/null; then
|
||||
echo "Stopping MCP server..."
|
||||
kill $MCP_PID 2>/dev/null || true
|
||||
wait $MCP_PID 2>/dev/null || true
|
||||
fi
|
||||
echo -e "${GREEN}✅ Cleanup complete${NC}"
|
||||
}
|
||||
|
||||
trap cleanup EXIT INT TERM
|
||||
|
||||
# Check if we're in the right directory
|
||||
if [ ! -f "package.json" ] || [ ! -d "dist" ]; then
|
||||
echo -e "${RED}❌ Error: Must run from n8n-mcp directory${NC}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Build the project (our fixes)
|
||||
echo -e "${YELLOW}📦 Building project with fixes...${NC}"
|
||||
npm run build
|
||||
|
||||
# Start MCP server in n8n mode
|
||||
echo -e "\n${GREEN}🚀 Starting MCP server in n8n mode...${NC}"
|
||||
N8N_MODE=true \
|
||||
MCP_MODE=http \
|
||||
AUTH_TOKEN="${AUTH_TOKEN}" \
|
||||
PORT=${MCP_PORT} \
|
||||
DEBUG_MCP=true \
|
||||
node dist/mcp/index.js > /tmp/mcp-n8n-test.log 2>&1 &
|
||||
|
||||
MCP_PID=$!
|
||||
echo -e "${YELLOW}📄 MCP server logs: /tmp/mcp-n8n-test.log${NC}"
|
||||
|
||||
# Wait for server to start
|
||||
echo -e "${YELLOW}⏳ Waiting for MCP server to start...${NC}"
|
||||
for i in {1..15}; do
|
||||
if curl -s http://localhost:${MCP_PORT}/health >/dev/null 2>&1; then
|
||||
echo -e "${GREEN}✅ MCP server is ready!${NC}"
|
||||
break
|
||||
fi
|
||||
if [ $i -eq 15 ]; then
|
||||
echo -e "${RED}❌ MCP server failed to start${NC}"
|
||||
echo "Server logs:"
|
||||
cat /tmp/mcp-n8n-test.log
|
||||
exit 1
|
||||
fi
|
||||
sleep 1
|
||||
done
|
||||
|
||||
# Test the protocol fixes
|
||||
echo -e "\n${BLUE}🧪 Testing protocol fixes...${NC}"
|
||||
|
||||
# Run our debug script
|
||||
echo -e "${YELLOW}Running comprehensive MCP protocol tests...${NC}"
|
||||
node scripts/debug-n8n-mode.js
|
||||
|
||||
echo -e "\n${GREEN}🎉 Test complete!${NC}"
|
||||
echo -e "\n📋 Summary of fixes applied:"
|
||||
echo -e " ✅ Fixed protocol version mismatch (now using 2025-03-26)"
|
||||
echo -e " ✅ Enhanced tool response formatting and size validation"
|
||||
echo -e " ✅ Added comprehensive parameter validation"
|
||||
echo -e " ✅ Improved error handling and logging"
|
||||
echo -e " ✅ Added initialization request debugging"
|
||||
|
||||
echo -e "\n📝 Next steps:"
|
||||
echo -e " 1. If tests pass, the n8n schema validation errors should be resolved"
|
||||
echo -e " 2. Test with actual n8n MCP Client Tool node"
|
||||
echo -e " 3. Monitor logs at /tmp/mcp-n8n-test.log for any remaining issues"
|
||||
|
||||
echo -e "\n${YELLOW}Press any key to view recent server logs, or Ctrl+C to exit...${NC}"
|
||||
read -n 1
|
||||
|
||||
echo -e "\n${BLUE}📄 Recent server logs:${NC}"
|
||||
tail -50 /tmp/mcp-n8n-test.log
|
||||
@@ -1,428 +0,0 @@
|
||||
#!/usr/bin/env ts-node
|
||||
|
||||
/**
|
||||
* TypeScript test script for n8n MCP integration fixes
|
||||
* Tests the protocol changes and identifies any remaining issues
|
||||
*/
|
||||
|
||||
import http from 'http';
|
||||
import { spawn, ChildProcess } from 'child_process';
|
||||
import path from 'path';
|
||||
|
||||
interface TestResult {
|
||||
name: string;
|
||||
passed: boolean;
|
||||
error?: string;
|
||||
response?: any;
|
||||
}
|
||||
|
||||
class N8nMcpTester {
|
||||
private mcpProcess: ChildProcess | null = null;
|
||||
private readonly mcpPort = 3001;
|
||||
private readonly authToken = 'test-token-for-n8n-testing-minimum-32-chars';
|
||||
private sessionId: string | null = null;
|
||||
|
||||
async start(): Promise<void> {
|
||||
console.log('🔧 Testing n8n MCP Integration Fixes');
|
||||
console.log('====================================\n');
|
||||
|
||||
try {
|
||||
await this.startMcpServer();
|
||||
await this.runTests();
|
||||
} finally {
|
||||
await this.cleanup();
|
||||
}
|
||||
}
|
||||
|
||||
private async startMcpServer(): Promise<void> {
|
||||
console.log('📦 Starting MCP server in n8n mode...');
|
||||
|
||||
const projectRoot = path.resolve(__dirname, '..');
|
||||
|
||||
this.mcpProcess = spawn('node', ['dist/mcp/index.js'], {
|
||||
cwd: projectRoot,
|
||||
env: {
|
||||
...process.env,
|
||||
N8N_MODE: 'true',
|
||||
MCP_MODE: 'http',
|
||||
AUTH_TOKEN: this.authToken,
|
||||
PORT: this.mcpPort.toString(),
|
||||
DEBUG_MCP: 'true'
|
||||
},
|
||||
stdio: ['ignore', 'pipe', 'pipe']
|
||||
});
|
||||
|
||||
// Log server output
|
||||
this.mcpProcess.stdout?.on('data', (data) => {
|
||||
console.log(`[MCP] ${data.toString().trim()}`);
|
||||
});
|
||||
|
||||
this.mcpProcess.stderr?.on('data', (data) => {
|
||||
console.error(`[MCP ERROR] ${data.toString().trim()}`);
|
||||
});
|
||||
|
||||
// Wait for server to be ready
|
||||
await this.waitForServer();
|
||||
}
|
||||
|
||||
private async waitForServer(): Promise<void> {
|
||||
console.log('⏳ Waiting for MCP server to be ready...');
|
||||
|
||||
for (let i = 0; i < 30; i++) {
|
||||
try {
|
||||
await this.makeHealthCheck();
|
||||
console.log('✅ MCP server is ready!\n');
|
||||
return;
|
||||
} catch (error) {
|
||||
if (i === 29) {
|
||||
throw new Error('MCP server failed to start within 30 seconds');
|
||||
}
|
||||
await this.sleep(1000);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private makeHealthCheck(): Promise<void> {
|
||||
return new Promise((resolve, reject) => {
|
||||
const req = http.get(`http://localhost:${this.mcpPort}/health`, (res) => {
|
||||
if (res.statusCode === 200) {
|
||||
resolve();
|
||||
} else {
|
||||
reject(new Error(`Health check failed: ${res.statusCode}`));
|
||||
}
|
||||
});
|
||||
|
||||
req.on('error', reject);
|
||||
req.setTimeout(5000, () => {
|
||||
req.destroy();
|
||||
reject(new Error('Health check timeout'));
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
private async runTests(): Promise<void> {
|
||||
const tests: TestResult[] = [];
|
||||
|
||||
// Test 1: Initialize with correct protocol version
|
||||
tests.push(await this.testInitialize());
|
||||
|
||||
// Test 2: List tools
|
||||
tests.push(await this.testListTools());
|
||||
|
||||
// Test 3: Call tools_documentation
|
||||
tests.push(await this.testToolCall('tools_documentation', {}));
|
||||
|
||||
// Test 4: Call get_node_essentials with parameters
|
||||
tests.push(await this.testToolCall('get_node_essentials', {
|
||||
nodeType: 'nodes-base.httpRequest'
|
||||
}));
|
||||
|
||||
// Test 5: Call with invalid parameters (should handle gracefully)
|
||||
tests.push(await this.testToolCallInvalid());
|
||||
|
||||
this.printResults(tests);
|
||||
}
|
||||
|
||||
private async testInitialize(): Promise<TestResult> {
|
||||
console.log('🧪 Testing MCP Initialize...');
|
||||
|
||||
try {
|
||||
const response = await this.makeRequest('POST', '/mcp', {
|
||||
jsonrpc: '2.0',
|
||||
method: 'initialize',
|
||||
params: {
|
||||
protocolVersion: '2025-03-26',
|
||||
capabilities: { tools: {} },
|
||||
clientInfo: { name: 'n8n-test', version: '1.0.0' }
|
||||
},
|
||||
id: 1
|
||||
});
|
||||
|
||||
if (response.statusCode !== 200) {
|
||||
return {
|
||||
name: 'Initialize',
|
||||
passed: false,
|
||||
error: `HTTP ${response.statusCode}`
|
||||
};
|
||||
}
|
||||
|
||||
const data = JSON.parse(response.body);
|
||||
|
||||
// Extract session ID
|
||||
this.sessionId = response.headers['mcp-session-id'] as string;
|
||||
|
||||
if (data.result?.protocolVersion === '2025-03-26') {
|
||||
return {
|
||||
name: 'Initialize',
|
||||
passed: true,
|
||||
response: data
|
||||
};
|
||||
} else {
|
||||
return {
|
||||
name: 'Initialize',
|
||||
passed: false,
|
||||
error: `Wrong protocol version: ${data.result?.protocolVersion}`,
|
||||
response: data
|
||||
};
|
||||
}
|
||||
} catch (error) {
|
||||
return {
|
||||
name: 'Initialize',
|
||||
passed: false,
|
||||
error: error instanceof Error ? error.message : 'Unknown error'
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
private async testListTools(): Promise<TestResult> {
|
||||
console.log('🧪 Testing Tools List...');
|
||||
|
||||
try {
|
||||
const response = await this.makeRequest('POST', '/mcp', {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools/list',
|
||||
params: {},
|
||||
id: 2
|
||||
}, this.sessionId);
|
||||
|
||||
if (response.statusCode !== 200) {
|
||||
return {
|
||||
name: 'List Tools',
|
||||
passed: false,
|
||||
error: `HTTP ${response.statusCode}`
|
||||
};
|
||||
}
|
||||
|
||||
const data = JSON.parse(response.body);
|
||||
|
||||
if (data.result?.tools && Array.isArray(data.result.tools)) {
|
||||
return {
|
||||
name: 'List Tools',
|
||||
passed: true,
|
||||
response: { toolCount: data.result.tools.length }
|
||||
};
|
||||
} else {
|
||||
return {
|
||||
name: 'List Tools',
|
||||
passed: false,
|
||||
error: 'Missing or invalid tools array',
|
||||
response: data
|
||||
};
|
||||
}
|
||||
} catch (error) {
|
||||
return {
|
||||
name: 'List Tools',
|
||||
passed: false,
|
||||
error: error instanceof Error ? error.message : 'Unknown error'
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
private async testToolCall(toolName: string, args: any): Promise<TestResult> {
|
||||
console.log(`🧪 Testing Tool Call: ${toolName}...`);
|
||||
|
||||
try {
|
||||
const response = await this.makeRequest('POST', '/mcp', {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools/call',
|
||||
params: {
|
||||
name: toolName,
|
||||
arguments: args
|
||||
},
|
||||
id: 3
|
||||
}, this.sessionId);
|
||||
|
||||
if (response.statusCode !== 200) {
|
||||
return {
|
||||
name: `Tool Call: ${toolName}`,
|
||||
passed: false,
|
||||
error: `HTTP ${response.statusCode}`
|
||||
};
|
||||
}
|
||||
|
||||
const data = JSON.parse(response.body);
|
||||
|
||||
if (data.result?.content && Array.isArray(data.result.content)) {
|
||||
return {
|
||||
name: `Tool Call: ${toolName}`,
|
||||
passed: true,
|
||||
response: { contentItems: data.result.content.length }
|
||||
};
|
||||
} else {
|
||||
return {
|
||||
name: `Tool Call: ${toolName}`,
|
||||
passed: false,
|
||||
error: 'Missing or invalid content array',
|
||||
response: data
|
||||
};
|
||||
}
|
||||
} catch (error) {
|
||||
return {
|
||||
name: `Tool Call: ${toolName}`,
|
||||
passed: false,
|
||||
error: error instanceof Error ? error.message : 'Unknown error'
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
private async testToolCallInvalid(): Promise<TestResult> {
|
||||
console.log('🧪 Testing Tool Call with invalid parameters...');
|
||||
|
||||
try {
|
||||
const response = await this.makeRequest('POST', '/mcp', {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools/call',
|
||||
params: {
|
||||
name: 'get_node_essentials',
|
||||
arguments: {} // Missing required nodeType parameter
|
||||
},
|
||||
id: 4
|
||||
}, this.sessionId);
|
||||
|
||||
if (response.statusCode !== 200) {
|
||||
return {
|
||||
name: 'Tool Call: Invalid Params',
|
||||
passed: false,
|
||||
error: `HTTP ${response.statusCode}`
|
||||
};
|
||||
}
|
||||
|
||||
const data = JSON.parse(response.body);
|
||||
|
||||
// Should either return an error response or handle gracefully
|
||||
if (data.error || (data.result?.isError && data.result?.content)) {
|
||||
return {
|
||||
name: 'Tool Call: Invalid Params',
|
||||
passed: true,
|
||||
response: { handledGracefully: true }
|
||||
};
|
||||
} else {
|
||||
return {
|
||||
name: 'Tool Call: Invalid Params',
|
||||
passed: false,
|
||||
error: 'Did not handle invalid parameters properly',
|
||||
response: data
|
||||
};
|
||||
}
|
||||
} catch (error) {
|
||||
return {
|
||||
name: 'Tool Call: Invalid Params',
|
||||
passed: false,
|
||||
error: error instanceof Error ? error.message : 'Unknown error'
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
private makeRequest(method: string, path: string, data?: any, sessionId?: string | null): Promise<{
|
||||
statusCode: number;
|
||||
headers: http.IncomingHttpHeaders;
|
||||
body: string;
|
||||
}> {
|
||||
return new Promise((resolve, reject) => {
|
||||
const postData = data ? JSON.stringify(data) : '';
|
||||
|
||||
const options: http.RequestOptions = {
|
||||
hostname: 'localhost',
|
||||
port: this.mcpPort,
|
||||
path,
|
||||
method,
|
||||
headers: {
|
||||
'Content-Type': 'application/json',
|
||||
'Authorization': `Bearer ${this.authToken}`,
|
||||
...(postData && { 'Content-Length': Buffer.byteLength(postData) }),
|
||||
...(sessionId && { 'Mcp-Session-Id': sessionId })
|
||||
}
|
||||
};
|
||||
|
||||
const req = http.request(options, (res) => {
|
||||
let body = '';
|
||||
res.on('data', (chunk) => body += chunk);
|
||||
res.on('end', () => {
|
||||
resolve({
|
||||
statusCode: res.statusCode || 0,
|
||||
headers: res.headers,
|
||||
body
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
req.on('error', reject);
|
||||
req.setTimeout(10000, () => {
|
||||
req.destroy();
|
||||
reject(new Error('Request timeout'));
|
||||
});
|
||||
|
||||
if (postData) {
|
||||
req.write(postData);
|
||||
}
|
||||
req.end();
|
||||
});
|
||||
}
|
||||
|
||||
private printResults(tests: TestResult[]): void {
|
||||
console.log('\n📊 TEST RESULTS');
|
||||
console.log('================');
|
||||
|
||||
const passed = tests.filter(t => t.passed).length;
|
||||
const total = tests.length;
|
||||
|
||||
tests.forEach(test => {
|
||||
const status = test.passed ? '✅' : '❌';
|
||||
console.log(`${status} ${test.name}`);
|
||||
if (!test.passed && test.error) {
|
||||
console.log(` Error: ${test.error}`);
|
||||
}
|
||||
if (test.response) {
|
||||
console.log(` Response: ${JSON.stringify(test.response, null, 2)}`);
|
||||
}
|
||||
});
|
||||
|
||||
console.log(`\n📈 Summary: ${passed}/${total} tests passed`);
|
||||
|
||||
if (passed === total) {
|
||||
console.log('🎉 All tests passed! The n8n integration fixes should resolve the schema validation errors.');
|
||||
} else {
|
||||
console.log('❌ Some tests failed. Please review the errors above.');
|
||||
}
|
||||
}
|
||||
|
||||
private async cleanup(): Promise<void> {
|
||||
console.log('\n🧹 Cleaning up...');
|
||||
|
||||
if (this.mcpProcess) {
|
||||
this.mcpProcess.kill('SIGTERM');
|
||||
|
||||
// Wait for graceful shutdown
|
||||
await new Promise<void>((resolve) => {
|
||||
if (!this.mcpProcess) {
|
||||
resolve();
|
||||
return;
|
||||
}
|
||||
|
||||
const timeout = setTimeout(() => {
|
||||
this.mcpProcess?.kill('SIGKILL');
|
||||
resolve();
|
||||
}, 5000);
|
||||
|
||||
this.mcpProcess.on('exit', () => {
|
||||
clearTimeout(timeout);
|
||||
resolve();
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
console.log('✅ Cleanup complete');
|
||||
}
|
||||
|
||||
private sleep(ms: number): Promise<void> {
|
||||
return new Promise(resolve => setTimeout(resolve, ms));
|
||||
}
|
||||
}
|
||||
|
||||
// Run the tests
|
||||
if (require.main === module) {
|
||||
const tester = new N8nMcpTester();
|
||||
tester.start().catch(console.error);
|
||||
}
|
||||
|
||||
export { N8nMcpTester };
|
||||
178
scripts/test-operation-validation.ts
Normal file
178
scripts/test-operation-validation.ts
Normal file
@@ -0,0 +1,178 @@
|
||||
/**
|
||||
* Test script for operation and resource validation with Google Drive example
|
||||
*/
|
||||
|
||||
import { DatabaseAdapter } from '../src/database/database-adapter';
|
||||
import { NodeRepository } from '../src/database/node-repository';
|
||||
import { EnhancedConfigValidator } from '../src/services/enhanced-config-validator';
|
||||
import { WorkflowValidator } from '../src/services/workflow-validator';
|
||||
import { createDatabaseAdapter } from '../src/database/database-adapter';
|
||||
import { logger } from '../src/utils/logger';
|
||||
import chalk from 'chalk';
|
||||
|
||||
async function testOperationValidation() {
|
||||
console.log(chalk.blue('Testing Operation and Resource Validation'));
|
||||
console.log('='.repeat(60));
|
||||
|
||||
// Initialize database
|
||||
const dbPath = process.env.NODE_DB_PATH || 'data/nodes.db';
|
||||
const db = await createDatabaseAdapter(dbPath);
|
||||
const repository = new NodeRepository(db);
|
||||
|
||||
// Initialize similarity services
|
||||
EnhancedConfigValidator.initializeSimilarityServices(repository);
|
||||
|
||||
// Test 1: Invalid operation "listFiles"
|
||||
console.log(chalk.yellow('\n📝 Test 1: Google Drive with invalid operation "listFiles"'));
|
||||
const invalidConfig = {
|
||||
resource: 'fileFolder',
|
||||
operation: 'listFiles'
|
||||
};
|
||||
|
||||
const node = repository.getNode('nodes-base.googleDrive');
|
||||
if (!node) {
|
||||
console.error(chalk.red('Google Drive node not found in database'));
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
const result1 = EnhancedConfigValidator.validateWithMode(
|
||||
'nodes-base.googleDrive',
|
||||
invalidConfig,
|
||||
node.properties,
|
||||
'operation',
|
||||
'ai-friendly'
|
||||
);
|
||||
|
||||
console.log(`Valid: ${result1.valid ? chalk.green('✓') : chalk.red('✗')}`);
|
||||
if (result1.errors.length > 0) {
|
||||
console.log(chalk.red('Errors:'));
|
||||
result1.errors.forEach(error => {
|
||||
console.log(` - ${error.property}: ${error.message}`);
|
||||
if (error.fix) {
|
||||
console.log(chalk.cyan(` Fix: ${error.fix}`));
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Test 2: Invalid resource "files" (should be singular)
|
||||
console.log(chalk.yellow('\n📝 Test 2: Google Drive with invalid resource "files"'));
|
||||
const pluralResourceConfig = {
|
||||
resource: 'files',
|
||||
operation: 'download'
|
||||
};
|
||||
|
||||
const result2 = EnhancedConfigValidator.validateWithMode(
|
||||
'nodes-base.googleDrive',
|
||||
pluralResourceConfig,
|
||||
node.properties,
|
||||
'operation',
|
||||
'ai-friendly'
|
||||
);
|
||||
|
||||
console.log(`Valid: ${result2.valid ? chalk.green('✓') : chalk.red('✗')}`);
|
||||
if (result2.errors.length > 0) {
|
||||
console.log(chalk.red('Errors:'));
|
||||
result2.errors.forEach(error => {
|
||||
console.log(` - ${error.property}: ${error.message}`);
|
||||
if (error.fix) {
|
||||
console.log(chalk.cyan(` Fix: ${error.fix}`));
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Test 3: Valid configuration
|
||||
console.log(chalk.yellow('\n📝 Test 3: Google Drive with valid configuration'));
|
||||
const validConfig = {
|
||||
resource: 'file',
|
||||
operation: 'download'
|
||||
};
|
||||
|
||||
const result3 = EnhancedConfigValidator.validateWithMode(
|
||||
'nodes-base.googleDrive',
|
||||
validConfig,
|
||||
node.properties,
|
||||
'operation',
|
||||
'ai-friendly'
|
||||
);
|
||||
|
||||
console.log(`Valid: ${result3.valid ? chalk.green('✓') : chalk.red('✗')}`);
|
||||
if (result3.errors.length > 0) {
|
||||
console.log(chalk.red('Errors:'));
|
||||
result3.errors.forEach(error => {
|
||||
console.log(` - ${error.property}: ${error.message}`);
|
||||
});
|
||||
} else {
|
||||
console.log(chalk.green('No errors - configuration is valid!'));
|
||||
}
|
||||
|
||||
// Test 4: Test in workflow context
|
||||
console.log(chalk.yellow('\n📝 Test 4: Full workflow with invalid Google Drive node'));
|
||||
const workflow = {
|
||||
name: 'Test Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Google Drive',
|
||||
type: 'n8n-nodes-base.googleDrive',
|
||||
position: [100, 100] as [number, number],
|
||||
parameters: {
|
||||
resource: 'fileFolder',
|
||||
operation: 'listFiles' // Invalid operation
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const validator = new WorkflowValidator(repository, EnhancedConfigValidator);
|
||||
const workflowResult = await validator.validateWorkflow(workflow, {
|
||||
validateNodes: true,
|
||||
profile: 'ai-friendly'
|
||||
});
|
||||
|
||||
console.log(`Workflow Valid: ${workflowResult.valid ? chalk.green('✓') : chalk.red('✗')}`);
|
||||
if (workflowResult.errors.length > 0) {
|
||||
console.log(chalk.red('Errors:'));
|
||||
workflowResult.errors.forEach(error => {
|
||||
console.log(` - ${error.nodeName || 'Workflow'}: ${error.message}`);
|
||||
if (error.details?.fix) {
|
||||
console.log(chalk.cyan(` Fix: ${error.details.fix}`));
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
// Test 5: Typo in operation
|
||||
console.log(chalk.yellow('\n📝 Test 5: Typo in operation "downlod"'));
|
||||
const typoConfig = {
|
||||
resource: 'file',
|
||||
operation: 'downlod' // Typo
|
||||
};
|
||||
|
||||
const result5 = EnhancedConfigValidator.validateWithMode(
|
||||
'nodes-base.googleDrive',
|
||||
typoConfig,
|
||||
node.properties,
|
||||
'operation',
|
||||
'ai-friendly'
|
||||
);
|
||||
|
||||
console.log(`Valid: ${result5.valid ? chalk.green('✓') : chalk.red('✗')}`);
|
||||
if (result5.errors.length > 0) {
|
||||
console.log(chalk.red('Errors:'));
|
||||
result5.errors.forEach(error => {
|
||||
console.log(` - ${error.property}: ${error.message}`);
|
||||
if (error.fix) {
|
||||
console.log(chalk.cyan(` Fix: ${error.fix}`));
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
console.log(chalk.green('\n✅ All tests completed!'));
|
||||
db.close();
|
||||
}
|
||||
|
||||
// Run tests
|
||||
testOperationValidation().catch(error => {
|
||||
console.error(chalk.red('Error running tests:'), error);
|
||||
process.exit(1);
|
||||
});
|
||||
560
scripts/test-release-automation.js
Executable file
560
scripts/test-release-automation.js
Executable file
@@ -0,0 +1,560 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Test script for release automation
|
||||
* Validates the release workflow components locally
|
||||
*/
|
||||
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
const { execSync } = require('child_process');
|
||||
|
||||
// Color codes for output
|
||||
const colors = {
|
||||
reset: '\x1b[0m',
|
||||
red: '\x1b[31m',
|
||||
green: '\x1b[32m',
|
||||
yellow: '\x1b[33m',
|
||||
blue: '\x1b[34m',
|
||||
magenta: '\x1b[35m',
|
||||
cyan: '\x1b[36m'
|
||||
};
|
||||
|
||||
function log(message, color = 'reset') {
|
||||
console.log(`${colors[color]}${message}${colors.reset}`);
|
||||
}
|
||||
|
||||
function header(title) {
|
||||
log(`\n${'='.repeat(60)}`, 'cyan');
|
||||
log(`🧪 ${title}`, 'cyan');
|
||||
log(`${'='.repeat(60)}`, 'cyan');
|
||||
}
|
||||
|
||||
function section(title) {
|
||||
log(`\n📋 ${title}`, 'blue');
|
||||
log(`${'-'.repeat(40)}`, 'blue');
|
||||
}
|
||||
|
||||
function success(message) {
|
||||
log(`✅ ${message}`, 'green');
|
||||
}
|
||||
|
||||
function warning(message) {
|
||||
log(`⚠️ ${message}`, 'yellow');
|
||||
}
|
||||
|
||||
function error(message) {
|
||||
log(`❌ ${message}`, 'red');
|
||||
}
|
||||
|
||||
function info(message) {
|
||||
log(`ℹ️ ${message}`, 'blue');
|
||||
}
|
||||
|
||||
class ReleaseAutomationTester {
|
||||
constructor() {
|
||||
this.rootDir = path.resolve(__dirname, '..');
|
||||
this.errors = [];
|
||||
this.warnings = [];
|
||||
}
|
||||
|
||||
/**
|
||||
* Test if required files exist
|
||||
*/
|
||||
testFileExistence() {
|
||||
section('Testing File Existence');
|
||||
|
||||
const requiredFiles = [
|
||||
'package.json',
|
||||
'package.runtime.json',
|
||||
'docs/CHANGELOG.md',
|
||||
'.github/workflows/release.yml',
|
||||
'scripts/sync-runtime-version.js',
|
||||
'scripts/publish-npm.sh'
|
||||
];
|
||||
|
||||
for (const file of requiredFiles) {
|
||||
const filePath = path.join(this.rootDir, file);
|
||||
if (fs.existsSync(filePath)) {
|
||||
success(`Found: ${file}`);
|
||||
} else {
|
||||
error(`Missing: ${file}`);
|
||||
this.errors.push(`Missing required file: ${file}`);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test version detection logic
|
||||
*/
|
||||
testVersionDetection() {
|
||||
section('Testing Version Detection');
|
||||
|
||||
try {
|
||||
const packageJson = require(path.join(this.rootDir, 'package.json'));
|
||||
const runtimeJson = require(path.join(this.rootDir, 'package.runtime.json'));
|
||||
|
||||
success(`Package.json version: ${packageJson.version}`);
|
||||
success(`Runtime package version: ${runtimeJson.version}`);
|
||||
|
||||
if (packageJson.version === runtimeJson.version) {
|
||||
success('Version sync: Both versions match');
|
||||
} else {
|
||||
warning('Version sync: Versions do not match - run sync:runtime-version');
|
||||
this.warnings.push('Package versions are not synchronized');
|
||||
}
|
||||
|
||||
// Test semantic version format
|
||||
const semverRegex = /^\d+\.\d+\.\d+(?:-[\w\.-]+)?(?:\+[\w\.-]+)?$/;
|
||||
if (semverRegex.test(packageJson.version)) {
|
||||
success(`Version format: Valid semantic version (${packageJson.version})`);
|
||||
} else {
|
||||
error(`Version format: Invalid semantic version (${packageJson.version})`);
|
||||
this.errors.push('Invalid semantic version format');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Version detection failed: ${err.message}`);
|
||||
this.errors.push(`Version detection error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test changelog parsing
|
||||
*/
|
||||
testChangelogParsing() {
|
||||
section('Testing Changelog Parsing');
|
||||
|
||||
try {
|
||||
const changelogPath = path.join(this.rootDir, 'docs/CHANGELOG.md');
|
||||
|
||||
if (!fs.existsSync(changelogPath)) {
|
||||
error('Changelog file not found');
|
||||
this.errors.push('Missing changelog file');
|
||||
return;
|
||||
}
|
||||
|
||||
const changelogContent = fs.readFileSync(changelogPath, 'utf8');
|
||||
const packageJson = require(path.join(this.rootDir, 'package.json'));
|
||||
const currentVersion = packageJson.version;
|
||||
|
||||
// Check if current version exists in changelog
|
||||
const versionRegex = new RegExp(`^## \\[${currentVersion.replace(/[.*+?^${}()|[\]\\]/g, '\\$&')}\\]`, 'm');
|
||||
|
||||
if (versionRegex.test(changelogContent)) {
|
||||
success(`Changelog entry found for version ${currentVersion}`);
|
||||
|
||||
// Test extraction logic (simplified version of the GitHub Actions script)
|
||||
const lines = changelogContent.split('\n');
|
||||
let startIndex = -1;
|
||||
let endIndex = -1;
|
||||
|
||||
for (let i = 0; i < lines.length; i++) {
|
||||
if (versionRegex.test(lines[i])) {
|
||||
startIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (startIndex !== -1) {
|
||||
// Find the end of this version's section
|
||||
for (let i = startIndex + 1; i < lines.length; i++) {
|
||||
if (lines[i].startsWith('## [') && !lines[i].includes('Unreleased')) {
|
||||
endIndex = i;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (endIndex === -1) {
|
||||
endIndex = lines.length;
|
||||
}
|
||||
|
||||
const sectionLines = lines.slice(startIndex + 1, endIndex);
|
||||
const contentLines = sectionLines.filter(line => line.trim() !== '');
|
||||
|
||||
if (contentLines.length > 0) {
|
||||
success(`Changelog content extracted: ${contentLines.length} lines`);
|
||||
info(`Preview: ${contentLines[0].substring(0, 100)}...`);
|
||||
} else {
|
||||
warning('Changelog section appears to be empty');
|
||||
this.warnings.push(`Empty changelog section for version ${currentVersion}`);
|
||||
}
|
||||
}
|
||||
|
||||
} else {
|
||||
warning(`No changelog entry found for current version ${currentVersion}`);
|
||||
this.warnings.push(`Missing changelog entry for version ${currentVersion}`);
|
||||
}
|
||||
|
||||
// Check changelog format
|
||||
if (changelogContent.includes('## [Unreleased]')) {
|
||||
success('Changelog format: Contains Unreleased section');
|
||||
} else {
|
||||
warning('Changelog format: Missing Unreleased section');
|
||||
}
|
||||
|
||||
if (changelogContent.includes('Keep a Changelog')) {
|
||||
success('Changelog format: Follows Keep a Changelog format');
|
||||
} else {
|
||||
warning('Changelog format: Does not reference Keep a Changelog');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Changelog parsing failed: ${err.message}`);
|
||||
this.errors.push(`Changelog parsing error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test build process
|
||||
*/
|
||||
testBuildProcess() {
|
||||
section('Testing Build Process');
|
||||
|
||||
try {
|
||||
// Check if dist directory exists
|
||||
const distPath = path.join(this.rootDir, 'dist');
|
||||
if (fs.existsSync(distPath)) {
|
||||
success('Build output: dist directory exists');
|
||||
|
||||
// Check for key build files
|
||||
const keyFiles = [
|
||||
'dist/index.js',
|
||||
'dist/mcp/index.js',
|
||||
'dist/mcp/server.js'
|
||||
];
|
||||
|
||||
for (const file of keyFiles) {
|
||||
const filePath = path.join(this.rootDir, file);
|
||||
if (fs.existsSync(filePath)) {
|
||||
success(`Build file: ${file} exists`);
|
||||
} else {
|
||||
warning(`Build file: ${file} missing - run 'npm run build'`);
|
||||
this.warnings.push(`Missing build file: ${file}`);
|
||||
}
|
||||
}
|
||||
|
||||
} else {
|
||||
warning('Build output: dist directory missing - run "npm run build"');
|
||||
this.warnings.push('Missing build output');
|
||||
}
|
||||
|
||||
// Check database
|
||||
const dbPath = path.join(this.rootDir, 'data/nodes.db');
|
||||
if (fs.existsSync(dbPath)) {
|
||||
const stats = fs.statSync(dbPath);
|
||||
success(`Database: nodes.db exists (${Math.round(stats.size / 1024 / 1024)}MB)`);
|
||||
} else {
|
||||
warning('Database: nodes.db missing - run "npm run rebuild"');
|
||||
this.warnings.push('Missing database file');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Build process test failed: ${err.message}`);
|
||||
this.errors.push(`Build process error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test npm publish preparation
|
||||
*/
|
||||
testNpmPublishPrep() {
|
||||
section('Testing NPM Publish Preparation');
|
||||
|
||||
try {
|
||||
const packageJson = require(path.join(this.rootDir, 'package.json'));
|
||||
const runtimeJson = require(path.join(this.rootDir, 'package.runtime.json'));
|
||||
|
||||
// Check package.json fields
|
||||
const requiredFields = ['name', 'version', 'description', 'main', 'bin'];
|
||||
for (const field of requiredFields) {
|
||||
if (packageJson[field]) {
|
||||
success(`Package field: ${field} is present`);
|
||||
} else {
|
||||
error(`Package field: ${field} is missing`);
|
||||
this.errors.push(`Missing package.json field: ${field}`);
|
||||
}
|
||||
}
|
||||
|
||||
// Check runtime dependencies
|
||||
if (runtimeJson.dependencies) {
|
||||
const depCount = Object.keys(runtimeJson.dependencies).length;
|
||||
success(`Runtime dependencies: ${depCount} packages`);
|
||||
|
||||
// List key dependencies
|
||||
const keyDeps = ['@modelcontextprotocol/sdk', 'express', 'sql.js'];
|
||||
for (const dep of keyDeps) {
|
||||
if (runtimeJson.dependencies[dep]) {
|
||||
success(`Key dependency: ${dep} (${runtimeJson.dependencies[dep]})`);
|
||||
} else {
|
||||
warning(`Key dependency: ${dep} is missing`);
|
||||
this.warnings.push(`Missing key dependency: ${dep}`);
|
||||
}
|
||||
}
|
||||
|
||||
} else {
|
||||
error('Runtime package has no dependencies');
|
||||
this.errors.push('Missing runtime dependencies');
|
||||
}
|
||||
|
||||
// Check files array
|
||||
if (packageJson.files && Array.isArray(packageJson.files)) {
|
||||
success(`Package files: ${packageJson.files.length} patterns specified`);
|
||||
info(`Files: ${packageJson.files.join(', ')}`);
|
||||
} else {
|
||||
warning('Package files: No files array specified');
|
||||
this.warnings.push('No files array in package.json');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`NPM publish prep test failed: ${err.message}`);
|
||||
this.errors.push(`NPM publish prep error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test Docker configuration
|
||||
*/
|
||||
testDockerConfig() {
|
||||
section('Testing Docker Configuration');
|
||||
|
||||
try {
|
||||
const dockerfiles = ['Dockerfile', 'Dockerfile.railway'];
|
||||
|
||||
for (const dockerfile of dockerfiles) {
|
||||
const dockerfilePath = path.join(this.rootDir, dockerfile);
|
||||
if (fs.existsSync(dockerfilePath)) {
|
||||
success(`Dockerfile: ${dockerfile} exists`);
|
||||
|
||||
const content = fs.readFileSync(dockerfilePath, 'utf8');
|
||||
|
||||
// Check for key instructions
|
||||
if (content.includes('FROM node:')) {
|
||||
success(`${dockerfile}: Uses Node.js base image`);
|
||||
} else {
|
||||
warning(`${dockerfile}: Does not use standard Node.js base image`);
|
||||
}
|
||||
|
||||
if (content.includes('COPY dist')) {
|
||||
success(`${dockerfile}: Copies build output`);
|
||||
} else {
|
||||
warning(`${dockerfile}: May not copy build output correctly`);
|
||||
}
|
||||
|
||||
} else {
|
||||
warning(`Dockerfile: ${dockerfile} not found`);
|
||||
this.warnings.push(`Missing Dockerfile: ${dockerfile}`);
|
||||
}
|
||||
}
|
||||
|
||||
// Check docker-compose files
|
||||
const composeFiles = ['docker-compose.yml', 'docker-compose.n8n.yml'];
|
||||
for (const composeFile of composeFiles) {
|
||||
const composePath = path.join(this.rootDir, composeFile);
|
||||
if (fs.existsSync(composePath)) {
|
||||
success(`Docker Compose: ${composeFile} exists`);
|
||||
} else {
|
||||
info(`Docker Compose: ${composeFile} not found (optional)`);
|
||||
}
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Docker config test failed: ${err.message}`);
|
||||
this.errors.push(`Docker config error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test workflow file syntax
|
||||
*/
|
||||
testWorkflowSyntax() {
|
||||
section('Testing Workflow Syntax');
|
||||
|
||||
try {
|
||||
const workflowPath = path.join(this.rootDir, '.github/workflows/release.yml');
|
||||
|
||||
if (!fs.existsSync(workflowPath)) {
|
||||
error('Release workflow file not found');
|
||||
this.errors.push('Missing release workflow file');
|
||||
return;
|
||||
}
|
||||
|
||||
const workflowContent = fs.readFileSync(workflowPath, 'utf8');
|
||||
|
||||
// Basic YAML structure checks
|
||||
if (workflowContent.includes('name: Automated Release')) {
|
||||
success('Workflow: Has correct name');
|
||||
} else {
|
||||
warning('Workflow: Name may be incorrect');
|
||||
}
|
||||
|
||||
if (workflowContent.includes('on:') && workflowContent.includes('push:')) {
|
||||
success('Workflow: Has push trigger');
|
||||
} else {
|
||||
error('Workflow: Missing push trigger');
|
||||
this.errors.push('Workflow missing push trigger');
|
||||
}
|
||||
|
||||
if (workflowContent.includes('branches: [main]')) {
|
||||
success('Workflow: Configured for main branch');
|
||||
} else {
|
||||
warning('Workflow: May not be configured for main branch');
|
||||
}
|
||||
|
||||
// Check for required jobs
|
||||
const requiredJobs = [
|
||||
'detect-version-change',
|
||||
'extract-changelog',
|
||||
'create-release',
|
||||
'publish-npm',
|
||||
'build-docker'
|
||||
];
|
||||
|
||||
for (const job of requiredJobs) {
|
||||
if (workflowContent.includes(`${job}:`)) {
|
||||
success(`Workflow job: ${job} defined`);
|
||||
} else {
|
||||
error(`Workflow job: ${job} missing`);
|
||||
this.errors.push(`Missing workflow job: ${job}`);
|
||||
}
|
||||
}
|
||||
|
||||
// Check for secrets usage
|
||||
if (workflowContent.includes('${{ secrets.NPM_TOKEN }}')) {
|
||||
success('Workflow: NPM_TOKEN secret configured');
|
||||
} else {
|
||||
warning('Workflow: NPM_TOKEN secret may be missing');
|
||||
this.warnings.push('NPM_TOKEN secret may need to be configured');
|
||||
}
|
||||
|
||||
if (workflowContent.includes('${{ secrets.GITHUB_TOKEN }}')) {
|
||||
success('Workflow: GITHUB_TOKEN secret configured');
|
||||
} else {
|
||||
warning('Workflow: GITHUB_TOKEN secret may be missing');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Workflow syntax test failed: ${err.message}`);
|
||||
this.errors.push(`Workflow syntax error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Test environment and dependencies
|
||||
*/
|
||||
testEnvironment() {
|
||||
section('Testing Environment');
|
||||
|
||||
try {
|
||||
// Check Node.js version
|
||||
const nodeVersion = process.version;
|
||||
success(`Node.js version: ${nodeVersion}`);
|
||||
|
||||
// Check if npm is available
|
||||
try {
|
||||
const npmVersion = execSync('npm --version', { encoding: 'utf8', stdio: 'pipe' }).trim();
|
||||
success(`NPM version: ${npmVersion}`);
|
||||
} catch (err) {
|
||||
error('NPM not available');
|
||||
this.errors.push('NPM not available');
|
||||
}
|
||||
|
||||
// Check if git is available
|
||||
try {
|
||||
const gitVersion = execSync('git --version', { encoding: 'utf8', stdio: 'pipe' }).trim();
|
||||
success(`Git available: ${gitVersion}`);
|
||||
} catch (err) {
|
||||
error('Git not available');
|
||||
this.errors.push('Git not available');
|
||||
}
|
||||
|
||||
// Check if we're in a git repository
|
||||
try {
|
||||
execSync('git rev-parse --git-dir', { stdio: 'pipe' });
|
||||
success('Git repository: Detected');
|
||||
|
||||
// Check current branch
|
||||
try {
|
||||
const branch = execSync('git branch --show-current', { encoding: 'utf8', stdio: 'pipe' }).trim();
|
||||
info(`Current branch: ${branch}`);
|
||||
} catch (err) {
|
||||
info('Could not determine current branch');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
warning('Not in a git repository');
|
||||
this.warnings.push('Not in a git repository');
|
||||
}
|
||||
|
||||
} catch (err) {
|
||||
error(`Environment test failed: ${err.message}`);
|
||||
this.errors.push(`Environment error: ${err.message}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Run all tests
|
||||
*/
|
||||
async runAllTests() {
|
||||
header('Release Automation Test Suite');
|
||||
|
||||
info('Testing release automation components...');
|
||||
|
||||
this.testFileExistence();
|
||||
this.testVersionDetection();
|
||||
this.testChangelogParsing();
|
||||
this.testBuildProcess();
|
||||
this.testNpmPublishPrep();
|
||||
this.testDockerConfig();
|
||||
this.testWorkflowSyntax();
|
||||
this.testEnvironment();
|
||||
|
||||
// Summary
|
||||
header('Test Summary');
|
||||
|
||||
if (this.errors.length === 0 && this.warnings.length === 0) {
|
||||
log('🎉 All tests passed! Release automation is ready.', 'green');
|
||||
} else {
|
||||
if (this.errors.length > 0) {
|
||||
log(`\n❌ ${this.errors.length} Error(s):`, 'red');
|
||||
this.errors.forEach(err => log(` • ${err}`, 'red'));
|
||||
}
|
||||
|
||||
if (this.warnings.length > 0) {
|
||||
log(`\n⚠️ ${this.warnings.length} Warning(s):`, 'yellow');
|
||||
this.warnings.forEach(warn => log(` • ${warn}`, 'yellow'));
|
||||
}
|
||||
|
||||
if (this.errors.length > 0) {
|
||||
log('\n🔧 Please fix the errors before running the release workflow.', 'red');
|
||||
process.exit(1);
|
||||
} else {
|
||||
log('\n✅ No critical errors found. Warnings should be reviewed but won\'t prevent releases.', 'yellow');
|
||||
}
|
||||
}
|
||||
|
||||
// Next steps
|
||||
log('\n📋 Next Steps:', 'cyan');
|
||||
log('1. Ensure all secrets are configured in GitHub repository settings:', 'cyan');
|
||||
log(' • NPM_TOKEN (required for npm publishing)', 'cyan');
|
||||
log(' • GITHUB_TOKEN (automatically available)', 'cyan');
|
||||
log('\n2. To trigger a release:', 'cyan');
|
||||
log(' • Update version in package.json', 'cyan');
|
||||
log(' • Update changelog in docs/CHANGELOG.md', 'cyan');
|
||||
log(' • Commit and push to main branch', 'cyan');
|
||||
log('\n3. Monitor the release workflow in GitHub Actions', 'cyan');
|
||||
|
||||
return this.errors.length === 0;
|
||||
}
|
||||
}
|
||||
|
||||
// Run the tests
|
||||
if (require.main === module) {
|
||||
const tester = new ReleaseAutomationTester();
|
||||
tester.runAllTests().catch(err => {
|
||||
console.error('Test suite failed:', err);
|
||||
process.exit(1);
|
||||
});
|
||||
}
|
||||
|
||||
module.exports = ReleaseAutomationTester;
|
||||
118
scripts/test-telemetry-debug.ts
Normal file
118
scripts/test-telemetry-debug.ts
Normal file
@@ -0,0 +1,118 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Debug script for telemetry integration
|
||||
* Tests direct Supabase connection
|
||||
*/
|
||||
|
||||
import { createClient } from '@supabase/supabase-js';
|
||||
import dotenv from 'dotenv';
|
||||
|
||||
// Load environment variables
|
||||
dotenv.config();
|
||||
|
||||
async function debugTelemetry() {
|
||||
console.log('🔍 Debugging Telemetry Integration\n');
|
||||
|
||||
const supabaseUrl = process.env.SUPABASE_URL;
|
||||
const supabaseAnonKey = process.env.SUPABASE_ANON_KEY;
|
||||
|
||||
if (!supabaseUrl || !supabaseAnonKey) {
|
||||
console.error('❌ Missing SUPABASE_URL or SUPABASE_ANON_KEY');
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
console.log('Environment:');
|
||||
console.log(' URL:', supabaseUrl);
|
||||
console.log(' Key:', supabaseAnonKey.substring(0, 30) + '...');
|
||||
|
||||
// Create Supabase client
|
||||
const supabase = createClient(supabaseUrl, supabaseAnonKey, {
|
||||
auth: {
|
||||
persistSession: false,
|
||||
autoRefreshToken: false,
|
||||
}
|
||||
});
|
||||
|
||||
// Test 1: Direct insert to telemetry_events
|
||||
console.log('\n📝 Test 1: Direct insert to telemetry_events...');
|
||||
const testEvent = {
|
||||
user_id: 'test-user-123',
|
||||
event: 'test_event',
|
||||
properties: {
|
||||
test: true,
|
||||
timestamp: new Date().toISOString()
|
||||
}
|
||||
};
|
||||
|
||||
const { data: eventData, error: eventError } = await supabase
|
||||
.from('telemetry_events')
|
||||
.insert([testEvent])
|
||||
.select();
|
||||
|
||||
if (eventError) {
|
||||
console.error('❌ Event insert failed:', eventError);
|
||||
} else {
|
||||
console.log('✅ Event inserted successfully:', eventData);
|
||||
}
|
||||
|
||||
// Test 2: Direct insert to telemetry_workflows
|
||||
console.log('\n📝 Test 2: Direct insert to telemetry_workflows...');
|
||||
const testWorkflow = {
|
||||
user_id: 'test-user-123',
|
||||
workflow_hash: 'test-hash-' + Date.now(),
|
||||
node_count: 3,
|
||||
node_types: ['webhook', 'http', 'slack'],
|
||||
has_trigger: true,
|
||||
has_webhook: true,
|
||||
complexity: 'simple',
|
||||
sanitized_workflow: {
|
||||
nodes: [],
|
||||
connections: {}
|
||||
}
|
||||
};
|
||||
|
||||
const { data: workflowData, error: workflowError } = await supabase
|
||||
.from('telemetry_workflows')
|
||||
.insert([testWorkflow])
|
||||
.select();
|
||||
|
||||
if (workflowError) {
|
||||
console.error('❌ Workflow insert failed:', workflowError);
|
||||
} else {
|
||||
console.log('✅ Workflow inserted successfully:', workflowData);
|
||||
}
|
||||
|
||||
// Test 3: Try to read data (should fail with anon key due to RLS)
|
||||
console.log('\n📖 Test 3: Attempting to read data (should fail due to RLS)...');
|
||||
const { data: readData, error: readError } = await supabase
|
||||
.from('telemetry_events')
|
||||
.select('*')
|
||||
.limit(1);
|
||||
|
||||
if (readError) {
|
||||
console.log('✅ Read correctly blocked by RLS:', readError.message);
|
||||
} else {
|
||||
console.log('⚠️ Unexpected: Read succeeded (RLS may not be working):', readData);
|
||||
}
|
||||
|
||||
// Test 4: Check table existence
|
||||
console.log('\n🔍 Test 4: Verifying tables exist...');
|
||||
const { data: tables, error: tablesError } = await supabase
|
||||
.rpc('get_tables', { schema_name: 'public' })
|
||||
.select('*');
|
||||
|
||||
if (tablesError) {
|
||||
// This is expected - the RPC function might not exist
|
||||
console.log('ℹ️ Cannot list tables (RPC function not available)');
|
||||
} else {
|
||||
console.log('Tables found:', tables);
|
||||
}
|
||||
|
||||
console.log('\n✨ Debug completed! Check your Supabase dashboard for the test data.');
|
||||
console.log('Dashboard: https://supabase.com/dashboard/project/ydyufsohxdfpopqbubwk/editor');
|
||||
}
|
||||
|
||||
debugTelemetry().catch(error => {
|
||||
console.error('❌ Debug failed:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
46
scripts/test-telemetry-direct.ts
Normal file
46
scripts/test-telemetry-direct.ts
Normal file
@@ -0,0 +1,46 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Direct telemetry test with hardcoded credentials
|
||||
*/
|
||||
|
||||
import { createClient } from '@supabase/supabase-js';
|
||||
|
||||
const TELEMETRY_BACKEND = {
|
||||
URL: 'https://ydyufsohxdfpopqbubwk.supabase.co',
|
||||
ANON_KEY: 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJzdXBhYmFzZSIsInJlZiI6InlkeXVmc29oeGRmcG9wcWJ1YndrIiwicm9sZSI6ImFub24iLCJpYXQiOjE3Mzc2MzAxMDgsImV4cCI6MjA1MzIwNjEwOH0.LsUTx9OsNtnqg-jxXaJPc84aBHVDehHiMaFoF2Ir8s0'
|
||||
};
|
||||
|
||||
async function testDirect() {
|
||||
console.log('🧪 Direct Telemetry Test\n');
|
||||
|
||||
const supabase = createClient(TELEMETRY_BACKEND.URL, TELEMETRY_BACKEND.ANON_KEY, {
|
||||
auth: {
|
||||
persistSession: false,
|
||||
autoRefreshToken: false,
|
||||
}
|
||||
});
|
||||
|
||||
const testEvent = {
|
||||
user_id: 'direct-test-' + Date.now(),
|
||||
event: 'direct_test',
|
||||
properties: {
|
||||
source: 'test-telemetry-direct.ts',
|
||||
timestamp: new Date().toISOString()
|
||||
}
|
||||
};
|
||||
|
||||
console.log('Sending event:', testEvent);
|
||||
|
||||
const { data, error } = await supabase
|
||||
.from('telemetry_events')
|
||||
.insert([testEvent]);
|
||||
|
||||
if (error) {
|
||||
console.error('❌ Failed:', error);
|
||||
} else {
|
||||
console.log('✅ Success! Event sent directly to Supabase');
|
||||
console.log('Response:', data);
|
||||
}
|
||||
}
|
||||
|
||||
testDirect().catch(console.error);
|
||||
62
scripts/test-telemetry-env.ts
Normal file
62
scripts/test-telemetry-env.ts
Normal file
@@ -0,0 +1,62 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Test telemetry environment variable override
|
||||
*/
|
||||
|
||||
import { TelemetryConfigManager } from '../src/telemetry/config-manager';
|
||||
import { telemetry } from '../src/telemetry/telemetry-manager';
|
||||
|
||||
async function testEnvOverride() {
|
||||
console.log('🧪 Testing Telemetry Environment Variable Override\n');
|
||||
|
||||
const configManager = TelemetryConfigManager.getInstance();
|
||||
|
||||
// Test 1: Check current status without env var
|
||||
console.log('Test 1: Without environment variable');
|
||||
console.log('Is Enabled:', configManager.isEnabled());
|
||||
console.log('Status:', configManager.getStatus());
|
||||
|
||||
// Test 2: Set environment variable and check again
|
||||
console.log('\n━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\n');
|
||||
console.log('Test 2: With N8N_MCP_TELEMETRY_DISABLED=true');
|
||||
process.env.N8N_MCP_TELEMETRY_DISABLED = 'true';
|
||||
|
||||
// Force reload by creating new instance (for testing)
|
||||
const newConfigManager = TelemetryConfigManager.getInstance();
|
||||
console.log('Is Enabled:', newConfigManager.isEnabled());
|
||||
console.log('Status:', newConfigManager.getStatus());
|
||||
|
||||
// Test 3: Try tracking with env disabled
|
||||
console.log('\n━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\n');
|
||||
console.log('Test 3: Attempting to track with telemetry disabled');
|
||||
telemetry.trackToolUsage('test_tool', true, 100);
|
||||
console.log('Tool usage tracking attempted (should be ignored)');
|
||||
|
||||
// Test 4: Alternative env vars
|
||||
console.log('\n━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\n');
|
||||
console.log('Test 4: Alternative environment variables');
|
||||
|
||||
delete process.env.N8N_MCP_TELEMETRY_DISABLED;
|
||||
process.env.TELEMETRY_DISABLED = 'true';
|
||||
console.log('With TELEMETRY_DISABLED=true:', newConfigManager.isEnabled());
|
||||
|
||||
delete process.env.TELEMETRY_DISABLED;
|
||||
process.env.DISABLE_TELEMETRY = 'true';
|
||||
console.log('With DISABLE_TELEMETRY=true:', newConfigManager.isEnabled());
|
||||
|
||||
// Test 5: Env var takes precedence over config
|
||||
console.log('\n━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\n');
|
||||
console.log('Test 5: Environment variable precedence');
|
||||
|
||||
// Enable via config
|
||||
newConfigManager.enable();
|
||||
console.log('After enabling via config:', newConfigManager.isEnabled());
|
||||
|
||||
// But env var should still override
|
||||
process.env.N8N_MCP_TELEMETRY_DISABLED = 'true';
|
||||
console.log('With env var set (should override config):', newConfigManager.isEnabled());
|
||||
|
||||
console.log('\n✅ All tests completed!');
|
||||
}
|
||||
|
||||
testEnvOverride().catch(console.error);
|
||||
94
scripts/test-telemetry-integration.ts
Normal file
94
scripts/test-telemetry-integration.ts
Normal file
@@ -0,0 +1,94 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Integration test for the telemetry manager
|
||||
*/
|
||||
|
||||
import { telemetry } from '../src/telemetry/telemetry-manager';
|
||||
|
||||
async function testIntegration() {
|
||||
console.log('🧪 Testing Telemetry Manager Integration\n');
|
||||
|
||||
// Check status
|
||||
console.log('Status:', telemetry.getStatus());
|
||||
|
||||
// Track session start
|
||||
console.log('\nTracking session start...');
|
||||
telemetry.trackSessionStart();
|
||||
|
||||
// Track tool usage
|
||||
console.log('Tracking tool usage...');
|
||||
telemetry.trackToolUsage('search_nodes', true, 150);
|
||||
telemetry.trackToolUsage('get_node_info', true, 75);
|
||||
telemetry.trackToolUsage('validate_workflow', false, 200);
|
||||
|
||||
// Track errors
|
||||
console.log('Tracking errors...');
|
||||
telemetry.trackError('ValidationError', 'workflow_validation', 'validate_workflow', 'Required field missing: nodes array is empty');
|
||||
|
||||
// Track a test workflow
|
||||
console.log('Tracking workflow creation...');
|
||||
const testWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
type: 'n8n-nodes-base.webhook',
|
||||
name: 'Webhook',
|
||||
position: [0, 0],
|
||||
parameters: {
|
||||
path: '/test-webhook',
|
||||
httpMethod: 'POST'
|
||||
}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
name: 'HTTP Request',
|
||||
position: [250, 0],
|
||||
parameters: {
|
||||
url: 'https://api.example.com/endpoint',
|
||||
method: 'POST',
|
||||
authentication: 'genericCredentialType',
|
||||
genericAuthType: 'httpHeaderAuth',
|
||||
sendHeaders: true,
|
||||
headerParameters: {
|
||||
parameters: [
|
||||
{
|
||||
name: 'Authorization',
|
||||
value: 'Bearer sk-1234567890abcdef'
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
type: 'n8n-nodes-base.slack',
|
||||
name: 'Slack',
|
||||
position: [500, 0],
|
||||
parameters: {
|
||||
channel: '#notifications',
|
||||
text: 'Workflow completed!'
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'1': {
|
||||
main: [[{ node: '2', type: 'main', index: 0 }]]
|
||||
},
|
||||
'2': {
|
||||
main: [[{ node: '3', type: 'main', index: 0 }]]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
telemetry.trackWorkflowCreation(testWorkflow, true);
|
||||
|
||||
// Force flush
|
||||
console.log('\nFlushing telemetry data...');
|
||||
await telemetry.flush();
|
||||
|
||||
console.log('\n✅ Telemetry integration test completed!');
|
||||
console.log('Check your Supabase dashboard for the telemetry data.');
|
||||
}
|
||||
|
||||
testIntegration().catch(console.error);
|
||||
68
scripts/test-telemetry-no-select.ts
Normal file
68
scripts/test-telemetry-no-select.ts
Normal file
@@ -0,0 +1,68 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Test telemetry without requesting data back
|
||||
*/
|
||||
|
||||
import { createClient } from '@supabase/supabase-js';
|
||||
import dotenv from 'dotenv';
|
||||
|
||||
dotenv.config();
|
||||
|
||||
async function testNoSelect() {
|
||||
const supabaseUrl = process.env.SUPABASE_URL!;
|
||||
const supabaseAnonKey = process.env.SUPABASE_ANON_KEY!;
|
||||
|
||||
console.log('🧪 Telemetry Test (No Select)\n');
|
||||
|
||||
const supabase = createClient(supabaseUrl, supabaseAnonKey, {
|
||||
auth: {
|
||||
persistSession: false,
|
||||
autoRefreshToken: false,
|
||||
}
|
||||
});
|
||||
|
||||
// Insert WITHOUT .select() - just fire and forget
|
||||
const testData = {
|
||||
user_id: 'test-' + Date.now(),
|
||||
event: 'test_event',
|
||||
properties: { test: true }
|
||||
};
|
||||
|
||||
console.log('Inserting:', testData);
|
||||
|
||||
const { error } = await supabase
|
||||
.from('telemetry_events')
|
||||
.insert([testData]); // No .select() here!
|
||||
|
||||
if (error) {
|
||||
console.error('❌ Failed:', error);
|
||||
} else {
|
||||
console.log('✅ Success! Data inserted (no response data)');
|
||||
}
|
||||
|
||||
// Test workflow insert too
|
||||
const testWorkflow = {
|
||||
user_id: 'test-' + Date.now(),
|
||||
workflow_hash: 'hash-' + Date.now(),
|
||||
node_count: 3,
|
||||
node_types: ['webhook', 'http', 'slack'],
|
||||
has_trigger: true,
|
||||
has_webhook: true,
|
||||
complexity: 'simple',
|
||||
sanitized_workflow: { nodes: [], connections: {} }
|
||||
};
|
||||
|
||||
console.log('\nInserting workflow:', testWorkflow);
|
||||
|
||||
const { error: workflowError } = await supabase
|
||||
.from('telemetry_workflows')
|
||||
.insert([testWorkflow]); // No .select() here!
|
||||
|
||||
if (workflowError) {
|
||||
console.error('❌ Workflow failed:', workflowError);
|
||||
} else {
|
||||
console.log('✅ Workflow inserted successfully!');
|
||||
}
|
||||
}
|
||||
|
||||
testNoSelect().catch(console.error);
|
||||
87
scripts/test-telemetry-security.ts
Normal file
87
scripts/test-telemetry-security.ts
Normal file
@@ -0,0 +1,87 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Test that RLS properly protects data
|
||||
*/
|
||||
|
||||
import { createClient } from '@supabase/supabase-js';
|
||||
import dotenv from 'dotenv';
|
||||
|
||||
dotenv.config();
|
||||
|
||||
async function testSecurity() {
|
||||
const supabaseUrl = process.env.SUPABASE_URL!;
|
||||
const supabaseAnonKey = process.env.SUPABASE_ANON_KEY!;
|
||||
|
||||
console.log('🔒 Testing Telemetry Security (RLS)\n');
|
||||
|
||||
const supabase = createClient(supabaseUrl, supabaseAnonKey, {
|
||||
auth: {
|
||||
persistSession: false,
|
||||
autoRefreshToken: false,
|
||||
}
|
||||
});
|
||||
|
||||
// Test 1: Verify anon can INSERT
|
||||
console.log('Test 1: Anonymous INSERT (should succeed)...');
|
||||
const testData = {
|
||||
user_id: 'security-test-' + Date.now(),
|
||||
event: 'security_test',
|
||||
properties: { test: true }
|
||||
};
|
||||
|
||||
const { error: insertError } = await supabase
|
||||
.from('telemetry_events')
|
||||
.insert([testData]);
|
||||
|
||||
if (insertError) {
|
||||
console.error('❌ Insert failed:', insertError.message);
|
||||
} else {
|
||||
console.log('✅ Insert succeeded (as expected)');
|
||||
}
|
||||
|
||||
// Test 2: Verify anon CANNOT SELECT
|
||||
console.log('\nTest 2: Anonymous SELECT (should fail)...');
|
||||
const { data, error: selectError } = await supabase
|
||||
.from('telemetry_events')
|
||||
.select('*')
|
||||
.limit(1);
|
||||
|
||||
if (selectError) {
|
||||
console.log('✅ Select blocked by RLS (as expected):', selectError.message);
|
||||
} else if (data && data.length > 0) {
|
||||
console.error('❌ SECURITY ISSUE: Anon can read data!', data);
|
||||
} else if (data && data.length === 0) {
|
||||
console.log('⚠️ Select returned empty array (might be RLS working)');
|
||||
}
|
||||
|
||||
// Test 3: Verify anon CANNOT UPDATE
|
||||
console.log('\nTest 3: Anonymous UPDATE (should fail)...');
|
||||
const { error: updateError } = await supabase
|
||||
.from('telemetry_events')
|
||||
.update({ event: 'hacked' })
|
||||
.eq('user_id', 'test');
|
||||
|
||||
if (updateError) {
|
||||
console.log('✅ Update blocked (as expected):', updateError.message);
|
||||
} else {
|
||||
console.error('❌ SECURITY ISSUE: Anon can update data!');
|
||||
}
|
||||
|
||||
// Test 4: Verify anon CANNOT DELETE
|
||||
console.log('\nTest 4: Anonymous DELETE (should fail)...');
|
||||
const { error: deleteError } = await supabase
|
||||
.from('telemetry_events')
|
||||
.delete()
|
||||
.eq('user_id', 'test');
|
||||
|
||||
if (deleteError) {
|
||||
console.log('✅ Delete blocked (as expected):', deleteError.message);
|
||||
} else {
|
||||
console.error('❌ SECURITY ISSUE: Anon can delete data!');
|
||||
}
|
||||
|
||||
console.log('\n✨ Security test completed!');
|
||||
console.log('Summary: Anonymous users can INSERT (for telemetry) but cannot READ/UPDATE/DELETE');
|
||||
}
|
||||
|
||||
testSecurity().catch(console.error);
|
||||
45
scripts/test-telemetry-simple.ts
Normal file
45
scripts/test-telemetry-simple.ts
Normal file
@@ -0,0 +1,45 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Simple test to verify telemetry works
|
||||
*/
|
||||
|
||||
import { createClient } from '@supabase/supabase-js';
|
||||
import dotenv from 'dotenv';
|
||||
|
||||
dotenv.config();
|
||||
|
||||
async function testSimple() {
|
||||
const supabaseUrl = process.env.SUPABASE_URL!;
|
||||
const supabaseAnonKey = process.env.SUPABASE_ANON_KEY!;
|
||||
|
||||
console.log('🧪 Simple Telemetry Test\n');
|
||||
|
||||
const supabase = createClient(supabaseUrl, supabaseAnonKey, {
|
||||
auth: {
|
||||
persistSession: false,
|
||||
autoRefreshToken: false,
|
||||
}
|
||||
});
|
||||
|
||||
// Simple insert
|
||||
const testData = {
|
||||
user_id: 'simple-test-' + Date.now(),
|
||||
event: 'test_event',
|
||||
properties: { test: true }
|
||||
};
|
||||
|
||||
console.log('Inserting:', testData);
|
||||
|
||||
const { data, error } = await supabase
|
||||
.from('telemetry_events')
|
||||
.insert([testData])
|
||||
.select();
|
||||
|
||||
if (error) {
|
||||
console.error('❌ Failed:', error);
|
||||
} else {
|
||||
console.log('✅ Success! Inserted:', data);
|
||||
}
|
||||
}
|
||||
|
||||
testSimple().catch(console.error);
|
||||
119
scripts/test-user-id-persistence.ts
Normal file
119
scripts/test-user-id-persistence.ts
Normal file
@@ -0,0 +1,119 @@
|
||||
/**
|
||||
* Test User ID Persistence
|
||||
* Verifies that user IDs are consistent across sessions and modes
|
||||
*/
|
||||
|
||||
import { TelemetryConfigManager } from '../src/telemetry/config-manager';
|
||||
import { hostname, platform, arch, homedir } from 'os';
|
||||
import { createHash } from 'crypto';
|
||||
|
||||
console.log('=== User ID Persistence Test ===\n');
|
||||
|
||||
// Test 1: Verify deterministic ID generation
|
||||
console.log('Test 1: Deterministic ID Generation');
|
||||
console.log('-----------------------------------');
|
||||
|
||||
const machineId = `${hostname()}-${platform()}-${arch()}-${homedir()}`;
|
||||
const expectedUserId = createHash('sha256')
|
||||
.update(machineId)
|
||||
.digest('hex')
|
||||
.substring(0, 16);
|
||||
|
||||
console.log('Machine characteristics:');
|
||||
console.log(' hostname:', hostname());
|
||||
console.log(' platform:', platform());
|
||||
console.log(' arch:', arch());
|
||||
console.log(' homedir:', homedir());
|
||||
console.log('\nGenerated machine ID:', machineId);
|
||||
console.log('Expected user ID:', expectedUserId);
|
||||
|
||||
// Test 2: Load actual config
|
||||
console.log('\n\nTest 2: Actual Config Manager');
|
||||
console.log('-----------------------------------');
|
||||
|
||||
const configManager = TelemetryConfigManager.getInstance();
|
||||
const actualUserId = configManager.getUserId();
|
||||
const config = configManager.loadConfig();
|
||||
|
||||
console.log('Actual user ID:', actualUserId);
|
||||
console.log('Config first run:', config.firstRun || 'Unknown');
|
||||
console.log('Config version:', config.version || 'Unknown');
|
||||
console.log('Telemetry enabled:', config.enabled);
|
||||
|
||||
// Test 3: Verify consistency
|
||||
console.log('\n\nTest 3: Consistency Check');
|
||||
console.log('-----------------------------------');
|
||||
|
||||
const match = actualUserId === expectedUserId;
|
||||
console.log('User IDs match:', match ? '✓ YES' : '✗ NO');
|
||||
|
||||
if (!match) {
|
||||
console.log('WARNING: User ID mismatch detected!');
|
||||
console.log('This could indicate an implementation issue.');
|
||||
}
|
||||
|
||||
// Test 4: Multiple loads (simulate multiple sessions)
|
||||
console.log('\n\nTest 4: Multiple Session Simulation');
|
||||
console.log('-----------------------------------');
|
||||
|
||||
const userId1 = configManager.getUserId();
|
||||
const userId2 = TelemetryConfigManager.getInstance().getUserId();
|
||||
const userId3 = configManager.getUserId();
|
||||
|
||||
console.log('Session 1 user ID:', userId1);
|
||||
console.log('Session 2 user ID:', userId2);
|
||||
console.log('Session 3 user ID:', userId3);
|
||||
|
||||
const consistent = userId1 === userId2 && userId2 === userId3;
|
||||
console.log('All sessions consistent:', consistent ? '✓ YES' : '✗ NO');
|
||||
|
||||
// Test 5: Docker environment simulation
|
||||
console.log('\n\nTest 5: Docker Environment Check');
|
||||
console.log('-----------------------------------');
|
||||
|
||||
const isDocker = process.env.IS_DOCKER === 'true';
|
||||
console.log('Running in Docker:', isDocker);
|
||||
|
||||
if (isDocker) {
|
||||
console.log('\n⚠️ DOCKER MODE DETECTED');
|
||||
console.log('In Docker, user IDs may change across container recreations because:');
|
||||
console.log(' 1. Container hostname changes each time');
|
||||
console.log(' 2. Config file is not persisted (no volume mount)');
|
||||
console.log(' 3. Each container gets a new ephemeral filesystem');
|
||||
console.log('\nRecommendation: Mount ~/.n8n-mcp as a volume for persistent user IDs');
|
||||
}
|
||||
|
||||
// Test 6: Environment variable override check
|
||||
console.log('\n\nTest 6: Environment Variable Override');
|
||||
console.log('-----------------------------------');
|
||||
|
||||
const telemetryDisabledVars = [
|
||||
'N8N_MCP_TELEMETRY_DISABLED',
|
||||
'TELEMETRY_DISABLED',
|
||||
'DISABLE_TELEMETRY'
|
||||
];
|
||||
|
||||
telemetryDisabledVars.forEach(varName => {
|
||||
const value = process.env[varName];
|
||||
if (value !== undefined) {
|
||||
console.log(`${varName}:`, value);
|
||||
}
|
||||
});
|
||||
|
||||
console.log('\nTelemetry status:', configManager.isEnabled() ? 'ENABLED' : 'DISABLED');
|
||||
|
||||
// Summary
|
||||
console.log('\n\n=== SUMMARY ===');
|
||||
console.log('User ID:', actualUserId);
|
||||
console.log('Deterministic:', match ? 'YES ✓' : 'NO ✗');
|
||||
console.log('Persistent across sessions:', consistent ? 'YES ✓' : 'NO ✗');
|
||||
console.log('Telemetry enabled:', config.enabled ? 'YES' : 'NO');
|
||||
console.log('Docker mode:', isDocker ? 'YES' : 'NO');
|
||||
|
||||
if (isDocker && !process.env.N8N_MCP_CONFIG_VOLUME) {
|
||||
console.log('\n⚠️ WARNING: Running in Docker without persistent volume!');
|
||||
console.log('User IDs will change on container recreation.');
|
||||
console.log('Mount /home/nodejs/.n8n-mcp to persist telemetry config.');
|
||||
}
|
||||
|
||||
console.log('\n');
|
||||
55
scripts/test-workflow-insert.ts
Normal file
55
scripts/test-workflow-insert.ts
Normal file
@@ -0,0 +1,55 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Test direct workflow insert to Supabase
|
||||
*/
|
||||
|
||||
import { createClient } from '@supabase/supabase-js';
|
||||
|
||||
const TELEMETRY_BACKEND = {
|
||||
URL: 'https://ydyufsohxdfpopqbubwk.supabase.co',
|
||||
ANON_KEY: 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJzdXBhYmFzZSIsInJlZiI6InlkeXVmc29oeGRmcG9wcWJ1YndrIiwicm9sZSI6ImFub24iLCJpYXQiOjE3NTg3OTYyMDAsImV4cCI6MjA3NDM3MjIwMH0.xESphg6h5ozaDsm4Vla3QnDJGc6Nc_cpfoqTHRynkCk'
|
||||
};
|
||||
|
||||
async function testWorkflowInsert() {
|
||||
const supabase = createClient(TELEMETRY_BACKEND.URL, TELEMETRY_BACKEND.ANON_KEY, {
|
||||
auth: {
|
||||
persistSession: false,
|
||||
autoRefreshToken: false,
|
||||
}
|
||||
});
|
||||
|
||||
const testWorkflow = {
|
||||
user_id: 'direct-test-' + Date.now(),
|
||||
workflow_hash: 'hash-direct-' + Date.now(),
|
||||
node_count: 2,
|
||||
node_types: ['webhook', 'http'],
|
||||
has_trigger: true,
|
||||
has_webhook: true,
|
||||
complexity: 'simple' as const,
|
||||
sanitized_workflow: {
|
||||
nodes: [
|
||||
{ id: '1', type: 'webhook', parameters: {} },
|
||||
{ id: '2', type: 'http', parameters: {} }
|
||||
],
|
||||
connections: {}
|
||||
}
|
||||
};
|
||||
|
||||
console.log('Attempting direct insert to telemetry_workflows...');
|
||||
console.log('Data:', JSON.stringify(testWorkflow, null, 2));
|
||||
|
||||
const { data, error } = await supabase
|
||||
.from('telemetry_workflows')
|
||||
.insert([testWorkflow]);
|
||||
|
||||
if (error) {
|
||||
console.error('\n❌ Error:', error);
|
||||
} else {
|
||||
console.log('\n✅ Success! Workflow inserted');
|
||||
if (data) {
|
||||
console.log('Response:', data);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
testWorkflowInsert().catch(console.error);
|
||||
67
scripts/test-workflow-sanitizer.ts
Normal file
67
scripts/test-workflow-sanitizer.ts
Normal file
@@ -0,0 +1,67 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Test workflow sanitizer
|
||||
*/
|
||||
|
||||
import { WorkflowSanitizer } from '../src/telemetry/workflow-sanitizer';
|
||||
|
||||
const testWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: 'webhook1',
|
||||
type: 'n8n-nodes-base.webhook',
|
||||
name: 'Webhook',
|
||||
position: [0, 0],
|
||||
parameters: {
|
||||
path: '/test-webhook',
|
||||
httpMethod: 'POST'
|
||||
}
|
||||
},
|
||||
{
|
||||
id: 'http1',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
name: 'HTTP Request',
|
||||
position: [250, 0],
|
||||
parameters: {
|
||||
url: 'https://api.example.com/endpoint',
|
||||
method: 'GET',
|
||||
authentication: 'genericCredentialType',
|
||||
sendHeaders: true,
|
||||
headerParameters: {
|
||||
parameters: [
|
||||
{
|
||||
name: 'Authorization',
|
||||
value: 'Bearer sk-1234567890abcdef'
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'webhook1': {
|
||||
main: [[{ node: 'http1', type: 'main', index: 0 }]]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
console.log('🧪 Testing Workflow Sanitizer\n');
|
||||
console.log('Original workflow has', testWorkflow.nodes.length, 'nodes');
|
||||
|
||||
try {
|
||||
const sanitized = WorkflowSanitizer.sanitizeWorkflow(testWorkflow);
|
||||
|
||||
console.log('\n✅ Sanitization successful!');
|
||||
console.log('\nSanitized output:');
|
||||
console.log(JSON.stringify(sanitized, null, 2));
|
||||
|
||||
console.log('\n📊 Metrics:');
|
||||
console.log('- Workflow Hash:', sanitized.workflowHash);
|
||||
console.log('- Node Count:', sanitized.nodeCount);
|
||||
console.log('- Node Types:', sanitized.nodeTypes);
|
||||
console.log('- Has Trigger:', sanitized.hasTrigger);
|
||||
console.log('- Has Webhook:', sanitized.hasWebhook);
|
||||
console.log('- Complexity:', sanitized.complexity);
|
||||
} catch (error) {
|
||||
console.error('❌ Sanitization failed:', error);
|
||||
}
|
||||
71
scripts/test-workflow-tracking-debug.ts
Normal file
71
scripts/test-workflow-tracking-debug.ts
Normal file
@@ -0,0 +1,71 @@
|
||||
#!/usr/bin/env npx tsx
|
||||
/**
|
||||
* Debug workflow tracking in telemetry manager
|
||||
*/
|
||||
|
||||
import { TelemetryManager } from '../src/telemetry/telemetry-manager';
|
||||
|
||||
// Get the singleton instance
|
||||
const telemetry = TelemetryManager.getInstance();
|
||||
|
||||
const testWorkflow = {
|
||||
nodes: [
|
||||
{
|
||||
id: 'webhook1',
|
||||
type: 'n8n-nodes-base.webhook',
|
||||
name: 'Webhook',
|
||||
position: [0, 0],
|
||||
parameters: {
|
||||
path: '/test-' + Date.now(),
|
||||
httpMethod: 'POST'
|
||||
}
|
||||
},
|
||||
{
|
||||
id: 'http1',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
name: 'HTTP Request',
|
||||
position: [250, 0],
|
||||
parameters: {
|
||||
url: 'https://api.example.com/data',
|
||||
method: 'GET'
|
||||
}
|
||||
},
|
||||
{
|
||||
id: 'slack1',
|
||||
type: 'n8n-nodes-base.slack',
|
||||
name: 'Slack',
|
||||
position: [500, 0],
|
||||
parameters: {
|
||||
channel: '#general',
|
||||
text: 'Workflow complete!'
|
||||
}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'webhook1': {
|
||||
main: [[{ node: 'http1', type: 'main', index: 0 }]]
|
||||
},
|
||||
'http1': {
|
||||
main: [[{ node: 'slack1', type: 'main', index: 0 }]]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
console.log('🧪 Testing Workflow Tracking\n');
|
||||
console.log('Workflow has', testWorkflow.nodes.length, 'nodes');
|
||||
|
||||
// Track the workflow
|
||||
console.log('Calling trackWorkflowCreation...');
|
||||
telemetry.trackWorkflowCreation(testWorkflow, true);
|
||||
|
||||
console.log('Waiting for async processing...');
|
||||
|
||||
// Wait for setImmediate to process
|
||||
setTimeout(async () => {
|
||||
console.log('\nForcing flush...');
|
||||
await telemetry.flush();
|
||||
console.log('✅ Flush complete!');
|
||||
|
||||
console.log('\nWorkflow should now be in the telemetry_workflows table.');
|
||||
console.log('Check with: SELECT * FROM telemetry_workflows ORDER BY created_at DESC LIMIT 1;');
|
||||
}, 2000);
|
||||
@@ -48,5 +48,27 @@ export function isN8nApiConfigured(): boolean {
|
||||
return config !== null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Create n8n API configuration from instance context
|
||||
* Used for flexible instance configuration support
|
||||
*/
|
||||
export function getN8nApiConfigFromContext(context: {
|
||||
n8nApiUrl?: string;
|
||||
n8nApiKey?: string;
|
||||
n8nApiTimeout?: number;
|
||||
n8nApiMaxRetries?: number;
|
||||
}): N8nApiConfig | null {
|
||||
if (!context.n8nApiUrl || !context.n8nApiKey) {
|
||||
return null;
|
||||
}
|
||||
|
||||
return {
|
||||
baseUrl: context.n8nApiUrl,
|
||||
apiKey: context.n8nApiKey,
|
||||
timeout: context.n8nApiTimeout ?? 30000,
|
||||
maxRetries: context.n8nApiMaxRetries ?? 3,
|
||||
};
|
||||
}
|
||||
|
||||
// Type export
|
||||
export type N8nApiConfig = NonNullable<ReturnType<typeof getN8nApiConfig>>;
|
||||
310
src/data/canonical-ai-tool-examples.json
Normal file
310
src/data/canonical-ai-tool-examples.json
Normal file
@@ -0,0 +1,310 @@
|
||||
{
|
||||
"description": "Canonical configuration examples for critical AI tools based on FINAL_AI_VALIDATION_SPEC.md",
|
||||
"version": "1.0.0",
|
||||
"examples": [
|
||||
{
|
||||
"node_type": "@n8n/n8n-nodes-langchain.toolHttpRequest",
|
||||
"display_name": "HTTP Request Tool",
|
||||
"examples": [
|
||||
{
|
||||
"name": "Weather API Tool",
|
||||
"use_case": "Fetch current weather data for AI Agent",
|
||||
"complexity": "simple",
|
||||
"parameters": {
|
||||
"method": "GET",
|
||||
"url": "https://api.weatherapi.com/v1/current.json?key={{$credentials.weatherApiKey}}&q={city}",
|
||||
"toolDescription": "Get current weather conditions for a city. Provide the city name (e.g., 'London', 'New York') and receive temperature, humidity, wind speed, and conditions.",
|
||||
"placeholderDefinitions": {
|
||||
"values": [
|
||||
{
|
||||
"name": "city",
|
||||
"description": "Name of the city to get weather for",
|
||||
"type": "string"
|
||||
}
|
||||
]
|
||||
},
|
||||
"authentication": "predefinedCredentialType",
|
||||
"nodeCredentialType": "weatherApiApi"
|
||||
},
|
||||
"credentials": {
|
||||
"weatherApiApi": {
|
||||
"id": "1",
|
||||
"name": "Weather API account"
|
||||
}
|
||||
},
|
||||
"notes": "Example shows proper toolDescription, URL with placeholder, and credential configuration"
|
||||
},
|
||||
{
|
||||
"name": "GitHub Issues Tool",
|
||||
"use_case": "Create GitHub issues from AI Agent conversations",
|
||||
"complexity": "medium",
|
||||
"parameters": {
|
||||
"method": "POST",
|
||||
"url": "https://api.github.com/repos/{owner}/{repo}/issues",
|
||||
"toolDescription": "Create a new GitHub issue. Requires owner (repo owner username), repo (repository name), title, and body. Returns the created issue URL and number.",
|
||||
"placeholderDefinitions": {
|
||||
"values": [
|
||||
{
|
||||
"name": "owner",
|
||||
"description": "GitHub repository owner username",
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"name": "repo",
|
||||
"description": "Repository name",
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"name": "title",
|
||||
"description": "Issue title",
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"name": "body",
|
||||
"description": "Issue description and details",
|
||||
"type": "string"
|
||||
}
|
||||
]
|
||||
},
|
||||
"sendBody": true,
|
||||
"specifyBody": "json",
|
||||
"jsonBody": "={{ { \"title\": $json.title, \"body\": $json.body } }}",
|
||||
"authentication": "predefinedCredentialType",
|
||||
"nodeCredentialType": "githubApi"
|
||||
},
|
||||
"credentials": {
|
||||
"githubApi": {
|
||||
"id": "2",
|
||||
"name": "GitHub credentials"
|
||||
}
|
||||
},
|
||||
"notes": "Example shows POST request with JSON body, multiple placeholders, and expressions"
|
||||
},
|
||||
{
|
||||
"name": "Slack Message Tool",
|
||||
"use_case": "Send Slack messages from AI Agent",
|
||||
"complexity": "simple",
|
||||
"parameters": {
|
||||
"method": "POST",
|
||||
"url": "https://slack.com/api/chat.postMessage",
|
||||
"toolDescription": "Send a message to a Slack channel. Provide channel ID or name (e.g., '#general', 'C1234567890') and message text.",
|
||||
"placeholderDefinitions": {
|
||||
"values": [
|
||||
{
|
||||
"name": "channel",
|
||||
"description": "Channel ID or name (e.g., #general)",
|
||||
"type": "string"
|
||||
},
|
||||
{
|
||||
"name": "text",
|
||||
"description": "Message text to send",
|
||||
"type": "string"
|
||||
}
|
||||
]
|
||||
},
|
||||
"sendHeaders": true,
|
||||
"headerParameters": {
|
||||
"parameters": [
|
||||
{
|
||||
"name": "Content-Type",
|
||||
"value": "application/json; charset=utf-8"
|
||||
},
|
||||
{
|
||||
"name": "Authorization",
|
||||
"value": "=Bearer {{$credentials.slackApi.accessToken}}"
|
||||
}
|
||||
]
|
||||
},
|
||||
"sendBody": true,
|
||||
"specifyBody": "json",
|
||||
"jsonBody": "={{ { \"channel\": $json.channel, \"text\": $json.text } }}",
|
||||
"authentication": "predefinedCredentialType",
|
||||
"nodeCredentialType": "slackApi"
|
||||
},
|
||||
"credentials": {
|
||||
"slackApi": {
|
||||
"id": "3",
|
||||
"name": "Slack account"
|
||||
}
|
||||
},
|
||||
"notes": "Example shows headers with credential expressions and JSON body construction"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"node_type": "@n8n/n8n-nodes-langchain.toolCode",
|
||||
"display_name": "Code Tool",
|
||||
"examples": [
|
||||
{
|
||||
"name": "Calculate Shipping Cost",
|
||||
"use_case": "Calculate shipping costs based on weight and distance",
|
||||
"complexity": "simple",
|
||||
"parameters": {
|
||||
"name": "calculate_shipping_cost",
|
||||
"description": "Calculate shipping cost based on package weight (in kg) and distance (in km). Returns the cost in USD.",
|
||||
"language": "javaScript",
|
||||
"code": "const baseRate = 5;\nconst perKgRate = 2;\nconst perKmRate = 0.1;\n\nconst weight = $input.weight || 0;\nconst distance = $input.distance || 0;\n\nconst cost = baseRate + (weight * perKgRate) + (distance * perKmRate);\n\nreturn { cost: parseFloat(cost.toFixed(2)), currency: 'USD' };",
|
||||
"specifyInputSchema": true,
|
||||
"schemaType": "manual",
|
||||
"inputSchema": "{\n \"type\": \"object\",\n \"properties\": {\n \"weight\": {\n \"type\": \"number\",\n \"description\": \"Package weight in kilograms\"\n },\n \"distance\": {\n \"type\": \"number\",\n \"description\": \"Shipping distance in kilometers\"\n }\n },\n \"required\": [\"weight\", \"distance\"]\n}"
|
||||
},
|
||||
"notes": "Example shows proper function naming, detailed description, input schema, and return value"
|
||||
},
|
||||
{
|
||||
"name": "Format Customer Data",
|
||||
"use_case": "Transform and validate customer information",
|
||||
"complexity": "medium",
|
||||
"parameters": {
|
||||
"name": "format_customer_data",
|
||||
"description": "Format and validate customer data. Takes raw customer info (name, email, phone) and returns formatted object with validation status.",
|
||||
"language": "javaScript",
|
||||
"code": "const { name, email, phone } = $input;\n\n// Validation\nconst emailRegex = /^[^\\s@]+@[^\\s@]+\\.[^\\s@]+$/;\nconst phoneRegex = /^\\+?[1-9]\\d{1,14}$/;\n\nconst errors = [];\nif (!emailRegex.test(email)) errors.push('Invalid email format');\nif (!phoneRegex.test(phone)) errors.push('Invalid phone format');\n\n// Formatting\nconst formatted = {\n name: name.trim(),\n email: email.toLowerCase().trim(),\n phone: phone.replace(/\\s/g, ''),\n valid: errors.length === 0,\n errors: errors\n};\n\nreturn formatted;",
|
||||
"specifyInputSchema": true,
|
||||
"schemaType": "manual",
|
||||
"inputSchema": "{\n \"type\": \"object\",\n \"properties\": {\n \"name\": {\n \"type\": \"string\",\n \"description\": \"Customer full name\"\n },\n \"email\": {\n \"type\": \"string\",\n \"description\": \"Customer email address\"\n },\n \"phone\": {\n \"type\": \"string\",\n \"description\": \"Customer phone number\"\n }\n },\n \"required\": [\"name\", \"email\", \"phone\"]\n}"
|
||||
},
|
||||
"notes": "Example shows data validation, formatting, and structured error handling"
|
||||
},
|
||||
{
|
||||
"name": "Parse Date Range",
|
||||
"use_case": "Convert natural language date ranges to ISO format",
|
||||
"complexity": "medium",
|
||||
"parameters": {
|
||||
"name": "parse_date_range",
|
||||
"description": "Parse natural language date ranges (e.g., 'last 7 days', 'this month', 'Q1 2024') into start and end dates in ISO format.",
|
||||
"language": "javaScript",
|
||||
"code": "const input = $input.dateRange || '';\nconst now = new Date();\nlet start, end;\n\nif (input.includes('last') && input.includes('days')) {\n const days = parseInt(input.match(/\\d+/)[0]);\n start = new Date(now.getTime() - (days * 24 * 60 * 60 * 1000));\n end = now;\n} else if (input === 'this month') {\n start = new Date(now.getFullYear(), now.getMonth(), 1);\n end = new Date(now.getFullYear(), now.getMonth() + 1, 0);\n} else if (input === 'this year') {\n start = new Date(now.getFullYear(), 0, 1);\n end = new Date(now.getFullYear(), 11, 31);\n} else {\n throw new Error('Unsupported date range format');\n}\n\nreturn {\n startDate: start.toISOString().split('T')[0],\n endDate: end.toISOString().split('T')[0],\n daysCount: Math.ceil((end - start) / (24 * 60 * 60 * 1000))\n};",
|
||||
"specifyInputSchema": true,
|
||||
"schemaType": "manual",
|
||||
"inputSchema": "{\n \"type\": \"object\",\n \"properties\": {\n \"dateRange\": {\n \"type\": \"string\",\n \"description\": \"Natural language date range (e.g., 'last 7 days', 'this month')\"\n }\n },\n \"required\": [\"dateRange\"]\n}"
|
||||
},
|
||||
"notes": "Example shows complex logic, error handling, and date manipulation"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"node_type": "@n8n/n8n-nodes-langchain.agentTool",
|
||||
"display_name": "AI Agent Tool",
|
||||
"examples": [
|
||||
{
|
||||
"name": "Research Specialist Agent",
|
||||
"use_case": "Specialized sub-agent for in-depth research tasks",
|
||||
"complexity": "medium",
|
||||
"parameters": {
|
||||
"name": "research_specialist",
|
||||
"description": "Expert research agent that can search multiple sources, synthesize information, and provide comprehensive analysis on any topic. Use this when you need detailed, well-researched information.",
|
||||
"promptType": "define",
|
||||
"text": "You are a research specialist. Your role is to:\n1. Search for relevant information from multiple sources\n2. Synthesize findings into a coherent analysis\n3. Cite your sources\n4. Highlight key insights and patterns\n\nProvide thorough, well-structured research that answers the user's question comprehensively.",
|
||||
"systemMessage": "You are a meticulous researcher focused on accuracy and completeness. Always cite sources and acknowledge limitations in available information."
|
||||
},
|
||||
"connections": {
|
||||
"ai_languageModel": [
|
||||
{
|
||||
"node": "OpenAI GPT-4",
|
||||
"type": "ai_languageModel",
|
||||
"index": 0
|
||||
}
|
||||
],
|
||||
"ai_tool": [
|
||||
{
|
||||
"node": "SerpApi Tool",
|
||||
"type": "ai_tool",
|
||||
"index": 0
|
||||
},
|
||||
{
|
||||
"node": "Wikipedia Tool",
|
||||
"type": "ai_tool",
|
||||
"index": 0
|
||||
}
|
||||
]
|
||||
},
|
||||
"notes": "Example shows specialized sub-agent with custom prompt, specific system message, and multiple search tools"
|
||||
},
|
||||
{
|
||||
"name": "Data Analysis Agent",
|
||||
"use_case": "Sub-agent for analyzing and visualizing data",
|
||||
"complexity": "complex",
|
||||
"parameters": {
|
||||
"name": "data_analyst",
|
||||
"description": "Data analysis specialist that can process datasets, calculate statistics, identify trends, and generate insights. Use for any data analysis or statistical questions.",
|
||||
"promptType": "auto",
|
||||
"systemMessage": "You are a data analyst with expertise in statistics and data interpretation. Break down complex datasets into understandable insights. Use the Code Tool to perform calculations when needed.",
|
||||
"maxIterations": 10
|
||||
},
|
||||
"connections": {
|
||||
"ai_languageModel": [
|
||||
{
|
||||
"node": "Anthropic Claude",
|
||||
"type": "ai_languageModel",
|
||||
"index": 0
|
||||
}
|
||||
],
|
||||
"ai_tool": [
|
||||
{
|
||||
"node": "Code Tool - Stats",
|
||||
"type": "ai_tool",
|
||||
"index": 0
|
||||
},
|
||||
{
|
||||
"node": "HTTP Request Tool - Data API",
|
||||
"type": "ai_tool",
|
||||
"index": 0
|
||||
}
|
||||
]
|
||||
},
|
||||
"notes": "Example shows auto prompt type with specialized system message and analytical tools"
|
||||
}
|
||||
]
|
||||
},
|
||||
{
|
||||
"node_type": "@n8n/n8n-nodes-langchain.mcpClientTool",
|
||||
"display_name": "MCP Client Tool",
|
||||
"examples": [
|
||||
{
|
||||
"name": "Filesystem MCP Tool",
|
||||
"use_case": "Access filesystem operations via MCP protocol",
|
||||
"complexity": "medium",
|
||||
"parameters": {
|
||||
"description": "Access file system operations through MCP. Can read files, list directories, create files, and search for content.",
|
||||
"mcpServer": {
|
||||
"transport": "stdio",
|
||||
"command": "npx",
|
||||
"args": ["-y", "@modelcontextprotocol/server-filesystem", "/path/to/allowed/directory"]
|
||||
},
|
||||
"tool": "read_file"
|
||||
},
|
||||
"notes": "Example shows stdio transport MCP server with filesystem access tool"
|
||||
},
|
||||
{
|
||||
"name": "Puppeteer MCP Tool",
|
||||
"use_case": "Browser automation via MCP for AI Agents",
|
||||
"complexity": "complex",
|
||||
"parameters": {
|
||||
"description": "Control a web browser to navigate pages, take screenshots, and extract content. Useful for web scraping and automated testing.",
|
||||
"mcpServer": {
|
||||
"transport": "stdio",
|
||||
"command": "npx",
|
||||
"args": ["-y", "@modelcontextprotocol/server-puppeteer"]
|
||||
},
|
||||
"tool": "puppeteer_navigate"
|
||||
},
|
||||
"notes": "Example shows Puppeteer MCP server for browser automation"
|
||||
},
|
||||
{
|
||||
"name": "Database MCP Tool",
|
||||
"use_case": "Query databases via MCP protocol",
|
||||
"complexity": "complex",
|
||||
"parameters": {
|
||||
"description": "Execute SQL queries and retrieve data from PostgreSQL databases. Supports SELECT, INSERT, UPDATE operations with proper escaping.",
|
||||
"mcpServer": {
|
||||
"transport": "sse",
|
||||
"url": "https://mcp-server.example.com/database"
|
||||
},
|
||||
"tool": "execute_query"
|
||||
},
|
||||
"notes": "Example shows SSE transport MCP server for remote database access"
|
||||
}
|
||||
]
|
||||
}
|
||||
]
|
||||
}
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user