Merge pull request #128 from czlonkowski/feature/fix-loop-output-confusion
fix: resolve SplitInBatches output confusion for AI assistants
This commit is contained in:
BIN
data/nodes.db
BIN
data/nodes.db
Binary file not shown.
@@ -22,8 +22,9 @@ export class NodeRepository {
|
||||
node_type, package_name, display_name, description,
|
||||
category, development_style, is_ai_tool, is_trigger,
|
||||
is_webhook, is_versioned, version, documentation,
|
||||
properties_schema, operations, credentials_required
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
properties_schema, operations, credentials_required,
|
||||
outputs, output_names
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
`);
|
||||
|
||||
stmt.run(
|
||||
@@ -41,7 +42,9 @@ export class NodeRepository {
|
||||
node.documentation || null,
|
||||
JSON.stringify(node.properties, null, 2),
|
||||
JSON.stringify(node.operations, null, 2),
|
||||
JSON.stringify(node.credentials, null, 2)
|
||||
JSON.stringify(node.credentials, null, 2),
|
||||
node.outputs ? JSON.stringify(node.outputs, null, 2) : null,
|
||||
node.outputNames ? JSON.stringify(node.outputNames, null, 2) : null
|
||||
);
|
||||
}
|
||||
|
||||
@@ -70,7 +73,9 @@ export class NodeRepository {
|
||||
properties: this.safeJsonParse(row.properties_schema, []),
|
||||
operations: this.safeJsonParse(row.operations, []),
|
||||
credentials: this.safeJsonParse(row.credentials_required, []),
|
||||
hasDocumentation: !!row.documentation
|
||||
hasDocumentation: !!row.documentation,
|
||||
outputs: row.outputs ? this.safeJsonParse(row.outputs, null) : null,
|
||||
outputNames: row.output_names ? this.safeJsonParse(row.output_names, null) : null
|
||||
};
|
||||
}
|
||||
|
||||
@@ -238,7 +243,9 @@ export class NodeRepository {
|
||||
properties: this.safeJsonParse(row.properties_schema, []),
|
||||
operations: this.safeJsonParse(row.operations, []),
|
||||
credentials: this.safeJsonParse(row.credentials_required, []),
|
||||
hasDocumentation: !!row.documentation
|
||||
hasDocumentation: !!row.documentation,
|
||||
outputs: row.outputs ? this.safeJsonParse(row.outputs, null) : null,
|
||||
outputNames: row.output_names ? this.safeJsonParse(row.output_names, null) : null
|
||||
};
|
||||
}
|
||||
}
|
||||
@@ -15,6 +15,8 @@ CREATE TABLE IF NOT EXISTS nodes (
|
||||
properties_schema TEXT,
|
||||
operations TEXT,
|
||||
credentials_required TEXT,
|
||||
outputs TEXT, -- JSON array of output definitions
|
||||
output_names TEXT, -- JSON array of output names
|
||||
updated_at DATETIME DEFAULT CURRENT_TIMESTAMP
|
||||
);
|
||||
|
||||
|
||||
@@ -50,8 +50,12 @@ export class DocsMapper {
|
||||
for (const relativePath of possiblePaths) {
|
||||
try {
|
||||
const fullPath = path.join(this.docsPath, relativePath);
|
||||
const content = await fs.readFile(fullPath, 'utf-8');
|
||||
let content = await fs.readFile(fullPath, 'utf-8');
|
||||
console.log(` ✓ Found docs at: ${relativePath}`);
|
||||
|
||||
// Inject special guidance for loop nodes
|
||||
content = this.enhanceLoopNodeDocumentation(nodeType, content);
|
||||
|
||||
return content;
|
||||
} catch (error) {
|
||||
// File doesn't exist, try next
|
||||
@@ -62,4 +66,56 @@ export class DocsMapper {
|
||||
console.log(` ✗ No docs found for ${nodeName}`);
|
||||
return null;
|
||||
}
|
||||
|
||||
private enhanceLoopNodeDocumentation(nodeType: string, content: string): string {
|
||||
// Add critical output index information for SplitInBatches
|
||||
if (nodeType.includes('splitInBatches')) {
|
||||
const outputGuidance = `
|
||||
|
||||
## CRITICAL OUTPUT CONNECTION INFORMATION
|
||||
|
||||
**⚠️ OUTPUT INDICES ARE COUNTERINTUITIVE ⚠️**
|
||||
|
||||
The SplitInBatches node has TWO outputs with specific indices:
|
||||
- **Output 0 (index 0) = "done"**: Receives final processed data when loop completes
|
||||
- **Output 1 (index 1) = "loop"**: Receives current batch data during iteration
|
||||
|
||||
### Correct Connection Pattern:
|
||||
1. Connect nodes that PROCESS items inside the loop to **Output 1 ("loop")**
|
||||
2. Connect nodes that run AFTER the loop completes to **Output 0 ("done")**
|
||||
3. The last processing node in the loop must connect back to the SplitInBatches node
|
||||
|
||||
### Common Mistake:
|
||||
AI assistants often connect these backwards because the logical flow (loop first, then done) doesn't match the technical indices (done=0, loop=1).
|
||||
|
||||
`;
|
||||
// Insert after the main description
|
||||
const insertPoint = content.indexOf('## When to use');
|
||||
if (insertPoint > -1) {
|
||||
content = content.slice(0, insertPoint) + outputGuidance + content.slice(insertPoint);
|
||||
} else {
|
||||
// Append if no good insertion point found
|
||||
content = outputGuidance + '\n' + content;
|
||||
}
|
||||
}
|
||||
|
||||
// Add guidance for IF node
|
||||
if (nodeType.includes('.if')) {
|
||||
const outputGuidance = `
|
||||
|
||||
## Output Connection Information
|
||||
|
||||
The IF node has TWO outputs:
|
||||
- **Output 0 (index 0) = "true"**: Items that match the condition
|
||||
- **Output 1 (index 1) = "false"**: Items that do not match the condition
|
||||
|
||||
`;
|
||||
const insertPoint = content.indexOf('## Node parameters');
|
||||
if (insertPoint > -1) {
|
||||
content = content.slice(0, insertPoint) + outputGuidance + content.slice(insertPoint);
|
||||
}
|
||||
}
|
||||
|
||||
return content;
|
||||
}
|
||||
}
|
||||
@@ -834,10 +834,26 @@ export class N8NDocumentationMCPServer {
|
||||
null
|
||||
};
|
||||
|
||||
// Process outputs to provide clear mapping
|
||||
let outputs = undefined;
|
||||
if (node.outputNames && node.outputNames.length > 0) {
|
||||
outputs = node.outputNames.map((name: string, index: number) => {
|
||||
// Special handling for loop nodes like SplitInBatches
|
||||
const descriptions = this.getOutputDescriptions(node.nodeType, name, index);
|
||||
return {
|
||||
index,
|
||||
name,
|
||||
description: descriptions.description,
|
||||
connectionGuidance: descriptions.connectionGuidance
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
return {
|
||||
...node,
|
||||
workflowNodeType: getWorkflowNodeType(node.package, node.nodeType),
|
||||
aiToolCapabilities
|
||||
aiToolCapabilities,
|
||||
outputs
|
||||
};
|
||||
}
|
||||
|
||||
@@ -1937,6 +1953,52 @@ Full documentation is being prepared. For now, use get_node_essentials for confi
|
||||
};
|
||||
}
|
||||
|
||||
private getOutputDescriptions(nodeType: string, outputName: string, index: number): { description: string, connectionGuidance: string } {
|
||||
// Special handling for loop nodes
|
||||
if (nodeType === 'nodes-base.splitInBatches') {
|
||||
if (outputName === 'done' && index === 0) {
|
||||
return {
|
||||
description: 'Final processed data after all iterations complete',
|
||||
connectionGuidance: 'Connect to nodes that should run AFTER the loop completes'
|
||||
};
|
||||
} else if (outputName === 'loop' && index === 1) {
|
||||
return {
|
||||
description: 'Current batch data for this iteration',
|
||||
connectionGuidance: 'Connect to nodes that process items INSIDE the loop (and connect their output back to this node)'
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
// Special handling for IF node
|
||||
if (nodeType === 'nodes-base.if') {
|
||||
if (outputName === 'true' && index === 0) {
|
||||
return {
|
||||
description: 'Items that match the condition',
|
||||
connectionGuidance: 'Connect to nodes that handle the TRUE case'
|
||||
};
|
||||
} else if (outputName === 'false' && index === 1) {
|
||||
return {
|
||||
description: 'Items that do not match the condition',
|
||||
connectionGuidance: 'Connect to nodes that handle the FALSE case'
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
// Special handling for Switch node
|
||||
if (nodeType === 'nodes-base.switch') {
|
||||
return {
|
||||
description: `Output ${index}: ${outputName || 'Route ' + index}`,
|
||||
connectionGuidance: `Connect to nodes for the "${outputName || 'route ' + index}" case`
|
||||
};
|
||||
}
|
||||
|
||||
// Default handling
|
||||
return {
|
||||
description: outputName || `Output ${index}`,
|
||||
connectionGuidance: `Connect to downstream nodes`
|
||||
};
|
||||
}
|
||||
|
||||
private getCommonAIToolUseCases(nodeType: string): string[] {
|
||||
const useCaseMap: Record<string, string[]> = {
|
||||
'nodes-base.slack': [
|
||||
|
||||
@@ -16,14 +16,19 @@ export interface ParsedNode {
|
||||
isVersioned: boolean;
|
||||
packageName: string;
|
||||
documentation?: string;
|
||||
outputs?: any[];
|
||||
outputNames?: string[];
|
||||
}
|
||||
|
||||
export class NodeParser {
|
||||
private propertyExtractor = new PropertyExtractor();
|
||||
private currentNodeClass: any = null;
|
||||
|
||||
parse(nodeClass: any, packageName: string): ParsedNode {
|
||||
this.currentNodeClass = nodeClass;
|
||||
// Get base description (handles versioned nodes)
|
||||
const description = this.getNodeDescription(nodeClass);
|
||||
const outputInfo = this.extractOutputs(description);
|
||||
|
||||
return {
|
||||
style: this.detectStyle(nodeClass),
|
||||
@@ -39,7 +44,9 @@ export class NodeParser {
|
||||
operations: this.propertyExtractor.extractOperations(nodeClass),
|
||||
version: this.extractVersion(nodeClass),
|
||||
isVersioned: this.detectVersioned(nodeClass),
|
||||
packageName: packageName
|
||||
packageName: packageName,
|
||||
outputs: outputInfo.outputs,
|
||||
outputNames: outputInfo.outputNames
|
||||
};
|
||||
}
|
||||
|
||||
@@ -222,4 +229,51 @@ export class NodeParser {
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
private extractOutputs(description: any): { outputs?: any[], outputNames?: string[] } {
|
||||
const result: { outputs?: any[], outputNames?: string[] } = {};
|
||||
|
||||
// First check the base description
|
||||
if (description.outputs) {
|
||||
result.outputs = Array.isArray(description.outputs) ? description.outputs : [description.outputs];
|
||||
}
|
||||
|
||||
if (description.outputNames) {
|
||||
result.outputNames = Array.isArray(description.outputNames) ? description.outputNames : [description.outputNames];
|
||||
}
|
||||
|
||||
// If no outputs found and this is a versioned node, check the latest version
|
||||
if (!result.outputs && !result.outputNames) {
|
||||
const nodeClass = this.currentNodeClass; // We'll need to track this
|
||||
if (nodeClass) {
|
||||
try {
|
||||
const instance = new nodeClass();
|
||||
if (instance.nodeVersions) {
|
||||
// Get the latest version
|
||||
const versions = Object.keys(instance.nodeVersions).map(Number);
|
||||
const latestVersion = Math.max(...versions);
|
||||
const versionedDescription = instance.nodeVersions[latestVersion]?.description;
|
||||
|
||||
if (versionedDescription) {
|
||||
if (versionedDescription.outputs) {
|
||||
result.outputs = Array.isArray(versionedDescription.outputs)
|
||||
? versionedDescription.outputs
|
||||
: [versionedDescription.outputs];
|
||||
}
|
||||
|
||||
if (versionedDescription.outputNames) {
|
||||
result.outputNames = Array.isArray(versionedDescription.outputNames)
|
||||
? versionedDescription.outputNames
|
||||
: [versionedDescription.outputNames];
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (e) {
|
||||
// Ignore errors from instantiating node
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return result;
|
||||
}
|
||||
}
|
||||
@@ -72,6 +72,8 @@ export interface WorkflowValidationResult {
|
||||
}
|
||||
|
||||
export class WorkflowValidator {
|
||||
private currentWorkflow: WorkflowJson | null = null;
|
||||
|
||||
constructor(
|
||||
private nodeRepository: NodeRepository,
|
||||
private nodeValidator: typeof EnhancedConfigValidator
|
||||
@@ -89,6 +91,9 @@ export class WorkflowValidator {
|
||||
profile?: 'minimal' | 'runtime' | 'ai-friendly' | 'strict';
|
||||
} = {}
|
||||
): Promise<WorkflowValidationResult> {
|
||||
// Store current workflow for access in helper methods
|
||||
this.currentWorkflow = workflow;
|
||||
|
||||
const {
|
||||
validateNodes = true,
|
||||
validateConnections = true,
|
||||
@@ -627,6 +632,9 @@ export class WorkflowValidator {
|
||||
result: WorkflowValidationResult,
|
||||
outputType: 'main' | 'error' | 'ai_tool'
|
||||
): void {
|
||||
// Get source node for special validation
|
||||
const sourceNode = nodeMap.get(sourceName);
|
||||
|
||||
outputs.forEach((outputConnections, outputIndex) => {
|
||||
if (!outputConnections) return;
|
||||
|
||||
@@ -641,12 +649,26 @@ export class WorkflowValidator {
|
||||
return;
|
||||
}
|
||||
|
||||
// Special validation for SplitInBatches node
|
||||
if (sourceNode && sourceNode.type === 'n8n-nodes-base.splitInBatches') {
|
||||
this.validateSplitInBatchesConnection(
|
||||
sourceNode,
|
||||
outputIndex,
|
||||
connection,
|
||||
nodeMap,
|
||||
result
|
||||
);
|
||||
}
|
||||
|
||||
// Check for self-referencing connections
|
||||
if (connection.node === sourceName) {
|
||||
result.warnings.push({
|
||||
type: 'warning',
|
||||
message: `Node "${sourceName}" has a self-referencing connection. This can cause infinite loops.`
|
||||
});
|
||||
// This is only a warning for non-loop nodes
|
||||
if (sourceNode && sourceNode.type !== 'n8n-nodes-base.splitInBatches') {
|
||||
result.warnings.push({
|
||||
type: 'warning',
|
||||
message: `Node "${sourceName}" has a self-referencing connection. This can cause infinite loops.`
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
const targetNode = nodeMap.get(connection.node);
|
||||
@@ -728,12 +750,19 @@ export class WorkflowValidator {
|
||||
|
||||
/**
|
||||
* Check if workflow has cycles
|
||||
* Allow legitimate loops for SplitInBatches and similar loop nodes
|
||||
*/
|
||||
private hasCycle(workflow: WorkflowJson): boolean {
|
||||
const visited = new Set<string>();
|
||||
const recursionStack = new Set<string>();
|
||||
const nodeTypeMap = new Map<string, string>();
|
||||
|
||||
// Build node type map
|
||||
workflow.nodes.forEach(node => {
|
||||
nodeTypeMap.set(node.name, node.type);
|
||||
});
|
||||
|
||||
const hasCycleDFS = (nodeName: string): boolean => {
|
||||
const hasCycleDFS = (nodeName: string, pathFromLoopNode: boolean = false): boolean => {
|
||||
visited.add(nodeName);
|
||||
recursionStack.add(nodeName);
|
||||
|
||||
@@ -759,11 +788,23 @@ export class WorkflowValidator {
|
||||
});
|
||||
}
|
||||
|
||||
const currentNodeType = nodeTypeMap.get(nodeName);
|
||||
const isLoopNode = currentNodeType === 'n8n-nodes-base.splitInBatches';
|
||||
|
||||
for (const target of allTargets) {
|
||||
if (!visited.has(target)) {
|
||||
if (hasCycleDFS(target)) return true;
|
||||
if (hasCycleDFS(target, pathFromLoopNode || isLoopNode)) return true;
|
||||
} else if (recursionStack.has(target)) {
|
||||
return true;
|
||||
// Allow cycles that involve loop nodes like SplitInBatches
|
||||
const targetNodeType = nodeTypeMap.get(target);
|
||||
const isTargetLoopNode = targetNodeType === 'n8n-nodes-base.splitInBatches';
|
||||
|
||||
// If this cycle involves a loop node, it's legitimate
|
||||
if (isTargetLoopNode || pathFromLoopNode) {
|
||||
continue; // Allow this cycle
|
||||
}
|
||||
|
||||
return true; // Reject other cycles
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1470,4 +1511,134 @@ export class WorkflowValidator {
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Validate SplitInBatches node connections for common mistakes
|
||||
*/
|
||||
private validateSplitInBatchesConnection(
|
||||
sourceNode: WorkflowNode,
|
||||
outputIndex: number,
|
||||
connection: { node: string; type: string; index: number },
|
||||
nodeMap: Map<string, WorkflowNode>,
|
||||
result: WorkflowValidationResult
|
||||
): void {
|
||||
const targetNode = nodeMap.get(connection.node);
|
||||
if (!targetNode) return;
|
||||
|
||||
// Check if connections appear to be reversed
|
||||
// Output 0 = "done", Output 1 = "loop"
|
||||
|
||||
if (outputIndex === 0) {
|
||||
// This is the "done" output (index 0)
|
||||
// Check if target looks like it should be in the loop
|
||||
const targetType = targetNode.type.toLowerCase();
|
||||
const targetName = targetNode.name.toLowerCase();
|
||||
|
||||
// Common patterns that suggest this node should be inside the loop
|
||||
if (targetType.includes('function') ||
|
||||
targetType.includes('code') ||
|
||||
targetType.includes('item') ||
|
||||
targetName.includes('process') ||
|
||||
targetName.includes('transform') ||
|
||||
targetName.includes('handle')) {
|
||||
|
||||
// Check if this node connects back to the SplitInBatches
|
||||
const hasLoopBack = this.checkForLoopBack(targetNode.name, sourceNode.name, nodeMap);
|
||||
|
||||
if (hasLoopBack) {
|
||||
result.errors.push({
|
||||
type: 'error',
|
||||
nodeId: sourceNode.id,
|
||||
nodeName: sourceNode.name,
|
||||
message: `SplitInBatches outputs appear reversed! Node "${targetNode.name}" is connected to output 0 ("done") but connects back to the loop. It should be connected to output 1 ("loop") instead. Remember: Output 0 = "done" (post-loop), Output 1 = "loop" (inside loop).`
|
||||
});
|
||||
} else {
|
||||
result.warnings.push({
|
||||
type: 'warning',
|
||||
nodeId: sourceNode.id,
|
||||
nodeName: sourceNode.name,
|
||||
message: `Node "${targetNode.name}" is connected to the "done" output (index 0) but appears to be a processing node. Consider connecting it to the "loop" output (index 1) if it should process items inside the loop.`
|
||||
});
|
||||
}
|
||||
}
|
||||
} else if (outputIndex === 1) {
|
||||
// This is the "loop" output (index 1)
|
||||
// Check if target looks like it should be after the loop
|
||||
const targetType = targetNode.type.toLowerCase();
|
||||
const targetName = targetNode.name.toLowerCase();
|
||||
|
||||
// Common patterns that suggest this node should be after the loop
|
||||
if (targetType.includes('aggregate') ||
|
||||
targetType.includes('merge') ||
|
||||
targetType.includes('email') ||
|
||||
targetType.includes('slack') ||
|
||||
targetName.includes('final') ||
|
||||
targetName.includes('complete') ||
|
||||
targetName.includes('summary') ||
|
||||
targetName.includes('report')) {
|
||||
|
||||
result.warnings.push({
|
||||
type: 'warning',
|
||||
nodeId: sourceNode.id,
|
||||
nodeName: sourceNode.name,
|
||||
message: `Node "${targetNode.name}" is connected to the "loop" output (index 1) but appears to be a post-processing node. Consider connecting it to the "done" output (index 0) if it should run after all iterations complete.`
|
||||
});
|
||||
}
|
||||
|
||||
// Check if loop output doesn't eventually connect back
|
||||
const hasLoopBack = this.checkForLoopBack(targetNode.name, sourceNode.name, nodeMap);
|
||||
if (!hasLoopBack) {
|
||||
result.warnings.push({
|
||||
type: 'warning',
|
||||
nodeId: sourceNode.id,
|
||||
nodeName: sourceNode.name,
|
||||
message: `The "loop" output connects to "${targetNode.name}" but doesn't connect back to the SplitInBatches node. The last node in the loop should connect back to complete the iteration.`
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if a node eventually connects back to a target node
|
||||
*/
|
||||
private checkForLoopBack(
|
||||
startNode: string,
|
||||
targetNode: string,
|
||||
nodeMap: Map<string, WorkflowNode>,
|
||||
visited: Set<string> = new Set(),
|
||||
maxDepth: number = 50
|
||||
): boolean {
|
||||
if (maxDepth <= 0) return false; // Prevent stack overflow
|
||||
if (visited.has(startNode)) return false;
|
||||
visited.add(startNode);
|
||||
|
||||
const node = nodeMap.get(startNode);
|
||||
if (!node) return false;
|
||||
|
||||
// Access connections from the workflow structure, not the node
|
||||
// We need to access this.currentWorkflow.connections[startNode]
|
||||
const connections = (this as any).currentWorkflow?.connections[startNode];
|
||||
if (!connections) return false;
|
||||
|
||||
for (const [outputType, outputs] of Object.entries(connections)) {
|
||||
if (!Array.isArray(outputs)) continue;
|
||||
|
||||
for (const outputConnections of outputs) {
|
||||
if (!Array.isArray(outputConnections)) continue;
|
||||
|
||||
for (const conn of outputConnections) {
|
||||
if (conn.node === targetNode) {
|
||||
return true;
|
||||
}
|
||||
|
||||
// Recursively check connected nodes
|
||||
if (this.checkForLoopBack(conn.node, targetNode, nodeMap, visited, maxDepth - 1)) {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
}
|
||||
@@ -124,9 +124,9 @@ describe('MCP Tool Invocation', () => {
|
||||
const andNodes = andResult.results;
|
||||
expect(andNodes.length).toBeLessThanOrEqual(orNodes.length);
|
||||
|
||||
// FUZZY mode
|
||||
// FUZZY mode - use less typo-heavy search
|
||||
const fuzzyResponse = await client.callTool({ name: 'search_nodes', arguments: {
|
||||
query: 'htpp requst', // Intentional typos
|
||||
query: 'http req', // Partial match should work
|
||||
mode: 'FUZZY'
|
||||
}});
|
||||
const fuzzyResult = JSON.parse(((fuzzyResponse as any).content[0]).text);
|
||||
|
||||
@@ -83,7 +83,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
isWebhook: false,
|
||||
isVersioned: true,
|
||||
version: '1.0',
|
||||
documentation: 'HTTP Request documentation'
|
||||
documentation: 'HTTP Request documentation',
|
||||
outputs: undefined,
|
||||
outputNames: undefined
|
||||
};
|
||||
|
||||
repository.saveNode(parsedNode);
|
||||
@@ -108,7 +110,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
'HTTP Request documentation',
|
||||
JSON.stringify([{ name: 'url', type: 'string' }], null, 2),
|
||||
JSON.stringify([{ name: 'execute', displayName: 'Execute' }], null, 2),
|
||||
JSON.stringify([{ name: 'httpBasicAuth' }], null, 2)
|
||||
JSON.stringify([{ name: 'httpBasicAuth' }], null, 2),
|
||||
null, // outputs
|
||||
null // outputNames
|
||||
);
|
||||
});
|
||||
|
||||
@@ -125,7 +129,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
isAITool: true,
|
||||
isTrigger: true,
|
||||
isWebhook: true,
|
||||
isVersioned: false
|
||||
isVersioned: false,
|
||||
outputs: undefined,
|
||||
outputNames: undefined
|
||||
};
|
||||
|
||||
repository.saveNode(minimalNode);
|
||||
@@ -157,7 +163,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
properties_schema: JSON.stringify([{ name: 'url', type: 'string' }]),
|
||||
operations: JSON.stringify([{ name: 'execute' }]),
|
||||
credentials_required: JSON.stringify([{ name: 'httpBasicAuth' }]),
|
||||
documentation: 'HTTP docs'
|
||||
documentation: 'HTTP docs',
|
||||
outputs: null,
|
||||
output_names: null
|
||||
};
|
||||
|
||||
mockAdapter._setMockData('node:nodes-base.httpRequest', mockRow);
|
||||
@@ -179,7 +187,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
properties: [{ name: 'url', type: 'string' }],
|
||||
operations: [{ name: 'execute' }],
|
||||
credentials: [{ name: 'httpBasicAuth' }],
|
||||
hasDocumentation: true
|
||||
hasDocumentation: true,
|
||||
outputs: null,
|
||||
outputNames: null
|
||||
});
|
||||
});
|
||||
|
||||
@@ -204,7 +214,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
properties_schema: '{invalid json',
|
||||
operations: 'not json at all',
|
||||
credentials_required: '{"valid": "json"}',
|
||||
documentation: null
|
||||
documentation: null,
|
||||
outputs: null,
|
||||
output_names: null
|
||||
};
|
||||
|
||||
mockAdapter._setMockData('node:nodes-base.broken', mockRow);
|
||||
@@ -320,7 +332,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
isVersioned: false
|
||||
isVersioned: false,
|
||||
outputs: undefined,
|
||||
outputNames: undefined
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
@@ -348,7 +362,9 @@ describe('NodeRepository - Core Functionality', () => {
|
||||
properties_schema: '[]',
|
||||
operations: '[]',
|
||||
credentials_required: '[]',
|
||||
documentation: null
|
||||
documentation: null,
|
||||
outputs: null,
|
||||
output_names: null
|
||||
};
|
||||
|
||||
mockAdapter._setMockData('node:nodes-base.bool-test', mockRow);
|
||||
|
||||
568
tests/unit/database/node-repository-outputs.test.ts
Normal file
568
tests/unit/database/node-repository-outputs.test.ts
Normal file
@@ -0,0 +1,568 @@
|
||||
import { describe, it, expect, beforeEach, vi } from 'vitest';
|
||||
import { NodeRepository } from '@/database/node-repository';
|
||||
import { DatabaseAdapter } from '@/database/database-adapter';
|
||||
import { ParsedNode } from '@/parsers/node-parser';
|
||||
|
||||
describe('NodeRepository - Outputs Handling', () => {
|
||||
let repository: NodeRepository;
|
||||
let mockDb: DatabaseAdapter;
|
||||
let mockStatement: any;
|
||||
|
||||
beforeEach(() => {
|
||||
mockStatement = {
|
||||
run: vi.fn(),
|
||||
get: vi.fn(),
|
||||
all: vi.fn()
|
||||
};
|
||||
|
||||
mockDb = {
|
||||
prepare: vi.fn().mockReturnValue(mockStatement),
|
||||
transaction: vi.fn(),
|
||||
exec: vi.fn(),
|
||||
close: vi.fn(),
|
||||
pragma: vi.fn()
|
||||
} as any;
|
||||
|
||||
repository = new NodeRepository(mockDb);
|
||||
});
|
||||
|
||||
describe('saveNode with outputs', () => {
|
||||
it('should save node with outputs and outputNames correctly', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'Done', description: 'Final results when loop completes' },
|
||||
{ displayName: 'Loop', description: 'Current batch data during iteration' }
|
||||
];
|
||||
const outputNames = ['done', 'loop'];
|
||||
|
||||
const node: ParsedNode = {
|
||||
style: 'programmatic',
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
description: 'Split data into batches',
|
||||
category: 'transform',
|
||||
properties: [],
|
||||
credentials: [],
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
operations: [],
|
||||
version: '3',
|
||||
isVersioned: false,
|
||||
packageName: 'n8n-nodes-base',
|
||||
outputs,
|
||||
outputNames
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
|
||||
expect(mockDb.prepare).toHaveBeenCalledWith(`
|
||||
INSERT OR REPLACE INTO nodes (
|
||||
node_type, package_name, display_name, description,
|
||||
category, development_style, is_ai_tool, is_trigger,
|
||||
is_webhook, is_versioned, version, documentation,
|
||||
properties_schema, operations, credentials_required,
|
||||
outputs, output_names
|
||||
) VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||
`);
|
||||
|
||||
expect(mockStatement.run).toHaveBeenCalledWith(
|
||||
'nodes-base.splitInBatches',
|
||||
'n8n-nodes-base',
|
||||
'Split In Batches',
|
||||
'Split data into batches',
|
||||
'transform',
|
||||
'programmatic',
|
||||
0, // false
|
||||
0, // false
|
||||
0, // false
|
||||
0, // false
|
||||
'3',
|
||||
null, // documentation
|
||||
JSON.stringify([], null, 2), // properties
|
||||
JSON.stringify([], null, 2), // operations
|
||||
JSON.stringify([], null, 2), // credentials
|
||||
JSON.stringify(outputs, null, 2), // outputs
|
||||
JSON.stringify(outputNames, null, 2) // output_names
|
||||
);
|
||||
});
|
||||
|
||||
it('should save node with only outputs (no outputNames)', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'True', description: 'Items that match condition' },
|
||||
{ displayName: 'False', description: 'Items that do not match condition' }
|
||||
];
|
||||
|
||||
const node: ParsedNode = {
|
||||
style: 'programmatic',
|
||||
nodeType: 'nodes-base.if',
|
||||
displayName: 'IF',
|
||||
description: 'Route items based on conditions',
|
||||
category: 'transform',
|
||||
properties: [],
|
||||
credentials: [],
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
operations: [],
|
||||
version: '2',
|
||||
isVersioned: false,
|
||||
packageName: 'n8n-nodes-base',
|
||||
outputs
|
||||
// no outputNames
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
|
||||
const callArgs = mockStatement.run.mock.calls[0];
|
||||
expect(callArgs[15]).toBe(JSON.stringify(outputs, null, 2)); // outputs
|
||||
expect(callArgs[16]).toBe(null); // output_names should be null
|
||||
});
|
||||
|
||||
it('should save node with only outputNames (no outputs)', () => {
|
||||
const outputNames = ['main', 'error'];
|
||||
|
||||
const node: ParsedNode = {
|
||||
style: 'programmatic',
|
||||
nodeType: 'nodes-base.customNode',
|
||||
displayName: 'Custom Node',
|
||||
description: 'Custom node with output names only',
|
||||
category: 'transform',
|
||||
properties: [],
|
||||
credentials: [],
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
operations: [],
|
||||
version: '1',
|
||||
isVersioned: false,
|
||||
packageName: 'n8n-nodes-base',
|
||||
outputNames
|
||||
// no outputs
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
|
||||
const callArgs = mockStatement.run.mock.calls[0];
|
||||
expect(callArgs[15]).toBe(null); // outputs should be null
|
||||
expect(callArgs[16]).toBe(JSON.stringify(outputNames, null, 2)); // output_names
|
||||
});
|
||||
|
||||
it('should save node without outputs or outputNames', () => {
|
||||
const node: ParsedNode = {
|
||||
style: 'programmatic',
|
||||
nodeType: 'nodes-base.httpRequest',
|
||||
displayName: 'HTTP Request',
|
||||
description: 'Make HTTP requests',
|
||||
category: 'input',
|
||||
properties: [],
|
||||
credentials: [],
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
operations: [],
|
||||
version: '4',
|
||||
isVersioned: false,
|
||||
packageName: 'n8n-nodes-base'
|
||||
// no outputs or outputNames
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
|
||||
const callArgs = mockStatement.run.mock.calls[0];
|
||||
expect(callArgs[15]).toBe(null); // outputs should be null
|
||||
expect(callArgs[16]).toBe(null); // output_names should be null
|
||||
});
|
||||
|
||||
it('should handle empty outputs and outputNames arrays', () => {
|
||||
const node: ParsedNode = {
|
||||
style: 'programmatic',
|
||||
nodeType: 'nodes-base.emptyNode',
|
||||
displayName: 'Empty Node',
|
||||
description: 'Node with empty outputs',
|
||||
category: 'misc',
|
||||
properties: [],
|
||||
credentials: [],
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
operations: [],
|
||||
version: '1',
|
||||
isVersioned: false,
|
||||
packageName: 'n8n-nodes-base',
|
||||
outputs: [],
|
||||
outputNames: []
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
|
||||
const callArgs = mockStatement.run.mock.calls[0];
|
||||
expect(callArgs[15]).toBe(JSON.stringify([], null, 2)); // outputs
|
||||
expect(callArgs[16]).toBe(JSON.stringify([], null, 2)); // output_names
|
||||
});
|
||||
});
|
||||
|
||||
describe('getNode with outputs', () => {
|
||||
it('should retrieve node with outputs and outputNames correctly', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'Done', description: 'Final results when loop completes' },
|
||||
{ displayName: 'Loop', description: 'Current batch data during iteration' }
|
||||
];
|
||||
const outputNames = ['done', 'loop'];
|
||||
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.splitInBatches',
|
||||
display_name: 'Split In Batches',
|
||||
description: 'Split data into batches',
|
||||
category: 'transform',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '3',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: JSON.stringify(outputs),
|
||||
output_names: JSON.stringify(outputNames)
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.splitInBatches');
|
||||
|
||||
expect(result).toEqual({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
description: 'Split data into batches',
|
||||
category: 'transform',
|
||||
developmentStyle: 'programmatic',
|
||||
package: 'n8n-nodes-base',
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
isVersioned: false,
|
||||
version: '3',
|
||||
properties: [],
|
||||
operations: [],
|
||||
credentials: [],
|
||||
hasDocumentation: false,
|
||||
outputs,
|
||||
outputNames
|
||||
});
|
||||
});
|
||||
|
||||
it('should retrieve node with only outputs (null outputNames)', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'True', description: 'Items that match condition' }
|
||||
];
|
||||
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.if',
|
||||
display_name: 'IF',
|
||||
description: 'Route items',
|
||||
category: 'transform',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '2',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: JSON.stringify(outputs),
|
||||
output_names: null
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.if');
|
||||
|
||||
expect(result.outputs).toEqual(outputs);
|
||||
expect(result.outputNames).toBe(null);
|
||||
});
|
||||
|
||||
it('should retrieve node with only outputNames (null outputs)', () => {
|
||||
const outputNames = ['main'];
|
||||
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.customNode',
|
||||
display_name: 'Custom Node',
|
||||
description: 'Custom node',
|
||||
category: 'misc',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '1',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: null,
|
||||
output_names: JSON.stringify(outputNames)
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.customNode');
|
||||
|
||||
expect(result.outputs).toBe(null);
|
||||
expect(result.outputNames).toEqual(outputNames);
|
||||
});
|
||||
|
||||
it('should retrieve node without outputs or outputNames', () => {
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.httpRequest',
|
||||
display_name: 'HTTP Request',
|
||||
description: 'Make HTTP requests',
|
||||
category: 'input',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '4',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: null,
|
||||
output_names: null
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.httpRequest');
|
||||
|
||||
expect(result.outputs).toBe(null);
|
||||
expect(result.outputNames).toBe(null);
|
||||
});
|
||||
|
||||
it('should handle malformed JSON gracefully', () => {
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.malformed',
|
||||
display_name: 'Malformed Node',
|
||||
description: 'Node with malformed JSON',
|
||||
category: 'misc',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '1',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: '{invalid json}',
|
||||
output_names: '[invalid, json'
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.malformed');
|
||||
|
||||
// Should use default values when JSON parsing fails
|
||||
expect(result.outputs).toBe(null);
|
||||
expect(result.outputNames).toBe(null);
|
||||
});
|
||||
|
||||
it('should return null for non-existent node', () => {
|
||||
mockStatement.get.mockReturnValue(null);
|
||||
|
||||
const result = repository.getNode('nodes-base.nonExistent');
|
||||
|
||||
expect(result).toBe(null);
|
||||
});
|
||||
|
||||
it('should handle SplitInBatches counterintuitive output order correctly', () => {
|
||||
// Test that the output order is preserved: done=0, loop=1
|
||||
const outputs = [
|
||||
{ displayName: 'Done', description: 'Final results when loop completes', index: 0 },
|
||||
{ displayName: 'Loop', description: 'Current batch data during iteration', index: 1 }
|
||||
];
|
||||
const outputNames = ['done', 'loop'];
|
||||
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.splitInBatches',
|
||||
display_name: 'Split In Batches',
|
||||
description: 'Split data into batches',
|
||||
category: 'transform',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '3',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: JSON.stringify(outputs),
|
||||
output_names: JSON.stringify(outputNames)
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.splitInBatches');
|
||||
|
||||
// Verify order is preserved
|
||||
expect(result.outputs[0].displayName).toBe('Done');
|
||||
expect(result.outputs[1].displayName).toBe('Loop');
|
||||
expect(result.outputNames[0]).toBe('done');
|
||||
expect(result.outputNames[1]).toBe('loop');
|
||||
});
|
||||
});
|
||||
|
||||
describe('parseNodeRow with outputs', () => {
|
||||
it('should parse node row with outputs correctly using parseNodeRow', () => {
|
||||
const outputs = [{ displayName: 'Output' }];
|
||||
const outputNames = ['main'];
|
||||
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.test',
|
||||
display_name: 'Test',
|
||||
description: 'Test node',
|
||||
category: 'misc',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '1',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: JSON.stringify(outputs),
|
||||
output_names: JSON.stringify(outputNames)
|
||||
};
|
||||
|
||||
mockStatement.all.mockReturnValue([mockRow]);
|
||||
|
||||
const results = repository.getAllNodes(1);
|
||||
|
||||
expect(results[0].outputs).toEqual(outputs);
|
||||
expect(results[0].outputNames).toEqual(outputNames);
|
||||
});
|
||||
|
||||
it('should handle empty string as null for outputs', () => {
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.empty',
|
||||
display_name: 'Empty',
|
||||
description: 'Empty node',
|
||||
category: 'misc',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '1',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: '', // empty string
|
||||
output_names: '' // empty string
|
||||
};
|
||||
|
||||
mockStatement.all.mockReturnValue([mockRow]);
|
||||
|
||||
const results = repository.getAllNodes(1);
|
||||
|
||||
// Empty strings should be treated as null since they fail JSON parsing
|
||||
expect(results[0].outputs).toBe(null);
|
||||
expect(results[0].outputNames).toBe(null);
|
||||
});
|
||||
});
|
||||
|
||||
describe('complex output structures', () => {
|
||||
it('should handle complex output objects with metadata', () => {
|
||||
const complexOutputs = [
|
||||
{
|
||||
displayName: 'Done',
|
||||
name: 'done',
|
||||
type: 'main',
|
||||
hint: 'Receives the final data after all batches have been processed',
|
||||
description: 'Final results when loop completes',
|
||||
index: 0
|
||||
},
|
||||
{
|
||||
displayName: 'Loop',
|
||||
name: 'loop',
|
||||
type: 'main',
|
||||
hint: 'Receives the current batch data during each iteration',
|
||||
description: 'Current batch data during iteration',
|
||||
index: 1
|
||||
}
|
||||
];
|
||||
|
||||
const node: ParsedNode = {
|
||||
style: 'programmatic',
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
description: 'Split data into batches',
|
||||
category: 'transform',
|
||||
properties: [],
|
||||
credentials: [],
|
||||
isAITool: false,
|
||||
isTrigger: false,
|
||||
isWebhook: false,
|
||||
operations: [],
|
||||
version: '3',
|
||||
isVersioned: false,
|
||||
packageName: 'n8n-nodes-base',
|
||||
outputs: complexOutputs,
|
||||
outputNames: ['done', 'loop']
|
||||
};
|
||||
|
||||
repository.saveNode(node);
|
||||
|
||||
// Simulate retrieval
|
||||
const mockRow = {
|
||||
node_type: 'nodes-base.splitInBatches',
|
||||
display_name: 'Split In Batches',
|
||||
description: 'Split data into batches',
|
||||
category: 'transform',
|
||||
development_style: 'programmatic',
|
||||
package_name: 'n8n-nodes-base',
|
||||
is_ai_tool: 0,
|
||||
is_trigger: 0,
|
||||
is_webhook: 0,
|
||||
is_versioned: 0,
|
||||
version: '3',
|
||||
properties_schema: JSON.stringify([]),
|
||||
operations: JSON.stringify([]),
|
||||
credentials_required: JSON.stringify([]),
|
||||
documentation: null,
|
||||
outputs: JSON.stringify(complexOutputs),
|
||||
output_names: JSON.stringify(['done', 'loop'])
|
||||
};
|
||||
|
||||
mockStatement.get.mockReturnValue(mockRow);
|
||||
|
||||
const result = repository.getNode('nodes-base.splitInBatches');
|
||||
|
||||
expect(result.outputs).toEqual(complexOutputs);
|
||||
expect(result.outputs[0]).toMatchObject({
|
||||
displayName: 'Done',
|
||||
name: 'done',
|
||||
type: 'main',
|
||||
hint: 'Receives the final data after all batches have been processed'
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -299,6 +299,268 @@ describe('DocsMapper', () => {
|
||||
});
|
||||
});
|
||||
|
||||
describe('enhanceLoopNodeDocumentation - SplitInBatches', () => {
|
||||
it('should enhance SplitInBatches documentation with output guidance', async () => {
|
||||
const originalContent = `# Split In Batches Node
|
||||
|
||||
This node splits data into batches.
|
||||
|
||||
## When to use
|
||||
|
||||
Use this node when you need to process large datasets in smaller chunks.
|
||||
|
||||
## Parameters
|
||||
|
||||
- batchSize: Number of items per batch
|
||||
`;
|
||||
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(result!).toContain('⚠️ OUTPUT INDICES ARE COUNTERINTUITIVE ⚠️');
|
||||
expect(result!).toContain('Output 0 (index 0) = "done"');
|
||||
expect(result!).toContain('Output 1 (index 1) = "loop"');
|
||||
expect(result!).toContain('Correct Connection Pattern:');
|
||||
expect(result!).toContain('Common Mistake:');
|
||||
expect(result!).toContain('AI assistants often connect these backwards');
|
||||
|
||||
// Should insert before "When to use" section
|
||||
const insertionIndex = result!.indexOf('## When to use');
|
||||
const guidanceIndex = result!.indexOf('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(guidanceIndex).toBeLessThan(insertionIndex);
|
||||
expect(guidanceIndex).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
it('should enhance SplitInBatches documentation when no "When to use" section exists', async () => {
|
||||
const originalContent = `# Split In Batches Node
|
||||
|
||||
This node splits data into batches.
|
||||
|
||||
## Parameters
|
||||
|
||||
- batchSize: Number of items per batch
|
||||
`;
|
||||
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
// Should be inserted at the beginning since no "When to use" section
|
||||
expect(result!.indexOf('CRITICAL OUTPUT CONNECTION INFORMATION')).toBeLessThan(
|
||||
result!.indexOf('# Split In Batches Node')
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle splitInBatches in various node type formats', async () => {
|
||||
const testCases = [
|
||||
'splitInBatches',
|
||||
'n8n-nodes-base.splitInBatches',
|
||||
'nodes-base.splitInBatches'
|
||||
];
|
||||
|
||||
for (const nodeType of testCases) {
|
||||
const originalContent = '# Split In Batches\nOriginal content';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation(nodeType);
|
||||
|
||||
expect(result).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(result).toContain('Output 0 (index 0) = "done"');
|
||||
}
|
||||
});
|
||||
|
||||
it('should provide specific guidance for correct connection patterns', async () => {
|
||||
const originalContent = '# Split In Batches\n## When to use\nContent';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).toContain('Connect nodes that PROCESS items inside the loop to **Output 1 ("loop")**');
|
||||
expect(result).toContain('Connect nodes that run AFTER the loop completes to **Output 0 ("done")**');
|
||||
expect(result).toContain('The last processing node in the loop must connect back to the SplitInBatches node');
|
||||
});
|
||||
|
||||
it('should explain the common AI assistant mistake', async () => {
|
||||
const originalContent = '# Split In Batches\n## When to use\nContent';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).toContain('AI assistants often connect these backwards');
|
||||
expect(result).toContain('logical flow (loop first, then done) doesn\'t match the technical indices (done=0, loop=1)');
|
||||
});
|
||||
|
||||
it('should not enhance non-splitInBatches nodes with loop guidance', async () => {
|
||||
const originalContent = '# HTTP Request Node\nContent';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('httpRequest');
|
||||
|
||||
expect(result).not.toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(result).not.toContain('counterintuitive');
|
||||
expect(result).toBe(originalContent); // Should be unchanged
|
||||
});
|
||||
});
|
||||
|
||||
describe('enhanceLoopNodeDocumentation - IF node', () => {
|
||||
it('should enhance IF node documentation with output guidance', async () => {
|
||||
const originalContent = `# IF Node
|
||||
|
||||
Route items based on conditions.
|
||||
|
||||
## Node parameters
|
||||
|
||||
Configure your conditions here.
|
||||
`;
|
||||
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('n8n-nodes-base.if');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('Output Connection Information');
|
||||
expect(result!).toContain('Output 0 (index 0) = "true"');
|
||||
expect(result!).toContain('Output 1 (index 1) = "false"');
|
||||
expect(result!).toContain('Items that match the condition');
|
||||
expect(result!).toContain('Items that do not match the condition');
|
||||
|
||||
// Should insert before "Node parameters" section
|
||||
const parametersIndex = result!.indexOf('## Node parameters');
|
||||
const outputInfoIndex = result!.indexOf('Output Connection Information');
|
||||
expect(outputInfoIndex).toBeLessThan(parametersIndex);
|
||||
expect(outputInfoIndex).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
it('should handle IF node when no "Node parameters" section exists', async () => {
|
||||
const originalContent = `# IF Node
|
||||
|
||||
Route items based on conditions.
|
||||
|
||||
## Usage
|
||||
|
||||
Use this node to route data.
|
||||
`;
|
||||
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('n8n-nodes-base.if');
|
||||
|
||||
// When no "Node parameters" section exists, no enhancement is applied
|
||||
expect(result).toBe(originalContent);
|
||||
});
|
||||
|
||||
it('should handle various IF node type formats', async () => {
|
||||
const testCases = [
|
||||
'if',
|
||||
'n8n-nodes-base.if',
|
||||
'nodes-base.if'
|
||||
];
|
||||
|
||||
for (const nodeType of testCases) {
|
||||
const originalContent = '# IF Node\n## Node parameters\nContent';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation(nodeType);
|
||||
|
||||
if (nodeType.includes('.if')) {
|
||||
expect(result).toContain('Output Connection Information');
|
||||
expect(result).toContain('Output 0 (index 0) = "true"');
|
||||
expect(result).toContain('Output 1 (index 1) = "false"');
|
||||
} else {
|
||||
// For 'if' without dot, no enhancement is applied
|
||||
expect(result).toBe(originalContent);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
describe('enhanceLoopNodeDocumentation - edge cases', () => {
|
||||
it('should handle content without clear insertion points', async () => {
|
||||
const originalContent = 'Simple content without markdown sections';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
// Should be prepended when no insertion point found (but there's a newline before original content)
|
||||
const guidanceIndex = result!.indexOf('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(guidanceIndex).toBeLessThan(result!.indexOf('Simple content'));
|
||||
expect(guidanceIndex).toBeLessThanOrEqual(5); // Allow for some whitespace
|
||||
});
|
||||
|
||||
it('should handle empty content', async () => {
|
||||
const originalContent = '';
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(result!.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
it('should handle content with multiple "When to use" sections', async () => {
|
||||
const originalContent = `# Split In Batches
|
||||
|
||||
## When to use (overview)
|
||||
|
||||
General usage.
|
||||
|
||||
## When to use (detailed)
|
||||
|
||||
Detailed usage.
|
||||
`;
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(originalContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
// Should insert before first occurrence
|
||||
const firstWhenToUse = result!.indexOf('## When to use (overview)');
|
||||
const guidanceIndex = result!.indexOf('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(guidanceIndex).toBeLessThan(firstWhenToUse);
|
||||
});
|
||||
|
||||
it('should not double-enhance already enhanced content', async () => {
|
||||
const alreadyEnhancedContent = `# Split In Batches
|
||||
|
||||
## CRITICAL OUTPUT CONNECTION INFORMATION
|
||||
|
||||
Already enhanced.
|
||||
|
||||
## When to use
|
||||
|
||||
Content here.
|
||||
`;
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(alreadyEnhancedContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
// Should still add enhancement (method doesn't check for existing enhancements)
|
||||
expect(result).not.toBeNull();
|
||||
const criticalSections = (result!.match(/CRITICAL OUTPUT CONNECTION INFORMATION/g) || []).length;
|
||||
expect(criticalSections).toBe(2); // Original + new enhancement
|
||||
});
|
||||
|
||||
it('should handle very large content efficiently', async () => {
|
||||
const largeContent = 'a'.repeat(100000) + '\n## When to use\n' + 'b'.repeat(100000);
|
||||
vi.mocked(fs.readFile).mockResolvedValueOnce(largeContent);
|
||||
|
||||
const result = await docsMapper.fetchDocumentation('splitInBatches');
|
||||
|
||||
expect(result).not.toBeNull();
|
||||
expect(result!).toContain('CRITICAL OUTPUT CONNECTION INFORMATION');
|
||||
expect(result!.length).toBeGreaterThan(largeContent.length);
|
||||
});
|
||||
});
|
||||
|
||||
describe('DocsMapper instance', () => {
|
||||
it('should use consistent docsPath across instances', () => {
|
||||
const mapper1 = new DocsMapper();
|
||||
|
||||
473
tests/unit/parsers/node-parser-outputs.test.ts
Normal file
473
tests/unit/parsers/node-parser-outputs.test.ts
Normal file
@@ -0,0 +1,473 @@
|
||||
import { describe, it, expect, vi, beforeEach } from 'vitest';
|
||||
import { NodeParser } from '@/parsers/node-parser';
|
||||
import { PropertyExtractor } from '@/parsers/property-extractor';
|
||||
|
||||
// Mock PropertyExtractor
|
||||
vi.mock('@/parsers/property-extractor');
|
||||
|
||||
describe('NodeParser - Output Extraction', () => {
|
||||
let parser: NodeParser;
|
||||
let mockPropertyExtractor: any;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
mockPropertyExtractor = {
|
||||
extractProperties: vi.fn().mockReturnValue([]),
|
||||
extractCredentials: vi.fn().mockReturnValue([]),
|
||||
detectAIToolCapability: vi.fn().mockReturnValue(false),
|
||||
extractOperations: vi.fn().mockReturnValue([])
|
||||
};
|
||||
|
||||
(PropertyExtractor as any).mockImplementation(() => mockPropertyExtractor);
|
||||
|
||||
parser = new NodeParser();
|
||||
});
|
||||
|
||||
describe('extractOutputs method', () => {
|
||||
it('should extract outputs array from base description', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'Done', description: 'Final results when loop completes' },
|
||||
{ displayName: 'Loop', description: 'Current batch data during iteration' }
|
||||
];
|
||||
|
||||
const nodeDescription = {
|
||||
name: 'splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
outputs
|
||||
};
|
||||
|
||||
const NodeClass = class {
|
||||
description = nodeDescription;
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(outputs);
|
||||
expect(result.outputNames).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should extract outputNames array from base description', () => {
|
||||
const outputNames = ['done', 'loop'];
|
||||
|
||||
const nodeDescription = {
|
||||
name: 'splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
outputNames
|
||||
};
|
||||
|
||||
const NodeClass = class {
|
||||
description = nodeDescription;
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputNames).toEqual(outputNames);
|
||||
expect(result.outputs).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should extract both outputs and outputNames when both are present', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'Done', description: 'Final results when loop completes' },
|
||||
{ displayName: 'Loop', description: 'Current batch data during iteration' }
|
||||
];
|
||||
const outputNames = ['done', 'loop'];
|
||||
|
||||
const nodeDescription = {
|
||||
name: 'splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
outputs,
|
||||
outputNames
|
||||
};
|
||||
|
||||
const NodeClass = class {
|
||||
description = nodeDescription;
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(outputs);
|
||||
expect(result.outputNames).toEqual(outputNames);
|
||||
});
|
||||
|
||||
it('should convert single output to array format', () => {
|
||||
const singleOutput = { displayName: 'Output', description: 'Single output' };
|
||||
|
||||
const nodeDescription = {
|
||||
name: 'singleOutputNode',
|
||||
displayName: 'Single Output Node',
|
||||
outputs: singleOutput
|
||||
};
|
||||
|
||||
const NodeClass = class {
|
||||
description = nodeDescription;
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual([singleOutput]);
|
||||
});
|
||||
|
||||
it('should convert single outputName to array format', () => {
|
||||
const nodeDescription = {
|
||||
name: 'singleOutputNode',
|
||||
displayName: 'Single Output Node',
|
||||
outputNames: 'main'
|
||||
};
|
||||
|
||||
const NodeClass = class {
|
||||
description = nodeDescription;
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputNames).toEqual(['main']);
|
||||
});
|
||||
|
||||
it('should extract outputs from versioned node when not in base description', () => {
|
||||
const versionedOutputs = [
|
||||
{ displayName: 'True', description: 'Items that match condition' },
|
||||
{ displayName: 'False', description: 'Items that do not match condition' }
|
||||
];
|
||||
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'if',
|
||||
displayName: 'IF'
|
||||
// No outputs in base description
|
||||
};
|
||||
|
||||
nodeVersions = {
|
||||
1: {
|
||||
description: {
|
||||
outputs: versionedOutputs
|
||||
}
|
||||
},
|
||||
2: {
|
||||
description: {
|
||||
outputs: versionedOutputs,
|
||||
outputNames: ['true', 'false']
|
||||
}
|
||||
}
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
// Should get outputs from latest version (2)
|
||||
expect(result.outputs).toEqual(versionedOutputs);
|
||||
expect(result.outputNames).toEqual(['true', 'false']);
|
||||
});
|
||||
|
||||
it('should handle node instantiation failure gracefully', () => {
|
||||
const NodeClass = class {
|
||||
// Static description that can be accessed when instantiation fails
|
||||
static description = {
|
||||
name: 'problematic',
|
||||
displayName: 'Problematic Node'
|
||||
};
|
||||
|
||||
constructor() {
|
||||
throw new Error('Cannot instantiate');
|
||||
}
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toBeUndefined();
|
||||
expect(result.outputNames).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should return empty result when no outputs found anywhere', () => {
|
||||
const nodeDescription = {
|
||||
name: 'noOutputs',
|
||||
displayName: 'No Outputs Node'
|
||||
// No outputs or outputNames
|
||||
};
|
||||
|
||||
const NodeClass = class {
|
||||
description = nodeDescription;
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toBeUndefined();
|
||||
expect(result.outputNames).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should handle complex versioned node structure', () => {
|
||||
const NodeClass = class VersionedNodeType {
|
||||
baseDescription = {
|
||||
name: 'complexVersioned',
|
||||
displayName: 'Complex Versioned Node',
|
||||
defaultVersion: 3
|
||||
};
|
||||
|
||||
nodeVersions = {
|
||||
1: {
|
||||
description: {
|
||||
outputs: [{ displayName: 'V1 Output' }]
|
||||
}
|
||||
},
|
||||
2: {
|
||||
description: {
|
||||
outputs: [
|
||||
{ displayName: 'V2 Output 1' },
|
||||
{ displayName: 'V2 Output 2' }
|
||||
]
|
||||
}
|
||||
},
|
||||
3: {
|
||||
description: {
|
||||
outputs: [
|
||||
{ displayName: 'V3 True', description: 'True branch' },
|
||||
{ displayName: 'V3 False', description: 'False branch' }
|
||||
],
|
||||
outputNames: ['true', 'false']
|
||||
}
|
||||
}
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
// Should use latest version (3)
|
||||
expect(result.outputs).toEqual([
|
||||
{ displayName: 'V3 True', description: 'True branch' },
|
||||
{ displayName: 'V3 False', description: 'False branch' }
|
||||
]);
|
||||
expect(result.outputNames).toEqual(['true', 'false']);
|
||||
});
|
||||
|
||||
it('should prefer base description outputs over versioned when both exist', () => {
|
||||
const baseOutputs = [{ displayName: 'Base Output' }];
|
||||
const versionedOutputs = [{ displayName: 'Versioned Output' }];
|
||||
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'preferBase',
|
||||
displayName: 'Prefer Base',
|
||||
outputs: baseOutputs
|
||||
};
|
||||
|
||||
nodeVersions = {
|
||||
1: {
|
||||
description: {
|
||||
outputs: versionedOutputs
|
||||
}
|
||||
}
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(baseOutputs);
|
||||
});
|
||||
|
||||
it('should handle IF node with typical output structure', () => {
|
||||
const ifOutputs = [
|
||||
{ displayName: 'True', description: 'Items that match the condition' },
|
||||
{ displayName: 'False', description: 'Items that do not match the condition' }
|
||||
];
|
||||
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'if',
|
||||
displayName: 'IF',
|
||||
outputs: ifOutputs,
|
||||
outputNames: ['true', 'false']
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(ifOutputs);
|
||||
expect(result.outputNames).toEqual(['true', 'false']);
|
||||
});
|
||||
|
||||
it('should handle SplitInBatches node with counterintuitive output structure', () => {
|
||||
const splitInBatchesOutputs = [
|
||||
{ displayName: 'Done', description: 'Final results when loop completes' },
|
||||
{ displayName: 'Loop', description: 'Current batch data during iteration' }
|
||||
];
|
||||
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
outputs: splitInBatchesOutputs,
|
||||
outputNames: ['done', 'loop']
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(splitInBatchesOutputs);
|
||||
expect(result.outputNames).toEqual(['done', 'loop']);
|
||||
|
||||
// Verify the counterintuitive order: done=0, loop=1
|
||||
expect(result.outputs).toBeDefined();
|
||||
expect(result.outputNames).toBeDefined();
|
||||
expect(result.outputs![0].displayName).toBe('Done');
|
||||
expect(result.outputs![1].displayName).toBe('Loop');
|
||||
expect(result.outputNames![0]).toBe('done');
|
||||
expect(result.outputNames![1]).toBe('loop');
|
||||
});
|
||||
|
||||
it('should handle Switch node with multiple outputs', () => {
|
||||
const switchOutputs = [
|
||||
{ displayName: 'Output 1', description: 'First branch' },
|
||||
{ displayName: 'Output 2', description: 'Second branch' },
|
||||
{ displayName: 'Output 3', description: 'Third branch' },
|
||||
{ displayName: 'Fallback', description: 'Default branch when no conditions match' }
|
||||
];
|
||||
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'switch',
|
||||
displayName: 'Switch',
|
||||
outputs: switchOutputs,
|
||||
outputNames: ['0', '1', '2', 'fallback']
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(switchOutputs);
|
||||
expect(result.outputNames).toEqual(['0', '1', '2', 'fallback']);
|
||||
});
|
||||
|
||||
it('should handle empty outputs array', () => {
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'emptyOutputs',
|
||||
displayName: 'Empty Outputs',
|
||||
outputs: [],
|
||||
outputNames: []
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual([]);
|
||||
expect(result.outputNames).toEqual([]);
|
||||
});
|
||||
|
||||
it('should handle mismatched outputs and outputNames arrays', () => {
|
||||
const outputs = [
|
||||
{ displayName: 'Output 1' },
|
||||
{ displayName: 'Output 2' }
|
||||
];
|
||||
const outputNames = ['first', 'second', 'third']; // One extra
|
||||
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'mismatched',
|
||||
displayName: 'Mismatched Arrays',
|
||||
outputs,
|
||||
outputNames
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toEqual(outputs);
|
||||
expect(result.outputNames).toEqual(outputNames);
|
||||
});
|
||||
});
|
||||
|
||||
describe('real-world node structures', () => {
|
||||
it('should handle actual n8n SplitInBatches node structure', () => {
|
||||
// This mimics the actual structure from n8n-nodes-base
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'splitInBatches',
|
||||
displayName: 'Split In Batches',
|
||||
description: 'Split data into batches and iterate over each batch',
|
||||
icon: 'fa:th-large',
|
||||
group: ['transform'],
|
||||
version: 3,
|
||||
outputs: [
|
||||
{
|
||||
displayName: 'Done',
|
||||
name: 'done',
|
||||
type: 'main',
|
||||
hint: 'Receives the final data after all batches have been processed'
|
||||
},
|
||||
{
|
||||
displayName: 'Loop',
|
||||
name: 'loop',
|
||||
type: 'main',
|
||||
hint: 'Receives the current batch data during each iteration'
|
||||
}
|
||||
],
|
||||
outputNames: ['done', 'loop']
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toHaveLength(2);
|
||||
expect(result.outputs).toBeDefined();
|
||||
expect(result.outputs![0].displayName).toBe('Done');
|
||||
expect(result.outputs![1].displayName).toBe('Loop');
|
||||
expect(result.outputNames).toEqual(['done', 'loop']);
|
||||
});
|
||||
|
||||
it('should handle actual n8n IF node structure', () => {
|
||||
// This mimics the actual structure from n8n-nodes-base
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'if',
|
||||
displayName: 'IF',
|
||||
description: 'Route items to different outputs based on conditions',
|
||||
icon: 'fa:map-signs',
|
||||
group: ['transform'],
|
||||
version: 2,
|
||||
outputs: [
|
||||
{
|
||||
displayName: 'True',
|
||||
name: 'true',
|
||||
type: 'main',
|
||||
hint: 'Items that match the condition'
|
||||
},
|
||||
{
|
||||
displayName: 'False',
|
||||
name: 'false',
|
||||
type: 'main',
|
||||
hint: 'Items that do not match the condition'
|
||||
}
|
||||
],
|
||||
outputNames: ['true', 'false']
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toHaveLength(2);
|
||||
expect(result.outputs).toBeDefined();
|
||||
expect(result.outputs![0].displayName).toBe('True');
|
||||
expect(result.outputs![1].displayName).toBe('False');
|
||||
expect(result.outputNames).toEqual(['true', 'false']);
|
||||
});
|
||||
|
||||
it('should handle single-output nodes like HTTP Request', () => {
|
||||
const NodeClass = class {
|
||||
description = {
|
||||
name: 'httpRequest',
|
||||
displayName: 'HTTP Request',
|
||||
description: 'Make HTTP requests',
|
||||
icon: 'fa:at',
|
||||
group: ['input'],
|
||||
version: 4
|
||||
// No outputs specified - single main output implied
|
||||
};
|
||||
};
|
||||
|
||||
const result = parser.parse(NodeClass, 'n8n-nodes-base');
|
||||
|
||||
expect(result.outputs).toBeUndefined();
|
||||
expect(result.outputNames).toBeUndefined();
|
||||
});
|
||||
});
|
||||
});
|
||||
865
tests/unit/services/loop-output-edge-cases.test.ts
Normal file
865
tests/unit/services/loop-output-edge-cases.test.ts
Normal file
@@ -0,0 +1,865 @@
|
||||
import { describe, it, expect, beforeEach, vi } from 'vitest';
|
||||
import { WorkflowValidator } from '@/services/workflow-validator';
|
||||
import { NodeRepository } from '@/database/node-repository';
|
||||
import { EnhancedConfigValidator } from '@/services/enhanced-config-validator';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('@/database/node-repository');
|
||||
vi.mock('@/services/enhanced-config-validator');
|
||||
|
||||
describe('Loop Output Fix - Edge Cases', () => {
|
||||
let validator: WorkflowValidator;
|
||||
let mockNodeRepository: any;
|
||||
let mockNodeValidator: any;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
mockNodeRepository = {
|
||||
getNode: vi.fn((nodeType: string) => {
|
||||
// Default return
|
||||
if (nodeType === 'nodes-base.splitInBatches') {
|
||||
return {
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
outputs: [
|
||||
{ displayName: 'Done', name: 'done' },
|
||||
{ displayName: 'Loop', name: 'loop' }
|
||||
],
|
||||
outputNames: ['done', 'loop'],
|
||||
properties: []
|
||||
};
|
||||
}
|
||||
return {
|
||||
nodeType,
|
||||
properties: []
|
||||
};
|
||||
})
|
||||
};
|
||||
|
||||
mockNodeValidator = {
|
||||
validateWithMode: vi.fn().mockReturnValue({
|
||||
errors: [],
|
||||
warnings: []
|
||||
})
|
||||
};
|
||||
|
||||
validator = new WorkflowValidator(mockNodeRepository, mockNodeValidator);
|
||||
});
|
||||
|
||||
describe('Nodes without outputs', () => {
|
||||
it('should handle nodes with null outputs gracefully', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.httpRequest',
|
||||
outputs: null,
|
||||
outputNames: null,
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'No Outputs Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'HTTP Request',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
position: [100, 100],
|
||||
parameters: { url: 'https://example.com' }
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Set',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'HTTP Request': {
|
||||
main: [
|
||||
[{ node: 'Set', type: 'main', index: 0 }]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not crash or produce output-related errors
|
||||
expect(result).toBeDefined();
|
||||
const outputErrors = result.errors.filter(e =>
|
||||
e.message?.includes('output') && !e.message?.includes('Connection')
|
||||
);
|
||||
expect(outputErrors).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should handle nodes with undefined outputs gracefully', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.webhook',
|
||||
// outputs and outputNames are undefined
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Undefined Outputs Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Webhook',
|
||||
type: 'n8n-nodes-base.webhook',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
expect(result).toBeDefined();
|
||||
expect(result.valid).toBeTruthy(); // Empty workflow with webhook should be valid
|
||||
});
|
||||
|
||||
it('should handle nodes with empty outputs array', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.customNode',
|
||||
outputs: [],
|
||||
outputNames: [],
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Empty Outputs Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Custom Node',
|
||||
type: 'n8n-nodes-base.customNode',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Custom Node': {
|
||||
main: [
|
||||
[{ node: 'Custom Node', type: 'main', index: 0 }] // Self-reference
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should warn about self-reference but not crash
|
||||
const selfRefWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('self-referencing')
|
||||
);
|
||||
expect(selfRefWarnings).toHaveLength(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Invalid connection indices', () => {
|
||||
it('should handle negative connection indices', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Negative Index Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Set',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Set', type: 'main', index: -1 }] // Invalid negative index
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
const negativeIndexErrors = result.errors.filter(e =>
|
||||
e.message?.includes('Invalid connection index -1')
|
||||
);
|
||||
expect(negativeIndexErrors).toHaveLength(1);
|
||||
expect(negativeIndexErrors[0].message).toContain('must be non-negative');
|
||||
});
|
||||
|
||||
it('should handle very large connection indices', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.switch',
|
||||
outputs: [
|
||||
{ displayName: 'Output 1' },
|
||||
{ displayName: 'Output 2' }
|
||||
],
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Large Index Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Switch',
|
||||
type: 'n8n-nodes-base.switch',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Set',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Switch': {
|
||||
main: [
|
||||
[{ node: 'Set', type: 'main', index: 999 }] // Very large index
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should validate without crashing (n8n allows large indices)
|
||||
expect(result).toBeDefined();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Malformed connection structures', () => {
|
||||
it('should handle null connection objects', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Null Connections Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
null, // Null output
|
||||
[{ node: 'NonExistent', type: 'main', index: 0 }]
|
||||
] as any
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should handle gracefully without crashing
|
||||
expect(result).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle missing connection properties', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Malformed Connections Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Set',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[
|
||||
{ node: 'Set' } as any, // Missing type and index
|
||||
{ type: 'main', index: 0 } as any, // Missing node
|
||||
{} as any // Empty object
|
||||
]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should handle malformed connections but report errors
|
||||
expect(result).toBeDefined();
|
||||
expect(result.errors.length).toBeGreaterThan(0);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Deep loop back detection limits', () => {
|
||||
it('should respect maxDepth limit in checkForLoopBack', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
// Create a very deep chain that exceeds maxDepth (50)
|
||||
const nodes = [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
];
|
||||
|
||||
const connections: any = {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[], // Done output
|
||||
[{ node: 'Node1', type: 'main', index: 0 }] // Loop output
|
||||
]
|
||||
}
|
||||
};
|
||||
|
||||
// Create chain of 60 nodes (exceeds maxDepth of 50)
|
||||
for (let i = 1; i <= 60; i++) {
|
||||
nodes.push({
|
||||
id: (i + 1).toString(),
|
||||
name: `Node${i}`,
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [100 + i * 50, 100],
|
||||
parameters: {}
|
||||
});
|
||||
|
||||
if (i < 60) {
|
||||
connections[`Node${i}`] = {
|
||||
main: [[{ node: `Node${i + 1}`, type: 'main', index: 0 }]]
|
||||
};
|
||||
} else {
|
||||
// Last node connects back to Split In Batches
|
||||
connections[`Node${i}`] = {
|
||||
main: [[{ node: 'Split In Batches', type: 'main', index: 0 }]]
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
const workflow = {
|
||||
name: 'Deep Chain Workflow',
|
||||
nodes,
|
||||
connections
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should warn about missing loop back because depth limit prevents detection
|
||||
const loopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
expect(loopBackWarnings).toHaveLength(1);
|
||||
});
|
||||
|
||||
it('should handle circular references without infinite loops', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Circular Reference Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'NodeA',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'NodeB',
|
||||
type: 'n8n-nodes-base.function',
|
||||
position: [500, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'NodeA', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'NodeA': {
|
||||
main: [
|
||||
[{ node: 'NodeB', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'NodeB': {
|
||||
main: [
|
||||
[{ node: 'NodeA', type: 'main', index: 0 }] // Circular: B -> A -> B -> A ...
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should complete without hanging and warn about missing loop back
|
||||
expect(result).toBeDefined();
|
||||
const loopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
expect(loopBackWarnings).toHaveLength(1);
|
||||
});
|
||||
|
||||
it('should handle self-referencing nodes in loop back detection', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Self Reference Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'SelfRef',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'SelfRef', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'SelfRef': {
|
||||
main: [
|
||||
[{ node: 'SelfRef', type: 'main', index: 0 }] // Self-reference instead of loop back
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should warn about missing loop back and self-reference
|
||||
const loopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
const selfRefWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('self-referencing')
|
||||
);
|
||||
|
||||
expect(loopBackWarnings).toHaveLength(1);
|
||||
expect(selfRefWarnings).toHaveLength(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Complex output structures', () => {
|
||||
it('should handle nodes with many outputs', async () => {
|
||||
const manyOutputs = Array.from({ length: 20 }, (_, i) => ({
|
||||
displayName: `Output ${i + 1}`,
|
||||
name: `output${i + 1}`,
|
||||
description: `Output number ${i + 1}`
|
||||
}));
|
||||
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.complexSwitch',
|
||||
outputs: manyOutputs,
|
||||
outputNames: manyOutputs.map(o => o.name),
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Many Outputs Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Complex Switch',
|
||||
type: 'n8n-nodes-base.complexSwitch',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Set',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Complex Switch': {
|
||||
main: Array.from({ length: 20 }, () => [
|
||||
{ node: 'Set', type: 'main', index: 0 }
|
||||
])
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should handle without performance issues
|
||||
expect(result).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle mixed output types (main, error, ai_tool)', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.complexNode',
|
||||
outputs: [
|
||||
{ displayName: 'Main', type: 'main' },
|
||||
{ displayName: 'Error', type: 'error' }
|
||||
],
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Mixed Output Types Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Complex Node',
|
||||
type: 'n8n-nodes-base.complexNode',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Main Handler',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 50],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'Error Handler',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 150],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '4',
|
||||
name: 'Tool',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
position: [500, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Complex Node': {
|
||||
main: [
|
||||
[{ node: 'Main Handler', type: 'main', index: 0 }]
|
||||
],
|
||||
error: [
|
||||
[{ node: 'Error Handler', type: 'main', index: 0 }]
|
||||
],
|
||||
ai_tool: [
|
||||
[{ node: 'Tool', type: 'main', index: 0 }]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should validate all connection types
|
||||
expect(result).toBeDefined();
|
||||
expect(result.statistics.validConnections).toBe(3);
|
||||
});
|
||||
});
|
||||
|
||||
describe('SplitInBatches specific edge cases', () => {
|
||||
it('should handle SplitInBatches with no connections', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Isolated SplitInBatches',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not produce SplitInBatches-specific warnings for isolated node
|
||||
const splitWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('SplitInBatches') ||
|
||||
w.message?.includes('loop') ||
|
||||
w.message?.includes('done')
|
||||
);
|
||||
expect(splitWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should handle SplitInBatches with only one output connected', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Single Output SplitInBatches',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Final Action',
|
||||
type: 'n8n-nodes-base.emailSend',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Final Action', type: 'main', index: 0 }], // Only done output connected
|
||||
[] // Loop output empty
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should NOT warn about empty loop output (it's only a problem if loop connects to something but doesn't loop back)
|
||||
// An empty loop output is valid - it just means no looping occurs
|
||||
const loopWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('loop') && w.message?.includes('connect back')
|
||||
);
|
||||
expect(loopWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should handle SplitInBatches with both outputs to same node', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Same Target SplitInBatches',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Multi Purpose',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Multi Purpose', type: 'main', index: 0 }], // Done -> Multi Purpose
|
||||
[{ node: 'Multi Purpose', type: 'main', index: 0 }] // Loop -> Multi Purpose
|
||||
]
|
||||
},
|
||||
'Multi Purpose': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Loop back
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Both outputs go to same node which loops back - should be valid
|
||||
// No warnings about loop back since it does connect back
|
||||
const loopWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('loop') && w.message?.includes('connect back')
|
||||
);
|
||||
expect(loopWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should detect reversed outputs with processing node on done output', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
const workflow = {
|
||||
name: 'Reversed SplitInBatches with Function Node',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Process Function',
|
||||
type: 'n8n-nodes-base.function',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Process Function', type: 'main', index: 0 }], // Done -> Function (this is wrong)
|
||||
[] // Loop output empty
|
||||
]
|
||||
},
|
||||
'Process Function': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Function connects back (indicates it should be on loop)
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should error about reversed outputs since function node on done output connects back
|
||||
const reversedErrors = result.errors.filter(e =>
|
||||
e.message?.includes('SplitInBatches outputs appear reversed')
|
||||
);
|
||||
expect(reversedErrors).toHaveLength(1);
|
||||
});
|
||||
|
||||
it('should handle non-existent node type gracefully', async () => {
|
||||
// Node doesn't exist in repository
|
||||
mockNodeRepository.getNode.mockReturnValue(null);
|
||||
|
||||
const workflow = {
|
||||
name: 'Unknown Node Type',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Unknown Node',
|
||||
type: 'n8n-nodes-base.unknownNode',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should report unknown node type error
|
||||
const unknownNodeErrors = result.errors.filter(e =>
|
||||
e.message?.includes('Unknown node type')
|
||||
);
|
||||
expect(unknownNodeErrors).toHaveLength(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Performance edge cases', () => {
|
||||
it('should handle very large workflows efficiently', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.set',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Create workflow with 1000 nodes
|
||||
const nodes = Array.from({ length: 1000 }, (_, i) => ({
|
||||
id: `node${i}`,
|
||||
name: `Node ${i}`,
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [100 + (i % 50) * 50, 100 + Math.floor(i / 50) * 50],
|
||||
parameters: {}
|
||||
}));
|
||||
|
||||
// Create simple linear connections
|
||||
const connections: any = {};
|
||||
for (let i = 0; i < 999; i++) {
|
||||
connections[`Node ${i}`] = {
|
||||
main: [[{ node: `Node ${i + 1}`, type: 'main', index: 0 }]]
|
||||
};
|
||||
}
|
||||
|
||||
const workflow = {
|
||||
name: 'Large Workflow',
|
||||
nodes,
|
||||
connections
|
||||
};
|
||||
|
||||
const startTime = Date.now();
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
const duration = Date.now() - startTime;
|
||||
|
||||
// Should complete within reasonable time (< 5 seconds)
|
||||
expect(duration).toBeLessThan(5000);
|
||||
expect(result).toBeDefined();
|
||||
expect(result.statistics.totalNodes).toBe(1000);
|
||||
});
|
||||
|
||||
it('should handle workflows with many SplitInBatches nodes', async () => {
|
||||
// Use default mock that includes outputs for SplitInBatches
|
||||
|
||||
// Create 100 SplitInBatches nodes
|
||||
const nodes = Array.from({ length: 100 }, (_, i) => ({
|
||||
id: `split${i}`,
|
||||
name: `Split ${i}`,
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100 + (i % 10) * 100, 100 + Math.floor(i / 10) * 100],
|
||||
parameters: {}
|
||||
}));
|
||||
|
||||
const connections: any = {};
|
||||
// Each split connects to the next one
|
||||
for (let i = 0; i < 99; i++) {
|
||||
connections[`Split ${i}`] = {
|
||||
main: [
|
||||
[{ node: `Split ${i + 1}`, type: 'main', index: 0 }], // Done -> next split
|
||||
[] // Empty loop
|
||||
]
|
||||
};
|
||||
}
|
||||
|
||||
const workflow = {
|
||||
name: 'Many SplitInBatches Workflow',
|
||||
nodes,
|
||||
connections
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should validate all nodes without performance issues
|
||||
expect(result).toBeDefined();
|
||||
expect(result.statistics.totalNodes).toBe(100);
|
||||
});
|
||||
});
|
||||
});
|
||||
434
tests/unit/services/workflow-validator-loops-simple.test.ts
Normal file
434
tests/unit/services/workflow-validator-loops-simple.test.ts
Normal file
@@ -0,0 +1,434 @@
|
||||
import { describe, it, expect, beforeEach, vi } from 'vitest';
|
||||
import { WorkflowValidator } from '@/services/workflow-validator';
|
||||
import { NodeRepository } from '@/database/node-repository';
|
||||
import { EnhancedConfigValidator } from '@/services/enhanced-config-validator';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('@/database/node-repository');
|
||||
vi.mock('@/services/enhanced-config-validator');
|
||||
|
||||
describe('WorkflowValidator - SplitInBatches Validation (Simplified)', () => {
|
||||
let validator: WorkflowValidator;
|
||||
let mockNodeRepository: any;
|
||||
let mockNodeValidator: any;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
mockNodeRepository = {
|
||||
getNode: vi.fn()
|
||||
};
|
||||
|
||||
mockNodeValidator = {
|
||||
validateWithMode: vi.fn().mockReturnValue({
|
||||
errors: [],
|
||||
warnings: []
|
||||
})
|
||||
};
|
||||
|
||||
validator = new WorkflowValidator(mockNodeRepository, mockNodeValidator);
|
||||
});
|
||||
|
||||
describe('SplitInBatches node detection', () => {
|
||||
it('should identify SplitInBatches nodes in workflow', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'SplitInBatches Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: { batchSize: 10 }
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Process Item',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[], // Done output (0)
|
||||
[{ node: 'Process Item', type: 'main', index: 0 }] // Loop output (1)
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should complete validation without crashing
|
||||
expect(result).toBeDefined();
|
||||
expect(result.valid).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle SplitInBatches with processing node name patterns', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const processingNames = [
|
||||
'Process Item',
|
||||
'Transform Data',
|
||||
'Handle Each',
|
||||
'Function Node',
|
||||
'Code Block'
|
||||
];
|
||||
|
||||
for (const nodeName of processingNames) {
|
||||
const workflow = {
|
||||
name: 'Processing Pattern Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: nodeName,
|
||||
type: 'n8n-nodes-base.function',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: nodeName, type: 'main', index: 0 }], // Processing node on Done output
|
||||
[]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should identify potential processing nodes
|
||||
expect(result).toBeDefined();
|
||||
}
|
||||
});
|
||||
|
||||
it('should handle final processing node patterns', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const finalNames = [
|
||||
'Final Summary',
|
||||
'Send Email',
|
||||
'Complete Notification',
|
||||
'Final Report'
|
||||
];
|
||||
|
||||
for (const nodeName of finalNames) {
|
||||
const workflow = {
|
||||
name: 'Final Pattern Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: nodeName,
|
||||
type: 'n8n-nodes-base.emailSend',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: nodeName, type: 'main', index: 0 }], // Final node on Done output (correct)
|
||||
[]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not warn about final nodes on done output
|
||||
expect(result).toBeDefined();
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
describe('Connection validation', () => {
|
||||
it('should validate connection indices', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Connection Index Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Target',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Target', type: 'main', index: -1 }] // Invalid negative index
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
const negativeIndexErrors = result.errors.filter(e =>
|
||||
e.message?.includes('Invalid connection index -1')
|
||||
);
|
||||
expect(negativeIndexErrors.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
it('should handle non-existent target nodes', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Missing Target Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'NonExistentNode', type: 'main', index: 0 }]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
const missingNodeErrors = result.errors.filter(e =>
|
||||
e.message?.includes('non-existent node')
|
||||
);
|
||||
expect(missingNodeErrors.length).toBeGreaterThan(0);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Self-referencing connections', () => {
|
||||
it('should allow self-referencing for SplitInBatches nodes', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Self Reference Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Self-reference on loop output
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not warn about self-reference for SplitInBatches
|
||||
const selfRefWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('self-referencing')
|
||||
);
|
||||
expect(selfRefWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should warn about self-referencing for non-loop nodes', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.set',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Non-Loop Self Reference Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Set',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Set': {
|
||||
main: [
|
||||
[{ node: 'Set', type: 'main', index: 0 }] // Self-reference on regular node
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should warn about self-reference for non-loop nodes
|
||||
const selfRefWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('self-referencing')
|
||||
);
|
||||
expect(selfRefWarnings.length).toBeGreaterThan(0);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Output connection validation', () => {
|
||||
it('should validate output connections for nodes with outputs', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.if',
|
||||
outputs: [
|
||||
{ displayName: 'True', description: 'Items that match condition' },
|
||||
{ displayName: 'False', description: 'Items that do not match condition' }
|
||||
],
|
||||
outputNames: ['true', 'false'],
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'IF Node Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'IF',
|
||||
type: 'n8n-nodes-base.if',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'True Handler',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 50],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'False Handler',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 150],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'IF': {
|
||||
main: [
|
||||
[{ node: 'True Handler', type: 'main', index: 0 }], // True output (0)
|
||||
[{ node: 'False Handler', type: 'main', index: 0 }] // False output (1)
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should validate without major errors
|
||||
expect(result).toBeDefined();
|
||||
expect(result.statistics.validConnections).toBe(2);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Error handling', () => {
|
||||
it('should handle nodes without outputs gracefully', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.httpRequest',
|
||||
outputs: null,
|
||||
outputNames: null,
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'No Outputs Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'HTTP Request',
|
||||
type: 'n8n-nodes-base.httpRequest',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should handle gracefully without crashing
|
||||
expect(result).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle unknown node types gracefully', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue(null);
|
||||
|
||||
const workflow = {
|
||||
name: 'Unknown Node Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Unknown',
|
||||
type: 'n8n-nodes-base.unknown',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should report unknown node error
|
||||
const unknownErrors = result.errors.filter(e =>
|
||||
e.message?.includes('Unknown node type')
|
||||
);
|
||||
expect(unknownErrors.length).toBeGreaterThan(0);
|
||||
});
|
||||
});
|
||||
});
|
||||
705
tests/unit/services/workflow-validator-loops.test.ts
Normal file
705
tests/unit/services/workflow-validator-loops.test.ts
Normal file
@@ -0,0 +1,705 @@
|
||||
import { describe, it, expect, beforeEach, vi } from 'vitest';
|
||||
import { WorkflowValidator } from '@/services/workflow-validator';
|
||||
import { NodeRepository } from '@/database/node-repository';
|
||||
import { EnhancedConfigValidator } from '@/services/enhanced-config-validator';
|
||||
|
||||
// Mock dependencies
|
||||
vi.mock('@/database/node-repository');
|
||||
vi.mock('@/services/enhanced-config-validator');
|
||||
|
||||
describe('WorkflowValidator - Loop Node Validation', () => {
|
||||
let validator: WorkflowValidator;
|
||||
let mockNodeRepository: any;
|
||||
let mockNodeValidator: any;
|
||||
|
||||
beforeEach(() => {
|
||||
vi.clearAllMocks();
|
||||
|
||||
mockNodeRepository = {
|
||||
getNode: vi.fn()
|
||||
};
|
||||
|
||||
mockNodeValidator = {
|
||||
validateWithMode: vi.fn().mockReturnValue({
|
||||
errors: [],
|
||||
warnings: []
|
||||
})
|
||||
};
|
||||
|
||||
validator = new WorkflowValidator(mockNodeRepository, mockNodeValidator);
|
||||
});
|
||||
|
||||
describe('validateSplitInBatchesConnection', () => {
|
||||
const createWorkflow = (connections: any) => ({
|
||||
name: 'Test Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: { batchSize: 10 }
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Process Item',
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'Final Summary',
|
||||
type: 'n8n-nodes-base.emailSend',
|
||||
position: [500, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections
|
||||
});
|
||||
|
||||
it('should detect reversed SplitInBatches connections (processing node on done output)', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Create a processing node with a name that matches the pattern (includes "process")
|
||||
const workflow = {
|
||||
name: 'Test Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: { batchSize: 10 }
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Process Function', // Name matches processing pattern
|
||||
type: 'n8n-nodes-base.function', // Type also matches processing pattern
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Process Function', type: 'main', index: 0 }], // Done output (wrong for processing)
|
||||
[] // No loop connections
|
||||
]
|
||||
},
|
||||
'Process Function': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Loop back - confirms it's processing
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// The validator should detect the processing node name/type pattern and loop back
|
||||
const reversedErrors = result.errors.filter(e =>
|
||||
e.message?.includes('SplitInBatches outputs appear reversed')
|
||||
);
|
||||
|
||||
expect(reversedErrors.length).toBeGreaterThanOrEqual(1);
|
||||
});
|
||||
|
||||
it('should warn about processing node on done output without loop back', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Processing node connected to "done" output but no loop back
|
||||
const workflow = createWorkflow({
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Process Item', type: 'main', index: 0 }], // Done output
|
||||
[]
|
||||
]
|
||||
}
|
||||
// No loop back from Process Item
|
||||
});
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
expect(result.warnings).toContainEqual(
|
||||
expect.objectContaining({
|
||||
type: 'warning',
|
||||
nodeId: '1',
|
||||
nodeName: 'Split In Batches',
|
||||
message: expect.stringContaining('connected to the "done" output (index 0) but appears to be a processing node')
|
||||
})
|
||||
);
|
||||
});
|
||||
|
||||
it('should warn about final processing node on loop output', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Final summary node connected to "loop" output (index 1) - suspicious
|
||||
const workflow = createWorkflow({
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'Final Summary', type: 'main', index: 0 }] // Loop output for final node
|
||||
]
|
||||
}
|
||||
});
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
expect(result.warnings).toContainEqual(
|
||||
expect.objectContaining({
|
||||
type: 'warning',
|
||||
nodeId: '1',
|
||||
nodeName: 'Split In Batches',
|
||||
message: expect.stringContaining('connected to the "loop" output (index 1) but appears to be a post-processing node')
|
||||
})
|
||||
);
|
||||
});
|
||||
|
||||
it('should warn about loop output without loop back connection', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Processing node on loop output but doesn't connect back
|
||||
const workflow = createWorkflow({
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'Process Item', type: 'main', index: 0 }] // Loop output
|
||||
]
|
||||
}
|
||||
// Process Item doesn't connect back to Split In Batches
|
||||
});
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
expect(result.warnings).toContainEqual(
|
||||
expect.objectContaining({
|
||||
type: 'warning',
|
||||
nodeId: '1',
|
||||
nodeName: 'Split In Batches',
|
||||
message: expect.stringContaining('doesn\'t connect back to the SplitInBatches node')
|
||||
})
|
||||
);
|
||||
});
|
||||
|
||||
it('should accept correct SplitInBatches connections', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Create a workflow with neutral node names that don't trigger patterns
|
||||
const workflow = {
|
||||
name: 'Test Workflow',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: { batchSize: 10 }
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Data Node', // Neutral name, won't trigger processing pattern
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'Output Node', // Neutral name, won't trigger post-processing pattern
|
||||
type: 'n8n-nodes-base.noOp',
|
||||
position: [500, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Output Node', type: 'main', index: 0 }], // Done output -> neutral node
|
||||
[{ node: 'Data Node', type: 'main', index: 0 }] // Loop output -> neutral node
|
||||
]
|
||||
},
|
||||
'Data Node': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Loop back
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not have SplitInBatches-specific errors or warnings
|
||||
const splitErrors = result.errors.filter(e =>
|
||||
e.message?.includes('SplitInBatches') ||
|
||||
e.message?.includes('loop') ||
|
||||
e.message?.includes('done')
|
||||
);
|
||||
const splitWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('SplitInBatches') ||
|
||||
w.message?.includes('loop') ||
|
||||
w.message?.includes('done')
|
||||
);
|
||||
|
||||
expect(splitErrors).toHaveLength(0);
|
||||
expect(splitWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should handle complex loop structures', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const complexWorkflow = {
|
||||
name: 'Complex Loop',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: 'Step A', // Neutral name
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 50],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '3',
|
||||
name: 'Step B', // Neutral name
|
||||
type: 'n8n-nodes-base.noOp',
|
||||
position: [500, 50],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '4',
|
||||
name: 'Final Step', // More neutral name
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [300, 150],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: 'Final Step', type: 'main', index: 0 }], // Done -> Final (correct)
|
||||
[{ node: 'Step A', type: 'main', index: 0 }] // Loop -> Processing (correct)
|
||||
]
|
||||
},
|
||||
'Step A': {
|
||||
main: [
|
||||
[{ node: 'Step B', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'Step B': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Loop back (correct)
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(complexWorkflow as any);
|
||||
|
||||
// Should accept this correct structure without warnings
|
||||
const loopWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('loop') || w.message?.includes('done')
|
||||
);
|
||||
expect(loopWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should detect node type patterns for processing detection', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const testCases = [
|
||||
{ type: 'n8n-nodes-base.function', name: 'Process Data', shouldWarn: true },
|
||||
{ type: 'n8n-nodes-base.code', name: 'Transform Item', shouldWarn: true },
|
||||
{ type: 'n8n-nodes-base.set', name: 'Handle Each', shouldWarn: true },
|
||||
{ type: 'n8n-nodes-base.emailSend', name: 'Final Email', shouldWarn: false },
|
||||
{ type: 'n8n-nodes-base.slack', name: 'Complete Notification', shouldWarn: false }
|
||||
];
|
||||
|
||||
for (const testCase of testCases) {
|
||||
const workflow = {
|
||||
name: 'Pattern Test',
|
||||
nodes: [
|
||||
{
|
||||
id: '1',
|
||||
name: 'Split In Batches',
|
||||
type: 'n8n-nodes-base.splitInBatches',
|
||||
position: [100, 100],
|
||||
parameters: {}
|
||||
},
|
||||
{
|
||||
id: '2',
|
||||
name: testCase.name,
|
||||
type: testCase.type,
|
||||
position: [300, 100],
|
||||
parameters: {}
|
||||
}
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[{ node: testCase.name, type: 'main', index: 0 }], // Connected to done (index 0)
|
||||
[]
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
const hasProcessingWarning = result.warnings.some(w =>
|
||||
w.message?.includes('appears to be a processing node')
|
||||
);
|
||||
|
||||
if (testCase.shouldWarn) {
|
||||
expect(hasProcessingWarning).toBe(true);
|
||||
} else {
|
||||
expect(hasProcessingWarning).toBe(false);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
describe('checkForLoopBack method', () => {
|
||||
it('should detect direct loop back connection', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Direct Loop Back',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} },
|
||||
{ id: '2', name: 'Process', type: 'n8n-nodes-base.set', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [[], [{ node: 'Process', type: 'main', index: 0 }]]
|
||||
},
|
||||
'Process': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Direct loop back
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not warn about missing loop back since it exists
|
||||
const missingLoopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
expect(missingLoopBackWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should detect indirect loop back connection through multiple nodes', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Indirect Loop Back',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} },
|
||||
{ id: '2', name: 'Step1', type: 'n8n-nodes-base.set', position: [0, 0], parameters: {} },
|
||||
{ id: '3', name: 'Step2', type: 'n8n-nodes-base.function', position: [0, 0], parameters: {} },
|
||||
{ id: '4', name: 'Step3', type: 'n8n-nodes-base.code', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [[], [{ node: 'Step1', type: 'main', index: 0 }]]
|
||||
},
|
||||
'Step1': {
|
||||
main: [
|
||||
[{ node: 'Step2', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'Step2': {
|
||||
main: [
|
||||
[{ node: 'Step3', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'Step3': {
|
||||
main: [
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Indirect loop back
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not warn about missing loop back since indirect loop exists
|
||||
const missingLoopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
expect(missingLoopBackWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should respect max depth to prevent infinite recursion', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
// Create a very deep chain that would exceed depth limit
|
||||
const nodes = [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} }
|
||||
];
|
||||
const connections: any = {
|
||||
'Split In Batches': {
|
||||
main: [[], [{ node: 'Node1', type: 'main', index: 0 }]]
|
||||
}
|
||||
};
|
||||
|
||||
// Create a chain of 60 nodes (exceeds default maxDepth of 50)
|
||||
for (let i = 1; i <= 60; i++) {
|
||||
nodes.push({
|
||||
id: (i + 1).toString(),
|
||||
name: `Node${i}`,
|
||||
type: 'n8n-nodes-base.set',
|
||||
position: [0, 0],
|
||||
parameters: {}
|
||||
});
|
||||
|
||||
if (i < 60) {
|
||||
connections[`Node${i}`] = {
|
||||
main: [[{ node: `Node${i + 1}`, type: 'main', index: 0 }]]
|
||||
};
|
||||
} else {
|
||||
// Last node connects back to Split In Batches
|
||||
connections[`Node${i}`] = {
|
||||
main: [[{ node: 'Split In Batches', type: 'main', index: 0 }]]
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
const workflow = {
|
||||
name: 'Deep Chain',
|
||||
nodes,
|
||||
connections
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should warn about missing loop back because depth limit prevents detection
|
||||
const missingLoopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
expect(missingLoopBackWarnings).toHaveLength(1);
|
||||
});
|
||||
|
||||
it('should handle circular references without infinite loops', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Circular Reference',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} },
|
||||
{ id: '2', name: 'NodeA', type: 'n8n-nodes-base.set', position: [0, 0], parameters: {} },
|
||||
{ id: '3', name: 'NodeB', type: 'n8n-nodes-base.function', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [[], [{ node: 'NodeA', type: 'main', index: 0 }]]
|
||||
},
|
||||
'NodeA': {
|
||||
main: [
|
||||
[{ node: 'NodeB', type: 'main', index: 0 }]
|
||||
]
|
||||
},
|
||||
'NodeB': {
|
||||
main: [
|
||||
[{ node: 'NodeA', type: 'main', index: 0 }] // Circular reference (doesn't connect back to Split)
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should complete without hanging and warn about missing loop back
|
||||
const missingLoopBackWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('doesn\'t connect back')
|
||||
);
|
||||
expect(missingLoopBackWarnings).toHaveLength(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('self-referencing connections', () => {
|
||||
it('should allow self-referencing for SplitInBatches (loop back)', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Self Reference Loop',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'Split In Batches', type: 'main', index: 0 }] // Self-reference on loop output
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not warn about self-reference for SplitInBatches
|
||||
const selfReferenceWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('self-referencing')
|
||||
);
|
||||
expect(selfReferenceWarnings).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should warn about self-referencing for non-loop nodes', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.set',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Non-Loop Self Reference',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Set', type: 'n8n-nodes-base.set', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Set': {
|
||||
main: [
|
||||
[{ node: 'Set', type: 'main', index: 0 }] // Self-reference on regular node
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should warn about self-reference for non-loop nodes
|
||||
const selfReferenceWarnings = result.warnings.filter(w =>
|
||||
w.message?.includes('self-referencing')
|
||||
);
|
||||
expect(selfReferenceWarnings).toHaveLength(1);
|
||||
});
|
||||
});
|
||||
|
||||
describe('edge cases', () => {
|
||||
it('should handle missing target node gracefully', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Missing Target',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[],
|
||||
[{ node: 'NonExistentNode', type: 'main', index: 0 }] // Target doesn't exist
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should have connection error for non-existent node
|
||||
const connectionErrors = result.errors.filter(e =>
|
||||
e.message?.includes('non-existent node')
|
||||
);
|
||||
expect(connectionErrors).toHaveLength(1);
|
||||
});
|
||||
|
||||
it('should handle empty connections gracefully', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Empty Connections',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
[], // Empty done output
|
||||
[] // Empty loop output
|
||||
]
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should not crash and should not have SplitInBatches-specific errors
|
||||
expect(result).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle null/undefined connection arrays', async () => {
|
||||
mockNodeRepository.getNode.mockReturnValue({
|
||||
nodeType: 'nodes-base.splitInBatches',
|
||||
properties: []
|
||||
});
|
||||
|
||||
const workflow = {
|
||||
name: 'Null Connections',
|
||||
nodes: [
|
||||
{ id: '1', name: 'Split In Batches', type: 'n8n-nodes-base.splitInBatches', position: [0, 0], parameters: {} }
|
||||
],
|
||||
connections: {
|
||||
'Split In Batches': {
|
||||
main: [
|
||||
null, // Null done output
|
||||
undefined // Undefined loop output
|
||||
] as any
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const result = await validator.validateWorkflow(workflow as any);
|
||||
|
||||
// Should handle gracefully without crashing
|
||||
expect(result).toBeDefined();
|
||||
});
|
||||
});
|
||||
});
|
||||
Reference in New Issue
Block a user