mirror of
https://github.com/n8n-io/n8n.git
synced 2025-01-07 02:47:32 -08:00
61696c3db3
Co-authored-by: Michael Kret <michael.k@radency.com>
1967 lines
62 KiB
TypeScript
1967 lines
62 KiB
TypeScript
/* eslint-disable @typescript-eslint/prefer-optional-chain */
|
|
/* eslint-disable @typescript-eslint/no-unsafe-member-access */
|
|
/* eslint-disable @typescript-eslint/no-unsafe-assignment */
|
|
/* eslint-disable @typescript-eslint/prefer-nullish-coalescing */
|
|
import * as assert from 'assert/strict';
|
|
import { setMaxListeners } from 'events';
|
|
import get from 'lodash/get';
|
|
import type {
|
|
ExecutionBaseError,
|
|
ExecutionStatus,
|
|
GenericValue,
|
|
IConnection,
|
|
IDataObject,
|
|
IExecuteData,
|
|
INode,
|
|
INodeConnections,
|
|
INodeExecutionData,
|
|
IPairedItemData,
|
|
IPinData,
|
|
IRun,
|
|
IRunData,
|
|
ISourceData,
|
|
ITaskData,
|
|
ITaskDataConnections,
|
|
ITaskDataConnectionsSource,
|
|
ITaskMetadata,
|
|
IWaitingForExecution,
|
|
IWaitingForExecutionSource,
|
|
NodeApiError,
|
|
NodeOperationError,
|
|
Workflow,
|
|
IRunExecutionData,
|
|
IWorkflowExecuteAdditionalData,
|
|
WorkflowExecuteMode,
|
|
CloseFunction,
|
|
StartNodeData,
|
|
NodeExecutionHint,
|
|
} from 'n8n-workflow';
|
|
import {
|
|
LoggerProxy as Logger,
|
|
WorkflowOperationError,
|
|
NodeHelpers,
|
|
NodeConnectionType,
|
|
ApplicationError,
|
|
NodeExecutionOutput,
|
|
sleep,
|
|
ErrorReporterProxy,
|
|
ExecutionCancelledError,
|
|
} from 'n8n-workflow';
|
|
import PCancelable from 'p-cancelable';
|
|
|
|
import * as NodeExecuteFunctions from './NodeExecuteFunctions';
|
|
import {
|
|
DirectedGraph,
|
|
findStartNodes,
|
|
findSubgraph,
|
|
findTriggerForPartialExecution,
|
|
cleanRunData,
|
|
recreateNodeExecutionStack,
|
|
handleCycles,
|
|
} from './PartialExecutionUtils';
|
|
|
|
export class WorkflowExecute {
|
|
private status: ExecutionStatus = 'new';
|
|
|
|
private readonly abortController = new AbortController();
|
|
|
|
constructor(
|
|
private readonly additionalData: IWorkflowExecuteAdditionalData,
|
|
private readonly mode: WorkflowExecuteMode,
|
|
private runExecutionData: IRunExecutionData = {
|
|
startData: {},
|
|
resultData: {
|
|
runData: {},
|
|
pinData: {},
|
|
},
|
|
executionData: {
|
|
contextData: {},
|
|
nodeExecutionStack: [],
|
|
metadata: {},
|
|
waitingExecution: {},
|
|
waitingExecutionSource: {},
|
|
},
|
|
},
|
|
) {}
|
|
|
|
/**
|
|
* Executes the given workflow.
|
|
*
|
|
* @param {Workflow} workflow The workflow to execute
|
|
* @param {INode[]} [startNode] Node to start execution from
|
|
* @param {string} [destinationNode] Node to stop execution at
|
|
*/
|
|
// IMPORTANT: Do not add "async" to this function, it will then convert the
|
|
// PCancelable to a regular Promise and does so not allow canceling
|
|
// active executions anymore
|
|
// eslint-disable-next-line @typescript-eslint/promise-function-async
|
|
run(
|
|
workflow: Workflow,
|
|
startNode?: INode,
|
|
destinationNode?: string,
|
|
pinData?: IPinData,
|
|
): PCancelable<IRun> {
|
|
this.status = 'running';
|
|
|
|
// Get the nodes to start workflow execution from
|
|
startNode = startNode || workflow.getStartNode(destinationNode);
|
|
|
|
if (startNode === undefined) {
|
|
throw new ApplicationError('No node to start the workflow from could be found');
|
|
}
|
|
|
|
// If a destination node is given we only run the direct parent nodes and no others
|
|
let runNodeFilter: string[] | undefined;
|
|
if (destinationNode) {
|
|
runNodeFilter = workflow.getParentNodes(destinationNode);
|
|
runNodeFilter.push(destinationNode);
|
|
}
|
|
|
|
// Initialize the data of the start nodes
|
|
const nodeExecutionStack: IExecuteData[] = [
|
|
{
|
|
node: startNode,
|
|
data: {
|
|
main: [
|
|
[
|
|
{
|
|
json: {},
|
|
},
|
|
],
|
|
],
|
|
},
|
|
source: null,
|
|
},
|
|
];
|
|
|
|
this.runExecutionData = {
|
|
startData: {
|
|
destinationNode,
|
|
runNodeFilter,
|
|
},
|
|
resultData: {
|
|
runData: {},
|
|
pinData,
|
|
},
|
|
executionData: {
|
|
contextData: {},
|
|
nodeExecutionStack,
|
|
metadata: {},
|
|
waitingExecution: {},
|
|
waitingExecutionSource: {},
|
|
},
|
|
};
|
|
|
|
return this.processRunExecutionData(workflow);
|
|
}
|
|
|
|
forceInputNodeExecution(workflow: Workflow): boolean {
|
|
return workflow.settings.executionOrder !== 'v1';
|
|
}
|
|
|
|
/**
|
|
* Executes the given workflow but only
|
|
*
|
|
* @param {Workflow} workflow The workflow to execute
|
|
* @param {string[]} startNodes Nodes to start execution from
|
|
* @param {string} destinationNode Node to stop execution at
|
|
*/
|
|
// IMPORTANT: Do not add "async" to this function, it will then convert the
|
|
// PCancelable to a regular Promise and does so not allow canceling
|
|
// active executions anymore
|
|
// eslint-disable-next-line @typescript-eslint/promise-function-async, complexity
|
|
runPartialWorkflow(
|
|
workflow: Workflow,
|
|
runData: IRunData,
|
|
startNodes: StartNodeData[],
|
|
destinationNode?: string,
|
|
pinData?: IPinData,
|
|
): PCancelable<IRun> {
|
|
let incomingNodeConnections: INodeConnections | undefined;
|
|
let connection: IConnection;
|
|
|
|
this.status = 'running';
|
|
|
|
const runIndex = 0;
|
|
let runNodeFilter: string[] | undefined;
|
|
|
|
// Initialize the nodeExecutionStack and waitingExecution with
|
|
// the data from runData
|
|
const nodeExecutionStack: IExecuteData[] = [];
|
|
const waitingExecution: IWaitingForExecution = {};
|
|
const waitingExecutionSource: IWaitingForExecutionSource = {};
|
|
for (const startNode of startNodes) {
|
|
incomingNodeConnections = workflow.connectionsByDestinationNode[startNode.name];
|
|
|
|
const incomingData: INodeExecutionData[][] = [];
|
|
let incomingSourceData: ITaskDataConnectionsSource | null = null;
|
|
|
|
if (incomingNodeConnections === undefined) {
|
|
incomingData.push([
|
|
{
|
|
json: {},
|
|
},
|
|
]);
|
|
} else {
|
|
// Get the data of the incoming connections
|
|
incomingSourceData = { main: [] };
|
|
for (const connections of incomingNodeConnections.main) {
|
|
for (let inputIndex = 0; inputIndex < connections.length; inputIndex++) {
|
|
connection = connections[inputIndex];
|
|
|
|
const node = workflow.getNode(connection.node);
|
|
|
|
if (node?.disabled) continue;
|
|
|
|
if (node && pinData && pinData[node.name]) {
|
|
incomingData.push(pinData[node.name]);
|
|
} else {
|
|
if (!runData[connection.node]) {
|
|
continue;
|
|
}
|
|
const nodeIncomingData =
|
|
runData[connection.node]?.[runIndex]?.data?.[connection.type]?.[connection.index];
|
|
if (nodeIncomingData) {
|
|
incomingData.push(nodeIncomingData);
|
|
}
|
|
}
|
|
|
|
incomingSourceData.main.push(startNode.sourceData ?? { previousNode: connection.node });
|
|
}
|
|
}
|
|
}
|
|
|
|
const executeData: IExecuteData = {
|
|
node: workflow.getNode(startNode.name) as INode,
|
|
data: {
|
|
main: incomingData,
|
|
},
|
|
source: incomingSourceData,
|
|
};
|
|
|
|
nodeExecutionStack.push(executeData);
|
|
|
|
if (destinationNode) {
|
|
// Check if the destinationNode has to be added as waiting
|
|
// because some input data is already fully available
|
|
incomingNodeConnections = workflow.connectionsByDestinationNode[destinationNode];
|
|
if (incomingNodeConnections !== undefined) {
|
|
for (const connections of incomingNodeConnections.main) {
|
|
for (let inputIndex = 0; inputIndex < connections.length; inputIndex++) {
|
|
connection = connections[inputIndex];
|
|
|
|
if (waitingExecution[destinationNode] === undefined) {
|
|
waitingExecution[destinationNode] = {};
|
|
waitingExecutionSource[destinationNode] = {};
|
|
}
|
|
if (waitingExecution[destinationNode][runIndex] === undefined) {
|
|
waitingExecution[destinationNode][runIndex] = {};
|
|
waitingExecutionSource[destinationNode][runIndex] = {};
|
|
}
|
|
if (waitingExecution[destinationNode][runIndex][connection.type] === undefined) {
|
|
waitingExecution[destinationNode][runIndex][connection.type] = [];
|
|
waitingExecutionSource[destinationNode][runIndex][connection.type] = [];
|
|
}
|
|
|
|
if (runData[connection.node] !== undefined) {
|
|
// Input data exists so add as waiting
|
|
// incomingDataDestination.push(runData[connection.node!][runIndex].data![connection.type][connection.index]);
|
|
waitingExecution[destinationNode][runIndex][connection.type].push(
|
|
runData[connection.node][runIndex].data![connection.type][connection.index],
|
|
);
|
|
waitingExecutionSource[destinationNode][runIndex][connection.type].push({
|
|
previousNode: connection.node,
|
|
previousNodeOutput: connection.index || undefined,
|
|
previousNodeRun: runIndex || undefined,
|
|
} as ISourceData);
|
|
} else {
|
|
waitingExecution[destinationNode][runIndex][connection.type].push(null);
|
|
waitingExecutionSource[destinationNode][runIndex][connection.type].push(null);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Only run the parent nodes and no others
|
|
runNodeFilter = workflow
|
|
.getParentNodes(destinationNode)
|
|
.filter((parentNodeName) => !workflow.getNode(parentNodeName)?.disabled);
|
|
|
|
runNodeFilter.push(destinationNode);
|
|
}
|
|
}
|
|
|
|
this.runExecutionData = {
|
|
startData: {
|
|
destinationNode,
|
|
runNodeFilter,
|
|
},
|
|
resultData: {
|
|
runData,
|
|
pinData,
|
|
},
|
|
executionData: {
|
|
contextData: {},
|
|
nodeExecutionStack,
|
|
metadata: {},
|
|
waitingExecution,
|
|
waitingExecutionSource,
|
|
},
|
|
};
|
|
|
|
return this.processRunExecutionData(workflow);
|
|
}
|
|
|
|
// IMPORTANT: Do not add "async" to this function, it will then convert the
|
|
// PCancelable to a regular Promise and does so not allow canceling
|
|
// active executions anymore
|
|
// eslint-disable-next-line @typescript-eslint/promise-function-async
|
|
runPartialWorkflow2(
|
|
workflow: Workflow,
|
|
runData: IRunData,
|
|
destinationNodeName?: string,
|
|
pinData?: IPinData,
|
|
): PCancelable<IRun> {
|
|
// TODO: Refactor the call-site to make `destinationNodeName` a required
|
|
// after removing the old partial execution flow.
|
|
assert.ok(
|
|
destinationNodeName,
|
|
'a destinationNodeName is required for the new partial execution flow',
|
|
);
|
|
|
|
const destination = workflow.getNode(destinationNodeName);
|
|
assert.ok(
|
|
destination,
|
|
`Could not find a node with the name ${destinationNodeName} in the workflow.`,
|
|
);
|
|
|
|
// 1. Find the Trigger
|
|
const trigger = findTriggerForPartialExecution(workflow, destinationNodeName);
|
|
if (trigger === undefined) {
|
|
throw new ApplicationError(
|
|
'The destination node is not connected to any trigger. Partial executions need a trigger.',
|
|
);
|
|
}
|
|
|
|
// 2. Find the Subgraph
|
|
const graph = DirectedGraph.fromWorkflow(workflow);
|
|
const subgraph = findSubgraph({ graph, destination, trigger });
|
|
const filteredNodes = subgraph.getNodes();
|
|
|
|
// 3. Find the Start Nodes
|
|
let startNodes = findStartNodes({ graph: subgraph, trigger, destination, runData });
|
|
|
|
// 4. Detect Cycles
|
|
// 5. Handle Cycles
|
|
startNodes = handleCycles(graph, startNodes, trigger);
|
|
|
|
// 6. Clean Run Data
|
|
const newRunData: IRunData = cleanRunData(runData, graph, startNodes);
|
|
|
|
// 7. Recreate Execution Stack
|
|
const { nodeExecutionStack, waitingExecution, waitingExecutionSource } =
|
|
recreateNodeExecutionStack(subgraph, new Set(startNodes), runData, pinData ?? {});
|
|
|
|
// 8. Execute
|
|
this.status = 'running';
|
|
this.runExecutionData = {
|
|
startData: {
|
|
destinationNode: destinationNodeName,
|
|
runNodeFilter: Array.from(filteredNodes.values()).map((node) => node.name),
|
|
},
|
|
resultData: {
|
|
runData: newRunData,
|
|
pinData,
|
|
},
|
|
executionData: {
|
|
contextData: {},
|
|
nodeExecutionStack,
|
|
metadata: {},
|
|
waitingExecution,
|
|
waitingExecutionSource,
|
|
},
|
|
};
|
|
|
|
return this.processRunExecutionData(subgraph.toWorkflow({ ...workflow }));
|
|
}
|
|
|
|
/**
|
|
* Executes the hook with the given name
|
|
*
|
|
*/
|
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
async executeHook(hookName: string, parameters: any[]): Promise<void> {
|
|
if (this.additionalData.hooks === undefined) {
|
|
return;
|
|
}
|
|
|
|
return await this.additionalData.hooks.executeHookFunctions(hookName, parameters);
|
|
}
|
|
|
|
moveNodeMetadata(): void {
|
|
const metadata = get(this.runExecutionData, 'executionData.metadata');
|
|
|
|
if (metadata) {
|
|
const runData = get(this.runExecutionData, 'resultData.runData');
|
|
|
|
let index: number;
|
|
let metaRunData: ITaskMetadata;
|
|
for (const nodeName of Object.keys(metadata)) {
|
|
for ([index, metaRunData] of metadata[nodeName].entries()) {
|
|
runData[nodeName][index].metadata = {
|
|
...(runData[nodeName][index].metadata ?? {}),
|
|
...metaRunData,
|
|
};
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Checks the incoming connection does not receive any data
|
|
*/
|
|
incomingConnectionIsEmpty(
|
|
runData: IRunData,
|
|
inputConnections: IConnection[],
|
|
runIndex: number,
|
|
): boolean {
|
|
// for (const inputConnection of workflow.connectionsByDestinationNode[nodeToAdd].main[0]) {
|
|
for (const inputConnection of inputConnections) {
|
|
const nodeIncomingData = get(runData, [
|
|
inputConnection.node,
|
|
runIndex,
|
|
'data',
|
|
'main',
|
|
inputConnection.index,
|
|
]);
|
|
if (nodeIncomingData !== undefined && (nodeIncomingData as object[]).length !== 0) {
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
prepareWaitingToExecution(nodeName: string, numberOfConnections: number, runIndex: number) {
|
|
if (!this.runExecutionData.executionData!.waitingExecutionSource) {
|
|
this.runExecutionData.executionData!.waitingExecutionSource = {};
|
|
}
|
|
|
|
this.runExecutionData.executionData!.waitingExecution[nodeName][runIndex] = {
|
|
main: [],
|
|
};
|
|
this.runExecutionData.executionData!.waitingExecutionSource[nodeName][runIndex] = {
|
|
main: [],
|
|
};
|
|
|
|
for (let i = 0; i < numberOfConnections; i++) {
|
|
this.runExecutionData.executionData!.waitingExecution[nodeName][runIndex].main.push(null);
|
|
|
|
this.runExecutionData.executionData!.waitingExecutionSource[nodeName][runIndex].main.push(
|
|
null,
|
|
);
|
|
}
|
|
}
|
|
|
|
// eslint-disable-next-line complexity
|
|
addNodeToBeExecuted(
|
|
workflow: Workflow,
|
|
connectionData: IConnection,
|
|
outputIndex: number,
|
|
parentNodeName: string,
|
|
nodeSuccessData: INodeExecutionData[][],
|
|
runIndex: number,
|
|
): void {
|
|
let stillDataMissing = false;
|
|
const enqueueFn = workflow.settings.executionOrder === 'v1' ? 'unshift' : 'push';
|
|
let waitingNodeIndex: number | undefined;
|
|
|
|
// Check if node has multiple inputs as then we have to wait for all input data
|
|
// to be present before we can add it to the node-execution-stack
|
|
if (workflow.connectionsByDestinationNode[connectionData.node].main.length > 1) {
|
|
// Node has multiple inputs
|
|
let nodeWasWaiting = true;
|
|
|
|
if (!this.runExecutionData.executionData!.waitingExecutionSource) {
|
|
this.runExecutionData.executionData!.waitingExecutionSource = {};
|
|
}
|
|
|
|
// Check if there is already data for the node
|
|
if (
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node] === undefined
|
|
) {
|
|
// Node does not have data yet so create a new empty one
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node] = {};
|
|
this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node] = {};
|
|
nodeWasWaiting = false;
|
|
}
|
|
|
|
// Figure out if the node is already waiting with partial data to which to add the
|
|
// data to or if a new entry has to get created
|
|
let createNewWaitingEntry = true;
|
|
|
|
if (
|
|
Object.keys(this.runExecutionData.executionData!.waitingExecution[connectionData.node])
|
|
.length > 0
|
|
) {
|
|
// Check if there is already data for the input on all of the waiting nodes
|
|
for (const index of Object.keys(
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node],
|
|
)) {
|
|
if (
|
|
!this.runExecutionData.executionData!.waitingExecution[connectionData.node][
|
|
parseInt(index)
|
|
].main[connectionData.index]
|
|
) {
|
|
// Data for the input is missing so we can add it to the existing entry
|
|
createNewWaitingEntry = false;
|
|
waitingNodeIndex = parseInt(index);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (waitingNodeIndex === undefined) {
|
|
waitingNodeIndex = Object.values(
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node],
|
|
).length;
|
|
}
|
|
|
|
if (createNewWaitingEntry) {
|
|
// There is currently no node waiting that does not already have data for
|
|
// the given input, so create a new entry
|
|
|
|
this.prepareWaitingToExecution(
|
|
connectionData.node,
|
|
workflow.connectionsByDestinationNode[connectionData.node].main.length,
|
|
waitingNodeIndex,
|
|
);
|
|
}
|
|
|
|
// Add the new data
|
|
if (nodeSuccessData === null) {
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node][
|
|
waitingNodeIndex
|
|
].main[connectionData.index] = null;
|
|
this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node][
|
|
waitingNodeIndex
|
|
].main[connectionData.index] = null;
|
|
} else {
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node][
|
|
waitingNodeIndex
|
|
].main[connectionData.index] = nodeSuccessData[outputIndex];
|
|
|
|
this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node][
|
|
waitingNodeIndex
|
|
].main[connectionData.index] = {
|
|
previousNode: parentNodeName,
|
|
previousNodeOutput: outputIndex || undefined,
|
|
previousNodeRun: runIndex || undefined,
|
|
};
|
|
}
|
|
|
|
// Check if all data exists now
|
|
let thisExecutionData: INodeExecutionData[] | null;
|
|
let allDataFound = true;
|
|
for (
|
|
let i = 0;
|
|
i <
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node][waitingNodeIndex]
|
|
.main.length;
|
|
i++
|
|
) {
|
|
thisExecutionData =
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node][
|
|
waitingNodeIndex
|
|
].main[i];
|
|
if (thisExecutionData === null) {
|
|
allDataFound = false;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (allDataFound) {
|
|
// All data exists for node to be executed
|
|
// So add it to the execution stack
|
|
|
|
const executionStackItem = {
|
|
node: workflow.nodes[connectionData.node],
|
|
data: this.runExecutionData.executionData!.waitingExecution[connectionData.node][
|
|
waitingNodeIndex
|
|
],
|
|
source:
|
|
this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node][
|
|
waitingNodeIndex
|
|
],
|
|
} as IExecuteData;
|
|
|
|
if (
|
|
this.runExecutionData.executionData!.waitingExecutionSource !== null &&
|
|
this.runExecutionData.executionData!.waitingExecutionSource !== undefined
|
|
) {
|
|
executionStackItem.source =
|
|
this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node][
|
|
waitingNodeIndex
|
|
];
|
|
}
|
|
|
|
this.runExecutionData.executionData!.nodeExecutionStack[enqueueFn](executionStackItem);
|
|
|
|
// Remove the data from waiting
|
|
delete this.runExecutionData.executionData!.waitingExecution[connectionData.node][
|
|
waitingNodeIndex
|
|
];
|
|
delete this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node][
|
|
waitingNodeIndex
|
|
];
|
|
|
|
if (
|
|
Object.keys(this.runExecutionData.executionData!.waitingExecution[connectionData.node])
|
|
.length === 0
|
|
) {
|
|
// No more data left for the node so also delete that one
|
|
delete this.runExecutionData.executionData!.waitingExecution[connectionData.node];
|
|
delete this.runExecutionData.executionData!.waitingExecutionSource[connectionData.node];
|
|
}
|
|
return;
|
|
}
|
|
stillDataMissing = true;
|
|
|
|
if (!nodeWasWaiting) {
|
|
// Get a list of all the output nodes that we can check for siblings easier
|
|
const checkOutputNodes = [];
|
|
// eslint-disable-next-line @typescript-eslint/no-for-in-array
|
|
for (const outputIndexParent in workflow.connectionsBySourceNode[parentNodeName].main) {
|
|
if (
|
|
!workflow.connectionsBySourceNode[parentNodeName].main.hasOwnProperty(outputIndexParent)
|
|
) {
|
|
continue;
|
|
}
|
|
for (const connectionDataCheck of workflow.connectionsBySourceNode[parentNodeName].main[
|
|
outputIndexParent
|
|
]) {
|
|
checkOutputNodes.push(connectionDataCheck.node);
|
|
}
|
|
}
|
|
|
|
// Node was not on "waitingExecution" so it is the first time it gets
|
|
// checked. So we have to go through all the inputs and check if they
|
|
// are already on the list to be processed.
|
|
// If that is not the case add it.
|
|
|
|
const forceInputNodeExecution = this.forceInputNodeExecution(workflow);
|
|
|
|
for (
|
|
let inputIndex = 0;
|
|
inputIndex < workflow.connectionsByDestinationNode[connectionData.node].main.length;
|
|
inputIndex++
|
|
) {
|
|
for (const inputData of workflow.connectionsByDestinationNode[connectionData.node].main[
|
|
inputIndex
|
|
]) {
|
|
if (inputData.node === parentNodeName) {
|
|
// Is the node we come from so its data will be available for sure
|
|
continue;
|
|
}
|
|
|
|
const executionStackNodes = this.runExecutionData.executionData!.nodeExecutionStack.map(
|
|
(stackData) => stackData.node.name,
|
|
);
|
|
|
|
// Check if that node is also an output connection of the
|
|
// previously processed one
|
|
if (inputData.node !== parentNodeName && checkOutputNodes.includes(inputData.node)) {
|
|
// So the parent node will be added anyway which
|
|
// will then process this node next. So nothing to do
|
|
// unless the incoming data of the node is empty
|
|
// because then it would not be executed
|
|
if (
|
|
!this.incomingConnectionIsEmpty(
|
|
this.runExecutionData.resultData.runData,
|
|
workflow.connectionsByDestinationNode[inputData.node].main[0],
|
|
runIndex,
|
|
)
|
|
) {
|
|
continue;
|
|
}
|
|
}
|
|
|
|
// Check if it is already in the execution stack
|
|
if (executionStackNodes.includes(inputData.node)) {
|
|
// Node is already on the list to be executed
|
|
// so there is nothing to do
|
|
continue;
|
|
}
|
|
|
|
// Check if node got processed already
|
|
if (this.runExecutionData.resultData.runData[inputData.node] !== undefined) {
|
|
// Node got processed already so no need to add it
|
|
continue;
|
|
}
|
|
|
|
if (!forceInputNodeExecution) {
|
|
// Do not automatically follow all incoming nodes and force them
|
|
// to execute
|
|
continue;
|
|
}
|
|
|
|
// Check if any of the parent nodes does not have any inputs. That
|
|
// would mean that it has to get added to the list of nodes to process.
|
|
const parentNodes = workflow.getParentNodes(
|
|
inputData.node,
|
|
NodeConnectionType.Main,
|
|
-1,
|
|
);
|
|
let nodeToAdd: string | undefined = inputData.node;
|
|
parentNodes.push(inputData.node);
|
|
parentNodes.reverse();
|
|
|
|
for (const parentNode of parentNodes) {
|
|
// Check if that node is also an output connection of the
|
|
// previously processed one
|
|
if (inputData.node !== parentNode && checkOutputNodes.includes(parentNode)) {
|
|
// So the parent node will be added anyway which
|
|
// will then process this node next. So nothing to do.
|
|
nodeToAdd = undefined;
|
|
break;
|
|
}
|
|
|
|
// Check if it is already in the execution stack
|
|
if (executionStackNodes.includes(parentNode)) {
|
|
// Node is already on the list to be executed
|
|
// so there is nothing to do
|
|
nodeToAdd = undefined;
|
|
break;
|
|
}
|
|
|
|
// Check if node got processed already
|
|
if (this.runExecutionData.resultData.runData[parentNode] !== undefined) {
|
|
// Node got processed already so we can use the
|
|
// output data as input of this node
|
|
break;
|
|
}
|
|
|
|
nodeToAdd = parentNode;
|
|
}
|
|
const parentNodesNodeToAdd = workflow.getParentNodes(nodeToAdd as string);
|
|
if (
|
|
parentNodesNodeToAdd.includes(parentNodeName) &&
|
|
nodeSuccessData[outputIndex].length === 0
|
|
) {
|
|
// We do not add the node if there is no input data and the node that should be connected
|
|
// is a child of the parent node. Because else it would run a node even though it should be
|
|
// specifically not run, as it did not receive any data.
|
|
nodeToAdd = undefined;
|
|
}
|
|
|
|
if (nodeToAdd === undefined) {
|
|
// No node has to get added so process
|
|
continue;
|
|
}
|
|
|
|
let addEmptyItem = false;
|
|
|
|
if (workflow.connectionsByDestinationNode[nodeToAdd] === undefined) {
|
|
// Add empty item if the node does not have any input connections
|
|
addEmptyItem = true;
|
|
} else if (
|
|
this.incomingConnectionIsEmpty(
|
|
this.runExecutionData.resultData.runData,
|
|
workflow.connectionsByDestinationNode[nodeToAdd].main[0],
|
|
runIndex,
|
|
)
|
|
) {
|
|
// Add empty item also if the input data is empty
|
|
addEmptyItem = true;
|
|
}
|
|
|
|
if (addEmptyItem) {
|
|
// Add only node if it does not have any inputs because else it will
|
|
// be added by its input node later anyway.
|
|
this.runExecutionData.executionData!.nodeExecutionStack[enqueueFn]({
|
|
node: workflow.getNode(nodeToAdd) as INode,
|
|
data: {
|
|
main: [
|
|
[
|
|
{
|
|
json: {},
|
|
},
|
|
],
|
|
],
|
|
},
|
|
source: {
|
|
main: [
|
|
{
|
|
previousNode: parentNodeName,
|
|
previousNodeOutput: outputIndex || undefined,
|
|
previousNodeRun: runIndex || undefined,
|
|
},
|
|
],
|
|
},
|
|
});
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
let connectionDataArray: Array<INodeExecutionData[] | null> = get(
|
|
this.runExecutionData,
|
|
[
|
|
'executionData',
|
|
'waitingExecution',
|
|
connectionData.node,
|
|
waitingNodeIndex!,
|
|
NodeConnectionType.Main,
|
|
],
|
|
null,
|
|
);
|
|
|
|
if (connectionDataArray === null) {
|
|
connectionDataArray = [];
|
|
for (let i: number = connectionData.index; i >= 0; i--) {
|
|
connectionDataArray[i] = null;
|
|
}
|
|
}
|
|
|
|
// Add the data of the current execution
|
|
if (nodeSuccessData === null) {
|
|
connectionDataArray[connectionData.index] = null;
|
|
} else {
|
|
connectionDataArray[connectionData.index] = nodeSuccessData[outputIndex];
|
|
}
|
|
|
|
if (stillDataMissing) {
|
|
waitingNodeIndex = waitingNodeIndex!;
|
|
|
|
// Additional data is needed to run node so add it to waiting
|
|
this.prepareWaitingToExecution(
|
|
connectionData.node,
|
|
workflow.connectionsByDestinationNode[connectionData.node].main.length,
|
|
waitingNodeIndex,
|
|
);
|
|
|
|
this.runExecutionData.executionData!.waitingExecution[connectionData.node][waitingNodeIndex] =
|
|
{
|
|
main: connectionDataArray,
|
|
};
|
|
|
|
this.runExecutionData.executionData!.waitingExecutionSource![connectionData.node][
|
|
waitingNodeIndex
|
|
].main[connectionData.index] = {
|
|
previousNode: parentNodeName,
|
|
previousNodeOutput: outputIndex || undefined,
|
|
previousNodeRun: runIndex || undefined,
|
|
};
|
|
} else {
|
|
// All data is there so add it directly to stack
|
|
this.runExecutionData.executionData!.nodeExecutionStack[enqueueFn]({
|
|
node: workflow.nodes[connectionData.node],
|
|
data: {
|
|
main: connectionDataArray,
|
|
},
|
|
source: {
|
|
main: [
|
|
{
|
|
previousNode: parentNodeName,
|
|
previousNodeOutput: outputIndex || undefined,
|
|
previousNodeRun: runIndex || undefined,
|
|
},
|
|
],
|
|
},
|
|
});
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Runs the given execution data.
|
|
*
|
|
*/
|
|
// IMPORTANT: Do not add "async" to this function, it will then convert the
|
|
// PCancelable to a regular Promise and does so not allow canceling
|
|
// active executions anymore
|
|
// eslint-disable-next-line @typescript-eslint/promise-function-async
|
|
processRunExecutionData(workflow: Workflow): PCancelable<IRun> {
|
|
Logger.debug('Workflow execution started', { workflowId: workflow.id });
|
|
|
|
const startedAt = new Date();
|
|
const forceInputNodeExecution = this.forceInputNodeExecution(workflow);
|
|
|
|
this.status = 'running';
|
|
|
|
const startNode = this.runExecutionData.executionData!.nodeExecutionStack[0].node.name;
|
|
|
|
let destinationNode: string | undefined;
|
|
if (this.runExecutionData.startData && this.runExecutionData.startData.destinationNode) {
|
|
destinationNode = this.runExecutionData.startData.destinationNode;
|
|
}
|
|
|
|
const pinDataNodeNames = Object.keys(this.runExecutionData.resultData.pinData ?? {});
|
|
|
|
const workflowIssues = workflow.checkReadyForExecution({
|
|
startNode,
|
|
destinationNode,
|
|
pinDataNodeNames,
|
|
});
|
|
if (workflowIssues !== null) {
|
|
throw new WorkflowOperationError(
|
|
'The workflow has issues and can for that reason not be executed. Please fix them first.',
|
|
);
|
|
}
|
|
|
|
// Variables which hold temporary data for each node-execution
|
|
let executionData: IExecuteData;
|
|
let executionError: ExecutionBaseError | undefined;
|
|
let executionHints: NodeExecutionHint[] = [];
|
|
let executionNode: INode;
|
|
let nodeSuccessData: INodeExecutionData[][] | null | undefined;
|
|
let runIndex: number;
|
|
let startTime: number;
|
|
|
|
if (this.runExecutionData.startData === undefined) {
|
|
this.runExecutionData.startData = {};
|
|
}
|
|
|
|
if (this.runExecutionData.waitTill) {
|
|
const lastNodeExecuted = this.runExecutionData.resultData.lastNodeExecuted as string;
|
|
this.runExecutionData.executionData!.nodeExecutionStack[0].node.disabled = true;
|
|
this.runExecutionData.waitTill = undefined;
|
|
this.runExecutionData.resultData.runData[lastNodeExecuted].pop();
|
|
}
|
|
|
|
let currentExecutionTry = '';
|
|
let lastExecutionTry = '';
|
|
let closeFunction: Promise<void> | undefined;
|
|
|
|
return new PCancelable(async (resolve, _reject, onCancel) => {
|
|
// Let as many nodes listen to the abort signal, without getting the MaxListenersExceededWarning
|
|
setMaxListeners(Infinity, this.abortController.signal);
|
|
|
|
onCancel.shouldReject = false;
|
|
onCancel(() => {
|
|
this.status = 'canceled';
|
|
this.abortController.abort();
|
|
const fullRunData = this.getFullRunData(startedAt);
|
|
void this.executeHook('workflowExecuteAfter', [fullRunData]);
|
|
});
|
|
|
|
// eslint-disable-next-line complexity
|
|
const returnPromise = (async () => {
|
|
try {
|
|
if (!this.additionalData.restartExecutionId) {
|
|
await this.executeHook('workflowExecuteBefore', [workflow]);
|
|
}
|
|
} catch (error) {
|
|
const e = error as unknown as ExecutionBaseError;
|
|
|
|
// Set the error that it can be saved correctly
|
|
executionError = {
|
|
...e,
|
|
message: e.message,
|
|
stack: e.stack,
|
|
};
|
|
|
|
// Set the incoming data of the node that it can be saved correctly
|
|
|
|
executionData = this.runExecutionData.executionData!.nodeExecutionStack[0];
|
|
this.runExecutionData.resultData = {
|
|
runData: {
|
|
[executionData.node.name]: [
|
|
{
|
|
startTime,
|
|
executionTime: new Date().getTime() - startTime,
|
|
data: {
|
|
main: executionData.data.main,
|
|
} as ITaskDataConnections,
|
|
source: [],
|
|
executionStatus: 'error',
|
|
},
|
|
],
|
|
},
|
|
lastNodeExecuted: executionData.node.name,
|
|
error: executionError,
|
|
};
|
|
|
|
throw error;
|
|
}
|
|
|
|
executionLoop: while (
|
|
this.runExecutionData.executionData!.nodeExecutionStack.length !== 0
|
|
) {
|
|
if (
|
|
this.additionalData.executionTimeoutTimestamp !== undefined &&
|
|
Date.now() >= this.additionalData.executionTimeoutTimestamp
|
|
) {
|
|
this.status = 'canceled';
|
|
}
|
|
|
|
if (this.status === 'canceled') {
|
|
return;
|
|
}
|
|
|
|
nodeSuccessData = null;
|
|
executionError = undefined;
|
|
executionHints = [];
|
|
executionData =
|
|
this.runExecutionData.executionData!.nodeExecutionStack.shift() as IExecuteData;
|
|
executionNode = executionData.node;
|
|
|
|
// Update the pairedItem information on items
|
|
const newTaskDataConnections: ITaskDataConnections = {};
|
|
for (const inputName of Object.keys(executionData.data)) {
|
|
newTaskDataConnections[inputName] = executionData.data[inputName].map(
|
|
(input, inputIndex) => {
|
|
if (input === null) {
|
|
return input;
|
|
}
|
|
|
|
return input.map((item, itemIndex) => {
|
|
return {
|
|
...item,
|
|
pairedItem: {
|
|
item: itemIndex,
|
|
input: inputIndex || undefined,
|
|
},
|
|
};
|
|
});
|
|
},
|
|
);
|
|
}
|
|
executionData.data = newTaskDataConnections;
|
|
|
|
Logger.debug(`Start processing node "${executionNode.name}"`, {
|
|
node: executionNode.name,
|
|
workflowId: workflow.id,
|
|
});
|
|
await this.executeHook('nodeExecuteBefore', [executionNode.name]);
|
|
|
|
// Get the index of the current run
|
|
runIndex = 0;
|
|
if (this.runExecutionData.resultData.runData.hasOwnProperty(executionNode.name)) {
|
|
runIndex = this.runExecutionData.resultData.runData[executionNode.name].length;
|
|
}
|
|
|
|
currentExecutionTry = `${executionNode.name}:${runIndex}`;
|
|
|
|
if (currentExecutionTry === lastExecutionTry) {
|
|
throw new ApplicationError(
|
|
'Stopped execution because it seems to be in an endless loop',
|
|
);
|
|
}
|
|
|
|
if (
|
|
this.runExecutionData.startData!.runNodeFilter !== undefined &&
|
|
this.runExecutionData.startData!.runNodeFilter.indexOf(executionNode.name) === -1
|
|
) {
|
|
// If filter is set and node is not on filter skip it, that avoids the problem that it executes
|
|
// leaves that are parallel to a selected destinationNode. Normally it would execute them because
|
|
// they have the same parent and it executes all child nodes.
|
|
continue;
|
|
}
|
|
|
|
// Check if all the data which is needed to run the node is available
|
|
if (workflow.connectionsByDestinationNode.hasOwnProperty(executionNode.name)) {
|
|
// Check if the node has incoming connections
|
|
if (workflow.connectionsByDestinationNode[executionNode.name].hasOwnProperty('main')) {
|
|
let inputConnections: IConnection[][];
|
|
let connectionIndex: number;
|
|
|
|
// eslint-disable-next-line prefer-const
|
|
inputConnections = workflow.connectionsByDestinationNode[executionNode.name].main;
|
|
|
|
for (
|
|
connectionIndex = 0;
|
|
connectionIndex < inputConnections.length;
|
|
connectionIndex++
|
|
) {
|
|
if (
|
|
workflow.getHighestNode(
|
|
executionNode.name,
|
|
NodeConnectionType.Main,
|
|
connectionIndex,
|
|
).length === 0
|
|
) {
|
|
// If there is no valid incoming node (if all are disabled)
|
|
// then ignore that it has inputs and simply execute it as it is without
|
|
// any data
|
|
continue;
|
|
}
|
|
|
|
if (!executionData.data.hasOwnProperty('main')) {
|
|
// ExecutionData does not even have the connection set up so can
|
|
// not have that data, so add it again to be executed later
|
|
this.runExecutionData.executionData!.nodeExecutionStack.push(executionData);
|
|
lastExecutionTry = currentExecutionTry;
|
|
continue executionLoop;
|
|
}
|
|
|
|
if (forceInputNodeExecution) {
|
|
// Check if it has the data for all the inputs
|
|
// The most nodes just have one but merge node for example has two and data
|
|
// of both inputs has to be available to be able to process the node.
|
|
if (
|
|
executionData.data.main.length < connectionIndex ||
|
|
executionData.data.main[connectionIndex] === null
|
|
) {
|
|
// Does not have the data of the connections so add back to stack
|
|
this.runExecutionData.executionData!.nodeExecutionStack.push(executionData);
|
|
lastExecutionTry = currentExecutionTry;
|
|
|
|
continue executionLoop;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
startTime = new Date().getTime();
|
|
|
|
let maxTries = 1;
|
|
if (executionData.node.retryOnFail === true) {
|
|
// TODO: Remove the hardcoded default-values here and also in NodeSettings.vue
|
|
maxTries = Math.min(5, Math.max(2, executionData.node.maxTries || 3));
|
|
}
|
|
|
|
let waitBetweenTries = 0;
|
|
if (executionData.node.retryOnFail === true) {
|
|
// TODO: Remove the hardcoded default-values here and also in NodeSettings.vue
|
|
waitBetweenTries = Math.min(
|
|
5000,
|
|
Math.max(0, executionData.node.waitBetweenTries || 1000),
|
|
);
|
|
}
|
|
|
|
for (let tryIndex = 0; tryIndex < maxTries; tryIndex++) {
|
|
try {
|
|
if (tryIndex !== 0) {
|
|
// Reset executionError from previous error try
|
|
executionError = undefined;
|
|
if (waitBetweenTries !== 0) {
|
|
// TODO: Improve that in the future and check if other nodes can
|
|
// be executed in the meantime
|
|
// eslint-disable-next-line @typescript-eslint/no-shadow
|
|
await new Promise((resolve) => {
|
|
setTimeout(() => {
|
|
resolve(undefined);
|
|
}, waitBetweenTries);
|
|
});
|
|
}
|
|
}
|
|
|
|
const { pinData } = this.runExecutionData.resultData;
|
|
|
|
if (pinData && !executionNode.disabled && pinData[executionNode.name] !== undefined) {
|
|
const nodePinData = pinData[executionNode.name];
|
|
|
|
nodeSuccessData = [nodePinData]; // always zeroth runIndex
|
|
} else {
|
|
Logger.debug(`Running node "${executionNode.name}" started`, {
|
|
node: executionNode.name,
|
|
workflowId: workflow.id,
|
|
});
|
|
|
|
let runNodeData = await workflow.runNode(
|
|
executionData,
|
|
this.runExecutionData,
|
|
runIndex,
|
|
this.additionalData,
|
|
NodeExecuteFunctions,
|
|
this.mode,
|
|
this.abortController.signal,
|
|
);
|
|
|
|
nodeSuccessData = runNodeData.data;
|
|
|
|
const didContinueOnFail = nodeSuccessData?.at(0)?.at(0)?.json?.error !== undefined;
|
|
|
|
while (didContinueOnFail && tryIndex !== maxTries - 1) {
|
|
await sleep(waitBetweenTries);
|
|
|
|
runNodeData = await workflow.runNode(
|
|
executionData,
|
|
this.runExecutionData,
|
|
runIndex,
|
|
this.additionalData,
|
|
NodeExecuteFunctions,
|
|
this.mode,
|
|
this.abortController.signal,
|
|
);
|
|
|
|
tryIndex++;
|
|
}
|
|
|
|
if (nodeSuccessData instanceof NodeExecutionOutput) {
|
|
const hints: NodeExecutionHint[] = nodeSuccessData.getHints();
|
|
|
|
executionHints.push(...hints);
|
|
}
|
|
|
|
if (nodeSuccessData && executionData.node.onError === 'continueErrorOutput') {
|
|
// If errorOutput is activated check all the output items for error data.
|
|
// If any is found, route them to the last output as that will be the
|
|
// error output.
|
|
|
|
const nodeType = workflow.nodeTypes.getByNameAndVersion(
|
|
executionData.node.type,
|
|
executionData.node.typeVersion,
|
|
);
|
|
const outputs = NodeHelpers.getNodeOutputs(
|
|
workflow,
|
|
executionData.node,
|
|
nodeType.description,
|
|
);
|
|
const outputTypes = NodeHelpers.getConnectionTypes(outputs);
|
|
const mainOutputTypes = outputTypes.filter(
|
|
(output) => output === NodeConnectionType.Main,
|
|
);
|
|
|
|
const errorItems: INodeExecutionData[] = [];
|
|
const closeFunctions: CloseFunction[] = [];
|
|
// Create a WorkflowDataProxy instance that we can get the data of the
|
|
// item which did error
|
|
const executeFunctions = NodeExecuteFunctions.getExecuteFunctions(
|
|
workflow,
|
|
this.runExecutionData,
|
|
runIndex,
|
|
[],
|
|
executionData.data,
|
|
executionData.node,
|
|
this.additionalData,
|
|
executionData,
|
|
this.mode,
|
|
closeFunctions,
|
|
this.abortController.signal,
|
|
);
|
|
const dataProxy = executeFunctions.getWorkflowDataProxy(0);
|
|
|
|
// Loop over all outputs except the error output as it would not contain data by default
|
|
for (
|
|
let outputIndex = 0;
|
|
outputIndex < mainOutputTypes.length - 1;
|
|
outputIndex++
|
|
) {
|
|
const successItems: INodeExecutionData[] = [];
|
|
const items = nodeSuccessData[outputIndex]?.length
|
|
? nodeSuccessData[outputIndex]
|
|
: [];
|
|
|
|
while (items.length) {
|
|
const item = items.shift();
|
|
if (item === undefined) {
|
|
continue;
|
|
}
|
|
|
|
let errorData: GenericValue | undefined;
|
|
if (item.error) {
|
|
errorData = item.error;
|
|
item.error = undefined;
|
|
} else if (item.json.error && Object.keys(item.json).length === 1) {
|
|
errorData = item.json.error;
|
|
} else if (
|
|
item.json.error &&
|
|
item.json.message &&
|
|
Object.keys(item.json).length === 2
|
|
) {
|
|
errorData = item.json.error;
|
|
}
|
|
|
|
if (errorData) {
|
|
const pairedItemData =
|
|
item.pairedItem && typeof item.pairedItem === 'object'
|
|
? Array.isArray(item.pairedItem)
|
|
? item.pairedItem[0]
|
|
: item.pairedItem
|
|
: undefined;
|
|
|
|
if (executionData!.source === null || pairedItemData === undefined) {
|
|
// Source data is missing for some reason so we can not figure out the item
|
|
errorItems.push(item);
|
|
} else {
|
|
const pairedItemInputIndex = pairedItemData.input || 0;
|
|
|
|
const sourceData =
|
|
executionData!.source[NodeConnectionType.Main][pairedItemInputIndex];
|
|
|
|
const constPairedItem = dataProxy.$getPairedItem(
|
|
sourceData!.previousNode,
|
|
sourceData,
|
|
pairedItemData,
|
|
);
|
|
|
|
if (constPairedItem === null) {
|
|
errorItems.push(item);
|
|
} else {
|
|
errorItems.push({
|
|
...item,
|
|
json: {
|
|
...constPairedItem.json,
|
|
...item.json,
|
|
},
|
|
});
|
|
}
|
|
}
|
|
} else {
|
|
successItems.push(item);
|
|
}
|
|
}
|
|
|
|
nodeSuccessData[outputIndex] = successItems;
|
|
}
|
|
|
|
nodeSuccessData[mainOutputTypes.length - 1] = errorItems;
|
|
}
|
|
|
|
if (runNodeData.closeFunction) {
|
|
// Explanation why we do this can be found in n8n-workflow/Workflow.ts -> runNode
|
|
|
|
closeFunction = runNodeData.closeFunction();
|
|
}
|
|
}
|
|
|
|
Logger.debug(`Running node "${executionNode.name}" finished successfully`, {
|
|
node: executionNode.name,
|
|
workflowId: workflow.id,
|
|
});
|
|
|
|
if (nodeSuccessData?.length) {
|
|
// Check if the output data contains pairedItem data and if not try
|
|
// to automatically fix it
|
|
|
|
const isSingleInputAndOutput =
|
|
executionData.data.main.length === 1 && executionData.data.main[0]?.length === 1;
|
|
|
|
const isSameNumberOfItems =
|
|
nodeSuccessData.length === 1 &&
|
|
executionData.data.main.length === 1 &&
|
|
executionData.data.main[0]?.length === nodeSuccessData[0].length;
|
|
|
|
checkOutputData: for (const outputData of nodeSuccessData) {
|
|
if (outputData === null) {
|
|
continue;
|
|
}
|
|
for (const [index, item] of outputData.entries()) {
|
|
if (item.pairedItem === undefined) {
|
|
// The pairedItem data is missing, so check if it can get automatically fixed
|
|
if (isSingleInputAndOutput) {
|
|
// The node has one input and one incoming item, so we know
|
|
// that all items must originate from that single
|
|
item.pairedItem = {
|
|
item: 0,
|
|
};
|
|
} else if (isSameNumberOfItems) {
|
|
// The number of oncoming and outcoming items is identical so we can
|
|
// make the reasonable assumption that each of the input items
|
|
// is the origin of the corresponding output items
|
|
item.pairedItem = {
|
|
item: index,
|
|
};
|
|
} else {
|
|
// In all other cases autofixing is not possible
|
|
break checkOutputData;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (nodeSuccessData === undefined) {
|
|
// Node did not get executed
|
|
nodeSuccessData = null;
|
|
} else {
|
|
this.runExecutionData.resultData.lastNodeExecuted = executionData.node.name;
|
|
}
|
|
|
|
if (nodeSuccessData === null || nodeSuccessData[0][0] === undefined) {
|
|
if (executionData.node.alwaysOutputData === true) {
|
|
const pairedItem: IPairedItemData[] = [];
|
|
|
|
// Get pairedItem from all input items
|
|
executionData.data.main.forEach((inputData, inputIndex) => {
|
|
if (!inputData) {
|
|
return;
|
|
}
|
|
inputData.forEach((_item, itemIndex) => {
|
|
pairedItem.push({
|
|
item: itemIndex,
|
|
input: inputIndex,
|
|
});
|
|
});
|
|
});
|
|
|
|
nodeSuccessData = nodeSuccessData || [];
|
|
nodeSuccessData[0] = [
|
|
{
|
|
json: {},
|
|
pairedItem,
|
|
},
|
|
];
|
|
}
|
|
}
|
|
|
|
if (nodeSuccessData === null && !this.runExecutionData.waitTill!) {
|
|
// If null gets returned it means that the node did succeed
|
|
// but did not have any data. So the branch should end
|
|
// (meaning the nodes afterwards should not be processed)
|
|
continue executionLoop;
|
|
}
|
|
|
|
break;
|
|
} catch (error) {
|
|
this.runExecutionData.resultData.lastNodeExecuted = executionData.node.name;
|
|
|
|
let toReport: Error | undefined;
|
|
if (error instanceof ApplicationError) {
|
|
// Report any unhandled errors that were wrapped in by one of our error classes
|
|
if (error.cause instanceof Error) toReport = error.cause;
|
|
} else {
|
|
// Report any unhandled and non-wrapped errors to Sentry
|
|
toReport = error;
|
|
}
|
|
if (toReport) {
|
|
ErrorReporterProxy.error(toReport, {
|
|
extra: {
|
|
nodeName: executionNode.name,
|
|
nodeType: executionNode.type,
|
|
nodeVersion: executionNode.typeVersion,
|
|
workflowId: workflow.id,
|
|
},
|
|
});
|
|
}
|
|
|
|
const e = error as unknown as ExecutionBaseError;
|
|
|
|
executionError = { ...e, message: e.message, stack: e.stack };
|
|
|
|
Logger.debug(`Running node "${executionNode.name}" finished with error`, {
|
|
node: executionNode.name,
|
|
workflowId: workflow.id,
|
|
});
|
|
}
|
|
}
|
|
|
|
// Add the data to return to the user
|
|
// (currently does not get cloned as data does not get changed, maybe later we should do that?!?!)
|
|
|
|
if (!this.runExecutionData.resultData.runData.hasOwnProperty(executionNode.name)) {
|
|
this.runExecutionData.resultData.runData[executionNode.name] = [];
|
|
}
|
|
|
|
const taskData: ITaskData = {
|
|
hints: executionHints,
|
|
startTime,
|
|
executionTime: new Date().getTime() - startTime,
|
|
source: !executionData.source ? [] : executionData.source.main,
|
|
metadata: executionData.metadata,
|
|
executionStatus: this.runExecutionData.waitTill ? 'waiting' : 'success',
|
|
};
|
|
|
|
if (executionError !== undefined) {
|
|
taskData.error = executionError;
|
|
taskData.executionStatus = 'error';
|
|
|
|
if (
|
|
executionData.node.continueOnFail === true ||
|
|
['continueRegularOutput', 'continueErrorOutput'].includes(
|
|
executionData.node.onError || '',
|
|
)
|
|
) {
|
|
// Workflow should continue running even if node errors
|
|
if (executionData.data.hasOwnProperty('main') && executionData.data.main.length > 0) {
|
|
// Simply get the input data of the node if it has any and pass it through
|
|
// to the next node
|
|
if (executionData.data.main[0] !== null) {
|
|
nodeSuccessData = [executionData.data.main[0]];
|
|
}
|
|
}
|
|
} else {
|
|
// Node execution did fail so add error and stop execution
|
|
this.runExecutionData.resultData.runData[executionNode.name].push(taskData);
|
|
|
|
// Add the execution data again so that it can get restarted
|
|
this.runExecutionData.executionData!.nodeExecutionStack.unshift(executionData);
|
|
// Only execute the nodeExecuteAfter hook if the node did not get aborted
|
|
if (!this.isCancelled) {
|
|
await this.executeHook('nodeExecuteAfter', [
|
|
executionNode.name,
|
|
taskData,
|
|
this.runExecutionData,
|
|
]);
|
|
}
|
|
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Merge error information to default output for now
|
|
// As the new nodes can report the errors in
|
|
// the `error` property.
|
|
for (const execution of nodeSuccessData!) {
|
|
for (const lineResult of execution) {
|
|
if (
|
|
lineResult.json !== undefined &&
|
|
lineResult.json.$error !== undefined &&
|
|
lineResult.json.$json !== undefined
|
|
) {
|
|
// eslint-disable-next-line @typescript-eslint/ban-ts-comment
|
|
// @ts-ignore
|
|
lineResult.error = lineResult.json.$error as NodeApiError | NodeOperationError;
|
|
lineResult.json = {
|
|
// eslint-disable-next-line @typescript-eslint/ban-ts-comment
|
|
// @ts-ignore
|
|
error: (lineResult.json.$error as NodeApiError | NodeOperationError).message,
|
|
};
|
|
} else if (lineResult.error !== undefined) {
|
|
// eslint-disable-next-line @typescript-eslint/ban-ts-comment
|
|
// @ts-ignore
|
|
lineResult.json = { error: lineResult.error.message };
|
|
}
|
|
}
|
|
}
|
|
|
|
// Node executed successfully. So add data and go on.
|
|
taskData.data = {
|
|
main: nodeSuccessData,
|
|
} as ITaskDataConnections;
|
|
|
|
this.runExecutionData.resultData.runData[executionNode.name].push(taskData);
|
|
|
|
if (this.runExecutionData.waitTill!) {
|
|
await this.executeHook('nodeExecuteAfter', [
|
|
executionNode.name,
|
|
taskData,
|
|
this.runExecutionData,
|
|
]);
|
|
|
|
// Add the node back to the stack that the workflow can start to execute again from that node
|
|
this.runExecutionData.executionData!.nodeExecutionStack.unshift(executionData);
|
|
|
|
break;
|
|
}
|
|
|
|
if (
|
|
this.runExecutionData.startData &&
|
|
this.runExecutionData.startData.destinationNode &&
|
|
this.runExecutionData.startData.destinationNode === executionNode.name
|
|
) {
|
|
// Before stopping, make sure we are executing hooks so
|
|
// That frontend is notified for example for manual executions.
|
|
await this.executeHook('nodeExecuteAfter', [
|
|
executionNode.name,
|
|
taskData,
|
|
this.runExecutionData,
|
|
]);
|
|
|
|
// If destination node is defined and got executed stop execution
|
|
continue;
|
|
}
|
|
|
|
// Add the nodes to which the current node has an output connection to that they can
|
|
// be executed next
|
|
if (workflow.connectionsBySourceNode.hasOwnProperty(executionNode.name)) {
|
|
if (workflow.connectionsBySourceNode[executionNode.name].hasOwnProperty('main')) {
|
|
let outputIndex: string;
|
|
let connectionData: IConnection;
|
|
// Iterate over all the outputs
|
|
|
|
const nodesToAdd: Array<{
|
|
position: [number, number];
|
|
connection: IConnection;
|
|
outputIndex: number;
|
|
}> = [];
|
|
|
|
// Add the nodes to be executed
|
|
// eslint-disable-next-line @typescript-eslint/no-for-in-array
|
|
for (outputIndex in workflow.connectionsBySourceNode[executionNode.name].main) {
|
|
if (
|
|
!workflow.connectionsBySourceNode[executionNode.name].main.hasOwnProperty(
|
|
outputIndex,
|
|
)
|
|
) {
|
|
continue;
|
|
}
|
|
|
|
// Iterate over all the different connections of this output
|
|
for (connectionData of workflow.connectionsBySourceNode[executionNode.name].main[
|
|
outputIndex
|
|
]) {
|
|
if (!workflow.nodes.hasOwnProperty(connectionData.node)) {
|
|
throw new ApplicationError('Destination node not found', {
|
|
extra: {
|
|
sourceNodeName: executionNode.name,
|
|
destinationNodeName: connectionData.node,
|
|
},
|
|
});
|
|
}
|
|
|
|
if (
|
|
nodeSuccessData![outputIndex] &&
|
|
(nodeSuccessData![outputIndex].length !== 0 ||
|
|
(connectionData.index > 0 && forceInputNodeExecution))
|
|
) {
|
|
// Add the node only if it did execute or if connected to second "optional" input
|
|
if (workflow.settings.executionOrder === 'v1') {
|
|
const nodeToAdd = workflow.getNode(connectionData.node);
|
|
nodesToAdd.push({
|
|
position: nodeToAdd?.position || [0, 0],
|
|
connection: connectionData,
|
|
outputIndex: parseInt(outputIndex, 10),
|
|
});
|
|
} else {
|
|
this.addNodeToBeExecuted(
|
|
workflow,
|
|
connectionData,
|
|
parseInt(outputIndex, 10),
|
|
executionNode.name,
|
|
nodeSuccessData!,
|
|
runIndex,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (workflow.settings.executionOrder === 'v1') {
|
|
// Always execute the node that is more to the top-left first
|
|
nodesToAdd.sort((a, b) => {
|
|
if (a.position[1] < b.position[1]) {
|
|
return 1;
|
|
}
|
|
if (a.position[1] > b.position[1]) {
|
|
return -1;
|
|
}
|
|
|
|
if (a.position[0] > b.position[0]) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
});
|
|
|
|
for (const nodeData of nodesToAdd) {
|
|
this.addNodeToBeExecuted(
|
|
workflow,
|
|
nodeData.connection,
|
|
nodeData.outputIndex,
|
|
executionNode.name,
|
|
nodeSuccessData!,
|
|
runIndex,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// If we got here, it means that we did not stop executing from manual executions / destination.
|
|
// Execute hooks now to make sure that all hooks are executed properly
|
|
// Await is needed to make sure that we don't fall into concurrency problems
|
|
// When saving node execution data
|
|
await this.executeHook('nodeExecuteAfter', [
|
|
executionNode.name,
|
|
taskData,
|
|
this.runExecutionData,
|
|
]);
|
|
|
|
let waitingNodes: string[] = Object.keys(
|
|
this.runExecutionData.executionData!.waitingExecution,
|
|
);
|
|
|
|
if (
|
|
this.runExecutionData.executionData!.nodeExecutionStack.length === 0 &&
|
|
waitingNodes.length
|
|
) {
|
|
// There are no more nodes in the execution stack. Check if there are
|
|
// waiting nodes that do not require data on all inputs and execute them,
|
|
// one by one.
|
|
|
|
// TODO: Should this also care about workflow position (top-left first?)
|
|
for (let i = 0; i < waitingNodes.length; i++) {
|
|
const nodeName = waitingNodes[i];
|
|
|
|
const checkNode = workflow.getNode(nodeName);
|
|
if (!checkNode) {
|
|
continue;
|
|
}
|
|
const nodeType = workflow.nodeTypes.getByNameAndVersion(
|
|
checkNode.type,
|
|
checkNode.typeVersion,
|
|
);
|
|
|
|
// Check if the node is only allowed execute if all inputs received data
|
|
let requiredInputs =
|
|
workflow.settings.executionOrder === 'v1'
|
|
? nodeType.description.requiredInputs
|
|
: undefined;
|
|
if (requiredInputs !== undefined) {
|
|
if (typeof requiredInputs === 'string') {
|
|
requiredInputs = workflow.expression.getSimpleParameterValue(
|
|
checkNode,
|
|
requiredInputs,
|
|
this.mode,
|
|
{ $version: checkNode.typeVersion },
|
|
undefined,
|
|
[],
|
|
) as number[];
|
|
}
|
|
|
|
if (
|
|
(requiredInputs !== undefined &&
|
|
Array.isArray(requiredInputs) &&
|
|
requiredInputs.length === nodeType.description.inputs.length) ||
|
|
requiredInputs === nodeType.description.inputs.length
|
|
) {
|
|
// All inputs are required, but not all have data so do not continue
|
|
continue;
|
|
}
|
|
}
|
|
|
|
const parentNodes = workflow.getParentNodes(nodeName);
|
|
|
|
// Check if input nodes (of same run) got already executed
|
|
// eslint-disable-next-line @typescript-eslint/no-loop-func
|
|
const parentIsWaiting = parentNodes.some((value) => waitingNodes.includes(value));
|
|
if (parentIsWaiting) {
|
|
// Execute node later as one of its dependencies is still outstanding
|
|
continue;
|
|
}
|
|
|
|
const runIndexes = Object.keys(
|
|
this.runExecutionData.executionData!.waitingExecution[nodeName],
|
|
).sort();
|
|
|
|
// The run-index of the earliest outstanding one
|
|
const firstRunIndex = parseInt(runIndexes[0]);
|
|
|
|
// Find all the inputs which received any kind of data, even if it was an empty
|
|
// array as this shows that the parent nodes executed but they did not have any
|
|
// data to pass on.
|
|
const inputsWithData = this.runExecutionData
|
|
.executionData!.waitingExecution[nodeName][firstRunIndex].main.map((data, index) =>
|
|
data === null ? null : index,
|
|
)
|
|
.filter((data) => data !== null);
|
|
|
|
if (requiredInputs !== undefined) {
|
|
// Certain inputs are required that the node can execute
|
|
|
|
if (Array.isArray(requiredInputs)) {
|
|
// Specific inputs are required (array of input indexes)
|
|
let inputDataMissing = false;
|
|
for (const requiredInput of requiredInputs) {
|
|
if (!inputsWithData.includes(requiredInput)) {
|
|
inputDataMissing = true;
|
|
break;
|
|
}
|
|
}
|
|
if (inputDataMissing) {
|
|
continue;
|
|
}
|
|
} else {
|
|
// A certain amount of inputs are required (amount of inputs)
|
|
if (inputsWithData.length < requiredInputs) {
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
const taskDataMain = this.runExecutionData.executionData!.waitingExecution[nodeName][
|
|
firstRunIndex
|
|
].main.map((data) => {
|
|
// For the inputs for which never any data got received set it to an empty array
|
|
return data === null ? [] : data;
|
|
});
|
|
|
|
if (taskDataMain.filter((data) => data.length).length !== 0) {
|
|
// Add the node to be executed
|
|
|
|
// Make sure that each input at least receives an empty array
|
|
if (taskDataMain.length < nodeType.description.inputs.length) {
|
|
for (; taskDataMain.length < nodeType.description.inputs.length; ) {
|
|
taskDataMain.push([]);
|
|
}
|
|
}
|
|
|
|
this.runExecutionData.executionData!.nodeExecutionStack.push({
|
|
node: workflow.nodes[nodeName],
|
|
data: {
|
|
main: taskDataMain,
|
|
},
|
|
source:
|
|
this.runExecutionData.executionData!.waitingExecutionSource![nodeName][
|
|
firstRunIndex
|
|
],
|
|
});
|
|
}
|
|
|
|
// Remove the node from waiting
|
|
delete this.runExecutionData.executionData!.waitingExecution[nodeName][firstRunIndex];
|
|
delete this.runExecutionData.executionData!.waitingExecutionSource![nodeName][
|
|
firstRunIndex
|
|
];
|
|
|
|
if (
|
|
Object.keys(this.runExecutionData.executionData!.waitingExecution[nodeName])
|
|
.length === 0
|
|
) {
|
|
// No more data left for the node so also delete that one
|
|
delete this.runExecutionData.executionData!.waitingExecution[nodeName];
|
|
delete this.runExecutionData.executionData!.waitingExecutionSource![nodeName];
|
|
}
|
|
|
|
if (taskDataMain.filter((data) => data.length).length !== 0) {
|
|
// Node to execute got found and added to stop
|
|
break;
|
|
} else {
|
|
// Node to add did not get found, rather an empty one removed so continue with search
|
|
waitingNodes = Object.keys(this.runExecutionData.executionData!.waitingExecution);
|
|
// Set counter to start again from the beginning. Set it to -1 as it auto increments
|
|
// after run. So only like that will we end up again at 0.
|
|
i = -1;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return;
|
|
})()
|
|
.then(async () => {
|
|
if (this.status === 'canceled' && executionError === undefined) {
|
|
return await this.processSuccessExecution(
|
|
startedAt,
|
|
workflow,
|
|
new ExecutionCancelledError(this.additionalData.executionId ?? 'unknown'),
|
|
closeFunction,
|
|
);
|
|
}
|
|
return await this.processSuccessExecution(
|
|
startedAt,
|
|
workflow,
|
|
executionError,
|
|
closeFunction,
|
|
);
|
|
})
|
|
.catch(async (error) => {
|
|
const fullRunData = this.getFullRunData(startedAt);
|
|
|
|
fullRunData.data.resultData.error = {
|
|
...error,
|
|
message: error.message,
|
|
stack: error.stack,
|
|
};
|
|
|
|
// Check if static data changed
|
|
let newStaticData: IDataObject | undefined;
|
|
|
|
if (workflow.staticData.__dataChanged === true) {
|
|
// Static data of workflow changed
|
|
newStaticData = workflow.staticData;
|
|
}
|
|
|
|
this.moveNodeMetadata();
|
|
|
|
await this.executeHook('workflowExecuteAfter', [fullRunData, newStaticData]).catch(
|
|
// eslint-disable-next-line @typescript-eslint/no-shadow
|
|
(error) => {
|
|
console.error('There was a problem running hook "workflowExecuteAfter"', error);
|
|
},
|
|
);
|
|
|
|
if (closeFunction) {
|
|
try {
|
|
await closeFunction;
|
|
} catch (errorClose) {
|
|
Logger.error(
|
|
`There was a problem deactivating trigger of workflow "${workflow.id}": "${errorClose.message}"`,
|
|
{
|
|
workflowId: workflow.id,
|
|
},
|
|
);
|
|
}
|
|
}
|
|
|
|
return fullRunData;
|
|
});
|
|
|
|
return await returnPromise.then(resolve);
|
|
});
|
|
}
|
|
|
|
async processSuccessExecution(
|
|
startedAt: Date,
|
|
workflow: Workflow,
|
|
executionError?: ExecutionBaseError,
|
|
closeFunction?: Promise<void>,
|
|
): Promise<IRun> {
|
|
const fullRunData = this.getFullRunData(startedAt);
|
|
|
|
if (executionError !== undefined) {
|
|
Logger.debug('Workflow execution finished with error', {
|
|
error: executionError,
|
|
workflowId: workflow.id,
|
|
});
|
|
fullRunData.data.resultData.error = {
|
|
...executionError,
|
|
message: executionError.message,
|
|
stack: executionError.stack,
|
|
} as ExecutionBaseError;
|
|
if (executionError.message?.includes('canceled')) {
|
|
fullRunData.status = 'canceled';
|
|
}
|
|
} else if (this.runExecutionData.waitTill!) {
|
|
// eslint-disable-next-line @typescript-eslint/restrict-template-expressions
|
|
Logger.debug(`Workflow execution will wait until ${this.runExecutionData.waitTill}`, {
|
|
workflowId: workflow.id,
|
|
});
|
|
fullRunData.waitTill = this.runExecutionData.waitTill;
|
|
fullRunData.status = 'waiting';
|
|
} else {
|
|
Logger.debug('Workflow execution finished successfully', { workflowId: workflow.id });
|
|
fullRunData.finished = true;
|
|
fullRunData.status = 'success';
|
|
}
|
|
|
|
// Check if static data changed
|
|
let newStaticData: IDataObject | undefined;
|
|
|
|
if (workflow.staticData.__dataChanged === true) {
|
|
// Static data of workflow changed
|
|
newStaticData = workflow.staticData;
|
|
}
|
|
|
|
this.moveNodeMetadata();
|
|
// Prevent from running the hook if the error is an abort error as it was already handled
|
|
if (!this.isCancelled) {
|
|
await this.executeHook('workflowExecuteAfter', [fullRunData, newStaticData]);
|
|
}
|
|
|
|
if (closeFunction) {
|
|
try {
|
|
await closeFunction;
|
|
} catch (error) {
|
|
Logger.error(
|
|
`There was a problem deactivating trigger of workflow "${workflow.id}": "${error.message}"`,
|
|
{
|
|
workflowId: workflow.id,
|
|
},
|
|
);
|
|
}
|
|
}
|
|
|
|
return fullRunData;
|
|
}
|
|
|
|
getFullRunData(startedAt: Date): IRun {
|
|
const fullRunData: IRun = {
|
|
data: this.runExecutionData,
|
|
mode: this.mode,
|
|
startedAt,
|
|
stoppedAt: new Date(),
|
|
status: this.status,
|
|
};
|
|
|
|
return fullRunData;
|
|
}
|
|
|
|
private get isCancelled() {
|
|
return this.abortController.signal.aborted;
|
|
}
|
|
}
|