/* eslint-disable @typescript-eslint/restrict-template-expressions */ /* eslint-disable @typescript-eslint/no-non-null-assertion */ /* eslint-disable @typescript-eslint/no-unsafe-assignment */ import { validate as jsonSchemaValidate } from 'jsonschema'; import { BinaryDataManager } from 'n8n-core'; import { deepCopy, IDataObject, LoggerProxy, JsonObject, jsonParse, Workflow } from 'n8n-workflow'; import { FindOperator, In, IsNull, LessThanOrEqual, Not, Raw } from 'typeorm'; import * as ActiveExecutions from '@/ActiveExecutions'; import config from '@/config'; import { User } from '@/databases/entities/User'; import { IExecutionFlattedResponse, IExecutionResponse, IExecutionsListResponse, IWorkflowBase, IWorkflowExecutionDataProcess, } from '@/Interfaces'; import { NodeTypes } from '@/NodeTypes'; import * as Queue from '@/Queue'; import type { ExecutionRequest } from '@/requests'; import * as ResponseHelper from '@/ResponseHelper'; import { getSharedWorkflowIds } from '@/WorkflowHelpers'; import { WorkflowRunner } from '@/WorkflowRunner'; import type { DatabaseType } from '@/Interfaces'; import * as Db from '@/Db'; import * as GenericHelpers from '@/GenericHelpers'; interface IGetExecutionsQueryFilter { id?: FindOperator; finished?: boolean; mode?: string; retryOf?: string; retrySuccessId?: string; workflowId?: number | string; // eslint-disable-next-line @typescript-eslint/no-explicit-any waitTill?: FindOperator | boolean; } const schemaGetExecutionsQueryFilter = { $id: '/IGetExecutionsQueryFilter', type: 'object', properties: { finished: { type: 'boolean' }, mode: { type: 'string' }, retryOf: { type: 'string' }, retrySuccessId: { type: 'string' }, waitTill: { type: 'boolean' }, workflowId: { anyOf: [{ type: 'integer' }, { type: 'string' }] }, }, }; const allowedExecutionsQueryFilterFields = Object.keys(schemaGetExecutionsQueryFilter.properties); export class ExecutionsService { /** * Function to get the workflow Ids for a User * Overridden in EE version to ignore roles */ static async getWorkflowIdsForUser(user: User): Promise { // Get all workflows using owner role return getSharedWorkflowIds(user, ['owner']); } /** * Helper function to retrieve count of Executions */ static async getExecutionsCount( countFilter: IDataObject, user: User, ): Promise<{ count: number; estimated: boolean }> { const dbType = (await GenericHelpers.getConfigValue('database.type')) as DatabaseType; const filteredFields = Object.keys(countFilter).filter((field) => field !== 'id'); // For databases other than Postgres, do a regular count // when filtering based on `workflowId` or `finished` fields. if (dbType !== 'postgresdb' || filteredFields.length > 0 || user.globalRole.name !== 'owner') { const sharedWorkflowIds = await this.getWorkflowIdsForUser(user); const countParams = { where: { workflowId: In(sharedWorkflowIds), ...countFilter } }; const count = await Db.collections.Execution.count(countParams); return { count, estimated: false }; } try { // Get an estimate of rows count. const estimateRowsNumberSql = "SELECT n_live_tup FROM pg_stat_all_tables WHERE relname = 'execution_entity';"; const rows: Array<{ n_live_tup: string }> = await Db.collections.Execution.query( estimateRowsNumberSql, ); const estimate = parseInt(rows[0].n_live_tup, 10); // If over 100k, return just an estimate. if (estimate > 100_000) { // if less than 100k, we get the real count as even a full // table scan should not take so long. return { count: estimate, estimated: true }; } } catch (error) { LoggerProxy.warn(`Failed to get executions count from Postgres: ${error}`); } const sharedWorkflowIds = await getSharedWorkflowIds(user); const count = await Db.collections.Execution.count({ where: { workflowId: In(sharedWorkflowIds), }, }); return { count, estimated: false }; } static massageFilters(filter: IDataObject): void { if (filter) { if (filter.waitTill === true) { filter.waitTill = Not(IsNull()); // eslint-disable-next-line @typescript-eslint/no-unnecessary-boolean-literal-compare } else if (filter.finished === false) { filter.waitTill = IsNull(); } else { delete filter.waitTill; } } } static async getExecutionsList(req: ExecutionRequest.GetAll): Promise { const sharedWorkflowIds = await this.getWorkflowIdsForUser(req.user); if (sharedWorkflowIds.length === 0) { // return early since without shared workflows there can be no hits // (note: getSharedWorkflowIds() returns _all_ workflow ids for global owners) return { count: 0, estimated: false, results: [], }; } // parse incoming filter object and remove non-valid fields let filter: IGetExecutionsQueryFilter | undefined = undefined; if (req.query.filter) { try { const filterJson: JsonObject = jsonParse(req.query.filter); if (filterJson) { Object.keys(filterJson).map((key) => { if (!allowedExecutionsQueryFilterFields.includes(key)) delete filterJson[key]; }); if (jsonSchemaValidate(filterJson, schemaGetExecutionsQueryFilter).valid) { filter = filterJson as IGetExecutionsQueryFilter; } } } catch (error) { LoggerProxy.error('Failed to parse filter', { userId: req.user.id, filter: req.query.filter, }); throw new ResponseHelper.InternalServerError( 'Parameter "filter" contained invalid JSON string.', ); } } // safeguard against querying workflowIds not shared with the user const workflowId = filter?.workflowId?.toString(); if (workflowId !== undefined && !sharedWorkflowIds.includes(workflowId)) { LoggerProxy.verbose( `User ${req.user.id} attempted to query non-shared workflow ${workflowId}`, ); return { count: 0, estimated: false, results: [], }; } const limit = req.query.limit ? parseInt(req.query.limit, 10) : GenericHelpers.DEFAULT_EXECUTIONS_GET_ALL_LIMIT; const executingWorkflowIds: string[] = []; if (config.getEnv('executions.mode') === 'queue') { const queue = await Queue.getInstance(); const currentJobs = await queue.getJobs(['active', 'waiting']); executingWorkflowIds.push(...currentJobs.map(({ data }) => data.executionId)); } // We may have manual executions even with queue so we must account for these. executingWorkflowIds.push( ...ActiveExecutions.getInstance() .getActiveExecutions() .map(({ id }) => id), ); const findWhere = { workflowId: In(sharedWorkflowIds) }; const rangeQuery: string[] = []; const rangeQueryParams: { lastId?: string; firstId?: string; executingWorkflowIds?: string[]; } = {}; if (req.query.lastId) { rangeQuery.push('id < :lastId'); rangeQueryParams.lastId = req.query.lastId; } if (req.query.firstId) { rangeQuery.push('id > :firstId'); rangeQueryParams.firstId = req.query.firstId; } if (executingWorkflowIds.length > 0) { rangeQuery.push('id NOT IN (:...executingWorkflowIds)'); rangeQueryParams.executingWorkflowIds = executingWorkflowIds; } if (rangeQuery.length) { Object.assign(findWhere, { id: Raw(() => rangeQuery.join(' and '), rangeQueryParams), }); } let query = Db.collections.Execution.createQueryBuilder() .select() .orderBy('id', 'DESC') .take(limit) .where(findWhere); const countFilter = deepCopy(filter ?? {}); if (filter) { this.massageFilters(filter as IDataObject); query = query.andWhere(filter); } this.massageFilters(countFilter as IDataObject); countFilter.id = Not(In(executingWorkflowIds)); const executions = await query.getMany(); const { count, estimated } = await this.getExecutionsCount( countFilter as IDataObject, req.user, ); const formattedExecutions = executions.map((execution) => { return { id: execution.id.toString(), finished: execution.finished, mode: execution.mode, retryOf: execution.retryOf?.toString(), retrySuccessId: execution?.retrySuccessId?.toString(), waitTill: execution.waitTill as Date | undefined, startedAt: execution.startedAt, stoppedAt: execution.stoppedAt, workflowId: execution.workflowData?.id?.toString() ?? '', workflowName: execution.workflowData?.name, }; }); return { count, results: formattedExecutions, estimated, }; } static async getExecution( req: ExecutionRequest.Get, ): Promise { const sharedWorkflowIds = await this.getWorkflowIdsForUser(req.user); if (!sharedWorkflowIds.length) return undefined; const { id: executionId } = req.params; const execution = await Db.collections.Execution.findOne({ where: { id: executionId, workflowId: In(sharedWorkflowIds), }, }); if (!execution) { LoggerProxy.info('Attempt to read execution was blocked due to insufficient permissions', { userId: req.user.id, executionId, }); return undefined; } if (req.query.unflattedResponse === 'true') { return ResponseHelper.unflattenExecutionData(execution); } const { id, ...rest } = execution; // @ts-ignore return { id: id.toString(), ...rest, }; } static async retryExecution(req: ExecutionRequest.Retry): Promise { const sharedWorkflowIds = await this.getWorkflowIdsForUser(req.user); if (!sharedWorkflowIds.length) return false; const { id: executionId } = req.params; const execution = await Db.collections.Execution.findOne({ where: { id: executionId, workflowId: In(sharedWorkflowIds), }, }); if (!execution) { LoggerProxy.info( 'Attempt to retry an execution was blocked due to insufficient permissions', { userId: req.user.id, executionId, }, ); throw new ResponseHelper.NotFoundError( `The execution with the ID "${executionId}" does not exist.`, ); } const fullExecutionData = ResponseHelper.unflattenExecutionData(execution); if (fullExecutionData.finished) { throw new Error('The execution succeeded, so it cannot be retried.'); } const executionMode = 'retry'; fullExecutionData.workflowData.active = false; // Start the workflow const data: IWorkflowExecutionDataProcess = { executionMode, executionData: fullExecutionData.data, retryOf: req.params.id, workflowData: fullExecutionData.workflowData, userId: req.user.id, }; const { lastNodeExecuted } = data.executionData!.resultData; if (lastNodeExecuted) { // Remove the old error and the data of the last run of the node that it can be replaced delete data.executionData!.resultData.error; const { length } = data.executionData!.resultData.runData[lastNodeExecuted]; if ( length > 0 && data.executionData!.resultData.runData[lastNodeExecuted][length - 1].error !== undefined ) { // Remove results only if it is an error. // If we are retrying due to a crash, the information is simply success info from last node data.executionData!.resultData.runData[lastNodeExecuted].pop(); // Stack will determine what to run next } } if (req.body.loadWorkflow) { // Loads the currently saved workflow to execute instead of the // one saved at the time of the execution. const workflowId = fullExecutionData.workflowData.id as string; const workflowData = (await Db.collections.Workflow.findOne(workflowId)) as IWorkflowBase; if (workflowData === undefined) { throw new Error( `The workflow with the ID "${workflowId}" could not be found and so the data not be loaded for the retry.`, ); } data.workflowData = workflowData; const nodeTypes = NodeTypes(); const workflowInstance = new Workflow({ id: workflowData.id as string, name: workflowData.name, nodes: workflowData.nodes, connections: workflowData.connections, active: false, nodeTypes, staticData: undefined, settings: workflowData.settings, }); // Replace all of the nodes in the execution stack with the ones of the new workflow for (const stack of data.executionData!.executionData!.nodeExecutionStack) { // Find the data of the last executed node in the new workflow const node = workflowInstance.getNode(stack.node.name); if (node === null) { LoggerProxy.error('Failed to retry an execution because a node could not be found', { userId: req.user.id, executionId, nodeName: stack.node.name, }); throw new Error( `Could not find the node "${stack.node.name}" in workflow. It probably got deleted or renamed. Without it the workflow can sadly not be retried.`, ); } // Replace the node data in the stack that it really uses the current data stack.node = node; } } const workflowRunner = new WorkflowRunner(); const retriedExecutionId = await workflowRunner.run(data); const executionData = await ActiveExecutions.getInstance().getPostExecutePromise( retriedExecutionId, ); if (!executionData) { throw new Error('The retry did not start for an unknown reason.'); } return !!executionData.finished; } static async deleteExecutions(req: ExecutionRequest.Delete): Promise { const sharedWorkflowIds = await this.getWorkflowIdsForUser(req.user); if (sharedWorkflowIds.length === 0) { // return early since without shared workflows there can be no hits // (note: getSharedWorkflowIds() returns _all_ workflow ids for global owners) return; } const { deleteBefore, ids, filters: requestFiltersRaw } = req.body; let requestFilters; if (requestFiltersRaw) { try { Object.keys(requestFiltersRaw).map((key) => { if (!allowedExecutionsQueryFilterFields.includes(key)) delete requestFiltersRaw[key]; }); if (jsonSchemaValidate(requestFiltersRaw, schemaGetExecutionsQueryFilter).valid) { requestFilters = requestFiltersRaw as IGetExecutionsQueryFilter; } } catch (error) { throw new ResponseHelper.InternalServerError( 'Parameter "filter" contained invalid JSON string.', ); } } if (!deleteBefore && !ids) { throw new Error('Either "deleteBefore" or "ids" must be present in the request body'); } const binaryDataManager = BinaryDataManager.getInstance(); // delete executions by date, if user may access the underlying workflows if (deleteBefore) { const filters: IDataObject = { startedAt: LessThanOrEqual(deleteBefore), }; let query = Db.collections.Execution.createQueryBuilder() .select() .where({ ...filters, workflowId: In(sharedWorkflowIds), }); if (requestFilters) { query = query.andWhere(requestFilters); } const executions = await query.getMany(); if (!executions.length) return; const idsToDelete = executions.map(({ id }) => id.toString()); await Promise.all( idsToDelete.map(async (id) => binaryDataManager.deleteBinaryDataByExecutionId(id)), ); await Db.collections.Execution.delete({ id: In(idsToDelete) }); return; } // delete executions by IDs, if user may access the underlying workflows if (ids) { const executions = await Db.collections.Execution.find({ where: { id: In(ids), workflowId: In(sharedWorkflowIds), }, }); if (!executions.length) { LoggerProxy.error('Failed to delete an execution due to insufficient permissions', { userId: req.user.id, executionIds: ids, }); return; } const idsToDelete = executions.map(({ id }) => id.toString()); await Promise.all( idsToDelete.map(async (id) => binaryDataManager.deleteBinaryDataByExecutionId(id)), ); await Db.collections.Execution.delete(idsToDelete); } } }