mirror of
https://github.com/n8n-io/n8n.git
synced 2025-03-05 20:50:17 -08:00
Merge branch 'testing-framework' into always-load-credentials-from-db
This commit is contained in:
commit
d0fa711aae
|
@ -40,6 +40,9 @@ export class Execute extends Command {
|
||||||
id: flags.string({
|
id: flags.string({
|
||||||
description: 'id of the workflow to execute',
|
description: 'id of the workflow to execute',
|
||||||
}),
|
}),
|
||||||
|
rawOutput: flags.boolean({
|
||||||
|
description: 'Outputs only JSON data, with no other text',
|
||||||
|
}),
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -173,9 +176,10 @@ export class Execute extends Command {
|
||||||
stack: error.stack,
|
stack: error.stack,
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
if (flags.rawOutput === undefined) {
|
||||||
this.log('Execution was successfull:');
|
this.log('Execution was successfull:');
|
||||||
this.log('====================================');
|
this.log('====================================');
|
||||||
|
}
|
||||||
this.log(JSON.stringify(data, null, 2));
|
this.log(JSON.stringify(data, null, 2));
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error('\nGOT ERROR');
|
console.error('\nGOT ERROR');
|
||||||
|
|
343
packages/cli/commands/executeAll.ts
Normal file
343
packages/cli/commands/executeAll.ts
Normal file
|
@ -0,0 +1,343 @@
|
||||||
|
import * as fs from 'fs';
|
||||||
|
import { Command, flags } from '@oclif/command';
|
||||||
|
import {
|
||||||
|
UserSettings,
|
||||||
|
} from 'n8n-core';
|
||||||
|
import {
|
||||||
|
INode,
|
||||||
|
} from 'n8n-workflow';
|
||||||
|
|
||||||
|
import {
|
||||||
|
ActiveExecutions,
|
||||||
|
CredentialsOverwrites,
|
||||||
|
CredentialTypes,
|
||||||
|
Db,
|
||||||
|
ExternalHooks,
|
||||||
|
IWorkflowExecutionDataProcess,
|
||||||
|
LoadNodesAndCredentials,
|
||||||
|
NodeTypes,
|
||||||
|
WorkflowCredentials,
|
||||||
|
WorkflowRunner,
|
||||||
|
} from "../src";
|
||||||
|
|
||||||
|
import {
|
||||||
|
sep
|
||||||
|
} from 'path';
|
||||||
|
|
||||||
|
import {
|
||||||
|
diff,
|
||||||
|
diffString
|
||||||
|
} from 'json-diff';
|
||||||
|
import { ObjectID } from 'typeorm';
|
||||||
|
|
||||||
|
export class ExecuteAll extends Command {
|
||||||
|
static description = '\nExecutes all workflows once';
|
||||||
|
|
||||||
|
static examples = [
|
||||||
|
`$ n8n executeAll`,
|
||||||
|
`$ n8n executeAll --debug --output=/data/output.json`,
|
||||||
|
`$ n8n executeAll --ids=10,13,15`,
|
||||||
|
`$ n8n executeAll --snapshot=/data/snapshots`,
|
||||||
|
`$ n8n executeAll --compare=/data/previousExecutionData`,
|
||||||
|
];
|
||||||
|
|
||||||
|
static flags = {
|
||||||
|
help: flags.help({ char: 'h' }),
|
||||||
|
debug: flags.boolean({
|
||||||
|
description: 'Toggles on displaying all errors and debug messages.',
|
||||||
|
}),
|
||||||
|
ids: flags.string({
|
||||||
|
description: 'Specifies workflow IDs to get executed, separated by a comma.',
|
||||||
|
}),
|
||||||
|
output: flags.string({
|
||||||
|
description: 'Enable execution saving, You must inform an existing folder to save execution via this param',
|
||||||
|
}),
|
||||||
|
snapshot: flags.string({
|
||||||
|
description: 'Enables snapshot saving. You must inform an existing folder to save snapshots via this param.',
|
||||||
|
}),
|
||||||
|
compare: flags.string({
|
||||||
|
description: 'Compares current execution with an existing snapshot. You must inform an existing folder where the snapshots are saved.',
|
||||||
|
}),
|
||||||
|
};
|
||||||
|
|
||||||
|
|
||||||
|
async run() {
|
||||||
|
const { flags } = this.parse(ExecuteAll);
|
||||||
|
|
||||||
|
const debug = flags.debug !== undefined;
|
||||||
|
const ids: number[] = [];
|
||||||
|
|
||||||
|
if (flags.snapshot !== undefined) {
|
||||||
|
if (fs.existsSync(flags.snapshot)) {
|
||||||
|
if (!fs.lstatSync(flags.snapshot).isDirectory()) {
|
||||||
|
console.log(`The parameter --snapshot must be an existing directory`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
console.log(`The parameter --snapshot must be an existing directory`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (flags.compare !== undefined) {
|
||||||
|
if (fs.existsSync(flags.compare)) {
|
||||||
|
if (!fs.lstatSync(flags.compare).isDirectory()) {
|
||||||
|
console.log(`The parameter --compare must be an existing directory`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
console.log(`The parameter --compare must be an existing directory`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (flags.output !== undefined) {
|
||||||
|
if (fs.existsSync(flags.output)) {
|
||||||
|
if (fs.lstatSync(flags.output).isDirectory()) {
|
||||||
|
console.log(`The parameter --output must be a writable file`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (flags.ids !== undefined) {
|
||||||
|
const paramIds = flags.ids.split(',');
|
||||||
|
const re = /\d+/;
|
||||||
|
const matchedIDs = paramIds.filter(id => id.match(re) ).map(id => parseInt(id, 10));
|
||||||
|
|
||||||
|
if (matchedIDs.length === 0) {
|
||||||
|
console.log(`The parameter --ids must be a list of numeric IDs separated by a comma.`);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
ids.push(...matchedIDs);
|
||||||
|
|
||||||
|
}
|
||||||
|
// Start directly with the init of the database to improve startup time
|
||||||
|
const startDbInitPromise = Db.init();
|
||||||
|
|
||||||
|
// Load all node and credential types
|
||||||
|
const loadNodesAndCredentials = LoadNodesAndCredentials();
|
||||||
|
const loadNodesAndCredentialsPromise = loadNodesAndCredentials.init();
|
||||||
|
|
||||||
|
// Wait till the database is ready
|
||||||
|
await startDbInitPromise;
|
||||||
|
|
||||||
|
let allWorkflows;
|
||||||
|
|
||||||
|
if (ids.length !== 0) {
|
||||||
|
allWorkflows = await Db.collections!.Workflow!.findByIds(ids);
|
||||||
|
} else {
|
||||||
|
allWorkflows = await Db.collections!.Workflow!.find();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (debug === true) {
|
||||||
|
this.log(`Found ${allWorkflows.length} workflows to execute.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Make sure the settings exist
|
||||||
|
await UserSettings.prepareUserSettings();
|
||||||
|
|
||||||
|
// Wait till the n8n-packages have been read
|
||||||
|
await loadNodesAndCredentialsPromise;
|
||||||
|
|
||||||
|
// Load the credentials overwrites if any exist
|
||||||
|
const credentialsOverwrites = CredentialsOverwrites();
|
||||||
|
await credentialsOverwrites.init();
|
||||||
|
|
||||||
|
// Load all external hooks
|
||||||
|
const externalHooks = ExternalHooks();
|
||||||
|
await externalHooks.init();
|
||||||
|
|
||||||
|
// Add the found types to an instance other parts of the application can use
|
||||||
|
const nodeTypes = NodeTypes();
|
||||||
|
await nodeTypes.init(loadNodesAndCredentials.nodeTypes);
|
||||||
|
const credentialTypes = CredentialTypes();
|
||||||
|
await credentialTypes.init(loadNodesAndCredentials.credentialTypes);
|
||||||
|
|
||||||
|
const result:IResult = {
|
||||||
|
workflowsNumber:allWorkflows.length,
|
||||||
|
summary:{
|
||||||
|
failedExecutions:0,
|
||||||
|
succeededExecution:0,
|
||||||
|
exceptions:0,
|
||||||
|
},
|
||||||
|
coveredNodes:{},
|
||||||
|
executions:[],
|
||||||
|
};
|
||||||
|
|
||||||
|
// Check if the workflow contains the required "Start" node
|
||||||
|
// "requiredNodeTypes" are also defined in editor-ui/views/NodeView.vue
|
||||||
|
const requiredNodeTypes = ['n8n-nodes-base.start'];
|
||||||
|
for (let i = 0; i < allWorkflows.length; i++) {
|
||||||
|
const workflowData = allWorkflows[i];
|
||||||
|
|
||||||
|
if (debug === true) {
|
||||||
|
this.log(`Starting execution of workflow ID ${workflowData.id}.`);
|
||||||
|
}
|
||||||
|
|
||||||
|
const executionResult:IExecutionResult = {
|
||||||
|
workflowId: workflowData.id,
|
||||||
|
executionTime: 0,
|
||||||
|
finished: false,
|
||||||
|
nodes: workflowData.nodes.map(node => ({
|
||||||
|
name: node.name,
|
||||||
|
typeVersion: node.typeVersion,
|
||||||
|
type: node.type,
|
||||||
|
})),
|
||||||
|
error:'',
|
||||||
|
changes:'',
|
||||||
|
};
|
||||||
|
|
||||||
|
let startNode: INode | undefined= undefined;
|
||||||
|
for (const node of workflowData!.nodes) {
|
||||||
|
if (requiredNodeTypes.includes(node.type)) {
|
||||||
|
startNode = node;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (startNode === undefined) {
|
||||||
|
// If the workflow does not contain a start-node we can not know what
|
||||||
|
// should be executed and with which data to start.
|
||||||
|
executionResult.error = 'Workflow cannot be started as it does not contain a "Start" node.';
|
||||||
|
result.summary.failedExecutions++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
const credentials = await WorkflowCredentials(workflowData!.nodes);
|
||||||
|
|
||||||
|
const runData: IWorkflowExecutionDataProcess = {
|
||||||
|
credentials,
|
||||||
|
executionMode: 'cli',
|
||||||
|
startNodes: [startNode.name],
|
||||||
|
workflowData: workflowData!,
|
||||||
|
};
|
||||||
|
|
||||||
|
const workflowRunner = new WorkflowRunner();
|
||||||
|
const executionId = await workflowRunner.run(runData);
|
||||||
|
|
||||||
|
const activeExecutions = ActiveExecutions.getInstance();
|
||||||
|
const data = await activeExecutions.getPostExecutePromise(executionId);
|
||||||
|
|
||||||
|
if (data === undefined) {
|
||||||
|
executionResult.error = 'Workflow did not return any data.';
|
||||||
|
result.summary.failedExecutions++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
workflowData.nodes.forEach(node => {
|
||||||
|
result.coveredNodes[node.type] = (result.coveredNodes[node.type] || 0) +1;
|
||||||
|
});
|
||||||
|
executionResult.executionTime = (Date.parse(data.stoppedAt as unknown as string) - Date.parse(data.startedAt as unknown as string))/1000;
|
||||||
|
executionResult.finished = (data?.finished !== undefined) as boolean;
|
||||||
|
|
||||||
|
if (data.data.resultData.error) {
|
||||||
|
executionResult.error = data.data.resultData.error.message;
|
||||||
|
result.summary.failedExecutions++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
if (debug === true) {
|
||||||
|
this.log(JSON.stringify(data, null, 2));
|
||||||
|
console.log(data.data.resultData.error);
|
||||||
|
}
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
const serializedData = JSON.stringify(data, null, 2);
|
||||||
|
if (flags.compare === undefined){
|
||||||
|
result.summary.succeededExecution++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
} else {
|
||||||
|
const fileName = (flags.compare.endsWith(sep) ? flags.compare : flags.compare + sep) + `${workflowData.id}-snapshot.json`;
|
||||||
|
if (fs.existsSync(fileName) === true) {
|
||||||
|
|
||||||
|
const contents = fs.readFileSync(fileName, {encoding: 'utf-8'});
|
||||||
|
|
||||||
|
//@ts-ignore
|
||||||
|
const changes = diff(JSON.parse(contents), data, {keysOnly: true}); // types are outdated here
|
||||||
|
|
||||||
|
if (changes !== undefined) {
|
||||||
|
// we have structural changes. Report them.
|
||||||
|
executionResult.error = `Workflow may contain breaking changes`;
|
||||||
|
executionResult.changes = changes;
|
||||||
|
result.summary.failedExecutions++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
if (debug === true) {
|
||||||
|
// @ts-ignore
|
||||||
|
console.log('Detailed changes: ', diffString(JSON.parse(contents), data, undefined, {keysOnly: true}));
|
||||||
|
}
|
||||||
|
}else{
|
||||||
|
result.summary.succeededExecution++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
executionResult.error = 'Snapshot for not found.';
|
||||||
|
result.summary.failedExecutions++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Save snapshots only after comparing - this is to make sure we're updating
|
||||||
|
// After comparing to existing verion.
|
||||||
|
if (flags.snapshot !== undefined) {
|
||||||
|
const fileName = (flags.snapshot.endsWith(sep) ? flags.snapshot : flags.snapshot + sep) + `${workflowData.id}-snapshot.json`;
|
||||||
|
fs.writeFileSync(fileName,serializedData);
|
||||||
|
}
|
||||||
|
|
||||||
|
} catch (e) {
|
||||||
|
executionResult.error = 'Workflow failed to execute.';
|
||||||
|
result.summary.exceptions++;
|
||||||
|
result.executions.push(executionResult);
|
||||||
|
if (debug === true) {
|
||||||
|
console.error(e.message);
|
||||||
|
console.error(e.stack);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
if(flags.output !== undefined){
|
||||||
|
fs.writeFileSync(flags.output,JSON.stringify(result, null, 2));
|
||||||
|
console.log('Execution finished.');
|
||||||
|
console.log('Summary:');
|
||||||
|
console.log(`\tSuccess: ${result.summary.succeededExecution}`);
|
||||||
|
console.log(`\tFailures: ${result.summary.failedExecutions}`);
|
||||||
|
console.log(`\tExceptions: ${result.summary.exceptions}`);
|
||||||
|
console.log('\nNodes covered:');
|
||||||
|
Object.entries(result.coveredNodes).map(entry => {
|
||||||
|
console.log(`\t${entry[0]}: ${entry[1]}`);
|
||||||
|
});
|
||||||
|
console.log('\nCheck the JSON file for more details.');
|
||||||
|
}else{
|
||||||
|
console.log(JSON.stringify(result, null, 2));
|
||||||
|
}
|
||||||
|
if(result.summary.succeededExecution !== result.workflowsNumber){
|
||||||
|
this.exit(1);
|
||||||
|
}
|
||||||
|
this.exit(0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
interface IResult {
|
||||||
|
workflowsNumber:number;
|
||||||
|
summary:{
|
||||||
|
failedExecutions:number,
|
||||||
|
succeededExecution:number,
|
||||||
|
exceptions:number,
|
||||||
|
};
|
||||||
|
coveredNodes:{
|
||||||
|
[key:string]:number
|
||||||
|
};
|
||||||
|
executions:IExecutionResult[];
|
||||||
|
}
|
||||||
|
interface IExecutionResult{
|
||||||
|
workflowId: string | number | ObjectID;
|
||||||
|
executionTime: number;
|
||||||
|
finished: boolean;
|
||||||
|
nodes: Array<{
|
||||||
|
name: string,
|
||||||
|
typeVersion: number,
|
||||||
|
type: string,
|
||||||
|
}>;
|
||||||
|
error: string;
|
||||||
|
changes: string;
|
||||||
|
}
|
66
packages/cli/commands/list/workflow.ts
Normal file
66
packages/cli/commands/list/workflow.ts
Normal file
|
@ -0,0 +1,66 @@
|
||||||
|
import {
|
||||||
|
Command, flags,
|
||||||
|
} from '@oclif/command';
|
||||||
|
|
||||||
|
import {
|
||||||
|
IDataObject
|
||||||
|
} from 'n8n-workflow';
|
||||||
|
|
||||||
|
import {
|
||||||
|
Db,
|
||||||
|
} from "../../src";
|
||||||
|
|
||||||
|
|
||||||
|
export class ListWorkflowCommand extends Command {
|
||||||
|
static description = '\nList workflows';
|
||||||
|
|
||||||
|
static examples = [
|
||||||
|
`$ n8n list:workflow`,
|
||||||
|
`$ n8n list:workflow --active=true --onlyId`,
|
||||||
|
`$ n8n list:workflow --active=false`,
|
||||||
|
];
|
||||||
|
|
||||||
|
static flags = {
|
||||||
|
help: flags.help({ char: 'h' }),
|
||||||
|
active: flags.string({
|
||||||
|
description: 'Filters workflows by active status. Can be true or false',
|
||||||
|
}),
|
||||||
|
onlyId: flags.boolean({
|
||||||
|
description: 'Outputs workflow IDs only, one per line.',
|
||||||
|
}),
|
||||||
|
};
|
||||||
|
|
||||||
|
async run() {
|
||||||
|
const { flags } = this.parse(ListWorkflowCommand);
|
||||||
|
|
||||||
|
if (flags.active !== undefined && !['true', 'false'].includes(flags.active)) {
|
||||||
|
this.error('The --active flag has to be passed using true or false');
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
await Db.init();
|
||||||
|
|
||||||
|
const findQuery: IDataObject = {};
|
||||||
|
if (flags.active !== undefined) {
|
||||||
|
findQuery.active = flags.active === 'true';
|
||||||
|
}
|
||||||
|
|
||||||
|
const workflows = await Db.collections.Workflow!.find(findQuery);
|
||||||
|
if (flags.onlyId) {
|
||||||
|
workflows.map(workflow => console.log(workflow.id));
|
||||||
|
} else {
|
||||||
|
workflows.map(workflow => console.log(workflow.id + "|" + workflow.name));
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
} catch (e) {
|
||||||
|
console.error('\nGOT ERROR');
|
||||||
|
console.log('====================================');
|
||||||
|
console.error(e.message);
|
||||||
|
console.error(e.stack);
|
||||||
|
this.exit(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
this.exit();
|
||||||
|
}
|
||||||
|
}
|
|
@ -82,6 +82,7 @@
|
||||||
"@node-rs/bcrypt": "^1.2.0",
|
"@node-rs/bcrypt": "^1.2.0",
|
||||||
"@oclif/command": "^1.5.18",
|
"@oclif/command": "^1.5.18",
|
||||||
"@oclif/errors": "^1.2.2",
|
"@oclif/errors": "^1.2.2",
|
||||||
|
"@types/json-diff": "^0.5.1",
|
||||||
"@types/jsonwebtoken": "^8.3.4",
|
"@types/jsonwebtoken": "^8.3.4",
|
||||||
"basic-auth": "^2.0.1",
|
"basic-auth": "^2.0.1",
|
||||||
"body-parser": "^1.18.3",
|
"body-parser": "^1.18.3",
|
||||||
|
@ -98,6 +99,7 @@
|
||||||
"glob-promise": "^3.4.0",
|
"glob-promise": "^3.4.0",
|
||||||
"google-timezones-json": "^1.0.2",
|
"google-timezones-json": "^1.0.2",
|
||||||
"inquirer": "^7.0.1",
|
"inquirer": "^7.0.1",
|
||||||
|
"json-diff": "^0.5.4",
|
||||||
"jsonwebtoken": "^8.5.1",
|
"jsonwebtoken": "^8.5.1",
|
||||||
"jwks-rsa": "~1.12.1",
|
"jwks-rsa": "~1.12.1",
|
||||||
"localtunnel": "^2.0.0",
|
"localtunnel": "^2.0.0",
|
||||||
|
|
Loading…
Reference in a new issue