import type Bull from 'bull'; import { Service } from 'typedi'; import type { ExecutionError, IExecuteResponsePromiseData } from 'n8n-workflow'; import { ActiveExecutions } from '@/ActiveExecutions'; import { decodeWebhookResponse } from '@/helpers/decodeWebhookResponse'; import { getRedisClusterClient, getRedisClusterNodes, getRedisPrefix, getRedisStandardClient, } from './services/redis/RedisServiceHelper'; import type { RedisClientType } from './services/redis/RedisServiceBaseClasses'; import config from '@/config'; export type JobId = Bull.JobId; export type Job = Bull.Job; export type JobQueue = Bull.Queue; export interface JobData { executionId: string; loadStaticData: boolean; } export interface JobResponse { success: boolean; error?: ExecutionError; } export interface WebhookResponse { executionId: string; response: IExecuteResponsePromiseData; } @Service() export class Queue { private jobQueue: JobQueue; constructor(private activeExecutions: ActiveExecutions) {} async init() { const bullPrefix = config.getEnv('queue.bull.prefix'); const prefix = getRedisPrefix(bullPrefix); const clusterNodes = getRedisClusterNodes(); const usesRedisCluster = clusterNodes.length > 0; // eslint-disable-next-line @typescript-eslint/naming-convention const { default: Bull } = await import('bull'); // eslint-disable-next-line @typescript-eslint/naming-convention const { default: Redis } = await import('ioredis'); // Disabling ready check is necessary as it allows worker to // quickly reconnect to Redis if Redis crashes or is unreachable // for some time. With it enabled, worker might take minutes to realize // redis is back up and resume working. // More here: https://github.com/OptimalBits/bull/issues/890 this.jobQueue = new Bull('jobs', { prefix, createClient: (type, clientConfig) => usesRedisCluster ? getRedisClusterClient(Redis, clientConfig, (type + '(bull)') as RedisClientType) : getRedisStandardClient(Redis, clientConfig, (type + '(bull)') as RedisClientType), }); this.jobQueue.on('global:progress', (jobId, progress: WebhookResponse) => { this.activeExecutions.resolveResponsePromise( progress.executionId, decodeWebhookResponse(progress.response), ); }); } async add(jobData: JobData, jobOptions: object): Promise { return this.jobQueue.add(jobData, jobOptions); } async getJob(jobId: JobId): Promise { return this.jobQueue.getJob(jobId); } async getJobs(jobTypes: Bull.JobStatus[]): Promise { return this.jobQueue.getJobs(jobTypes); } async process(concurrency: number, fn: Bull.ProcessCallbackFunction): Promise { return this.jobQueue.process(concurrency, fn); } async ping(): Promise { return this.jobQueue.client.ping(); } async pause(isLocal?: boolean): Promise { return this.jobQueue.pause(isLocal); } getBullObjectInstance(): JobQueue { if (this.jobQueue === undefined) { // if queue is not initialized yet throw an error, since we do not want to hand around an undefined queue throw new Error('Queue is not initialized yet!'); } return this.jobQueue; } /** * * @param job A Job instance * @returns boolean true if we were able to securely stop the job */ async stopJob(job: Job): Promise { if (await job.isActive()) { // Job is already running so tell it to stop await job.progress(-1); return true; } // Job did not get started yet so remove from queue try { await job.remove(); return true; } catch (e) { await job.progress(-1); } return false; } }