mirror of
https://github.com/n8n-io/n8n.git
synced 2025-01-11 12:57:29 -08:00
feat(Kafka Node): Overhaul Kafka and KafkaTrigger nodes
This commit is contained in:
parent
2e61722002
commit
e044b78049
37
packages/nodes-base/nodes/Kafka/GenericFunctions.ts
Normal file
37
packages/nodes-base/nodes/Kafka/GenericFunctions.ts
Normal file
|
@ -0,0 +1,37 @@
|
|||
import { logLevel, SASLOptions, type KafkaConfig } from 'kafkajs';
|
||||
import type { KafkaCredential } from './types';
|
||||
import {
|
||||
type ICredentialTestFunctions,
|
||||
NodeOperationError,
|
||||
type ITriggerFunctions,
|
||||
} from 'n8n-workflow';
|
||||
|
||||
export const getConnectionConfig = (
|
||||
context: ITriggerFunctions | ICredentialTestFunctions,
|
||||
credentials: KafkaCredential,
|
||||
): KafkaConfig => {
|
||||
const brokers = ((credentials.brokers as string) || '').split(',').map((item) => item.trim());
|
||||
|
||||
const config: KafkaConfig = {
|
||||
brokers,
|
||||
clientId: credentials.clientId,
|
||||
ssl: credentials.ssl,
|
||||
logLevel: logLevel.ERROR,
|
||||
};
|
||||
|
||||
if (credentials.authentication) {
|
||||
if (!(credentials.username && credentials.password)) {
|
||||
throw new NodeOperationError(
|
||||
context.getNode(),
|
||||
'Username and password are required for authentication',
|
||||
);
|
||||
}
|
||||
config.sasl = {
|
||||
username: credentials.username,
|
||||
password: credentials.password,
|
||||
mechanism: credentials.saslMechanism,
|
||||
} as SASLOptions;
|
||||
}
|
||||
|
||||
return config;
|
||||
};
|
|
@ -1,11 +1,8 @@
|
|||
import type { KafkaConfig, SASLOptions, TopicMessages } from 'kafkajs';
|
||||
import { CompressionTypes, Kafka as apacheKafka } from 'kafkajs';
|
||||
|
||||
import { SchemaRegistry } from '@kafkajs/confluent-schema-registry';
|
||||
|
||||
import type {
|
||||
IExecuteFunctions,
|
||||
ICredentialDataDecryptedObject,
|
||||
ICredentialsDecrypted,
|
||||
ICredentialTestFunctions,
|
||||
IDataObject,
|
||||
|
@ -14,8 +11,11 @@ import type {
|
|||
INodeType,
|
||||
INodeTypeDescription,
|
||||
} from 'n8n-workflow';
|
||||
import { ApplicationError, NodeConnectionType, NodeOperationError } from 'n8n-workflow';
|
||||
import { NodeConnectionType, NodeOperationError } from 'n8n-workflow';
|
||||
|
||||
import { generatePairedItemData } from '../../utils/utilities';
|
||||
import { KafkaCredential } from './types';
|
||||
import { getConnectionConfig } from './GenericFunctions';
|
||||
|
||||
export class Kafka implements INodeType {
|
||||
description: INodeTypeDescription = {
|
||||
|
@ -212,34 +212,9 @@ export class Kafka implements INodeType {
|
|||
this: ICredentialTestFunctions,
|
||||
credential: ICredentialsDecrypted,
|
||||
): Promise<INodeCredentialTestResult> {
|
||||
const credentials = credential.data as ICredentialDataDecryptedObject;
|
||||
const credentials = credential.data as KafkaCredential;
|
||||
try {
|
||||
const brokers = ((credentials.brokers as string) || '')
|
||||
.split(',')
|
||||
.map((item) => item.trim());
|
||||
|
||||
const clientId = credentials.clientId as string;
|
||||
|
||||
const ssl = credentials.ssl as boolean;
|
||||
|
||||
const config: KafkaConfig = {
|
||||
clientId,
|
||||
brokers,
|
||||
ssl,
|
||||
};
|
||||
if (credentials.authentication === true) {
|
||||
if (!(credentials.username && credentials.password)) {
|
||||
throw new ApplicationError('Username and password are required for authentication', {
|
||||
level: 'warning',
|
||||
});
|
||||
}
|
||||
config.sasl = {
|
||||
username: credentials.username as string,
|
||||
password: credentials.password as string,
|
||||
mechanism: credentials.saslMechanism as string,
|
||||
} as SASLOptions;
|
||||
}
|
||||
|
||||
const config = getConnectionConfig(this, credentials);
|
||||
const kafka = new apacheKafka(config);
|
||||
|
||||
await kafka.admin().connect();
|
||||
|
|
|
@ -1,19 +1,20 @@
|
|||
import type { KafkaConfig, SASLOptions } from 'kafkajs';
|
||||
import { Kafka as apacheKafka, logLevel } from 'kafkajs';
|
||||
|
||||
import type { KafkaMessage } from 'kafkajs';
|
||||
import { Kafka as apacheKafka } from 'kafkajs';
|
||||
import { SchemaRegistry } from '@kafkajs/confluent-schema-registry';
|
||||
|
||||
import type {
|
||||
ITriggerFunctions,
|
||||
IDataObject,
|
||||
INodeType,
|
||||
INodeTypeDescription,
|
||||
ITriggerResponse,
|
||||
IRun,
|
||||
INodeExecutionData,
|
||||
} from 'n8n-workflow';
|
||||
import { NodeConnectionType, NodeOperationError } from 'n8n-workflow';
|
||||
import { Node, NodeConnectionType } from 'n8n-workflow';
|
||||
|
||||
export class KafkaTrigger implements INodeType {
|
||||
import type { KafkaCredential, TriggerNodeOptions } from './types';
|
||||
import { getConnectionConfig } from './GenericFunctions';
|
||||
|
||||
export class KafkaTrigger extends Node {
|
||||
description: INodeTypeDescription = {
|
||||
displayName: 'Kafka Trigger',
|
||||
name: 'kafkaTrigger',
|
||||
|
@ -178,75 +179,13 @@ export class KafkaTrigger implements INodeType {
|
|||
],
|
||||
};
|
||||
|
||||
async trigger(this: ITriggerFunctions): Promise<ITriggerResponse> {
|
||||
const topic = this.getNodeParameter('topic') as string;
|
||||
|
||||
const groupId = this.getNodeParameter('groupId') as string;
|
||||
|
||||
const credentials = await this.getCredentials('kafka');
|
||||
|
||||
const brokers = ((credentials.brokers as string) || '').split(',').map((item) => item.trim());
|
||||
|
||||
const clientId = credentials.clientId as string;
|
||||
|
||||
const ssl = credentials.ssl as boolean;
|
||||
|
||||
const options = this.getNodeParameter('options', {}) as IDataObject;
|
||||
|
||||
options.nodeVersion = this.getNode().typeVersion;
|
||||
|
||||
const config: KafkaConfig = {
|
||||
clientId,
|
||||
brokers,
|
||||
ssl,
|
||||
logLevel: logLevel.ERROR,
|
||||
};
|
||||
|
||||
if (credentials.authentication === true) {
|
||||
if (!(credentials.username && credentials.password)) {
|
||||
throw new NodeOperationError(
|
||||
this.getNode(),
|
||||
'Username and password are required for authentication',
|
||||
);
|
||||
}
|
||||
config.sasl = {
|
||||
username: credentials.username as string,
|
||||
password: credentials.password as string,
|
||||
mechanism: credentials.saslMechanism as string,
|
||||
} as SASLOptions;
|
||||
}
|
||||
|
||||
const kafka = new apacheKafka(config);
|
||||
|
||||
const maxInFlightRequests = (
|
||||
this.getNodeParameter('options.maxInFlightRequests', null) === 0
|
||||
? null
|
||||
: this.getNodeParameter('options.maxInFlightRequests', null)
|
||||
) as number;
|
||||
|
||||
const consumer = kafka.consumer({
|
||||
groupId,
|
||||
maxInFlightRequests,
|
||||
sessionTimeout: this.getNodeParameter('options.sessionTimeout', 30000) as number,
|
||||
heartbeatInterval: this.getNodeParameter('options.heartbeatInterval', 3000) as number,
|
||||
});
|
||||
|
||||
const parallelProcessing = options.parallelProcessing as boolean;
|
||||
|
||||
await consumer.connect();
|
||||
|
||||
await consumer.subscribe({ topic, fromBeginning: options.fromBeginning ? true : false });
|
||||
|
||||
const useSchemaRegistry = this.getNodeParameter('useSchemaRegistry', 0) as boolean;
|
||||
|
||||
const schemaRegistryUrl = this.getNodeParameter('schemaRegistryUrl', 0) as string;
|
||||
|
||||
const startConsumer = async () => {
|
||||
await consumer.run({
|
||||
autoCommitInterval: (options.autoCommitInterval as number) || null,
|
||||
autoCommitThreshold: (options.autoCommitThreshold as number) || null,
|
||||
eachMessage: async ({ topic: messageTopic, message }) => {
|
||||
let data: IDataObject = {};
|
||||
async parsePayload(
|
||||
message: KafkaMessage,
|
||||
messageTopic: string,
|
||||
options: TriggerNodeOptions,
|
||||
context: ITriggerFunctions,
|
||||
): Promise<INodeExecutionData[][]> {
|
||||
const data: IDataObject = {};
|
||||
let value = message.value?.toString() as string;
|
||||
|
||||
if (options.jsonParseMessage) {
|
||||
|
@ -255,13 +194,19 @@ export class KafkaTrigger implements INodeType {
|
|||
} catch (error) {}
|
||||
}
|
||||
|
||||
const useSchemaRegistry = context.getNodeParameter('useSchemaRegistry', 0) as boolean;
|
||||
if (useSchemaRegistry) {
|
||||
const schemaRegistryUrl = context.getNodeParameter('schemaRegistryUrl', 0) as string;
|
||||
try {
|
||||
const registry = new SchemaRegistry({ host: schemaRegistryUrl });
|
||||
value = await registry.decode(message.value as Buffer);
|
||||
} catch (error) {}
|
||||
}
|
||||
|
||||
if (options.onlyMessage) {
|
||||
return [context.helpers.returnJsonArray([value as unknown as IDataObject])];
|
||||
}
|
||||
|
||||
if (options.returnHeaders && message.headers) {
|
||||
const headers: { [key: string]: string } = {};
|
||||
for (const key of Object.keys(message.headers)) {
|
||||
|
@ -275,42 +220,58 @@ export class KafkaTrigger implements INodeType {
|
|||
data.message = value;
|
||||
data.topic = messageTopic;
|
||||
|
||||
if (options.onlyMessage) {
|
||||
//@ts-ignore
|
||||
data = value;
|
||||
}
|
||||
let responsePromise = undefined;
|
||||
if (!parallelProcessing && (options.nodeVersion as number) > 1) {
|
||||
responsePromise = this.helpers.createDeferredPromise<IRun>();
|
||||
this.emit([this.helpers.returnJsonArray([data])], undefined, responsePromise);
|
||||
} else {
|
||||
this.emit([this.helpers.returnJsonArray([data])]);
|
||||
}
|
||||
if (responsePromise) {
|
||||
await responsePromise.promise;
|
||||
return [context.helpers.returnJsonArray([data])];
|
||||
}
|
||||
|
||||
async trigger(context: ITriggerFunctions): Promise<ITriggerResponse> {
|
||||
const topic = context.getNodeParameter('topic') as string;
|
||||
const groupId = context.getNodeParameter('groupId') as string;
|
||||
|
||||
const options = context.getNodeParameter('options', {}) as TriggerNodeOptions;
|
||||
const nodeVersion = context.getNode().typeVersion;
|
||||
|
||||
const credentials = await context.getCredentials<KafkaCredential>('kafka');
|
||||
const config = getConnectionConfig(context, credentials);
|
||||
const kafka = new apacheKafka(config);
|
||||
|
||||
const consumer = kafka.consumer({
|
||||
groupId,
|
||||
maxInFlightRequests: options.maxInFlightRequests,
|
||||
sessionTimeout: options.sessionTimeout ?? 30000,
|
||||
heartbeatInterval: options.heartbeatInterval ?? 3000,
|
||||
});
|
||||
|
||||
const startConsumer = async () => {
|
||||
await consumer.connect();
|
||||
await consumer.subscribe({ topic, fromBeginning: options.fromBeginning ? true : false });
|
||||
|
||||
await consumer.run({
|
||||
autoCommitInterval: options.autoCommitInterval || null,
|
||||
autoCommitThreshold: options.autoCommitThreshold || null,
|
||||
eachMessage: async ({ topic: messageTopic, message }) => {
|
||||
const data = await this.parsePayload(message, messageTopic, options, context);
|
||||
const donePromise =
|
||||
!options.parallelProcessing && nodeVersion > 1 && context.getMode() === 'trigger'
|
||||
? context.helpers.createDeferredPromise<IRun>()
|
||||
: undefined;
|
||||
context.emit(data, undefined, donePromise);
|
||||
await donePromise?.promise;
|
||||
},
|
||||
});
|
||||
};
|
||||
|
||||
await startConsumer();
|
||||
|
||||
// The "closeFunction" function gets called by n8n whenever
|
||||
// the workflow gets deactivated and can so clean up.
|
||||
async function closeFunction() {
|
||||
await consumer.disconnect();
|
||||
}
|
||||
|
||||
// The "manualTriggerFunction" function gets called by n8n
|
||||
// when a user is in the workflow editor and starts the
|
||||
// workflow manually. So the function has to make sure that
|
||||
// the emit() gets called with similar data like when it
|
||||
// would trigger by itself so that the user knows what data
|
||||
// to expect.
|
||||
async function manualTriggerFunction() {
|
||||
await startConsumer();
|
||||
}
|
||||
|
||||
if (context.getMode() === 'trigger') {
|
||||
await startConsumer();
|
||||
}
|
||||
|
||||
async function closeFunction() {
|
||||
await consumer.disconnect();
|
||||
}
|
||||
|
||||
return {
|
||||
closeFunction,
|
||||
manualTriggerFunction,
|
||||
|
|
32
packages/nodes-base/nodes/Kafka/types.ts
Normal file
32
packages/nodes-base/nodes/Kafka/types.ts
Normal file
|
@ -0,0 +1,32 @@
|
|||
import type { SASLMechanism } from 'kafkajs';
|
||||
|
||||
export type KafkaCredential = {
|
||||
clientId: string;
|
||||
brokers: string;
|
||||
ssl: boolean;
|
||||
authentication: boolean;
|
||||
} & (
|
||||
| {
|
||||
authentication: true;
|
||||
username: string;
|
||||
password: string;
|
||||
saslMechanism: SASLMechanism;
|
||||
}
|
||||
| {
|
||||
authentication: false;
|
||||
}
|
||||
);
|
||||
|
||||
export interface TriggerNodeOptions {
|
||||
allowAutoTopicCreation: boolean;
|
||||
autoCommitThreshold: number;
|
||||
autoCommitInterval: number;
|
||||
heartbeatInterval: number;
|
||||
maxInFlightRequests: number;
|
||||
fromBeginning: boolean;
|
||||
jsonParseMessage: boolean;
|
||||
parallelProcessing: boolean;
|
||||
onlyMessage: boolean;
|
||||
returnHeaders: boolean;
|
||||
sessionTimeout: number;
|
||||
}
|
|
@ -1661,6 +1661,7 @@ export abstract class Node {
|
|||
execute?(context: IExecuteFunctions): Promise<INodeExecutionData[][]>;
|
||||
webhook?(context: IWebhookFunctions): Promise<IWebhookResponseData>;
|
||||
poll?(context: IPollFunctions): Promise<INodeExecutionData[][] | null>;
|
||||
trigger?(context: ITriggerFunctions): Promise<ITriggerResponse | undefined>;
|
||||
}
|
||||
|
||||
export interface IVersionedNodeType {
|
||||
|
|
|
@ -1148,7 +1148,11 @@ export class Workflow {
|
|||
if (mode === 'manual') {
|
||||
// In manual mode we do not just start the trigger function we also
|
||||
// want to be able to get informed as soon as the first data got emitted
|
||||
const triggerResponse = await nodeType.trigger.call(triggerFunctions);
|
||||
|
||||
const triggerResponse =
|
||||
nodeType instanceof Node
|
||||
? await nodeType.trigger(triggerFunctions)
|
||||
: await nodeType.trigger.call(triggerFunctions);
|
||||
|
||||
// Add the manual trigger response which resolves when the first time data got emitted
|
||||
triggerResponse!.manualTriggerResponse = new Promise((resolve, reject) => {
|
||||
|
@ -1197,7 +1201,9 @@ export class Workflow {
|
|||
return triggerResponse;
|
||||
}
|
||||
// In all other modes simply start the trigger
|
||||
return await nodeType.trigger.call(triggerFunctions);
|
||||
return nodeType instanceof Node
|
||||
? await nodeType.trigger(triggerFunctions)
|
||||
: await nodeType.trigger.call(triggerFunctions);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
Loading…
Reference in a new issue