import { Kafka as apacheKafka, KafkaConfig, logLevel, SASLOptions } from 'kafkajs'; import { SchemaRegistry } from '@kafkajs/confluent-schema-registry'; import { ITriggerFunctions } from 'n8n-core'; import { IDataObject, INodeType, INodeTypeDescription, ITriggerResponse, NodeOperationError, } from 'n8n-workflow'; export class KafkaTrigger implements INodeType { description: INodeTypeDescription = { displayName: 'Kafka Trigger', name: 'kafkaTrigger', icon: 'file:kafka.svg', group: ['trigger'], version: 1, description: 'Consume messages from a Kafka topic', defaults: { name: 'Kafka Trigger', }, inputs: [], outputs: ['main'], credentials: [ { name: 'kafka', required: true, }, ], properties: [ { displayName: 'Topic', name: 'topic', type: 'string', default: '', required: true, placeholder: 'topic-name', description: 'Name of the queue of topic to consume from', }, { displayName: 'Group ID', name: 'groupId', type: 'string', default: '', required: true, placeholder: 'n8n-kafka', description: 'ID of the consumer group', }, { displayName: 'Use Schema Registry', name: 'useSchemaRegistry', type: 'boolean', default: false, description: 'Whether to use Confluent Schema Registry', }, { displayName: 'Schema Registry URL', name: 'schemaRegistryUrl', type: 'string', required: true, displayOptions: { show: { useSchemaRegistry: [true], }, }, placeholder: 'https://schema-registry-domain:8081', default: '', description: 'URL of the schema registry', }, { displayName: 'Options', name: 'options', type: 'collection', default: {}, placeholder: 'Add Option', options: [ { displayName: 'Allow Topic Creation', name: 'allowAutoTopicCreation', type: 'boolean', default: false, description: 'Whether to allow sending message to a previously non exisiting topic', }, { displayName: 'Auto Commit Threshold', name: 'autoCommitThreshold', type: 'number', default: 0, description: 'The consumer will commit offsets after resolving a given number of messages', }, { displayName: 'Auto Commit Interval', name: 'autoCommitInterval', type: 'number', default: 0, description: 'The consumer will commit offsets after a given period, for example, five seconds', hint: 'Value in milliseconds', }, { displayName: 'Heartbeat Interval', name: 'heartbeatInterval', type: 'number', default: 3000, description: "Heartbeats are used to ensure that the consumer's session stays active", hint: 'The value must be set lower than Session Timeout', }, { displayName: 'Max Number of Requests', name: 'maxInFlightRequests', type: 'number', default: 1, description: 'Max number of requests that may be in progress at any time. If falsey then no limit.', }, { displayName: 'Read Messages From Beginning', name: 'fromBeginning', type: 'boolean', default: true, description: 'Whether to read message from beginning', }, { displayName: 'JSON Parse Message', name: 'jsonParseMessage', type: 'boolean', default: false, description: 'Whether to try to parse the message to an object', }, { displayName: 'Only Message', name: 'onlyMessage', type: 'boolean', displayOptions: { show: { jsonParseMessage: [true], }, }, default: false, description: 'Whether to return only the message property', }, { displayName: 'Return Headers', name: 'returnHeaders', type: 'boolean', default: false, description: 'Whether to return the headers received from Kafka', }, { displayName: 'Session Timeout', name: 'sessionTimeout', type: 'number', default: 30000, description: 'The time to await a response in ms', hint: 'Value in milliseconds', }, ], }, ], }; async trigger(this: ITriggerFunctions): Promise { const topic = this.getNodeParameter('topic') as string; const groupId = this.getNodeParameter('groupId') as string; const credentials = await this.getCredentials('kafka'); const brokers = ((credentials.brokers as string) || '').split(',').map((item) => item.trim()); const clientId = credentials.clientId as string; const ssl = credentials.ssl as boolean; const config: KafkaConfig = { clientId, brokers, ssl, logLevel: logLevel.ERROR, }; if (credentials.authentication === true) { if (!(credentials.username && credentials.password)) { throw new NodeOperationError( this.getNode(), 'Username and password are required for authentication', ); } config.sasl = { username: credentials.username as string, password: credentials.password as string, mechanism: credentials.saslMechanism as string, } as SASLOptions; } const kafka = new apacheKafka(config); const maxInFlightRequests = ( this.getNodeParameter('options.maxInFlightRequests', null) === 0 ? null : this.getNodeParameter('options.maxInFlightRequests', null) ) as number; const consumer = kafka.consumer({ groupId, maxInFlightRequests, sessionTimeout: this.getNodeParameter('options.sessionTimeout', 30000) as number, heartbeatInterval: this.getNodeParameter('options.heartbeatInterval', 3000) as number, }); await consumer.connect(); const options = this.getNodeParameter('options', {}) as IDataObject; await consumer.subscribe({ topic, fromBeginning: options.fromBeginning ? true : false }); const useSchemaRegistry = this.getNodeParameter('useSchemaRegistry', 0) as boolean; const schemaRegistryUrl = this.getNodeParameter('schemaRegistryUrl', 0) as string; const startConsumer = async () => { await consumer.run({ autoCommitInterval: (options.autoCommitInterval as number) || null, autoCommitThreshold: (options.autoCommitThreshold as number) || null, eachMessage: async ({ topic: messageTopic, message }) => { let data: IDataObject = {}; let value = message.value?.toString() as string; if (options.jsonParseMessage) { try { value = JSON.parse(value); } catch (error) {} } if (useSchemaRegistry) { try { const registry = new SchemaRegistry({ host: schemaRegistryUrl }); value = await registry.decode(message.value as Buffer); } catch (error) {} } if (options.returnHeaders && message.headers) { const headers: { [key: string]: string } = {}; for (const key of Object.keys(message.headers)) { const header = message.headers[key]; headers[key] = header?.toString('utf8') ?? ''; } data.headers = headers; } data.message = value; data.topic = messageTopic; if (options.onlyMessage) { //@ts-ignore data = value; } this.emit([this.helpers.returnJsonArray([data])]); }, }); }; await startConsumer(); // The "closeFunction" function gets called by n8n whenever // the workflow gets deactivated and can so clean up. async function closeFunction() { await consumer.disconnect(); } // The "manualTriggerFunction" function gets called by n8n // when a user is in the workflow editor and starts the // workflow manually. So the function has to make sure that // the emit() gets called with similar data like when it // would trigger by itself so that the user knows what data // to expect. async function manualTriggerFunction() { await startConsumer(); } return { closeFunction, manualTriggerFunction, }; } }