2020-12-03 04:02:22 -08:00
|
|
|
import {
|
|
|
|
Kafka as apacheKafka,
|
|
|
|
KafkaConfig,
|
|
|
|
logLevel,
|
|
|
|
SASLOptions,
|
|
|
|
} from 'kafkajs';
|
|
|
|
|
2021-06-05 10:33:12 -07:00
|
|
|
import { SchemaRegistry } from '@kafkajs/confluent-schema-registry';
|
|
|
|
|
2020-12-03 04:02:22 -08:00
|
|
|
import {
|
|
|
|
ITriggerFunctions,
|
|
|
|
} from 'n8n-core';
|
|
|
|
|
|
|
|
import {
|
|
|
|
IDataObject,
|
|
|
|
INodeType,
|
|
|
|
INodeTypeDescription,
|
|
|
|
ITriggerResponse,
|
2021-04-16 09:33:36 -07:00
|
|
|
NodeOperationError,
|
2020-12-03 04:02:22 -08:00
|
|
|
} from 'n8n-workflow';
|
|
|
|
|
|
|
|
export class KafkaTrigger implements INodeType {
|
|
|
|
description: INodeTypeDescription = {
|
|
|
|
displayName: 'Kafka Trigger',
|
|
|
|
name: 'kafkaTrigger',
|
|
|
|
icon: 'file:kafka.svg',
|
|
|
|
group: ['trigger'],
|
|
|
|
version: 1,
|
|
|
|
description: 'Consume messages from a Kafka topic',
|
|
|
|
defaults: {
|
|
|
|
name: 'Kafka Trigger',
|
|
|
|
},
|
|
|
|
inputs: [],
|
|
|
|
outputs: ['main'],
|
|
|
|
credentials: [
|
|
|
|
{
|
|
|
|
name: 'kafka',
|
|
|
|
required: true,
|
|
|
|
},
|
|
|
|
],
|
|
|
|
properties: [
|
|
|
|
{
|
|
|
|
displayName: 'Topic',
|
|
|
|
name: 'topic',
|
|
|
|
type: 'string',
|
|
|
|
default: '',
|
|
|
|
required: true,
|
|
|
|
placeholder: 'topic-name',
|
2022-05-06 14:01:25 -07:00
|
|
|
description: 'Name of the queue of topic to consume from',
|
2020-12-03 04:02:22 -08:00
|
|
|
},
|
|
|
|
{
|
|
|
|
displayName: 'Group ID',
|
|
|
|
name: 'groupId',
|
|
|
|
type: 'string',
|
|
|
|
default: '',
|
|
|
|
required: true,
|
|
|
|
placeholder: 'n8n-kafka',
|
2022-05-06 14:01:25 -07:00
|
|
|
description: 'ID of the consumer group',
|
2020-12-03 04:02:22 -08:00
|
|
|
},
|
2021-06-05 14:48:00 -07:00
|
|
|
{
|
|
|
|
displayName: 'Use Schema Registry',
|
|
|
|
name: 'useSchemaRegistry',
|
|
|
|
type: 'boolean',
|
|
|
|
default: false,
|
2022-06-20 07:54:01 -07:00
|
|
|
description: 'Whether to use Confluent Schema Registry',
|
2021-06-05 14:48:00 -07:00
|
|
|
},
|
|
|
|
{
|
|
|
|
displayName: 'Schema Registry URL',
|
|
|
|
name: 'schemaRegistryUrl',
|
|
|
|
type: 'string',
|
|
|
|
required: true,
|
|
|
|
displayOptions: {
|
|
|
|
show: {
|
|
|
|
useSchemaRegistry: [
|
|
|
|
true,
|
|
|
|
],
|
|
|
|
},
|
|
|
|
},
|
2021-07-05 07:54:37 -07:00
|
|
|
placeholder: 'https://schema-registry-domain:8081',
|
|
|
|
default: '',
|
2022-05-06 14:01:25 -07:00
|
|
|
description: 'URL of the schema registry',
|
2021-06-05 14:48:00 -07:00
|
|
|
},
|
2020-12-03 04:02:22 -08:00
|
|
|
{
|
|
|
|
displayName: 'Options',
|
|
|
|
name: 'options',
|
|
|
|
type: 'collection',
|
|
|
|
default: {},
|
|
|
|
placeholder: 'Add Option',
|
|
|
|
options: [
|
|
|
|
{
|
|
|
|
displayName: 'Allow Topic Creation',
|
|
|
|
name: 'allowAutoTopicCreation',
|
|
|
|
type: 'boolean',
|
|
|
|
default: false,
|
2022-06-20 07:54:01 -07:00
|
|
|
description: 'Whether to allow sending message to a previously non exisiting topic',
|
2020-12-03 04:02:22 -08:00
|
|
|
},
|
2021-07-14 11:31:43 -07:00
|
|
|
{
|
2022-06-03 10:23:49 -07:00
|
|
|
displayName: 'Read Messages From Beginning',
|
2021-07-14 11:31:43 -07:00
|
|
|
name: 'fromBeginning',
|
|
|
|
type: 'boolean',
|
|
|
|
default: true,
|
2022-06-20 07:54:01 -07:00
|
|
|
description: 'Whether to read message from beginning',
|
2021-12-23 04:30:35 -08:00
|
|
|
},
|
2020-12-03 04:02:22 -08:00
|
|
|
{
|
|
|
|
displayName: 'JSON Parse Message',
|
|
|
|
name: 'jsonParseMessage',
|
|
|
|
type: 'boolean',
|
|
|
|
default: false,
|
2022-06-20 07:54:01 -07:00
|
|
|
description: 'Whether to try to parse the message to an object',
|
2020-12-03 04:02:22 -08:00
|
|
|
},
|
|
|
|
{
|
|
|
|
displayName: 'Only Message',
|
|
|
|
name: 'onlyMessage',
|
|
|
|
type: 'boolean',
|
|
|
|
displayOptions: {
|
|
|
|
show: {
|
|
|
|
jsonParseMessage: [
|
|
|
|
true,
|
|
|
|
],
|
|
|
|
},
|
|
|
|
},
|
|
|
|
default: false,
|
2022-06-20 07:54:01 -07:00
|
|
|
description: 'Whether to return only the message property',
|
2020-12-03 04:02:22 -08:00
|
|
|
},
|
|
|
|
{
|
|
|
|
displayName: 'Session Timeout',
|
|
|
|
name: 'sessionTimeout',
|
|
|
|
type: 'number',
|
|
|
|
default: 30000,
|
2022-05-06 14:01:25 -07:00
|
|
|
description: 'The time to await a response in ms',
|
2020-12-03 04:02:22 -08:00
|
|
|
},
|
2021-06-27 11:18:58 -07:00
|
|
|
{
|
2022-06-03 10:23:49 -07:00
|
|
|
displayName: 'Return Headers',
|
2021-06-27 11:18:58 -07:00
|
|
|
name: 'returnHeaders',
|
|
|
|
type: 'boolean',
|
|
|
|
default: false,
|
2022-06-20 07:54:01 -07:00
|
|
|
description: 'Whether to return the headers received from Kafka',
|
2021-06-27 11:18:58 -07:00
|
|
|
},
|
2020-12-03 04:02:22 -08:00
|
|
|
],
|
|
|
|
},
|
|
|
|
],
|
|
|
|
};
|
|
|
|
|
|
|
|
async trigger(this: ITriggerFunctions): Promise<ITriggerResponse> {
|
|
|
|
|
|
|
|
const topic = this.getNodeParameter('topic') as string;
|
|
|
|
|
|
|
|
const groupId = this.getNodeParameter('groupId') as string;
|
|
|
|
|
2022-04-14 23:00:47 -07:00
|
|
|
const credentials = await this.getCredentials('kafka');
|
2020-12-03 04:02:22 -08:00
|
|
|
|
|
|
|
const brokers = (credentials.brokers as string || '').split(',').map(item => item.trim()) as string[];
|
|
|
|
|
|
|
|
const clientId = credentials.clientId as string;
|
|
|
|
|
|
|
|
const ssl = credentials.ssl as boolean;
|
|
|
|
|
|
|
|
const config: KafkaConfig = {
|
|
|
|
clientId,
|
|
|
|
brokers,
|
|
|
|
ssl,
|
|
|
|
logLevel: logLevel.ERROR,
|
|
|
|
};
|
|
|
|
|
2021-03-11 05:25:02 -08:00
|
|
|
if (credentials.authentication === true) {
|
|
|
|
if(!(credentials.username && credentials.password)) {
|
2021-04-16 09:33:36 -07:00
|
|
|
throw new NodeOperationError(this.getNode(), 'Username and password are required for authentication');
|
2021-03-11 05:25:02 -08:00
|
|
|
}
|
2020-12-03 04:02:22 -08:00
|
|
|
config.sasl = {
|
|
|
|
username: credentials.username as string,
|
|
|
|
password: credentials.password as string,
|
2021-03-11 05:25:02 -08:00
|
|
|
mechanism: credentials.saslMechanism as string,
|
2020-12-03 04:02:22 -08:00
|
|
|
} as SASLOptions;
|
|
|
|
}
|
|
|
|
|
|
|
|
const kafka = new apacheKafka(config);
|
|
|
|
|
|
|
|
const consumer = kafka.consumer({ groupId });
|
|
|
|
|
|
|
|
await consumer.connect();
|
2021-12-23 04:30:35 -08:00
|
|
|
|
2021-07-14 11:31:43 -07:00
|
|
|
const options = this.getNodeParameter('options', {}) as IDataObject;
|
2020-12-03 04:02:22 -08:00
|
|
|
|
2021-07-14 11:31:43 -07:00
|
|
|
await consumer.subscribe({ topic, fromBeginning: (options.fromBeginning)? true : false });
|
2020-12-03 04:02:22 -08:00
|
|
|
|
|
|
|
const self = this;
|
|
|
|
|
2021-06-05 14:48:00 -07:00
|
|
|
const useSchemaRegistry = this.getNodeParameter('useSchemaRegistry', 0) as boolean;
|
|
|
|
|
|
|
|
const schemaRegistryUrl = this.getNodeParameter('schemaRegistryUrl', 0) as string;
|
|
|
|
|
2020-12-03 04:02:22 -08:00
|
|
|
const startConsumer = async () => {
|
|
|
|
await consumer.run({
|
|
|
|
eachMessage: async ({ topic, message }) => {
|
|
|
|
|
|
|
|
let data: IDataObject = {};
|
|
|
|
let value = message.value?.toString() as string;
|
|
|
|
|
|
|
|
if (options.jsonParseMessage) {
|
|
|
|
try {
|
|
|
|
value = JSON.parse(value);
|
2021-04-16 09:33:36 -07:00
|
|
|
} catch (error) { }
|
2020-12-03 04:02:22 -08:00
|
|
|
}
|
|
|
|
|
2021-06-05 17:22:32 -07:00
|
|
|
if (useSchemaRegistry) {
|
2021-06-05 10:33:12 -07:00
|
|
|
try {
|
2021-07-02 09:24:28 -07:00
|
|
|
const registry = new SchemaRegistry({ host: schemaRegistryUrl });
|
2021-06-05 10:33:12 -07:00
|
|
|
value = await registry.decode(message.value as Buffer);
|
|
|
|
} catch (error) { }
|
|
|
|
}
|
|
|
|
|
2022-01-08 02:47:24 -08:00
|
|
|
if (options.returnHeaders && message.headers) {
|
2021-06-27 11:18:58 -07:00
|
|
|
const headers: {[key: string]: string} = {};
|
2022-01-08 02:47:24 -08:00
|
|
|
for (const key of Object.keys(message.headers)) {
|
2021-06-27 11:18:58 -07:00
|
|
|
const header = message.headers[key];
|
|
|
|
headers[key] = header?.toString('utf8') || '';
|
|
|
|
}
|
|
|
|
|
|
|
|
data.headers = headers;
|
|
|
|
}
|
|
|
|
|
2020-12-03 04:02:22 -08:00
|
|
|
data.message = value;
|
|
|
|
data.topic = topic;
|
|
|
|
|
|
|
|
if (options.onlyMessage) {
|
|
|
|
//@ts-ignore
|
|
|
|
data = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
self.emit([self.helpers.returnJsonArray([data])]);
|
|
|
|
},
|
|
|
|
});
|
|
|
|
};
|
|
|
|
|
|
|
|
startConsumer();
|
|
|
|
|
|
|
|
// The "closeFunction" function gets called by n8n whenever
|
|
|
|
// the workflow gets deactivated and can so clean up.
|
|
|
|
async function closeFunction() {
|
|
|
|
await consumer.disconnect();
|
|
|
|
}
|
|
|
|
|
|
|
|
// The "manualTriggerFunction" function gets called by n8n
|
|
|
|
// when a user is in the workflow editor and starts the
|
|
|
|
// workflow manually. So the function has to make sure that
|
|
|
|
// the emit() gets called with similar data like when it
|
|
|
|
// would trigger by itself so that the user knows what data
|
|
|
|
// to expect.
|
|
|
|
async function manualTriggerFunction() {
|
|
|
|
startConsumer();
|
|
|
|
}
|
|
|
|
|
|
|
|
return {
|
|
|
|
closeFunction,
|
|
|
|
manualTriggerFunction,
|
|
|
|
};
|
|
|
|
}
|
|
|
|
}
|