import type { IExecuteFunctions } from 'n8n-core';
import type {
INodeExecutionData,
INodeType,
INodeTypeDescription,
IPairedItemData,
} from 'n8n-workflow';
import { deepCopy } from 'n8n-workflow';
export class SplitInBatches implements INodeType {
description: INodeTypeDescription = {
displayName: 'Split In Batches',
name: 'splitInBatches',
icon: 'fa:th-large',
group: ['organization'],
version: 1,
description: 'Split data into batches and iterate over each batch',
defaults: {
name: 'Split In Batches',
color: '#007755',
},
inputs: ['main'],
outputs: ['main'],
properties: [
{
displayName:
'You may not need this node — n8n nodes automatically run once for each input item. More info',
name: 'splitInBatchesNotice',
type: 'notice',
default: '',
},
{
displayName: 'Batch Size',
name: 'batchSize',
type: 'number',
typeOptions: {
minValue: 1,
},
default: 10,
description: 'The number of items to return with each call',
},
{
displayName: 'Options',
name: 'options',
type: 'collection',
placeholder: 'Add Option',
default: {},
options: [
{
displayName: 'Reset',
name: 'reset',
type: 'boolean',
default: false,
description:
'Whether the node will be reset and so with the current input-data newly initialized',
},
],
},
],
};
async execute(this: IExecuteFunctions): Promise {
// Get the input data and create a new array so that we can remove
// items without a problem
const items = this.getInputData().slice();
const nodeContext = this.getContext('node');
const batchSize = this.getNodeParameter('batchSize', 0) as number;
const returnItems: INodeExecutionData[] = [];
const options = this.getNodeParameter('options', 0, {});
if (nodeContext.items === undefined || options.reset === true) {
// Is the first time the node runs
const sourceData = this.getInputSourceData();
nodeContext.currentRunIndex = 0;
nodeContext.maxRunIndex = Math.ceil(items.length / batchSize);
nodeContext.sourceData = deepCopy(sourceData);
// Get the items which should be returned
returnItems.push.apply(returnItems, items.splice(0, batchSize));
// Set the other items to be saved in the context to return at later runs
nodeContext.items = [...items];
} else {
// The node has been called before. So return the next batch of items.
nodeContext.currentRunIndex += 1;
returnItems.push.apply(
returnItems,
(nodeContext.items as INodeExecutionData[]).splice(0, batchSize),
);
const addSourceOverwrite = (pairedItem: IPairedItemData | number): IPairedItemData => {
if (typeof pairedItem === 'number') {
return {
item: pairedItem,
sourceOverwrite: nodeContext.sourceData,
};
}
return {
...pairedItem,
sourceOverwrite: nodeContext.sourceData,
};
};
function getPairedItemInformation(
item: INodeExecutionData,
): IPairedItemData | IPairedItemData[] {
if (item.pairedItem === undefined) {
return {
item: 0,
sourceOverwrite: nodeContext.sourceData,
};
}
if (Array.isArray(item.pairedItem)) {
return item.pairedItem.map(addSourceOverwrite);
}
return addSourceOverwrite(item.pairedItem);
}
returnItems.map((item) => {
item.pairedItem = getPairedItemInformation(item);
});
}
nodeContext.noItemsLeft = nodeContext.items.length === 0;
if (returnItems.length === 0) {
// No data left to return so stop execution of the branch
return null;
}
return [returnItems];
}
}