mirror of
https://github.com/n8n-io/n8n.git
synced 2025-02-02 07:01:30 -08:00
fix(Postgres Node): Re-use connection pool across executions (#12346)
Co-authored-by: कारतोफ्फेलस्क्रिप्ट™ <aditya@netroy.in>
This commit is contained in:
parent
7b2630d1a0
commit
2ca37f5f7f
|
@ -102,7 +102,6 @@ export async function searchSchema(this: ILoadOptionsFunctions): Promise<INodeLi
|
||||||
name: s.schema_name as string,
|
name: s.schema_name as string,
|
||||||
value: s.schema_name as string,
|
value: s.schema_name as string,
|
||||||
}));
|
}));
|
||||||
await db.$pool.end();
|
|
||||||
return { results };
|
return { results };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -122,6 +121,5 @@ export async function searchTables(this: ILoadOptionsFunctions): Promise<INodeLi
|
||||||
name: s.table_name as string,
|
name: s.table_name as string,
|
||||||
value: s.table_name as string,
|
value: s.table_name as string,
|
||||||
}));
|
}));
|
||||||
await db.$pool.end();
|
|
||||||
return { results };
|
return { results };
|
||||||
}
|
}
|
||||||
|
|
|
@ -321,7 +321,6 @@ export class PostgresTrigger implements INodeType {
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
connection.client.removeListener('notification', onNotification);
|
connection.client.removeListener('notification', onNotification);
|
||||||
if (!db.$pool.ending) await db.$pool.end();
|
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -320,7 +320,6 @@ export class PostgresV1 implements INodeType {
|
||||||
|
|
||||||
const db = pgp(config);
|
const db = pgp(config);
|
||||||
await db.connect();
|
await db.connect();
|
||||||
await db.$pool.end();
|
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
return {
|
return {
|
||||||
status: 'Error',
|
status: 'Error',
|
||||||
|
@ -409,16 +408,12 @@ export class PostgresV1 implements INodeType {
|
||||||
|
|
||||||
returnItems = wrapData(updateItems);
|
returnItems = wrapData(updateItems);
|
||||||
} else {
|
} else {
|
||||||
await db.$pool.end();
|
|
||||||
throw new NodeOperationError(
|
throw new NodeOperationError(
|
||||||
this.getNode(),
|
this.getNode(),
|
||||||
`The operation "${operation}" is not supported!`,
|
`The operation "${operation}" is not supported!`,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// shuts down the connection pool associated with the db object to allow the process to finish
|
|
||||||
await db.$pool.end();
|
|
||||||
|
|
||||||
return [returnItems];
|
return [returnItems];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -35,7 +35,6 @@ export async function router(this: IExecuteFunctions): Promise<INodeExecutionDat
|
||||||
operation,
|
operation,
|
||||||
} as PostgresType;
|
} as PostgresType;
|
||||||
|
|
||||||
try {
|
|
||||||
switch (postgresNodeData.resource) {
|
switch (postgresNodeData.resource) {
|
||||||
case 'database':
|
case 'database':
|
||||||
returnData = await database[postgresNodeData.operation].execute.call(
|
returnData = await database[postgresNodeData.operation].execute.call(
|
||||||
|
@ -52,9 +51,6 @@ export async function router(this: IExecuteFunctions): Promise<INodeExecutionDat
|
||||||
`The operation "${operation}" is not supported!`,
|
`The operation "${operation}" is not supported!`,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
} finally {
|
|
||||||
if (!db.$pool.ending) await db.$pool.end();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (operation === 'select' && items.length > 1 && !node.executeOnce) {
|
if (operation === 'select' && items.length > 1 && !node.executeOnce) {
|
||||||
return new NodeExecutionOutput(
|
return new NodeExecutionOutput(
|
||||||
|
|
|
@ -9,7 +9,6 @@ export async function schemaSearch(this: ILoadOptionsFunctions): Promise<INodeLi
|
||||||
|
|
||||||
const { db } = await configurePostgres.call(this, credentials, options);
|
const { db } = await configurePostgres.call(this, credentials, options);
|
||||||
|
|
||||||
try {
|
|
||||||
const response = await db.any('SELECT schema_name FROM information_schema.schemata');
|
const response = await db.any('SELECT schema_name FROM information_schema.schemata');
|
||||||
|
|
||||||
return {
|
return {
|
||||||
|
@ -18,9 +17,6 @@ export async function schemaSearch(this: ILoadOptionsFunctions): Promise<INodeLi
|
||||||
value: schema.schema_name as string,
|
value: schema.schema_name as string,
|
||||||
})),
|
})),
|
||||||
};
|
};
|
||||||
} finally {
|
|
||||||
if (!db.$pool.ending) await db.$pool.end();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
export async function tableSearch(this: ILoadOptionsFunctions): Promise<INodeListSearchResult> {
|
export async function tableSearch(this: ILoadOptionsFunctions): Promise<INodeListSearchResult> {
|
||||||
const credentials = await this.getCredentials<PostgresNodeCredentials>('postgres');
|
const credentials = await this.getCredentials<PostgresNodeCredentials>('postgres');
|
||||||
|
@ -32,7 +28,6 @@ export async function tableSearch(this: ILoadOptionsFunctions): Promise<INodeLis
|
||||||
extractValue: true,
|
extractValue: true,
|
||||||
}) as string;
|
}) as string;
|
||||||
|
|
||||||
try {
|
|
||||||
const response = await db.any(
|
const response = await db.any(
|
||||||
'SELECT table_name FROM information_schema.tables WHERE table_schema=$1',
|
'SELECT table_name FROM information_schema.tables WHERE table_schema=$1',
|
||||||
[schema],
|
[schema],
|
||||||
|
@ -44,7 +39,4 @@ export async function tableSearch(this: ILoadOptionsFunctions): Promise<INodeLis
|
||||||
value: table.table_name as string,
|
value: table.table_name as string,
|
||||||
})),
|
})),
|
||||||
};
|
};
|
||||||
} finally {
|
|
||||||
if (!db.$pool.ending) await db.$pool.end();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -18,7 +18,6 @@ export async function getColumns(this: ILoadOptionsFunctions): Promise<INodeProp
|
||||||
extractValue: true,
|
extractValue: true,
|
||||||
}) as string;
|
}) as string;
|
||||||
|
|
||||||
try {
|
|
||||||
const columns = await getTableSchema(db, schema, table);
|
const columns = await getTableSchema(db, schema, table);
|
||||||
|
|
||||||
return columns.map((column) => ({
|
return columns.map((column) => ({
|
||||||
|
@ -26,9 +25,6 @@ export async function getColumns(this: ILoadOptionsFunctions): Promise<INodeProp
|
||||||
value: column.column_name,
|
value: column.column_name,
|
||||||
description: `Type: ${column.data_type.toUpperCase()}, Nullable: ${column.is_nullable}`,
|
description: `Type: ${column.data_type.toUpperCase()}, Nullable: ${column.is_nullable}`,
|
||||||
}));
|
}));
|
||||||
} finally {
|
|
||||||
if (!db.$pool.ending) await db.$pool.end();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export async function getColumnsMultiOptions(
|
export async function getColumnsMultiOptions(
|
||||||
|
|
|
@ -63,7 +63,6 @@ export async function getMappingColumns(
|
||||||
extractValue: true,
|
extractValue: true,
|
||||||
}) as string;
|
}) as string;
|
||||||
|
|
||||||
try {
|
|
||||||
const columns = await getTableSchema(db, schema, table, { getColumnsForResourceMapper: true });
|
const columns = await getTableSchema(db, schema, table, { getColumnsForResourceMapper: true });
|
||||||
const unique = operation === 'upsert' ? await uniqueColumns(db, table, schema) : [];
|
const unique = operation === 'upsert' ? await uniqueColumns(db, table, schema) : [];
|
||||||
const enumInfo = await getEnums(db);
|
const enumInfo = await getEnums(db);
|
||||||
|
@ -90,7 +89,4 @@ export async function getMappingColumns(
|
||||||
}),
|
}),
|
||||||
);
|
);
|
||||||
return { fields };
|
return { fields };
|
||||||
} finally {
|
|
||||||
if (!db.$pool.ending) await db.$pool.end();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -7,6 +7,7 @@ import type {
|
||||||
import { createServer, type AddressInfo } from 'node:net';
|
import { createServer, type AddressInfo } from 'node:net';
|
||||||
import pgPromise from 'pg-promise';
|
import pgPromise from 'pg-promise';
|
||||||
|
|
||||||
|
import { ConnectionPoolManager } from '@utils/connection-pool-manager';
|
||||||
import { LOCALHOST } from '@utils/constants';
|
import { LOCALHOST } from '@utils/constants';
|
||||||
import { formatPrivateKey } from '@utils/utilities';
|
import { formatPrivateKey } from '@utils/utilities';
|
||||||
|
|
||||||
|
@ -56,6 +57,9 @@ export async function configurePostgres(
|
||||||
credentials: PostgresNodeCredentials,
|
credentials: PostgresNodeCredentials,
|
||||||
options: PostgresNodeOptions = {},
|
options: PostgresNodeOptions = {},
|
||||||
): Promise<ConnectionsData> {
|
): Promise<ConnectionsData> {
|
||||||
|
const poolManager = ConnectionPoolManager.getInstance();
|
||||||
|
|
||||||
|
const fallBackHandler = async () => {
|
||||||
const pgp = pgPromise({
|
const pgp = pgPromise({
|
||||||
// prevent spam in console "WARNING: Creating a duplicate database object for the same connection."
|
// prevent spam in console "WARNING: Creating a duplicate database object for the same connection."
|
||||||
// duplicate connections created when auto loading parameters, they are closed immediately after, but several could be open at the same time
|
// duplicate connections created when auto loading parameters, they are closed immediately after, but several could be open at the same time
|
||||||
|
@ -90,6 +94,7 @@ export async function configurePostgres(
|
||||||
|
|
||||||
if (!credentials.sshTunnel) {
|
if (!credentials.sshTunnel) {
|
||||||
const db = pgp(dbConfig);
|
const db = pgp(dbConfig);
|
||||||
|
|
||||||
return { db, pgp };
|
return { db, pgp };
|
||||||
} else {
|
} else {
|
||||||
if (credentials.sshAuthenticateWith === 'privateKey' && credentials.privateKey) {
|
if (credentials.sshAuthenticateWith === 'privateKey' && credentials.privateKey) {
|
||||||
|
@ -172,4 +177,15 @@ export async function configurePostgres(
|
||||||
});
|
});
|
||||||
return { db, pgp };
|
return { db, pgp };
|
||||||
}
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
return await poolManager.getConnection({
|
||||||
|
credentials,
|
||||||
|
nodeType: 'postgres',
|
||||||
|
nodeVersion: options.nodeVersion as unknown as string,
|
||||||
|
fallBackHandler,
|
||||||
|
cleanUpHandler: async ({ db }) => {
|
||||||
|
await db.$pool.end();
|
||||||
|
},
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,179 @@
|
||||||
|
import { ConnectionPoolManager } from '@utils/connection-pool-manager';
|
||||||
|
|
||||||
|
const ttl = 5 * 60 * 1000;
|
||||||
|
const cleanUpInterval = 60 * 1000;
|
||||||
|
|
||||||
|
let cpm: ConnectionPoolManager;
|
||||||
|
|
||||||
|
beforeAll(() => {
|
||||||
|
jest.useFakeTimers();
|
||||||
|
cpm = ConnectionPoolManager.getInstance();
|
||||||
|
});
|
||||||
|
|
||||||
|
beforeEach(async () => {
|
||||||
|
await cpm.purgeConnections();
|
||||||
|
});
|
||||||
|
|
||||||
|
afterAll(() => {
|
||||||
|
cpm.onShutdown();
|
||||||
|
});
|
||||||
|
|
||||||
|
test('getInstance returns a singleton', () => {
|
||||||
|
const instance1 = ConnectionPoolManager.getInstance();
|
||||||
|
const instance2 = ConnectionPoolManager.getInstance();
|
||||||
|
|
||||||
|
expect(instance1).toBe(instance2);
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('getConnection', () => {
|
||||||
|
test('calls fallBackHandler only once and returns the first value', async () => {
|
||||||
|
// ARRANGE
|
||||||
|
const connectionType = {};
|
||||||
|
const fallBackHandler = jest.fn().mockResolvedValue(connectionType);
|
||||||
|
const cleanUpHandler = jest.fn();
|
||||||
|
const options = {
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '1',
|
||||||
|
fallBackHandler,
|
||||||
|
cleanUpHandler,
|
||||||
|
};
|
||||||
|
|
||||||
|
// ACT 1
|
||||||
|
const connection = await cpm.getConnection<string>(options);
|
||||||
|
|
||||||
|
// ASSERT 1
|
||||||
|
expect(fallBackHandler).toHaveBeenCalledTimes(1);
|
||||||
|
expect(connection).toBe(connectionType);
|
||||||
|
|
||||||
|
// ACT 2
|
||||||
|
const connection2 = await cpm.getConnection<string>(options);
|
||||||
|
// ASSERT 2
|
||||||
|
expect(fallBackHandler).toHaveBeenCalledTimes(1);
|
||||||
|
expect(connection2).toBe(connectionType);
|
||||||
|
});
|
||||||
|
|
||||||
|
test('creates different pools for different node versions', async () => {
|
||||||
|
// ARRANGE
|
||||||
|
const connectionType1 = {};
|
||||||
|
const fallBackHandler1 = jest.fn().mockResolvedValue(connectionType1);
|
||||||
|
const cleanUpHandler1 = jest.fn();
|
||||||
|
|
||||||
|
const connectionType2 = {};
|
||||||
|
const fallBackHandler2 = jest.fn().mockResolvedValue(connectionType2);
|
||||||
|
const cleanUpHandler2 = jest.fn();
|
||||||
|
|
||||||
|
// ACT 1
|
||||||
|
const connection1 = await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '1',
|
||||||
|
fallBackHandler: fallBackHandler1,
|
||||||
|
cleanUpHandler: cleanUpHandler1,
|
||||||
|
});
|
||||||
|
const connection2 = await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '2',
|
||||||
|
fallBackHandler: fallBackHandler2,
|
||||||
|
cleanUpHandler: cleanUpHandler2,
|
||||||
|
});
|
||||||
|
|
||||||
|
// ASSERT
|
||||||
|
expect(fallBackHandler1).toHaveBeenCalledTimes(1);
|
||||||
|
expect(connection1).toBe(connectionType1);
|
||||||
|
|
||||||
|
expect(fallBackHandler2).toHaveBeenCalledTimes(1);
|
||||||
|
expect(connection2).toBe(connectionType2);
|
||||||
|
|
||||||
|
expect(connection1).not.toBe(connection2);
|
||||||
|
});
|
||||||
|
|
||||||
|
test('calls cleanUpHandler after TTL expires', async () => {
|
||||||
|
// ARRANGE
|
||||||
|
const connectionType = {};
|
||||||
|
const fallBackHandler = jest.fn().mockResolvedValue(connectionType);
|
||||||
|
const cleanUpHandler = jest.fn();
|
||||||
|
await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '1',
|
||||||
|
fallBackHandler,
|
||||||
|
cleanUpHandler,
|
||||||
|
});
|
||||||
|
|
||||||
|
// ACT
|
||||||
|
jest.advanceTimersByTime(ttl + cleanUpInterval * 2);
|
||||||
|
|
||||||
|
// ASSERT
|
||||||
|
expect(cleanUpHandler).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
describe('onShutdown', () => {
|
||||||
|
test('calls all clean up handlers', async () => {
|
||||||
|
// ARRANGE
|
||||||
|
const connectionType1 = {};
|
||||||
|
const fallBackHandler1 = jest.fn().mockResolvedValue(connectionType1);
|
||||||
|
const cleanUpHandler1 = jest.fn();
|
||||||
|
await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '1',
|
||||||
|
fallBackHandler: fallBackHandler1,
|
||||||
|
cleanUpHandler: cleanUpHandler1,
|
||||||
|
});
|
||||||
|
|
||||||
|
const connectionType2 = {};
|
||||||
|
const fallBackHandler2 = jest.fn().mockResolvedValue(connectionType2);
|
||||||
|
const cleanUpHandler2 = jest.fn();
|
||||||
|
await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '2',
|
||||||
|
fallBackHandler: fallBackHandler2,
|
||||||
|
cleanUpHandler: cleanUpHandler2,
|
||||||
|
});
|
||||||
|
|
||||||
|
// ACT 1
|
||||||
|
cpm.onShutdown();
|
||||||
|
|
||||||
|
// ASSERT
|
||||||
|
expect(cleanUpHandler1).toHaveBeenCalledTimes(1);
|
||||||
|
expect(cleanUpHandler2).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
|
||||||
|
test('calls all clean up handlers when `exit` is emitted on process', async () => {
|
||||||
|
// ARRANGE
|
||||||
|
const connectionType1 = {};
|
||||||
|
const fallBackHandler1 = jest.fn().mockResolvedValue(connectionType1);
|
||||||
|
const cleanUpHandler1 = jest.fn();
|
||||||
|
await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '1',
|
||||||
|
fallBackHandler: fallBackHandler1,
|
||||||
|
cleanUpHandler: cleanUpHandler1,
|
||||||
|
});
|
||||||
|
|
||||||
|
const connectionType2 = {};
|
||||||
|
const fallBackHandler2 = jest.fn().mockResolvedValue(connectionType2);
|
||||||
|
const cleanUpHandler2 = jest.fn();
|
||||||
|
await cpm.getConnection<string>({
|
||||||
|
credentials: {},
|
||||||
|
nodeType: 'example',
|
||||||
|
nodeVersion: '2',
|
||||||
|
fallBackHandler: fallBackHandler2,
|
||||||
|
cleanUpHandler: cleanUpHandler2,
|
||||||
|
});
|
||||||
|
|
||||||
|
// ACT 1
|
||||||
|
// @ts-expect-error we're not supposed to emit `exit` so it's missing from
|
||||||
|
// the type definition
|
||||||
|
process.emit('exit');
|
||||||
|
|
||||||
|
// ASSERT
|
||||||
|
expect(cleanUpHandler1).toHaveBeenCalledTimes(1);
|
||||||
|
expect(cleanUpHandler2).toHaveBeenCalledTimes(1);
|
||||||
|
});
|
||||||
|
});
|
137
packages/nodes-base/utils/connection-pool-manager.ts
Normal file
137
packages/nodes-base/utils/connection-pool-manager.ts
Normal file
|
@ -0,0 +1,137 @@
|
||||||
|
import { createHash } from 'crypto';
|
||||||
|
|
||||||
|
let instance: ConnectionPoolManager;
|
||||||
|
|
||||||
|
// 5 minutes
|
||||||
|
const ttl = 5 * 60 * 1000;
|
||||||
|
|
||||||
|
// 1 minute
|
||||||
|
const cleanUpInterval = 60 * 1000;
|
||||||
|
|
||||||
|
type RegistrationOptions = {
|
||||||
|
credentials: unknown;
|
||||||
|
nodeType: string;
|
||||||
|
nodeVersion?: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
type GetConnectionOption<Pool> = RegistrationOptions & {
|
||||||
|
/** When a node requests for a connection pool, but none is available, this handler is called to create new instance of the pool, which then cached and re-used until it goes stale. */
|
||||||
|
fallBackHandler: () => Promise<Pool>;
|
||||||
|
|
||||||
|
/** When a pool hasn't been used in a while, or when the server is shutting down, this handler is invoked to close the pool */
|
||||||
|
cleanUpHandler: (pool: Pool) => Promise<void>;
|
||||||
|
};
|
||||||
|
|
||||||
|
type Registration<Pool> = {
|
||||||
|
/** This is an instance of a Connection Pool class, that gets reused across multiple executions */
|
||||||
|
pool: Pool;
|
||||||
|
|
||||||
|
/** @see GetConnectionOption['closeHandler'] */
|
||||||
|
cleanUpHandler: (pool: Pool) => Promise<void>;
|
||||||
|
|
||||||
|
/** We keep this timestamp to check if a pool hasn't been used in a while, and if it needs to be closed */
|
||||||
|
lastUsed: number;
|
||||||
|
};
|
||||||
|
|
||||||
|
export class ConnectionPoolManager {
|
||||||
|
/**
|
||||||
|
* Gets the singleton instance of the ConnectionPoolManager.
|
||||||
|
* Creates a new instance if one doesn't exist.
|
||||||
|
*/
|
||||||
|
static getInstance(): ConnectionPoolManager {
|
||||||
|
if (!instance) {
|
||||||
|
instance = new ConnectionPoolManager();
|
||||||
|
}
|
||||||
|
return instance;
|
||||||
|
}
|
||||||
|
|
||||||
|
private map = new Map<string, Registration<unknown>>();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Private constructor that initializes the connection pool manager.
|
||||||
|
* Sets up cleanup handlers for process exit and stale connections.
|
||||||
|
*/
|
||||||
|
private constructor() {
|
||||||
|
// Close all open pools when the process exits
|
||||||
|
process.on('exit', () => this.onShutdown());
|
||||||
|
|
||||||
|
// Regularly close stale pools
|
||||||
|
setInterval(() => this.cleanupStaleConnections(), cleanUpInterval);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Generates a unique key for connection pool identification.
|
||||||
|
* Hashes the credentials and node information for security.
|
||||||
|
*/
|
||||||
|
private makeKey({ credentials, nodeType, nodeVersion }: RegistrationOptions): string {
|
||||||
|
// The credential contains decrypted secrets, that's why we hash it.
|
||||||
|
return createHash('sha1')
|
||||||
|
.update(
|
||||||
|
JSON.stringify({
|
||||||
|
credentials,
|
||||||
|
nodeType,
|
||||||
|
nodeVersion,
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
.digest('base64');
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets or creates a connection pool for the given options.
|
||||||
|
* Updates the last used timestamp for existing connections.
|
||||||
|
*/
|
||||||
|
async getConnection<T>(options: GetConnectionOption<T>): Promise<T> {
|
||||||
|
const key = this.makeKey(options);
|
||||||
|
|
||||||
|
let value = this.map.get(key);
|
||||||
|
if (!value) {
|
||||||
|
value = {
|
||||||
|
pool: await options.fallBackHandler(),
|
||||||
|
cleanUpHandler: options.cleanUpHandler,
|
||||||
|
} as Registration<unknown>;
|
||||||
|
}
|
||||||
|
|
||||||
|
this.map.set(key, { ...value, lastUsed: Date.now() });
|
||||||
|
return value.pool as T;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Removes and cleans up connection pools that haven't been used within the
|
||||||
|
* TTL.
|
||||||
|
*/
|
||||||
|
private cleanupStaleConnections() {
|
||||||
|
const now = Date.now();
|
||||||
|
for (const [key, { cleanUpHandler, lastUsed, pool }] of this.map.entries()) {
|
||||||
|
if (now - lastUsed > ttl) {
|
||||||
|
void cleanUpHandler(pool);
|
||||||
|
this.map.delete(key);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Removes and cleans up all existing connection pools.
|
||||||
|
*/
|
||||||
|
async purgeConnections(): Promise<void> {
|
||||||
|
await Promise.all(
|
||||||
|
[...this.map.entries()].map(async ([key, value]) => {
|
||||||
|
this.map.delete(key);
|
||||||
|
|
||||||
|
return await value.cleanUpHandler(value.pool);
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Cleans up all connection pools when the process is shutting down.
|
||||||
|
* Does not wait for cleanup promises to resolve also does not remove the
|
||||||
|
* references from the pool.
|
||||||
|
*
|
||||||
|
* Only call this on process shutdown.
|
||||||
|
*/
|
||||||
|
onShutdown() {
|
||||||
|
for (const { cleanUpHandler, pool } of this.map.values()) {
|
||||||
|
void cleanUpHandler(pool);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in a new issue