mirror of
https://github.com/ToolJet/ToolJet
synced 2026-04-30 09:57:19 +00:00
* GUI mode abstraction and GUI mode for postgresql * CSS Issue and Query response fixes * GUI Mode for MySQL * Mysql changes similar to pgsql * Mysql GUI mode bug fixes * pagination for tables in mysql gui mode has been reapplied * styles issues for drop down has been fixed * Merge branch 'lts-3.16' into feature/mysql-gui-mode --------- Co-authored-by: Adish M <adish.madhu@gmail.com> Co-authored-by: gsmithun4 <gsmithun4@gmail.com>
925 lines
33 KiB
TypeScript
925 lines
33 KiB
TypeScript
import knex, { Knex } from 'knex';
|
|
import {
|
|
cacheConnectionWithConfiguration,
|
|
generateSourceOptionsHash,
|
|
getCachedConnection,
|
|
ConnectionTestResult,
|
|
QueryService,
|
|
QueryResult,
|
|
QueryError,
|
|
createQueryBuilder,
|
|
User,
|
|
App,
|
|
} from '@tooljet-plugins/common';
|
|
import { SourceOptions, QueryOptions } from './types';
|
|
import { isEmpty } from '@tooljet-plugins/common';
|
|
import { Client } from 'ssh2';
|
|
|
|
function createSSHStream(sourceOptions: SourceOptions): Promise<{ client: Client; stream: NodeJS.ReadWriteStream }> {
|
|
return new Promise((resolve, reject) => {
|
|
const sshClient = new Client();
|
|
|
|
sshClient.on('ready', () => {
|
|
sshClient.forwardOut('127.0.0.1', 0, sourceOptions.host, Number(sourceOptions.port), (err, stream) => {
|
|
if (err) {
|
|
sshClient.end();
|
|
return reject(err);
|
|
}
|
|
stream.on('error', (streamErr) => {
|
|
console.error('SSH stream error (suppressed):', streamErr.message);
|
|
});
|
|
|
|
stream.on('close', () => {
|
|
setImmediate(() => {
|
|
try {
|
|
if (sshClient) {
|
|
sshClient.destroy();
|
|
}
|
|
} catch (e) {
|
|
console.error('Error closing SSH client (suppressed):', e.message);
|
|
}
|
|
});
|
|
});
|
|
|
|
resolve({ client: sshClient, stream });
|
|
});
|
|
});
|
|
|
|
sshClient.on('error', (err) => {
|
|
reject(err);
|
|
});
|
|
|
|
sshClient.on('end', () => {
|
|
console.log('SSH connection ended');
|
|
});
|
|
|
|
sshClient.on('close', () => {
|
|
console.log('SSH connection closed');
|
|
});
|
|
|
|
sshClient.connect({
|
|
host: sourceOptions.ssh_host,
|
|
port: sourceOptions.ssh_port || 22,
|
|
username: sourceOptions.ssh_username,
|
|
...(sourceOptions.ssh_auth_type === 'password'
|
|
? { password: sourceOptions.ssh_password }
|
|
: {
|
|
privateKey: sourceOptions.ssh_private_key,
|
|
passphrase: sourceOptions.ssh_passphrase,
|
|
}),
|
|
readyTimeout: 20000,
|
|
keepaliveInterval: 10000,
|
|
});
|
|
});
|
|
}
|
|
|
|
export default class MysqlQueryService implements QueryService {
|
|
private static _instance: MysqlQueryService;
|
|
private STATEMENT_TIMEOUT;
|
|
|
|
constructor() {
|
|
this.STATEMENT_TIMEOUT =
|
|
process.env?.PLUGINS_SQL_DB_STATEMENT_TIMEOUT && !isNaN(Number(process.env?.PLUGINS_SQL_DB_STATEMENT_TIMEOUT))
|
|
? Number(process.env.PLUGINS_SQL_DB_STATEMENT_TIMEOUT)
|
|
: 120000;
|
|
|
|
if (MysqlQueryService._instance) {
|
|
return MysqlQueryService._instance;
|
|
}
|
|
MysqlQueryService._instance = this;
|
|
return MysqlQueryService._instance;
|
|
}
|
|
|
|
async run(
|
|
sourceOptions: SourceOptions,
|
|
queryOptions: QueryOptions,
|
|
dataSourceId: string,
|
|
dataSourceUpdatedAt: string
|
|
): Promise<QueryResult> {
|
|
let checkCache, knexInstance;
|
|
if (sourceOptions['allow_dynamic_connection_parameters']) {
|
|
if (sourceOptions.connection_type === 'hostname') {
|
|
sourceOptions['host'] = queryOptions['host'] ? queryOptions['host'] : sourceOptions['host'];
|
|
sourceOptions['database'] = queryOptions['database'] ? queryOptions['database'] : sourceOptions['database'];
|
|
} else if (sourceOptions.connection_type === 'socket_path') {
|
|
sourceOptions['database'] = queryOptions['database'] ? queryOptions['database'] : sourceOptions['database'];
|
|
}
|
|
}
|
|
|
|
try {
|
|
// If dynamic connection parameters is toggled on - We don't cache the connection also destroy the connection created.
|
|
checkCache = sourceOptions['allow_dynamic_connection_parameters'] ? false : true;
|
|
knexInstance = await this.getConnection(sourceOptions, {}, checkCache, dataSourceId, dataSourceUpdatedAt);
|
|
|
|
switch (queryOptions.mode) {
|
|
case 'sql':
|
|
return await this.handleRawQuery(knexInstance, queryOptions);
|
|
case 'gui':
|
|
return await this.handleGuiQuery(knexInstance, queryOptions);
|
|
default:
|
|
throw new Error("Invalid query mode. Must be either 'sql' or 'gui'.");
|
|
}
|
|
} catch (err) {
|
|
const errorMessage = err?.message || 'An unknown error occurred';
|
|
const errorDetails: any = {};
|
|
|
|
if (err instanceof Error) {
|
|
const mysqlError = err as any;
|
|
const { code, errno, sqlMessage, sqlState } = mysqlError;
|
|
|
|
errorDetails.code = code || null;
|
|
errorDetails.errno = errno || null;
|
|
errorDetails.sqlMessage = sqlMessage || null;
|
|
errorDetails.sqlState = sqlState || null;
|
|
}
|
|
|
|
throw new QueryError('Query could not be completed', errorMessage, errorDetails);
|
|
} finally {
|
|
if (!checkCache) await knexInstance.destroy();
|
|
}
|
|
}
|
|
|
|
async testConnection(sourceOptions: SourceOptions): Promise<ConnectionTestResult> {
|
|
let knexInstance;
|
|
try {
|
|
knexInstance = await this.getConnection(sourceOptions, {}, false);
|
|
await knexInstance.raw('select @@version;').timeout(this.STATEMENT_TIMEOUT);
|
|
if (knexInstance) {
|
|
try {
|
|
await knexInstance.destroy();
|
|
// eslint-disable-next-line no-empty
|
|
} catch (_) {}
|
|
}
|
|
return { status: 'ok' };
|
|
} catch (err) {
|
|
const errorMessage = err?.message || 'An unknown error occurred';
|
|
const errorDetails: any = {};
|
|
|
|
if (err instanceof Error) {
|
|
const mysqlError = err as any;
|
|
const { code, errno, sqlMessage, sqlState } = mysqlError;
|
|
|
|
errorDetails.code = code || null;
|
|
errorDetails.errno = errno || null;
|
|
errorDetails.sqlMessage = sqlMessage || null;
|
|
errorDetails.sqlState = sqlState || null;
|
|
}
|
|
|
|
throw new QueryError('Connection test failed', errorMessage, errorDetails);
|
|
}
|
|
}
|
|
|
|
private async handleGuiQuery(knexInstance: Knex, queryOptions: QueryOptions): Promise<QueryResult> {
|
|
const { operation, table } = queryOptions;
|
|
const queryBuilder = createQueryBuilder('mysql');
|
|
|
|
switch (operation) {
|
|
case 'list_rows': {
|
|
const { list_rows, limit, offset } = queryOptions;
|
|
const { where_filters, order_filters, aggregates, group_by } = list_rows || {};
|
|
const { query, params } = queryBuilder.listRows(table, {
|
|
where_filters,
|
|
order_filters,
|
|
aggregates,
|
|
group_by,
|
|
limit,
|
|
offset,
|
|
}) as { query: string; params: unknown[] };
|
|
const rows = await this.executeParameterizedQuery(knexInstance, query, params);
|
|
return { status: 'ok', data: rows };
|
|
}
|
|
|
|
case 'create_row': {
|
|
const { columns } = queryOptions.create_row || {};
|
|
const { query, params } = queryBuilder.createRow(table, undefined, columns) as {
|
|
query: string;
|
|
params: unknown[];
|
|
};
|
|
const insertedRows = await this.executeCreateRow(
|
|
knexInstance,
|
|
queryBuilder,
|
|
table,
|
|
undefined,
|
|
columns,
|
|
query,
|
|
params
|
|
);
|
|
return { status: 'ok', data: insertedRows };
|
|
}
|
|
|
|
case 'update_rows': {
|
|
const { allow_multiple_updates, zero_records_as_success } = queryOptions;
|
|
const { columns, where_filters } = queryOptions.update_rows || {};
|
|
const hasAtLeastOneFilter = Object.values(where_filters || {}).some((f: any) => f?.column?.trim());
|
|
if (!hasAtLeastOneFilter) {
|
|
throw new Error('At least one filter condition is required.');
|
|
}
|
|
const { query, params } = queryBuilder.updateRows(table, { columns, where_filters }) as {
|
|
query: string;
|
|
params: unknown[];
|
|
};
|
|
const { query: selectQuery, params: selectParams } = queryBuilder.listRows(table, {
|
|
where_filters,
|
|
}) as { query: string; params: unknown[] };
|
|
const rows = await this.executeWriteQuery(knexInstance, query, params, selectQuery, selectParams, {
|
|
allow_multiple_updates,
|
|
zero_records_as_success,
|
|
operationLabel: 'updated',
|
|
});
|
|
return { status: 'ok', data: rows };
|
|
}
|
|
|
|
case 'upsert_rows': {
|
|
const { primary_key_columns, allow_multiple_updates, zero_records_as_success } = queryOptions;
|
|
const { columns } = queryOptions.upsert_rows || {};
|
|
const { query, params } = queryBuilder.upsertRows(table, { primary_key_columns, columns }) as {
|
|
query: string;
|
|
params: unknown[];
|
|
};
|
|
const normalizedPrimaryKeys = Array.isArray(primary_key_columns)
|
|
? primary_key_columns.filter(Boolean)
|
|
: typeof primary_key_columns === 'string' && primary_key_columns.trim()
|
|
? [primary_key_columns.trim()]
|
|
: [];
|
|
const columnEntries = Object.values(columns || {});
|
|
const whereFiltersForSelect = Object.fromEntries(
|
|
normalizedPrimaryKeys
|
|
.map((keyColumn) => {
|
|
const keyEntry = columnEntries.find((entry: any) => entry.column === keyColumn) as any;
|
|
if (!keyEntry) return null;
|
|
return [keyColumn, { column: keyColumn, operator: 'eq', value: keyEntry.value }];
|
|
})
|
|
.filter((entry): entry is [string, any] => entry !== null)
|
|
);
|
|
const { query: selectQuery, params: selectParams } = queryBuilder.listRows(table, {
|
|
where_filters: whereFiltersForSelect,
|
|
}) as { query: string; params: unknown[] };
|
|
const rows = await this.executeWriteQuery(knexInstance, query, params, selectQuery, selectParams, {
|
|
allow_multiple_updates,
|
|
zero_records_as_success,
|
|
operationLabel: 'upserted',
|
|
});
|
|
return { status: 'ok', data: rows };
|
|
}
|
|
|
|
case 'delete_rows': {
|
|
const { limit, zero_records_as_success } = queryOptions;
|
|
const { where_filters } = queryOptions.delete_rows || {};
|
|
const hasAtLeastOneFilter = Object.values(where_filters || {}).some((f: any) => f?.column?.trim());
|
|
const hasLimit = limit != null && limit !== '';
|
|
if (!hasAtLeastOneFilter && !hasLimit) {
|
|
throw new Error(
|
|
'Delete requires at least one filter condition or a limit to prevent deleting all rows unintentionally.'
|
|
);
|
|
}
|
|
const { query, params } = queryBuilder.deleteRows(table, { where_filters, limit }) as {
|
|
query: string;
|
|
params: unknown[];
|
|
};
|
|
const [okPacket] = await knexInstance.raw(query, params as any[]).timeout(this.STATEMENT_TIMEOUT);
|
|
const deletedRecords = okPacket.affectedRows || 0;
|
|
|
|
if (zero_records_as_success === false && deletedRecords === 0) {
|
|
throw new Error('No rows were deleted.');
|
|
}
|
|
|
|
return { status: 'ok', data: { deletedRecords } };
|
|
}
|
|
|
|
case 'bulk_insert': {
|
|
const { records } = queryOptions;
|
|
const { query, params } = queryBuilder.bulkInsert(table, { rows_insert: records }) as {
|
|
query: string;
|
|
params: unknown[];
|
|
};
|
|
const [okPacket] = await knexInstance.raw(query, params as any[]).timeout(this.STATEMENT_TIMEOUT);
|
|
return { status: 'ok', data: okPacket };
|
|
}
|
|
|
|
case 'bulk_update_pkey': {
|
|
const { primary_key_columns, records } = queryOptions;
|
|
const { queries } = queryBuilder.bulkUpdateWithPrimaryKey(table, {
|
|
primary_key: primary_key_columns,
|
|
rows_update: records,
|
|
}) as { queries: { query: string; params: unknown[] }[] };
|
|
const data = await this.executeBulkQueriesInTransaction(knexInstance, queries);
|
|
return { status: 'ok', data, bulk_update_status: 'success' } as unknown as QueryResult;
|
|
}
|
|
|
|
case 'bulk_upsert_pkey': {
|
|
const { primary_key_columns, records } = queryOptions;
|
|
const { queries } = queryBuilder.bulkUpsertWithPrimaryKey(table, {
|
|
primary_key: primary_key_columns,
|
|
row_upsert: records,
|
|
}) as { queries: { query: string; params: unknown[] }[] };
|
|
const data = await this.executeBulkQueriesInTransaction(knexInstance, queries);
|
|
return { status: 'ok', data, bulk_upsert_status: 'success' } as unknown as QueryResult;
|
|
}
|
|
|
|
default:
|
|
throw new Error(`Unsupported GUI operation: "${operation}"`);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Executes a SELECT query and returns the rows array.
|
|
* MySQL raw() returns [rows, fields], so we destructure and return rows.
|
|
*/
|
|
private async executeParameterizedQuery(knexInstance: Knex, query: string, params: unknown[]): Promise<unknown[]> {
|
|
const [rows] = await knexInstance.raw(query, params as any[]).timeout(this.STATEMENT_TIMEOUT);
|
|
return rows;
|
|
}
|
|
|
|
/**
|
|
* Executes INSERT and then SELECTs back the inserted row(s) using the inserted column values as WHERE conditions.
|
|
* Wrapped in a transaction to ensure consistency.
|
|
*/
|
|
private async executeCreateRow(
|
|
knexInstance: Knex,
|
|
queryBuilder: ReturnType<typeof createQueryBuilder>,
|
|
tableName: string,
|
|
schema: string | undefined | null,
|
|
columns: Record<string, any> | undefined | null,
|
|
insertQuery: string,
|
|
insertParams: unknown[]
|
|
): Promise<unknown[]> {
|
|
return knexInstance.transaction(async (transaction) => {
|
|
await transaction.raw(insertQuery, insertParams as any[]);
|
|
|
|
const columnEntries = Object.values(columns || {});
|
|
if (columnEntries.length === 0) return [];
|
|
|
|
const whereFilters = Object.fromEntries(
|
|
columnEntries.map((entry: any) => [entry.column, { column: entry.column, operator: 'eq', value: entry.value }])
|
|
);
|
|
const { query: selectQuery, params: selectParams } = queryBuilder.listRows(tableName, {
|
|
schema,
|
|
where_filters: whereFilters,
|
|
}) as { query: string; params: unknown[] };
|
|
|
|
const [rows] = await transaction.raw(selectQuery, selectParams as any[]);
|
|
return rows;
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Executes a write query (UPDATE or upsert) and checks affectedRows constraints.
|
|
* When constraints apply, wraps in a transaction so violations automatically roll back.
|
|
* After the write, SELECTs and returns the affected rows.
|
|
*/
|
|
private async executeWriteQuery(
|
|
knexInstance: Knex,
|
|
writeQuery: string,
|
|
writeParams: unknown[],
|
|
selectQuery: string,
|
|
selectParams: unknown[],
|
|
options: { allow_multiple_updates?: boolean; zero_records_as_success?: boolean; operationLabel: string }
|
|
): Promise<unknown[]> {
|
|
const { allow_multiple_updates, zero_records_as_success, operationLabel } = options;
|
|
const hasConstraints = allow_multiple_updates === false || zero_records_as_success === false;
|
|
|
|
if (!hasConstraints) {
|
|
await knexInstance.raw(writeQuery, writeParams as any[]).timeout(this.STATEMENT_TIMEOUT);
|
|
const [rows] = await knexInstance.raw(selectQuery, selectParams as any[]).timeout(this.STATEMENT_TIMEOUT);
|
|
return rows;
|
|
}
|
|
|
|
return knexInstance.transaction(async (transaction) => {
|
|
const [okPacket] = await transaction.raw(writeQuery, writeParams as any[]);
|
|
const affectedRows = okPacket.affectedRows || 0;
|
|
|
|
if (allow_multiple_updates === false && affectedRows > 1) {
|
|
throw new Error(
|
|
'Query matches more than one row. Enable "Allow this Query to modify multiple rows" to permit this.'
|
|
);
|
|
}
|
|
|
|
if (zero_records_as_success === false && affectedRows === 0) {
|
|
throw new Error(`No rows were ${operationLabel}.`);
|
|
}
|
|
|
|
const [rows] = await transaction.raw(selectQuery, selectParams as any[]);
|
|
return rows;
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Executes multiple queries sequentially inside a single transaction.
|
|
* Returns the OkPacket from each query as an array.
|
|
*/
|
|
private async executeBulkQueriesInTransaction(
|
|
knexInstance: Knex,
|
|
queries: { query: string; params: unknown[] }[]
|
|
): Promise<unknown[]> {
|
|
const allResults: unknown[] = [];
|
|
await knexInstance.transaction(async (transaction) => {
|
|
for (const { query, params } of queries) {
|
|
const [okPacket] = await transaction.raw(query, params as any[]);
|
|
allResults.push(okPacket);
|
|
}
|
|
});
|
|
return allResults;
|
|
}
|
|
|
|
private async handleRawQuery(knexInstance: Knex, queryOptions: QueryOptions): Promise<QueryResult> {
|
|
const { query, query_params } = queryOptions;
|
|
const queryParams = query_params || [];
|
|
const sanitizedQueryParams: Record<string, any> = Object.fromEntries(queryParams.filter(([key]) => !isEmpty(key)));
|
|
const result = await this.executeQuery(knexInstance, query, sanitizedQueryParams);
|
|
|
|
return { status: 'ok', data: result[0] };
|
|
}
|
|
|
|
private async executeQuery(knexInstance: Knex, query: string, sanitizedQueryParams: Record<string, any> = {}) {
|
|
if (isEmpty(query)) throw new Error('Query is empty');
|
|
|
|
const result = await knexInstance.raw(query, sanitizedQueryParams).timeout(this.STATEMENT_TIMEOUT);
|
|
return result;
|
|
}
|
|
|
|
private connectionOptions(sourceOptions: SourceOptions) {
|
|
const _connectionOptions = (sourceOptions?.connection_options || []).filter((o) => o.some((e) => !isEmpty(e)));
|
|
const connectionOptions = Object.fromEntries(_connectionOptions);
|
|
Object.keys(connectionOptions).forEach((key) =>
|
|
connectionOptions[key] === '' ? delete connectionOptions[key] : {}
|
|
);
|
|
return connectionOptions;
|
|
}
|
|
|
|
private parseConnectionString(connectionString: string): Partial<SourceOptions> {
|
|
const parsed: Partial<SourceOptions & { params?: Record<string, string>; socketPath?: string }> = {};
|
|
|
|
if (!connectionString) return parsed;
|
|
|
|
const trimmed = connectionString.trim();
|
|
|
|
try {
|
|
const url = new URL(trimmed);
|
|
if (url.protocol && url.protocol.startsWith('mysql')) {
|
|
if (url.username) parsed.username = decodeURIComponent(url.username);
|
|
if (url.password) parsed.password = decodeURIComponent(url.password);
|
|
if (url.hostname) parsed.host = decodeURIComponent(url.hostname);
|
|
|
|
if (url.port) {
|
|
const p = parseInt(url.port, 10);
|
|
if (!isNaN(p)) parsed.port = String(p);
|
|
}
|
|
|
|
if (url.pathname && url.pathname !== '/') {
|
|
parsed.database = decodeURIComponent(url.pathname.slice(1));
|
|
}
|
|
|
|
const queryParams: Record<string, string> = {};
|
|
url.searchParams.forEach((v, k) => {
|
|
queryParams[k] = v;
|
|
});
|
|
|
|
if (Object.keys(queryParams).length) {
|
|
parsed.params = queryParams;
|
|
}
|
|
|
|
const socket = url.searchParams.get('socket') || url.searchParams.get('socketPath');
|
|
if (socket) parsed.socket_path = socket;
|
|
const sslParam =
|
|
url.searchParams.get('ssl') || url.searchParams.get('sslmode') || url.searchParams.get('ssl_mode');
|
|
if (sslParam !== null) {
|
|
parsed.ssl_enabled = ['true', '1', 'require', 'required', 'verify-ca', 'verify-full'].includes(
|
|
sslParam.toLowerCase()
|
|
) as any;
|
|
}
|
|
|
|
return parsed;
|
|
}
|
|
// eslint-disable-next-line no-empty
|
|
} catch (_) {}
|
|
|
|
const regex =
|
|
// eslint-disable-next-line no-useless-escape
|
|
/^mysql:\/\/(?:([^:@\/?#]+)(?::([^@\/?#]*))?@)?(\[[^\]]+\]|[^:\/?#]+)?(?::(\d+))?(?:\/([^?]*))?(?:\?(.*))?$/;
|
|
|
|
const m = trimmed.match(regex);
|
|
|
|
if (m) {
|
|
const user = m[1];
|
|
const pass = m[2];
|
|
const hostRaw = m[3];
|
|
const portRaw = m[4];
|
|
const db = m[5];
|
|
const query = m[6];
|
|
|
|
if (user) {
|
|
try {
|
|
parsed.username = decodeURIComponent(user);
|
|
} catch {
|
|
parsed.username = user;
|
|
}
|
|
}
|
|
if (pass !== undefined) {
|
|
try {
|
|
parsed.password = decodeURIComponent(pass);
|
|
} catch {
|
|
parsed.password = pass;
|
|
}
|
|
}
|
|
|
|
if (hostRaw) {
|
|
let host = hostRaw;
|
|
if (host.startsWith('[') && host.endsWith(']')) host = host.slice(1, -1);
|
|
try {
|
|
parsed.host = decodeURIComponent(host);
|
|
} catch {
|
|
parsed.host = host;
|
|
}
|
|
}
|
|
|
|
if (portRaw) {
|
|
const p = parseInt(portRaw, 10);
|
|
if (!isNaN(p)) parsed.port = String(p);
|
|
}
|
|
|
|
if (db) {
|
|
try {
|
|
parsed.database = decodeURIComponent(db);
|
|
} catch {
|
|
parsed.database = db;
|
|
}
|
|
}
|
|
|
|
if (query) {
|
|
const q: Record<string, string> = {};
|
|
query.split('&').forEach((part) => {
|
|
if (!part) return;
|
|
const idx = part.indexOf('=');
|
|
if (idx === -1) {
|
|
q[decodeURIComponent(part)] = '';
|
|
} else {
|
|
const k = part.slice(0, idx);
|
|
const v = part.slice(idx + 1);
|
|
try {
|
|
q[decodeURIComponent(k)] = decodeURIComponent(v);
|
|
} catch {
|
|
q[k] = v;
|
|
}
|
|
}
|
|
});
|
|
|
|
if (Object.keys(q).length) parsed.params = q;
|
|
if (q['socket'] || q['socketPath']) {
|
|
parsed.socket_path = q['socket'] || q['socketPath'];
|
|
const sslMode = q['ssl'] || q['sslmode'] || q['ssl_mode'];
|
|
if (sslMode !== undefined) {
|
|
parsed.ssl_enabled = ['true', '1', 'require', 'required', 'verify-ca', 'verify-full'].includes(
|
|
sslMode.toLowerCase()
|
|
) as any;
|
|
}
|
|
}
|
|
}
|
|
|
|
return parsed;
|
|
}
|
|
|
|
// Strategy 3: Key-value pair parsing (Server=host;Port=3306;...)
|
|
// Uses join('=') to correctly handle values that contain '=' characters
|
|
const pairs = trimmed.split(';').filter((p) => p.trim());
|
|
|
|
pairs.forEach((pair) => {
|
|
if (!pair.includes('=')) return;
|
|
|
|
const [key, ...valueParts] = pair.split('=');
|
|
const value = valueParts.join('=').trim();
|
|
const lowerKey = key.trim().toLowerCase();
|
|
|
|
if (lowerKey === 'server' || lowerKey === 'host') {
|
|
parsed.host = value;
|
|
} else if (lowerKey === 'port') {
|
|
const p = parseInt(value, 10);
|
|
if (!isNaN(p)) parsed.port = String(p);
|
|
} else if (lowerKey === 'database' || lowerKey === 'db') {
|
|
parsed.database = value;
|
|
} else if (lowerKey === 'uid' || lowerKey === 'user' || lowerKey === 'username') {
|
|
parsed.username = value;
|
|
} else if (lowerKey === 'pwd' || lowerKey === 'password') {
|
|
parsed.password = value;
|
|
} else if (lowerKey === 'socket' || lowerKey === 'socketpath') {
|
|
parsed.socket_path = value;
|
|
} else if (lowerKey === 'ssl' || lowerKey === 'sslmode' || lowerKey === 'ssl_mode' || lowerKey === 'usessl') {
|
|
const v = value.toLowerCase();
|
|
parsed.ssl_enabled = ['true', '1', 'require', 'required', 'verify-ca', 'verify-full'].includes(v) as any;
|
|
} else {
|
|
if (!parsed.params) parsed.params = {};
|
|
parsed.params[key.trim()] = value;
|
|
}
|
|
});
|
|
|
|
return parsed;
|
|
}
|
|
|
|
private async buildConnection(sourceOptions: SourceOptions): Promise<Knex> {
|
|
const effectiveOptions = { ...sourceOptions };
|
|
|
|
const shouldUseSSL = effectiveOptions.ssl_enabled;
|
|
let sslObject: any = null;
|
|
|
|
if (shouldUseSSL) {
|
|
sslObject = { rejectUnauthorized: (effectiveOptions.ssl_certificate ?? 'none') !== 'none' };
|
|
|
|
if (effectiveOptions.ssl_certificate === 'ca_certificate') {
|
|
sslObject.ca = effectiveOptions.ca_cert;
|
|
sslObject.key = effectiveOptions.client_key;
|
|
sslObject.cert = effectiveOptions.client_cert;
|
|
} else if (effectiveOptions.ssl_certificate === 'self_signed') {
|
|
sslObject.ca = effectiveOptions.root_cert;
|
|
sslObject.key = effectiveOptions.client_key;
|
|
sslObject.cert = effectiveOptions.client_cert;
|
|
}
|
|
}
|
|
let connectionConfig: any;
|
|
if (effectiveOptions.ssh_enabled == 'enabled') {
|
|
connectionConfig = async () => {
|
|
const ssh = await createSSHStream(effectiveOptions);
|
|
return {
|
|
stream: ssh.stream,
|
|
user: effectiveOptions.username,
|
|
password: effectiveOptions.password,
|
|
database: effectiveOptions.database,
|
|
multipleStatements: true,
|
|
...(shouldUseSSL ? { ssl: sslObject } : {}),
|
|
};
|
|
};
|
|
} else {
|
|
connectionConfig = {
|
|
user: effectiveOptions.username,
|
|
password: effectiveOptions.password,
|
|
database: effectiveOptions.database,
|
|
multipleStatements: true,
|
|
...(shouldUseSSL ? { ssl: sslObject } : {}),
|
|
};
|
|
|
|
if (effectiveOptions.socket_path) {
|
|
connectionConfig.socketPath = effectiveOptions.socket_path;
|
|
} else {
|
|
connectionConfig.host = effectiveOptions.host;
|
|
connectionConfig.port = Number(effectiveOptions.port);
|
|
}
|
|
}
|
|
|
|
const config: Knex.Config = {
|
|
client: 'mysql2',
|
|
connection: connectionConfig,
|
|
...this.connectionOptions(effectiveOptions),
|
|
};
|
|
|
|
const knexInstance = knex(config);
|
|
|
|
return knexInstance;
|
|
}
|
|
|
|
async getConnection(
|
|
sourceOptions: SourceOptions,
|
|
options: any,
|
|
checkCache: boolean,
|
|
dataSourceId?: string,
|
|
dataSourceUpdatedAt?: string
|
|
): Promise<Knex> {
|
|
if (checkCache) {
|
|
const optionsHash = generateSourceOptionsHash(sourceOptions);
|
|
const enhancedCacheKey = `${dataSourceId}_${optionsHash}`;
|
|
const cachedConnection = await getCachedConnection(enhancedCacheKey, dataSourceUpdatedAt);
|
|
if (cachedConnection) return cachedConnection;
|
|
|
|
const connection = await this.buildConnection(sourceOptions);
|
|
cacheConnectionWithConfiguration(dataSourceId, enhancedCacheKey, connection);
|
|
return connection;
|
|
}
|
|
|
|
return await this.buildConnection(sourceOptions);
|
|
}
|
|
|
|
buildBulkUpdateQuery(queryOptions: QueryOptions): string {
|
|
let queryText = '';
|
|
|
|
const { table: tableName, primary_key_column: primaryKey, records } = queryOptions;
|
|
|
|
for (const record of records) {
|
|
const primaryKeyValue = typeof record[primaryKey] === 'string' ? `'${record[primaryKey]}'` : record[primaryKey];
|
|
|
|
queryText = `${queryText} UPDATE ${tableName} SET`;
|
|
|
|
for (const key of Object.keys(record)) {
|
|
if (key !== primaryKey) {
|
|
queryText = ` ${queryText} ${key} = '${record[key]}',`;
|
|
}
|
|
}
|
|
|
|
queryText = queryText.slice(0, -1);
|
|
queryText = `${queryText} WHERE ${primaryKey} = ${primaryKeyValue};`;
|
|
}
|
|
|
|
return queryText.trim();
|
|
}
|
|
async listTables(
|
|
sourceOptions: SourceOptions,
|
|
queryOptions?: { search?: string; page?: number; limit?: number }
|
|
): Promise<QueryResult> {
|
|
let knexInstance;
|
|
try {
|
|
knexInstance = await this.buildConnection(sourceOptions);
|
|
|
|
const search = queryOptions?.search || '';
|
|
const searchPattern = `%${search}%`;
|
|
|
|
if (queryOptions?.limit) {
|
|
const limit = queryOptions.limit;
|
|
const page = queryOptions.page || 1;
|
|
const offset = (page - 1) * limit;
|
|
|
|
const [dataResult, countResult] = await Promise.all([
|
|
knexInstance.raw(
|
|
`SELECT TABLE_NAME FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_SCHEMA = DATABASE() AND TABLE_TYPE = 'BASE TABLE' AND TABLE_NAME LIKE ? ORDER BY TABLE_NAME LIMIT ? OFFSET ?`,
|
|
[searchPattern, limit, offset]
|
|
),
|
|
knexInstance.raw(
|
|
`SELECT COUNT(*) AS total FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_SCHEMA = DATABASE() AND TABLE_TYPE = 'BASE TABLE' AND TABLE_NAME LIKE ?`,
|
|
[searchPattern]
|
|
),
|
|
]);
|
|
|
|
const rows = (dataResult[0] || []).map((row: any) => ({ label: row.TABLE_NAME, value: row.TABLE_NAME }));
|
|
const totalCount = parseInt(countResult[0]?.[0]?.total ?? '0', 10);
|
|
|
|
return { status: 'ok', data: { rows, totalCount } };
|
|
}
|
|
|
|
const result = await knexInstance.raw(
|
|
`SELECT TABLE_NAME FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_SCHEMA = DATABASE() AND TABLE_TYPE = 'BASE TABLE' AND TABLE_NAME LIKE ? ORDER BY TABLE_NAME`,
|
|
[searchPattern]
|
|
);
|
|
|
|
const tables = (result[0] || []).map((row: any) => ({
|
|
label: row.TABLE_NAME,
|
|
value: row.TABLE_NAME,
|
|
}));
|
|
|
|
return { status: 'ok', data: tables };
|
|
} catch (err) {
|
|
const errorMessage = err instanceof Error ? err?.message : 'An unknown error occurred';
|
|
throw new QueryError('Could not fetch tables', errorMessage, {});
|
|
} finally {
|
|
if (knexInstance) {
|
|
await knexInstance.destroy();
|
|
}
|
|
}
|
|
}
|
|
|
|
private async _fetchTablesForDatabase(
|
|
sourceOptions: SourceOptions,
|
|
queryOptions?: { search?: string; page?: number; limit?: number }
|
|
): Promise<{ items: Array<{ value: string; label: string }>; totalCount: number }> {
|
|
let knexInstance;
|
|
try {
|
|
knexInstance = await this.buildConnection(sourceOptions);
|
|
const search = queryOptions?.search || '';
|
|
const searchPattern = `%${search}%`;
|
|
|
|
if (queryOptions?.limit) {
|
|
const limit = queryOptions.limit;
|
|
const page = queryOptions.page || 1;
|
|
const offset = (page - 1) * limit;
|
|
|
|
const [dataResult, countResult] = await Promise.all([
|
|
knexInstance.raw(
|
|
`SELECT table_name as table_name FROM information_schema.tables
|
|
WHERE table_schema = DATABASE() AND table_type = 'BASE TABLE' AND table_name LIKE ?
|
|
ORDER BY table_name LIMIT ? OFFSET ?`,
|
|
[searchPattern, limit, offset]
|
|
),
|
|
knexInstance.raw(
|
|
`SELECT COUNT(*) AS total FROM information_schema.tables
|
|
WHERE table_schema = DATABASE() AND table_type = 'BASE TABLE' AND table_name LIKE ?`,
|
|
[searchPattern]
|
|
),
|
|
]);
|
|
|
|
const items = (dataResult[0] || []).map((row: any) => ({ value: row.table_name, label: row.table_name }));
|
|
const totalCount = parseInt(countResult[0]?.[0]?.total ?? '0', 10);
|
|
return { items, totalCount };
|
|
}
|
|
|
|
const result = await knexInstance.raw(
|
|
`SELECT table_name as table_name FROM information_schema.tables
|
|
WHERE table_schema = DATABASE() AND table_type = 'BASE TABLE' AND table_name LIKE ?
|
|
ORDER BY table_name`,
|
|
[searchPattern]
|
|
);
|
|
|
|
const items = (result[0] || []).map((row: any) => ({ value: row.table_name, label: row.table_name }));
|
|
return { items, totalCount: items.length };
|
|
} catch (err) {
|
|
const errorMessage = err instanceof Error ? err.message : 'An unknown error occurred';
|
|
throw new QueryError('Could not fetch tables', errorMessage, {});
|
|
} finally {
|
|
if (knexInstance) await knexInstance.destroy();
|
|
}
|
|
}
|
|
|
|
private async _fetchColumnsForTable(
|
|
sourceOptions: SourceOptions,
|
|
table: string
|
|
): Promise<Array<{ value: string; label: string }>> {
|
|
let knexInstance;
|
|
try {
|
|
knexInstance = await this.buildConnection(sourceOptions);
|
|
const result = await knexInstance.raw(
|
|
`SELECT column_name as column_name FROM information_schema.columns
|
|
WHERE table_schema = DATABASE() AND table_name = ?
|
|
ORDER BY ordinal_position`,
|
|
[table]
|
|
);
|
|
const rows = result[0] || [];
|
|
return rows.map((row: any) => ({ value: row.column_name, label: row.column_name }));
|
|
} catch (err) {
|
|
const errorMessage = err instanceof Error ? err.message : 'An unknown error occurred';
|
|
throw new QueryError('Could not fetch columns', errorMessage, {});
|
|
} finally {
|
|
if (knexInstance) await knexInstance.destroy();
|
|
}
|
|
}
|
|
|
|
async invokeMethod(
|
|
methodName: string,
|
|
context: { user?: User; app?: App },
|
|
sourceOptions: SourceOptions,
|
|
args?: any
|
|
): Promise<any> {
|
|
try {
|
|
if (methodName === 'getTables') {
|
|
// return await this.listTables(sourceOptions);
|
|
const isPaginated = !!args?.limit;
|
|
const result = await this.listTables(sourceOptions, {
|
|
search: args?.search,
|
|
page: args?.page,
|
|
limit: args?.limit,
|
|
});
|
|
|
|
const payload = (result as any)?.data ?? [];
|
|
|
|
if (isPaginated) {
|
|
const rows = (payload as any)?.rows ?? [];
|
|
const totalCount = (payload as any)?.totalCount ?? 0;
|
|
return { items: rows, totalCount };
|
|
}
|
|
|
|
return { status: 'ok', data: Array.isArray(payload) ? payload : [] };
|
|
}
|
|
|
|
if (methodName === 'listTables') {
|
|
if (!sourceOptions.database)
|
|
throw new QueryError('Database is required', 'No database specified in source options', {});
|
|
const isPaginated = !!args?.limit;
|
|
const result = await this._fetchTablesForDatabase(sourceOptions, {
|
|
search: args?.search,
|
|
page: args?.page,
|
|
limit: args?.limit,
|
|
});
|
|
|
|
if (isPaginated) {
|
|
return { items: result.items, totalCount: result.totalCount };
|
|
}
|
|
|
|
return { status: 'ok', data: result.items };
|
|
}
|
|
|
|
if (methodName === 'listColumns') {
|
|
if (!sourceOptions.database)
|
|
throw new QueryError('Database is required', 'No database specified in source options', {});
|
|
const table = args?.values?.table || '';
|
|
if (!table) return { status: 'ok', data: [] };
|
|
const columns = await this._fetchColumnsForTable(sourceOptions, table);
|
|
return { status: 'ok', data: columns };
|
|
}
|
|
|
|
throw new QueryError('Method not found', `Method ${methodName} is not supported for MySQL plugin`, {
|
|
availableMethods: ['getTables', 'listTables', 'listColumns'],
|
|
});
|
|
} catch (err) {
|
|
if (err instanceof QueryError) {
|
|
throw err;
|
|
}
|
|
|
|
const errorMessage = err?.message || 'An unknown error occurred';
|
|
const errorDetails: any = {};
|
|
|
|
if (err instanceof Error) {
|
|
const mysqlError = err as any;
|
|
const { code, errno, sqlMessage, sqlState } = mysqlError;
|
|
|
|
errorDetails.code = code || null;
|
|
errorDetails.errno = errno || null;
|
|
errorDetails.sqlMessage = sqlMessage || null;
|
|
errorDetails.sqlState = sqlState || null;
|
|
}
|
|
|
|
throw new QueryError('Method invocation failed', errorMessage, errorDetails);
|
|
}
|
|
}
|
|
}
|