import knex, { Knex } from 'knex'; import { cacheConnectionWithConfiguration, generateSourceOptionsHash, getCachedConnection, ConnectionTestResult, QueryService, QueryResult, QueryError, createQueryBuilder, User, App, } from '@tooljet-plugins/common'; import { SourceOptions, QueryOptions } from './types'; import { isEmpty } from '@tooljet-plugins/common'; import { Client } from 'ssh2'; function createSSHStream(sourceOptions: SourceOptions): Promise<{ client: Client; stream: NodeJS.ReadWriteStream }> { return new Promise((resolve, reject) => { const sshClient = new Client(); sshClient.on('ready', () => { sshClient.forwardOut('127.0.0.1', 0, sourceOptions.host, Number(sourceOptions.port), (err, stream) => { if (err) { sshClient.end(); return reject(err); } stream.on('error', (streamErr) => { console.error('SSH stream error (suppressed):', streamErr.message); }); stream.on('close', () => { setImmediate(() => { try { if (sshClient) { sshClient.destroy(); } } catch (e) { console.error('Error closing SSH client (suppressed):', e.message); } }); }); resolve({ client: sshClient, stream }); }); }); sshClient.on('error', (err) => { reject(err); }); sshClient.on('end', () => { console.log('SSH connection ended'); }); sshClient.on('close', () => { console.log('SSH connection closed'); }); sshClient.connect({ host: sourceOptions.ssh_host, port: sourceOptions.ssh_port || 22, username: sourceOptions.ssh_username, ...(sourceOptions.ssh_auth_type === 'password' ? { password: sourceOptions.ssh_password } : { privateKey: sourceOptions.ssh_private_key, passphrase: sourceOptions.ssh_passphrase, }), readyTimeout: 20000, keepaliveInterval: 10000, }); }); } export default class MysqlQueryService implements QueryService { private static _instance: MysqlQueryService; private STATEMENT_TIMEOUT; constructor() { this.STATEMENT_TIMEOUT = process.env?.PLUGINS_SQL_DB_STATEMENT_TIMEOUT && !isNaN(Number(process.env?.PLUGINS_SQL_DB_STATEMENT_TIMEOUT)) ? Number(process.env.PLUGINS_SQL_DB_STATEMENT_TIMEOUT) : 120000; if (MysqlQueryService._instance) { return MysqlQueryService._instance; } MysqlQueryService._instance = this; return MysqlQueryService._instance; } async run( sourceOptions: SourceOptions, queryOptions: QueryOptions, dataSourceId: string, dataSourceUpdatedAt: string ): Promise { let checkCache, knexInstance; if (sourceOptions['allow_dynamic_connection_parameters']) { if (sourceOptions.connection_type === 'hostname') { sourceOptions['host'] = queryOptions['host'] ? queryOptions['host'] : sourceOptions['host']; sourceOptions['database'] = queryOptions['database'] ? queryOptions['database'] : sourceOptions['database']; } else if (sourceOptions.connection_type === 'socket_path') { sourceOptions['database'] = queryOptions['database'] ? queryOptions['database'] : sourceOptions['database']; } } try { // If dynamic connection parameters is toggled on - We don't cache the connection also destroy the connection created. checkCache = sourceOptions['allow_dynamic_connection_parameters'] ? false : true; knexInstance = await this.getConnection(sourceOptions, {}, checkCache, dataSourceId, dataSourceUpdatedAt); switch (queryOptions.mode) { case 'sql': return await this.handleRawQuery(knexInstance, queryOptions); case 'gui': return await this.handleGuiQuery(knexInstance, queryOptions); default: throw new Error("Invalid query mode. Must be either 'sql' or 'gui'."); } } catch (err) { const errorMessage = err?.message || 'An unknown error occurred'; const errorDetails: any = {}; if (err instanceof Error) { const mysqlError = err as any; const { code, errno, sqlMessage, sqlState } = mysqlError; errorDetails.code = code || null; errorDetails.errno = errno || null; errorDetails.sqlMessage = sqlMessage || null; errorDetails.sqlState = sqlState || null; } throw new QueryError('Query could not be completed', errorMessage, errorDetails); } finally { if (!checkCache) await knexInstance.destroy(); } } async testConnection(sourceOptions: SourceOptions): Promise { let knexInstance; try { knexInstance = await this.getConnection(sourceOptions, {}, false); await knexInstance.raw('select @@version;').timeout(this.STATEMENT_TIMEOUT); if (knexInstance) { try { await knexInstance.destroy(); // eslint-disable-next-line no-empty } catch (_) {} } return { status: 'ok' }; } catch (err) { const errorMessage = err?.message || 'An unknown error occurred'; const errorDetails: any = {}; if (err instanceof Error) { const mysqlError = err as any; const { code, errno, sqlMessage, sqlState } = mysqlError; errorDetails.code = code || null; errorDetails.errno = errno || null; errorDetails.sqlMessage = sqlMessage || null; errorDetails.sqlState = sqlState || null; } throw new QueryError('Connection test failed', errorMessage, errorDetails); } } private async handleGuiQuery(knexInstance: Knex, queryOptions: QueryOptions): Promise { const { operation, table } = queryOptions; const queryBuilder = createQueryBuilder('mysql'); switch (operation) { case 'list_rows': { const { list_rows, limit, offset } = queryOptions; const { where_filters, order_filters, aggregates, group_by } = list_rows || {}; const { query, params } = queryBuilder.listRows(table, { where_filters, order_filters, aggregates, group_by, limit, offset, }) as { query: string; params: unknown[] }; const rows = await this.executeParameterizedQuery(knexInstance, query, params); return { status: 'ok', data: rows }; } case 'create_row': { const { columns } = queryOptions.create_row || {}; const { query, params } = queryBuilder.createRow(table, undefined, columns) as { query: string; params: unknown[]; }; const insertedRows = await this.executeCreateRow( knexInstance, queryBuilder, table, undefined, columns, query, params ); return { status: 'ok', data: insertedRows }; } case 'update_rows': { const { allow_multiple_updates, zero_records_as_success } = queryOptions; const { columns, where_filters } = queryOptions.update_rows || {}; const hasAtLeastOneFilter = Object.values(where_filters || {}).some((f: any) => f?.column?.trim()); if (!hasAtLeastOneFilter) { throw new Error('At least one filter condition is required.'); } const { query, params } = queryBuilder.updateRows(table, { columns, where_filters }) as { query: string; params: unknown[]; }; const { query: selectQuery, params: selectParams } = queryBuilder.listRows(table, { where_filters, }) as { query: string; params: unknown[] }; const rows = await this.executeWriteQuery(knexInstance, query, params, selectQuery, selectParams, { allow_multiple_updates, zero_records_as_success, operationLabel: 'updated', }); return { status: 'ok', data: rows }; } case 'upsert_rows': { const { primary_key_columns, allow_multiple_updates, zero_records_as_success } = queryOptions; const { columns } = queryOptions.upsert_rows || {}; const { query, params } = queryBuilder.upsertRows(table, { primary_key_columns, columns }) as { query: string; params: unknown[]; }; const normalizedPrimaryKeys = Array.isArray(primary_key_columns) ? primary_key_columns.filter(Boolean) : typeof primary_key_columns === 'string' && primary_key_columns.trim() ? [primary_key_columns.trim()] : []; const columnEntries = Object.values(columns || {}); const whereFiltersForSelect = Object.fromEntries( normalizedPrimaryKeys .map((keyColumn) => { const keyEntry = columnEntries.find((entry: any) => entry.column === keyColumn) as any; if (!keyEntry) return null; return [keyColumn, { column: keyColumn, operator: 'eq', value: keyEntry.value }]; }) .filter((entry): entry is [string, any] => entry !== null) ); const { query: selectQuery, params: selectParams } = queryBuilder.listRows(table, { where_filters: whereFiltersForSelect, }) as { query: string; params: unknown[] }; const rows = await this.executeWriteQuery(knexInstance, query, params, selectQuery, selectParams, { allow_multiple_updates, zero_records_as_success, operationLabel: 'upserted', }); return { status: 'ok', data: rows }; } case 'delete_rows': { const { limit, zero_records_as_success } = queryOptions; const { where_filters } = queryOptions.delete_rows || {}; const hasAtLeastOneFilter = Object.values(where_filters || {}).some((f: any) => f?.column?.trim()); const hasLimit = limit != null && limit !== ''; if (!hasAtLeastOneFilter && !hasLimit) { throw new Error( 'Delete requires at least one filter condition or a limit to prevent deleting all rows unintentionally.' ); } const { query, params } = queryBuilder.deleteRows(table, { where_filters, limit }) as { query: string; params: unknown[]; }; const [okPacket] = await knexInstance.raw(query, params as any[]).timeout(this.STATEMENT_TIMEOUT); const deletedRecords = okPacket.affectedRows || 0; if (zero_records_as_success === false && deletedRecords === 0) { throw new Error('No rows were deleted.'); } return { status: 'ok', data: { deletedRecords } }; } case 'bulk_insert': { const { records } = queryOptions; const { query, params } = queryBuilder.bulkInsert(table, { rows_insert: records }) as { query: string; params: unknown[]; }; const [okPacket] = await knexInstance.raw(query, params as any[]).timeout(this.STATEMENT_TIMEOUT); return { status: 'ok', data: okPacket }; } case 'bulk_update_pkey': { const { primary_key_columns, records } = queryOptions; const { queries } = queryBuilder.bulkUpdateWithPrimaryKey(table, { primary_key: primary_key_columns, rows_update: records, }) as { queries: { query: string; params: unknown[] }[] }; const data = await this.executeBulkQueriesInTransaction(knexInstance, queries); return { status: 'ok', data, bulk_update_status: 'success' } as unknown as QueryResult; } case 'bulk_upsert_pkey': { const { primary_key_columns, records } = queryOptions; const { queries } = queryBuilder.bulkUpsertWithPrimaryKey(table, { primary_key: primary_key_columns, row_upsert: records, }) as { queries: { query: string; params: unknown[] }[] }; const data = await this.executeBulkQueriesInTransaction(knexInstance, queries); return { status: 'ok', data, bulk_upsert_status: 'success' } as unknown as QueryResult; } default: throw new Error(`Unsupported GUI operation: "${operation}"`); } } /** * Executes a SELECT query and returns the rows array. * MySQL raw() returns [rows, fields], so we destructure and return rows. */ private async executeParameterizedQuery(knexInstance: Knex, query: string, params: unknown[]): Promise { const [rows] = await knexInstance.raw(query, params as any[]).timeout(this.STATEMENT_TIMEOUT); return rows; } /** * Executes INSERT and then SELECTs back the inserted row(s) using the inserted column values as WHERE conditions. * Wrapped in a transaction to ensure consistency. */ private async executeCreateRow( knexInstance: Knex, queryBuilder: ReturnType, tableName: string, schema: string | undefined | null, columns: Record | undefined | null, insertQuery: string, insertParams: unknown[] ): Promise { return knexInstance.transaction(async (transaction) => { await transaction.raw(insertQuery, insertParams as any[]); const columnEntries = Object.values(columns || {}); if (columnEntries.length === 0) return []; const whereFilters = Object.fromEntries( columnEntries.map((entry: any) => [entry.column, { column: entry.column, operator: 'eq', value: entry.value }]) ); const { query: selectQuery, params: selectParams } = queryBuilder.listRows(tableName, { schema, where_filters: whereFilters, }) as { query: string; params: unknown[] }; const [rows] = await transaction.raw(selectQuery, selectParams as any[]); return rows; }); } /** * Executes a write query (UPDATE or upsert) and checks affectedRows constraints. * When constraints apply, wraps in a transaction so violations automatically roll back. * After the write, SELECTs and returns the affected rows. */ private async executeWriteQuery( knexInstance: Knex, writeQuery: string, writeParams: unknown[], selectQuery: string, selectParams: unknown[], options: { allow_multiple_updates?: boolean; zero_records_as_success?: boolean; operationLabel: string } ): Promise { const { allow_multiple_updates, zero_records_as_success, operationLabel } = options; const hasConstraints = allow_multiple_updates === false || zero_records_as_success === false; if (!hasConstraints) { await knexInstance.raw(writeQuery, writeParams as any[]).timeout(this.STATEMENT_TIMEOUT); const [rows] = await knexInstance.raw(selectQuery, selectParams as any[]).timeout(this.STATEMENT_TIMEOUT); return rows; } return knexInstance.transaction(async (transaction) => { const [okPacket] = await transaction.raw(writeQuery, writeParams as any[]); const affectedRows = okPacket.affectedRows || 0; if (allow_multiple_updates === false && affectedRows > 1) { throw new Error( 'Query matches more than one row. Enable "Allow this Query to modify multiple rows" to permit this.' ); } if (zero_records_as_success === false && affectedRows === 0) { throw new Error(`No rows were ${operationLabel}.`); } const [rows] = await transaction.raw(selectQuery, selectParams as any[]); return rows; }); } /** * Executes multiple queries sequentially inside a single transaction. * Returns the OkPacket from each query as an array. */ private async executeBulkQueriesInTransaction( knexInstance: Knex, queries: { query: string; params: unknown[] }[] ): Promise { const allResults: unknown[] = []; await knexInstance.transaction(async (transaction) => { for (const { query, params } of queries) { const [okPacket] = await transaction.raw(query, params as any[]); allResults.push(okPacket); } }); return allResults; } private async handleRawQuery(knexInstance: Knex, queryOptions: QueryOptions): Promise { const { query, query_params } = queryOptions; const queryParams = query_params || []; const sanitizedQueryParams: Record = Object.fromEntries(queryParams.filter(([key]) => !isEmpty(key))); const result = await this.executeQuery(knexInstance, query, sanitizedQueryParams); return { status: 'ok', data: result[0] }; } private async executeQuery(knexInstance: Knex, query: string, sanitizedQueryParams: Record = {}) { if (isEmpty(query)) throw new Error('Query is empty'); const result = await knexInstance.raw(query, sanitizedQueryParams).timeout(this.STATEMENT_TIMEOUT); return result; } private connectionOptions(sourceOptions: SourceOptions) { const _connectionOptions = (sourceOptions?.connection_options || []).filter((o) => o.some((e) => !isEmpty(e))); const connectionOptions = Object.fromEntries(_connectionOptions); Object.keys(connectionOptions).forEach((key) => connectionOptions[key] === '' ? delete connectionOptions[key] : {} ); return connectionOptions; } private parseConnectionString(connectionString: string): Partial { const parsed: Partial; socketPath?: string }> = {}; if (!connectionString) return parsed; const trimmed = connectionString.trim(); try { const url = new URL(trimmed); if (url.protocol && url.protocol.startsWith('mysql')) { if (url.username) parsed.username = decodeURIComponent(url.username); if (url.password) parsed.password = decodeURIComponent(url.password); if (url.hostname) parsed.host = decodeURIComponent(url.hostname); if (url.port) { const p = parseInt(url.port, 10); if (!isNaN(p)) parsed.port = String(p); } if (url.pathname && url.pathname !== '/') { parsed.database = decodeURIComponent(url.pathname.slice(1)); } const queryParams: Record = {}; url.searchParams.forEach((v, k) => { queryParams[k] = v; }); if (Object.keys(queryParams).length) { parsed.params = queryParams; } const socket = url.searchParams.get('socket') || url.searchParams.get('socketPath'); if (socket) parsed.socket_path = socket; const sslParam = url.searchParams.get('ssl') || url.searchParams.get('sslmode') || url.searchParams.get('ssl_mode'); if (sslParam !== null) { parsed.ssl_enabled = ['true', '1', 'require', 'required', 'verify-ca', 'verify-full'].includes( sslParam.toLowerCase() ) as any; } return parsed; } // eslint-disable-next-line no-empty } catch (_) {} const regex = // eslint-disable-next-line no-useless-escape /^mysql:\/\/(?:([^:@\/?#]+)(?::([^@\/?#]*))?@)?(\[[^\]]+\]|[^:\/?#]+)?(?::(\d+))?(?:\/([^?]*))?(?:\?(.*))?$/; const m = trimmed.match(regex); if (m) { const user = m[1]; const pass = m[2]; const hostRaw = m[3]; const portRaw = m[4]; const db = m[5]; const query = m[6]; if (user) { try { parsed.username = decodeURIComponent(user); } catch { parsed.username = user; } } if (pass !== undefined) { try { parsed.password = decodeURIComponent(pass); } catch { parsed.password = pass; } } if (hostRaw) { let host = hostRaw; if (host.startsWith('[') && host.endsWith(']')) host = host.slice(1, -1); try { parsed.host = decodeURIComponent(host); } catch { parsed.host = host; } } if (portRaw) { const p = parseInt(portRaw, 10); if (!isNaN(p)) parsed.port = String(p); } if (db) { try { parsed.database = decodeURIComponent(db); } catch { parsed.database = db; } } if (query) { const q: Record = {}; query.split('&').forEach((part) => { if (!part) return; const idx = part.indexOf('='); if (idx === -1) { q[decodeURIComponent(part)] = ''; } else { const k = part.slice(0, idx); const v = part.slice(idx + 1); try { q[decodeURIComponent(k)] = decodeURIComponent(v); } catch { q[k] = v; } } }); if (Object.keys(q).length) parsed.params = q; if (q['socket'] || q['socketPath']) { parsed.socket_path = q['socket'] || q['socketPath']; const sslMode = q['ssl'] || q['sslmode'] || q['ssl_mode']; if (sslMode !== undefined) { parsed.ssl_enabled = ['true', '1', 'require', 'required', 'verify-ca', 'verify-full'].includes( sslMode.toLowerCase() ) as any; } } } return parsed; } // Strategy 3: Key-value pair parsing (Server=host;Port=3306;...) // Uses join('=') to correctly handle values that contain '=' characters const pairs = trimmed.split(';').filter((p) => p.trim()); pairs.forEach((pair) => { if (!pair.includes('=')) return; const [key, ...valueParts] = pair.split('='); const value = valueParts.join('=').trim(); const lowerKey = key.trim().toLowerCase(); if (lowerKey === 'server' || lowerKey === 'host') { parsed.host = value; } else if (lowerKey === 'port') { const p = parseInt(value, 10); if (!isNaN(p)) parsed.port = String(p); } else if (lowerKey === 'database' || lowerKey === 'db') { parsed.database = value; } else if (lowerKey === 'uid' || lowerKey === 'user' || lowerKey === 'username') { parsed.username = value; } else if (lowerKey === 'pwd' || lowerKey === 'password') { parsed.password = value; } else if (lowerKey === 'socket' || lowerKey === 'socketpath') { parsed.socket_path = value; } else if (lowerKey === 'ssl' || lowerKey === 'sslmode' || lowerKey === 'ssl_mode' || lowerKey === 'usessl') { const v = value.toLowerCase(); parsed.ssl_enabled = ['true', '1', 'require', 'required', 'verify-ca', 'verify-full'].includes(v) as any; } else { if (!parsed.params) parsed.params = {}; parsed.params[key.trim()] = value; } }); return parsed; } private async buildConnection(sourceOptions: SourceOptions): Promise { const effectiveOptions = { ...sourceOptions }; const shouldUseSSL = effectiveOptions.ssl_enabled; let sslObject: any = null; if (shouldUseSSL) { sslObject = { rejectUnauthorized: (effectiveOptions.ssl_certificate ?? 'none') !== 'none' }; if (effectiveOptions.ssl_certificate === 'ca_certificate') { sslObject.ca = effectiveOptions.ca_cert; sslObject.key = effectiveOptions.client_key; sslObject.cert = effectiveOptions.client_cert; } else if (effectiveOptions.ssl_certificate === 'self_signed') { sslObject.ca = effectiveOptions.root_cert; sslObject.key = effectiveOptions.client_key; sslObject.cert = effectiveOptions.client_cert; } } let connectionConfig: any; if (effectiveOptions.ssh_enabled == 'enabled') { connectionConfig = async () => { const ssh = await createSSHStream(effectiveOptions); return { stream: ssh.stream, user: effectiveOptions.username, password: effectiveOptions.password, database: effectiveOptions.database, multipleStatements: true, ...(shouldUseSSL ? { ssl: sslObject } : {}), }; }; } else { connectionConfig = { user: effectiveOptions.username, password: effectiveOptions.password, database: effectiveOptions.database, multipleStatements: true, ...(shouldUseSSL ? { ssl: sslObject } : {}), }; if (effectiveOptions.socket_path) { connectionConfig.socketPath = effectiveOptions.socket_path; } else { connectionConfig.host = effectiveOptions.host; connectionConfig.port = Number(effectiveOptions.port); } } const config: Knex.Config = { client: 'mysql2', connection: connectionConfig, ...this.connectionOptions(effectiveOptions), }; const knexInstance = knex(config); return knexInstance; } async getConnection( sourceOptions: SourceOptions, options: any, checkCache: boolean, dataSourceId?: string, dataSourceUpdatedAt?: string ): Promise { if (checkCache) { const optionsHash = generateSourceOptionsHash(sourceOptions); const enhancedCacheKey = `${dataSourceId}_${optionsHash}`; const cachedConnection = await getCachedConnection(enhancedCacheKey, dataSourceUpdatedAt); if (cachedConnection) return cachedConnection; const connection = await this.buildConnection(sourceOptions); cacheConnectionWithConfiguration(dataSourceId, enhancedCacheKey, connection); return connection; } return await this.buildConnection(sourceOptions); } buildBulkUpdateQuery(queryOptions: QueryOptions): string { let queryText = ''; const { table: tableName, primary_key_column: primaryKey, records } = queryOptions; for (const record of records) { const primaryKeyValue = typeof record[primaryKey] === 'string' ? `'${record[primaryKey]}'` : record[primaryKey]; queryText = `${queryText} UPDATE ${tableName} SET`; for (const key of Object.keys(record)) { if (key !== primaryKey) { queryText = ` ${queryText} ${key} = '${record[key]}',`; } } queryText = queryText.slice(0, -1); queryText = `${queryText} WHERE ${primaryKey} = ${primaryKeyValue};`; } return queryText.trim(); } async listTables( sourceOptions: SourceOptions, queryOptions?: { search?: string; page?: number; limit?: number } ): Promise { let knexInstance; try { knexInstance = await this.buildConnection(sourceOptions); const search = queryOptions?.search || ''; const searchPattern = `%${search}%`; if (queryOptions?.limit) { const limit = queryOptions.limit; const page = queryOptions.page || 1; const offset = (page - 1) * limit; const [dataResult, countResult] = await Promise.all([ knexInstance.raw( `SELECT TABLE_NAME FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_SCHEMA = DATABASE() AND TABLE_TYPE = 'BASE TABLE' AND TABLE_NAME LIKE ? ORDER BY TABLE_NAME LIMIT ? OFFSET ?`, [searchPattern, limit, offset] ), knexInstance.raw( `SELECT COUNT(*) AS total FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_SCHEMA = DATABASE() AND TABLE_TYPE = 'BASE TABLE' AND TABLE_NAME LIKE ?`, [searchPattern] ), ]); const rows = (dataResult[0] || []).map((row: any) => ({ label: row.TABLE_NAME, value: row.TABLE_NAME })); const totalCount = parseInt(countResult[0]?.[0]?.total ?? '0', 10); return { status: 'ok', data: { rows, totalCount } }; } const result = await knexInstance.raw( `SELECT TABLE_NAME FROM INFORMATION_SCHEMA.TABLES WHERE TABLE_SCHEMA = DATABASE() AND TABLE_TYPE = 'BASE TABLE' AND TABLE_NAME LIKE ? ORDER BY TABLE_NAME`, [searchPattern] ); const tables = (result[0] || []).map((row: any) => ({ label: row.TABLE_NAME, value: row.TABLE_NAME, })); return { status: 'ok', data: tables }; } catch (err) { const errorMessage = err instanceof Error ? err?.message : 'An unknown error occurred'; throw new QueryError('Could not fetch tables', errorMessage, {}); } finally { if (knexInstance) { await knexInstance.destroy(); } } } private async _fetchTablesForDatabase( sourceOptions: SourceOptions, queryOptions?: { search?: string; page?: number; limit?: number } ): Promise<{ items: Array<{ value: string; label: string }>; totalCount: number }> { let knexInstance; try { knexInstance = await this.buildConnection(sourceOptions); const search = queryOptions?.search || ''; const searchPattern = `%${search}%`; if (queryOptions?.limit) { const limit = queryOptions.limit; const page = queryOptions.page || 1; const offset = (page - 1) * limit; const [dataResult, countResult] = await Promise.all([ knexInstance.raw( `SELECT table_name as table_name FROM information_schema.tables WHERE table_schema = DATABASE() AND table_type = 'BASE TABLE' AND table_name LIKE ? ORDER BY table_name LIMIT ? OFFSET ?`, [searchPattern, limit, offset] ), knexInstance.raw( `SELECT COUNT(*) AS total FROM information_schema.tables WHERE table_schema = DATABASE() AND table_type = 'BASE TABLE' AND table_name LIKE ?`, [searchPattern] ), ]); const items = (dataResult[0] || []).map((row: any) => ({ value: row.table_name, label: row.table_name })); const totalCount = parseInt(countResult[0]?.[0]?.total ?? '0', 10); return { items, totalCount }; } const result = await knexInstance.raw( `SELECT table_name as table_name FROM information_schema.tables WHERE table_schema = DATABASE() AND table_type = 'BASE TABLE' AND table_name LIKE ? ORDER BY table_name`, [searchPattern] ); const items = (result[0] || []).map((row: any) => ({ value: row.table_name, label: row.table_name })); return { items, totalCount: items.length }; } catch (err) { const errorMessage = err instanceof Error ? err.message : 'An unknown error occurred'; throw new QueryError('Could not fetch tables', errorMessage, {}); } finally { if (knexInstance) await knexInstance.destroy(); } } private async _fetchColumnsForTable( sourceOptions: SourceOptions, table: string ): Promise> { let knexInstance; try { knexInstance = await this.buildConnection(sourceOptions); const result = await knexInstance.raw( `SELECT column_name as column_name FROM information_schema.columns WHERE table_schema = DATABASE() AND table_name = ? ORDER BY ordinal_position`, [table] ); const rows = result[0] || []; return rows.map((row: any) => ({ value: row.column_name, label: row.column_name })); } catch (err) { const errorMessage = err instanceof Error ? err.message : 'An unknown error occurred'; throw new QueryError('Could not fetch columns', errorMessage, {}); } finally { if (knexInstance) await knexInstance.destroy(); } } async invokeMethod( methodName: string, context: { user?: User; app?: App }, sourceOptions: SourceOptions, args?: any ): Promise { try { if (methodName === 'getTables') { // return await this.listTables(sourceOptions); const isPaginated = !!args?.limit; const result = await this.listTables(sourceOptions, { search: args?.search, page: args?.page, limit: args?.limit, }); const payload = (result as any)?.data ?? []; if (isPaginated) { const rows = (payload as any)?.rows ?? []; const totalCount = (payload as any)?.totalCount ?? 0; return { items: rows, totalCount }; } return { status: 'ok', data: Array.isArray(payload) ? payload : [] }; } if (methodName === 'listTables') { if (!sourceOptions.database) throw new QueryError('Database is required', 'No database specified in source options', {}); const isPaginated = !!args?.limit; const result = await this._fetchTablesForDatabase(sourceOptions, { search: args?.search, page: args?.page, limit: args?.limit, }); if (isPaginated) { return { items: result.items, totalCount: result.totalCount }; } return { status: 'ok', data: result.items }; } if (methodName === 'listColumns') { if (!sourceOptions.database) throw new QueryError('Database is required', 'No database specified in source options', {}); const table = args?.values?.table || ''; if (!table) return { status: 'ok', data: [] }; const columns = await this._fetchColumnsForTable(sourceOptions, table); return { status: 'ok', data: columns }; } throw new QueryError('Method not found', `Method ${methodName} is not supported for MySQL plugin`, { availableMethods: ['getTables', 'listTables', 'listColumns'], }); } catch (err) { if (err instanceof QueryError) { throw err; } const errorMessage = err?.message || 'An unknown error occurred'; const errorDetails: any = {}; if (err instanceof Error) { const mysqlError = err as any; const { code, errno, sqlMessage, sqlState } = mysqlError; errorDetails.code = code || null; errorDetails.errno = errno || null; errorDetails.sqlMessage = sqlMessage || null; errorDetails.sqlState = sqlState || null; } throw new QueryError('Method invocation failed', errorMessage, errorDetails); } } }