npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2026 – Pkg Stats / Ryan Hefner

@polyglot-sql/sdk

v0.3.2

Published

SQL dialect translator powered by WebAssembly

Downloads

16,088

Readme

@polyglot-sql/sdk

Rust/Wasm-powered SQL transpiler for TypeScript. Parse, generate, transpile, format, and build SQL across 32 database dialects.

Part of the Polyglot project.

Installation

npm install @polyglot-sql/sdk

Quick Start

Transpile

import { transpile, Dialect } from '@polyglot-sql/sdk';

const result = transpile(
  'SELECT IFNULL(a, b) FROM t',
  Dialect.MySQL,
  Dialect.PostgreSQL,
);
console.log(result.sql[0]); // SELECT COALESCE(a, b) FROM t

Parse + Generate

import { parse, generate, Dialect } from '@polyglot-sql/sdk';

const { ast } = parse('SELECT 1 + 2', Dialect.Generic);
const { sql } = generate(ast, Dialect.PostgreSQL);
console.log(sql[0]); // SELECT 1 + 2

Format

import { format, Dialect } from '@polyglot-sql/sdk';

const { sql } = format('SELECT a,b FROM t WHERE x=1', Dialect.PostgreSQL);
console.log(sql[0]);
// SELECT
//   a,
//   b
// FROM t
// WHERE
//   x = 1

Format With Guard Options

Formatting guard defaults from Rust core:

  • maxInputBytes: 16 * 1024 * 1024
  • maxTokens: 1_000_000
  • maxAstNodes: 1_000_000
  • maxSetOpChain: 256
import { formatWithOptions, Dialect } from '@polyglot-sql/sdk';

const result = formatWithOptions(
  'SELECT a,b FROM t WHERE x=1',
  Dialect.PostgreSQL,
  {
    maxInputBytes: 2 * 1024 * 1024,
    maxTokens: 250_000,
    maxAstNodes: 250_000,
    maxSetOpChain: 128,
  },
);

if (!result.success) {
  // Includes one of:
  // E_GUARD_INPUT_TOO_LARGE
  // E_GUARD_TOKEN_BUDGET_EXCEEDED
  // E_GUARD_AST_BUDGET_EXCEEDED
  // E_GUARD_SET_OP_CHAIN_EXCEEDED
  console.error(result.error);
}

Fluent Query Builder

Build SQL queries programmatically with full type safety. All builder operations are backed by the Rust engine via WASM.

SELECT

import { select, col, lit } from '@polyglot-sql/sdk';

const sql = select('id', 'name')
  .from('users')
  .where(col('age').gt(lit(18)))
  .orderBy(col('name').asc())
  .limit(10)
  .toSql('postgresql');

// SELECT id, name FROM users WHERE age > 18 ORDER BY name ASC LIMIT 10

Joins

const sql = select('u.id', 'o.total')
  .from('users')
  .join('orders', col('u.id').eq(col('o.user_id')))
  .leftJoin('addresses', col('u.id').eq(col('a.user_id')))
  .toSql();

GROUP BY / HAVING

const sql = select(col('dept'), count())
  .from('employees')
  .groupBy('dept')
  .having(count().gt(lit(5)))
  .toSql();

DISTINCT / QUALIFY

const sql = select('id', 'name')
  .from('users')
  .distinct()
  .qualify(col('rn').eq(lit(1)))
  .toSql('snowflake');

Expression Helpers

import { col, lit, star, sqlNull, boolean, table, sqlExpr, func } from '@polyglot-sql/sdk';

col('users.id');         // Column reference (supports dotted names)
lit('hello');            // String literal
lit(42);                 // Numeric literal
star();                  // *
sqlNull();               // NULL
boolean(true);           // TRUE
table('schema.users');   // Table reference
sqlExpr('x + 1');        // Raw SQL fragment
func('MY_FUNC', col('a'), lit(1));  // Function call

Expression Operators

import { col, lit } from '@polyglot-sql/sdk';

// Comparison
col('age').eq(lit(18));
col('age').neq(lit(0));
col('age').gt(lit(18));
col('age').gte(lit(18));
col('age').lt(lit(100));
col('age').lte(lit(100));

// Logical
col('a').and(col('b'));
col('a').or(col('b'));
col('a').not();
col('a').xor(col('b'));

// Arithmetic
col('price').mul(lit(1.1));
col('a').add(col('b'));

// Pattern matching
col('name').like(lit('%Alice%'));
col('name').ilike(lit('%alice%'));

// Predicates
col('email').isNull();
col('email').isNotNull();
col('age').between(lit(18), lit(65));
col('status').inList(lit('active'), lit('pending'));
col('status').notIn(lit('deleted'), lit('banned'));

// Transform
col('total').alias('grand_total');  // or .as('grand_total')
col('id').cast('TEXT');
col('name').asc();
col('name').desc();

Convenience Functions

import {
  // Aggregate
  count, countDistinct, sum, avg, min, max,
  // String
  upper, lower, length, trim, ltrim, rtrim, reverse, initcap,
  substring, replace, concatWs,
  // Null handling
  coalesce, nullIf, ifNull,
  // Math
  abs, round, floor, ceil, power, sqrt, ln, exp, sign, greatest, least,
  // Date/time
  currentDate, currentTime, currentTimestamp, extract,
  // Window
  rowNumber, rank, denseRank,
  // Logical
  and, or, not, cast, alias,
} from '@polyglot-sql/sdk';

// Examples
count();                          // COUNT(*)
countDistinct(col('id'));         // COUNT(DISTINCT id)
coalesce(col('a'), col('b'));     // COALESCE(a, b)
upper(col('name'));               // UPPER(name)
round(col('price'), lit(2));      // ROUND(price, 2)
extract('YEAR', col('created'));  // EXTRACT(YEAR FROM created)

CASE Expressions

import { caseWhen, caseOf, col, lit } from '@polyglot-sql/sdk';

// Searched CASE
const expr = caseWhen()
  .when(col('x').gt(lit(0)), lit('positive'))
  .when(col('x').eq(lit(0)), lit('zero'))
  .else_(lit('negative'))
  .build();

// Simple CASE
const expr2 = caseOf(col('status'))
  .when(lit('A'), lit('Active'))
  .when(lit('I'), lit('Inactive'))
  .else_(lit('Unknown'))
  .build();

INSERT

import { insertInto, lit, select } from '@polyglot-sql/sdk';

// INSERT ... VALUES
const sql = insertInto('users')
  .columns('id', 'name')
  .values(lit(1), lit('Alice'))
  .toSql();

// INSERT ... SELECT
const sql2 = insertInto('archive')
  .query(select('*').from('users').where(col('active').eq(lit(false))))
  .toSql();

UPDATE

import { update, col, lit } from '@polyglot-sql/sdk';

const sql = update('users')
  .set('name', lit('Bob'))
  .set('updated_at', sqlExpr('NOW()'))
  .where(col('id').eq(lit(1)))
  .toSql();

DELETE

import { deleteFrom, col, lit } from '@polyglot-sql/sdk';

const sql = deleteFrom('users')
  .where(col('id').eq(lit(1)))
  .toSql();

MERGE

import { mergeInto, col } from '@polyglot-sql/sdk';

const sql = mergeInto('target')
  .using('source', col('target.id').eq(col('source.id')))
  .whenMatchedUpdate({ name: col('source.name') })
  .whenNotMatchedInsert(
    ['id', 'name'],
    [col('source.id'), col('source.name')]
  )
  .toSql();

Set Operations

import { select, union, unionAll, intersect, except } from '@polyglot-sql/sdk';

const q1 = select('id').from('a');
const q2 = select('id').from('b');

union(q1, q2).toSql();
unionAll(q1, q2).orderBy(col('id').asc()).limit(10).toSql();
intersect(q1, q2).toSql();
except(q1, q2).toSql();

AST Visitor

Walk, search, and transform parsed AST nodes.

import {
  parse, Dialect, walk, transform, findAll, findFirst, findByType,
  getColumns, getColumnNames, getTableNames, renameColumns, renameTables,
  addWhere, removeWhere, setLimit, setDistinct, qualifyColumns,
  getAggregateFunctions, hasSubqueries, nodeCount,
} from '@polyglot-sql/sdk';

const { ast } = parse('SELECT a, b FROM t WHERE x > 1', Dialect.Generic);

// Walk all nodes with visitor callbacks
walk(ast, {
  enter: (node) => console.log('Entering:', node),
  column: (node) => console.log('Found column:', node),
});

// Search for nodes
const columns = getColumns(ast);
const first = findFirst(ast, (node) => getExprType(node) === 'column');
const selects = findByType(ast, 'select');

// Get names as strings
const colNames = getColumnNames(ast);   // ['a', 'b']
const tableNames = getTableNames(ast);  // ['t']

// Check for specific constructs
const hasAggs = hasAggregates(ast);
const hasSubs = hasSubqueries(ast);
const count = nodeCount(ast);

// Transform AST nodes
const renamed = renameColumns(ast, { a: 'alpha', b: 'beta' });
const renamedTables = renameTables(ast, { t: 'users' });
const qualified = qualifyColumns(ast, 'users');

// Modify query structure
const withLimit = setLimit(ast, 100);
const distinct = setDistinct(ast, true);
const noWhere = removeWhere(ast);

Validation

Syntax Validation

import { validate } from '@polyglot-sql/sdk';

const result = validate('SELECT * FROM users', 'postgresql');
if (!result.valid) {
  for (const err of result.errors) {
    console.log(`${err.code}: ${err.message} (line ${err.line}, col ${err.column})`);
  }
}

Semantic Validation

const result = validate('SELECT * FROM users', 'postgresql', { semantic: true });
// May also report warnings like "SELECT * is discouraged"

Schema Validation

import { validateWithSchema } from '@polyglot-sql/sdk';

const schema = {
  tables: [
    {
      name: 'users',
      columns: [
        { name: 'id', type: 'integer', primaryKey: true },
        { name: 'name', type: 'varchar' },
        { name: 'email', type: 'varchar', unique: true },
      ],
      primaryKey: ['id'],
      uniqueKeys: [['email']],
    },
    {
      name: 'orders',
      columns: [
        { name: 'id', type: 'integer', primaryKey: true },
        {
          name: 'user_id',
          type: 'integer',
          references: { table: 'users', column: 'id' },
        },
        { name: 'total', type: 'decimal' },
      ],
      foreignKeys: [
        {
          columns: ['user_id'],
          references: { table: 'users', columns: ['id'] },
        },
      ],
    },
  ],
};

const result = validateWithSchema(
  'SELECT id, total FROM orders',
  schema,
  'postgresql',
  {
    checkTypes: true,
    checkReferences: true,
    semantic: true,
  },
);

validateWithSchema supports:

  • identifier checks (E200, E201) for unknown tables/columns
  • optional type checks (checkTypes) for comparisons, predicates, arithmetic, assignments, and set operations
  • optional reference checks (checkReferences) for schema FK integrity and query-level join/reference quality

Migration note: checkTypes and checkReferences are opt-in and default to false, so existing schema validation behavior stays backward-compatible until these options are enabled.

Common diagnostic codes:

| Code | Meaning | |------|---------| | E200 | Unknown table | | E201 | Unknown column | | E210-E217 | Type incompatibilities (strict mode errors) | | W210-W216 | Type coercion warnings (non-strict mode) | | E220 | Invalid foreign key/reference metadata in schema | | E221 | Ambiguous unqualified column in multi-table scope | | W220 | Cartesian join warning | | W221 | JOIN predicate does not use declared FK relationship | | W222 | Weak reference integrity warning (non-strict mode) |

// Ambiguous unqualified column in a join (E221 in strict mode)
const ambiguous = validateWithSchema(
  'SELECT id FROM users u JOIN orders o ON u.id = o.user_id',
  schema,
  'postgresql',
  { checkReferences: true },
);

// Cartesian join warning (W220)
const cartesian = validateWithSchema(
  'SELECT * FROM users u JOIN orders o',
  schema,
  'postgresql',
  { checkReferences: true },
);

Tokenize

Access the raw SQL token stream with full source position spans. Useful for syntax highlighting, custom linters, or editor integrations.

import { tokenize, Dialect } from '@polyglot-sql/sdk';

const result = tokenize('SELECT a, b FROM t', Dialect.Generic);
if (result.success) {
  for (const token of result.tokens!) {
    console.log(token.tokenType, token.text, token.span);
    // "Select" "SELECT" { start: 0, end: 6, line: 1, column: 1 }
    // "Var"    "a"      { start: 7, end: 8, line: 1, column: 8 }
    // ...
  }
}

Each token includes:

| Field | Type | Description | |-------|------|-------------| | tokenType | string | Token type name (e.g. "Select", "Var", "Comma") | | text | string | Raw source text of the token | | span | SpanInfo | Source position: start/end byte offsets, line/column (1-based) | | comments | string[] | Leading comments attached to this token | | trailingComments | string[] | Trailing comments attached to this token |

Error Reporting

Parse, transpile, and tokenize errors include source position information with both line/column and byte offset ranges, making it easy to highlight errors in editors or show precise error messages.

import { parse, transpile, Dialect } from '@polyglot-sql/sdk';

const result = parse('SELECT 1 +', Dialect.Generic);
if (!result.success) {
  console.log(result.error);       // "Parse error at line 1, column 11: ..."
  console.log(result.errorLine);   // 1
  console.log(result.errorColumn); // 11
  console.log(result.errorStart);  // 10 (byte offset)
  console.log(result.errorEnd);    // 11 (byte offset, exclusive)
}

ParseResult, TranspileResult, and TokenizeResult include optional position fields:

| Field | Type | Description | |-------|------|-------------| | errorLine | number \| undefined | 1-based line number where the error occurred | | errorColumn | number \| undefined | 1-based column number where the error occurred | | errorStart | number \| undefined | Start byte offset of the error range (0-based) | | errorEnd | number \| undefined | End byte offset of the error range (exclusive) |

These fields are only present when success is false. On success, they are undefined.

// Use with transpile errors too
const result = transpile('SELECT FROM WHERE', Dialect.MySQL, Dialect.PostgreSQL);
if (!result.success) {
  // Pinpoint the exact location in the source SQL
  console.log(`Error at ${result.errorLine}:${result.errorColumn}: ${result.error}`);
}

Column Lineage

Trace how columns flow through SQL queries, from source tables to the result set.

import { lineage, lineageWithSchema, getSourceTables } from '@polyglot-sql/sdk';

// Trace a column through joins, CTEs, and subqueries
const result = lineage('total', 'SELECT o.total FROM orders o JOIN users u ON o.user_id = u.id');
if (result.success) {
  console.log(result.lineage.name);        // 'total'
  console.log(result.lineage.downstream);  // source nodes
}

// Schema-aware lineage (same schema format as validateWithSchema)
const schema = {
  tables: [
    { name: 'users', columns: [{ name: 'id', type: 'INT' }] },
    { name: 'orders', columns: [{ name: 'user_id', type: 'INT' }] },
  ],
};
const schemaLineage = lineageWithSchema(
  'id',
  'SELECT id FROM users u JOIN orders o ON u.id = o.user_id',
  schema,
);

// Get all source tables that contribute to a column
const tables = getSourceTables('total', 'SELECT o.total FROM orders o JOIN users u ON o.user_id = u.id');
if (tables.success) {
  console.log(tables.tables);  // ['orders']
}

SQL Diff

Compare two SQL statements and get a list of edit operations using the ChangeDistiller algorithm.

import { diff, hasChanges, changesOnly } from '@polyglot-sql/sdk';

const result = diff(
  'SELECT a, b FROM t WHERE x > 1',
  'SELECT a, c FROM t WHERE x > 2',
);

if (result.success) {
  console.log(hasChanges(result.edits));     // true
  console.log(changesOnly(result.edits));    // only insert/remove/move/update edits

  for (const edit of result.edits) {
    // edit.type: 'insert' | 'remove' | 'move' | 'update' | 'keep'
    console.log(edit.type, edit.source, edit.target);
  }
}

Query Planner

Convert a SQL query into an execution plan represented as a DAG of steps.

import { plan } from '@polyglot-sql/sdk';

const result = plan('SELECT dept, SUM(salary) FROM employees GROUP BY dept');
if (result.success) {
  const { root, leaves } = result.plan;
  console.log(root.kind);       // 'aggregate'
  console.log(leaves[0].kind);  // 'scan'
  console.log(leaves[0].name);  // 'employees'
}

Class-Based API

For an object-oriented style, use the singleton Polyglot class:

import { Polyglot, Dialect } from '@polyglot-sql/sdk';

const pg = Polyglot.getInstance();
const result = pg.transpile('SELECT 1', Dialect.MySQL, Dialect.PostgreSQL);
const formatted = pg.format('SELECT a,b FROM t');
const formattedSafe = pg.formatWithOptions('SELECT a,b FROM t', Dialect.Generic, {
  maxInputBytes: 2 * 1024 * 1024,
  maxSetOpChain: 128,
});

API Reference

Core Functions

| Function | Description | |----------|-------------| | transpile(sql, read, write) | Transpile SQL between dialects | | parse(sql, dialect?) | Parse SQL into AST | | generate(ast, dialect?) | Generate SQL from AST | | format(sql, dialect?) | Pretty-print SQL | | formatWithOptions(sql, dialect?, options?) | Pretty-print SQL with guard overrides | | tokenize(sql, dialect?) | Tokenize SQL into a token stream with source spans | | validate(sql, dialect?, options?) | Validate SQL syntax/semantics | | validateWithSchema(sql, schema, dialect?, options?) | Validate against a database schema | | getDialects() | List supported dialect names | | getVersion() | Get library version |

Analysis Functions

| Function | Description | |----------|-------------| | lineage(column, sql, dialect?, trimSelects?) | Trace column lineage through a query | | lineageWithSchema(column, sql, schema, dialect?, trimSelects?) | Trace lineage with schema-based qualification | | getSourceTables(column, sql, dialect?) | Get source tables for a column | | diff(source, target, dialect?, options?) | Diff two SQL statements | | hasChanges(edits) | Check if diff has non-keep edits | | changesOnly(edits) | Filter to only change edits | | plan(sql, dialect?) | Build a query execution plan DAG |

Expression Helpers

| Function | Description | |----------|-------------| | col(name) | Column reference | | lit(value) | Literal value (string, number, boolean, null) | | star() | Star (*) expression | | sqlNull() | NULL literal | | boolean(value) | Boolean literal | | table(name) | Table reference | | sqlExpr(sql) | Parse raw SQL fragment | | condition(sql) | Alias for sqlExpr | | func(name, ...args) | Function call | | not(expr) | NOT expression | | cast(expr, type) | CAST expression | | alias(expr, name) | Alias expression | | and(...conditions) | Chain with AND | | or(...conditions) | Chain with OR |

Query Builders

| Builder | Constructor | Description | |---------|------------|-------------| | SelectBuilder | select(...cols) | SELECT queries | | InsertBuilder | insertInto(table) / insert(table) | INSERT statements | | UpdateBuilder | update(table) | UPDATE statements | | DeleteBuilder | deleteFrom(table) / del(table) | DELETE statements | | MergeBuilder | mergeInto(table) | MERGE statements | | CaseBuilder | caseWhen() / caseOf(expr) | CASE expressions | | SetOpBuilder | union() / unionAll() / intersect() / except() | Set operations |

AST Walker

| Function | Description | |----------|-------------| | walk(node, visitor) | Walk all AST nodes with visitor callbacks | | findAll(node, predicate) | Find nodes matching a predicate | | findByType(node, type) | Find all nodes of a specific type | | findFirst(node, predicate) | Find the first matching node | | some(node, predicate) | Check if any node matches | | every(node, predicate) | Check if all nodes match | | countNodes(node, predicate) | Count nodes matching a predicate | | getChildren(node) | Get direct children of a node | | getColumns(node) | Get all column expression nodes | | getTables(node) | Get all table expression nodes | | getIdentifiers(node) | Get all identifier nodes | | getFunctions(node) | Get all function call nodes | | getAggregateFunctions(node) | Get all aggregate function nodes | | getWindowFunctions(node) | Get all window function nodes | | getSubqueries(node) | Get all subquery nodes | | getLiterals(node) | Get all literal nodes | | getColumnNames(node) | Get column names as strings | | getTableNames(node) | Get table names as strings | | hasAggregates(node) | Check for aggregate functions | | hasWindowFunctions(node) | Check for window functions | | hasSubqueries(node) | Check for subqueries | | nodeCount(node) | Total number of AST nodes | | getDepth(node) | Max depth of the AST tree | | getParent(root, target) | Find parent of a node | | findAncestor(root, target, predicate) | Find matching ancestor | | getNodeDepth(root, target) | Depth of a specific node |

AST Transformer

| Function | Description | |----------|-------------| | transform(node, config) | Immutable tree transformation with callbacks | | replaceNodes(node, predicate, replacement) | Replace nodes matching a predicate | | replaceByType(node, type, replacement) | Replace nodes of a specific type | | renameColumns(node, mapping) | Rename columns in AST | | renameTables(node, mapping) | Rename tables in AST | | qualifyColumns(node, table) | Add table qualifier to columns | | addWhere(node, condition, operator?) | Add/extend WHERE clause (AND/OR) | | removeWhere(node) | Remove WHERE clause | | addSelectColumns(node, ...columns) | Add columns to SELECT | | removeSelectColumns(node, predicate) | Remove columns from SELECT | | setLimit(node, limit) | Set LIMIT clause | | setOffset(node, offset) | Set OFFSET clause | | removeLimitOffset(node) | Remove LIMIT and OFFSET | | setDistinct(node, distinct?) | Set SELECT DISTINCT | | clone(node) | Deep clone AST | | remove(node, predicate) | Remove nodes matching a predicate |

Supported Dialects

| Dialect | Enum Value | |---------|-----------| | Athena | Dialect.Athena | | BigQuery | Dialect.BigQuery | | ClickHouse | Dialect.ClickHouse | | CockroachDB | Dialect.CockroachDB | | Databricks | Dialect.Databricks | | Doris | Dialect.Doris | | Dremio | Dialect.Dremio | | Drill | Dialect.Drill | | Druid | Dialect.Druid | | DuckDB | Dialect.DuckDB | | Dune | Dialect.Dune | | Exasol | Dialect.Exasol | | Fabric | Dialect.Fabric | | Hive | Dialect.Hive | | Materialize | Dialect.Materialize | | MySQL | Dialect.MySQL | | Oracle | Dialect.Oracle | | PostgreSQL | Dialect.PostgreSQL | | Presto | Dialect.Presto | | Redshift | Dialect.Redshift | | RisingWave | Dialect.RisingWave | | SingleStore | Dialect.SingleStore | | Snowflake | Dialect.Snowflake | | Solr | Dialect.Solr | | Spark | Dialect.Spark | | SQLite | Dialect.SQLite | | StarRocks | Dialect.StarRocks | | Tableau | Dialect.Tableau | | Teradata | Dialect.Teradata | | TiDB | Dialect.TiDB | | Trino | Dialect.Trino | | TSQL | Dialect.TSQL |

CDN Usage

For browser use without a bundler:

<script type="module">
  import polyglot from 'https://unpkg.com/@polyglot-sql/sdk/dist/cdn/polyglot.esm.js';
  // or: https://cdn.jsdelivr.net/npm/@polyglot-sql/sdk/dist/cdn/polyglot.esm.js

  const { transpile, Dialect } = polyglot;
  const result = transpile('SELECT 1', Dialect.MySQL, Dialect.PostgreSQL);
  console.log(result.sql);
</script>

CommonJS (CJS) Usage

For Node.js projects using require(), the SDK ships a CJS build. Since WASM cannot be loaded synchronously, you must call init() before using any other function:

const { init, transpile, parse, select, col, lit, isInitialized } = require('@polyglot-sql/sdk');

async function main() {
  await init();

  // Now all functions work
  const result = transpile('SELECT IFNULL(a, b)', 'mysql', 'postgresql');
  console.log(result.sql[0]); // SELECT COALESCE(a, b)

  const parsed = parse('SELECT 1', 'generic');
  console.log(parsed.success); // true

  const sql = select('id', 'name').from('users')
    .where(col('id').eq(lit(1)))
    .toSql();
  console.log(sql); // SELECT id, name FROM users WHERE id = 1
}

main();

You can check initialization status with isInitialized():

const { init, isInitialized } = require('@polyglot-sql/sdk');

console.log(isInitialized()); // false
await init();
console.log(isInitialized()); // true

Note: The ESM build (import) auto-initializes via top-level await, so init() is not required there. The CJS build requires it because require() is synchronous.

Bundler Configuration

The ESM build uses top-level await (ES2022) and new URL("...", import.meta.url) for WASM loading. Both are web standards supported by all modern bundlers, but some require explicit configuration.

Vite

Vite works out of the box with vite-plugin-wasm. This is the setup used by the Polyglot Playground:

// vite.config.ts
import { defineConfig } from 'vite';
import wasm from 'vite-plugin-wasm';

export default defineConfig({
  plugins: [wasm()],
  build: {
    target: 'esnext', // required for top-level await
  },
  optimizeDeps: {
    exclude: ['@polyglot-sql/sdk'], // prevent esbuild from pre-bundling WASM
  },
});

Next.js

Next.js requires enabling WebAssembly experiments in the webpack config:

// next.config.js
module.exports = {
  webpack: (config) => {
    config.experiments = {
      ...config.experiments,
      asyncWebAssembly: true,
      topLevelAwait: true,
      layers: true,
    };
    return config;
  },
};

For production SSR builds, Next.js may emit .wasm files at a different path than where it tries to load them. A common workaround is to add a plugin that fixes the output path:

// next.config.js
class WasmChunksFixPlugin {
  apply(compiler) {
    compiler.hooks.thisCompilation.tap('WasmChunksFixPlugin', (compilation) => {
      compilation.hooks.processAssets.tap(
        { name: 'WasmChunksFixPlugin' },
        (assets) =>
          Object.entries(assets).forEach(([pathname, source]) => {
            if (!pathname.match(/\.wasm$/)) return;
            compilation.deleteAsset(pathname);
            const name = pathname.split('/')[1];
            const info = compilation.assetsInfo.get(pathname);
            compilation.emitAsset(name, source, info);
          })
      );
    });
  }
}

module.exports = {
  webpack: (config, { isServer, dev }) => {
    config.experiments = {
      ...config.experiments,
      asyncWebAssembly: true,
      topLevelAwait: true,
      layers: true,
    };
    if (!dev && isServer) {
      config.output.webassemblyModuleFilename = 'chunks/[id].wasm';
      config.plugins.push(new WasmChunksFixPlugin());
    }
    return config;
  },
};

webpack 5 (standalone)

Enable the required experiments in your webpack config:

// webpack.config.js
module.exports = {
  experiments: {
    asyncWebAssembly: true,
    topLevelAwait: true,
  },
  // ...
};

topLevelAwait is enabled by default since webpack 5.83.0. asyncWebAssembly must be set explicitly.

Troubleshooting

If you see undefined exports (e.g., Dialect is undefined) or errors about require("fs"):

  1. Make sure asyncWebAssembly and topLevelAwait experiments are enabled in your bundler config
  2. Ensure the bundler is resolving the ESM build (the "import" or "browser" export condition), not the CJS build which uses Node.js APIs
  3. For SSR frameworks (Next.js, Nuxt), check that the .wasm file is correctly included in the server build output

License

MIT