npm package discovery and stats viewer.

Discover Tips

  • General search

    [free text search, go nuts!]

  • Package details

    pkg:[package-name]

  • User packages

    @[username]

Sponsor

Optimize Toolset

I’ve always been into building performant and accessible sites, but lately I’ve been taking it extremely seriously. So much so that I’ve been building a tool to help me optimize and monitor the sites that I build to make sure that I’m making an attempt to offer the best experience to those who visit them. If you’re into performant, accessible and SEO friendly sites, you might like it too! You can check it out at Optimize Toolset.

About

Hi, 👋, I’m Ryan Hefner  and I built this site for me, and you! The goal of this site was to provide an easy way for me to check the stats on my npm packages, both for prioritizing issues and updates, and to give me a little kick in the pants to keep up on stuff.

As I was building it, I realized that I was actually using the tool to build the tool, and figured I might as well put this out there and hopefully others will find it to be a fast and useful way to search and browse npm packages as I have.

If you’re interested in other things I’m working on, follow me on Twitter or check out the open source projects I’ve been publishing on GitHub.

I am also working on a Twitter bot for this site to tweet the most popular, newest, random packages from npm. Please follow that account now and it will start sending out packages soon–ish.

Open Software & Tools

This site wouldn’t be possible without the immense generosity and tireless efforts from the people who make contributions to the world and share their work via open source initiatives. Thank you 🙏

© 2025 – Pkg Stats / Ryan Hefner

@sebastiansanioj/data-api-mapper

v1.0.8

Published

The Data API Mapper is a lightweight wrapper for Amazon Aurora Serverless Data API.

Downloads

5

Readme

Data API Mapper

The Data API Mapper is a lightweight wrapper for Amazon Aurora Serverless Data API.

This library uses: RDS Data Client - AWS SDK for JavaScript v3

Motivation

DataApi requires a special parsing of the fields sent to it. For example:

const select = await data.executeStatement({
  secretArn: <SECRET_ARN>,
  resourceArn: <SECRET_RESOURCE>,
  database: <DATABASE>,
  sql: 'SELECT * FROM myTable WHERE number=:number and name=:name',
  parameters: [
    { name: 'name', value: { stringValue: 'Roberto Perez' } },
    { number: 'age', value: { longValue: 10 } },
  ]
).promise()

Specifying all the mapping can be quite tedious and then the answer also comes with an unfriendly format and with its own data type of dataApi. For example:

[[{"longValue":1},{"stringValue":"first row"},{"stringValue":"{\"int_value\": 1, \"float_value\": 1.11, \"string_vale\": \"string1\"}"},{"stringValue":"1.12345"},{"doubleValue":1.11},{"longValue":1},{"stringValue":"1976-11-02 08:45:00"},{"isNull":true},{"isNull":true},{"isNull":true},{"isNull":true},{"stringValue":"2021-03-03 15:51:48.082288"},{"stringValue":"1976-11-02"}]]

So this library is in charge of doing all this mapping work, handling transactions, etc... for you.

How to use this module

dataApiClient = new DataApiClient(secret_arn, databaseArn, databaseName, region)

Running a query ⚙️

Once initialized, running a query is super simple. Use the query() method and pass in your SQL statement:

result = await dataApiClient.query('SELECT * FROM myTable')

By default, this will return the converted elements.

For example, for this database:

CREATE TABLE aurora_data_api_node_test (id SERIAL, a_name TEXT, doc JSONB DEFAULT '{}', num_numeric NUMERIC (10, 5) DEFAULT 0.0, num_float float, num_integer integer, ts TIMESTAMP WITH TIME ZONE, field_string_null TEXT NULL, field_long_null integer NULL, field_doc_null JSONB NULL, field_boolean BOOLEAN NULL, tz_notimezone TIMESTAMP, a_date DATE);
INSERT INTO aurora_data_api_node_test (a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES ('first row', '{"string_vale": "string1", "int_value": 1, "float_value": 1.11}', 1.12345, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');

this query

await dataApiClient.query("select * from aurora_data_api_node_test")

will return this elements:

  • Items:
  [{
    "id": 1,
    "a_name": "first row",
    "doc": {
      "int_value": 1,
      "float_value": 1.11,
      "string_vale": "string1"
    },
    "num_numeric": 1.12345,
    "num_float": 1.11,
    "num_integer": 1,
    "ts": "1976-11-02T11:45:00.000Z",
    "field_string_null": null,
    "field_long_null": null,
    "field_doc_null": null,
    "field_boolean": null,
    "tz_notimezone": "2021-03-03T18:51:48.082Z",
    "a_date": "1976-11-02"
  }]

Running a query with parameters ⚙️

To query with parameters, you can use named parameters in your SQL, and then provide an object containing your parameters as the second argument to the query() method and the client does the conversion for you:


result = await dataApiClient.query(
    'SELECT * FROM myTable WHERE id = :id',
    { 'id': 2 }
)

Transactions

You can initialize a transaction with this method:

const transaction = await dataApiClient.beginTransaction()

and then you can run queries (INSERT,UPDATE,DELETE,SELECT) and finally you can do commit/rollback.

For example:

const transaction = await dataApiClient.beginTransaction()
const thirdInsert = "INSERT INTO aurora_data_api_node_test (a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES ('first row', '{\"string_vale\": \"string1\", \"int_value\": 1, \"float_value\": 1.11}', 1.12345, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');"
const response = await transaction.query(thirdInsert)
console.log(response)
const responseThirdInsert = await transaction.query('SELECT * FROM aurora_data_api_node_test where id=:id', {id: 3})
const resultExpected = [{
    "id": 3,
    "a_name": "first row",
    "doc": {
        "int_value": 1,
        "float_value": 1.11,
        "string_vale": "string1"
    },
    "num_numeric": 1.12345,
    "num_float": 1.11,
    "num_integer": 1,
    "ts": "1976-11-02T11:45:00.000Z",
    "field_string_null": null,
    "field_long_null": null,
    "field_doc_null": null,
    "field_boolean": null,
    "tz_notimezone": "2021-03-03T18:51:48.082Z",
    "a_date": "1976-11-02"
}]
//Do rollback
await transaction.rollbackTransaction() // or await transaction.commitTransaction()
const responseThirdInsertDoesntExists = await dataApiClient.query('SELECT * FROM aurora_data_api_node_test where id=:id', {id: 3})

Running a batchQueries ⚙️

If you want, you can run batchQuery. You create a query with an array with the parameters and then you invoke the batchQuery method. The result is the final number of modified records.


const insert = "INSERT INTO aurora_data_api_node_test (id, a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES (:id, :a_name, '{\"string_vale\": \"string1\", \"int_value\": 1, \"float_value\": 1.11}',:num_numeric, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');"
const params = [{
    id: 9,
    a_name: 'a',
    num_numeric: 1.123
},
{
    id: 10,
    a_name: 'b',
    num_numeric: 2.123
},
{
    id: 11,
    a_name: 'c',
    num_numeric: 3.123
},
{
    id: 12,
    a_name: 'd',
    num_numeric: 4.432
}
]
const inserts = await dataApiClient.batchInsert(insert, params)

or with an update...


const update = "UPDATE aurora_data_api_node_test SET A_NAME = :a_name WHERE ID=:id;"
const params = [{
 id: 1,
 a_name: 'a',
 num_numeric: 1.123
},
{
  id: 2,
  a_name: 'b',
  num_numeric: 2.123
}
]
const inserts = await dataApiClient.batchQuery(update, params)

You can use batchQuery inside a transaction too. For example:


const transaction = await dataApiClient.beginTransaction()
            const insert = "INSERT INTO aurora_data_api_node_test (id, a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES (:id, :a_name, '{\"string_vale\": \"string1\", \"int_value\": 1, \"float_value\": 1.11}',:num_numeric, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');"
const params = [{
    id: 14,
    a_name: 'a',
    num_numeric: 1.123
},
{
    id: 15,
    a_name: 'b',
    num_numeric: 2.123
},
{
    id: 16,
    a_name: 'c',
    num_numeric: 3.123
},
{
    id: 17,
    a_name: 'd',
    num_numeric: 4.312
}
]
const inserts = await transaction.batchInsert(insert, params) // or await transaction.batchQuery(update, params)
await transaction.commitTransaction() // or await transaction.rollbackTransaction()

Running a query with pagination ⚙️

If you want, you can run a pagination query. You only need to use the queryPaginated method. This method receives the sql, the parameters and the pageSize. If you pass a pageSize with the value 50, the query only returns the first 50 elements. The default value for the pageSize is 100. An example of execution:

const response = await dataApiClient.queryPaginated('SELECT * FROM aurora_data_api_node_test', [],50)

Test ⚙️

For runs the test, you only need to configure a .env.json with your values.

DB_SECRET_ARN=<SECRET_ARN>
DB_RESOURCE_ARN=<RESOURCE_ARN>
DB_DATABASE=<DATABASE_NAME>
REGION=<REGION>

Authors ✒️

  • Carefull Engineering Team

Expressions of Gratitude 🎁

  • tell others about the project 📢
  • Invite someone of the team for a 🍺 or a coffee ☕ .
  • etc.