1.0.7 • Published 3 years ago

@lifs/data-api-mapper v1.0.7

Weekly downloads
-
License
MIT
Repository
github
Last release
3 years ago

Data API Mapper

The Data API Mapper is a lightweight wrapper for Amazon Aurora Serverless Data API.

This library uses: RDS Data Client - AWS SDK for JavaScript v3

Motivation

DataApi requires a special parsing of the fields sent to it. For example:

const select = await data.executeStatement({
  secretArn: <SECRET_ARN>,
  resourceArn: <SECRET_RESOURCE>,
  database: <DATABASE>,
  sql: 'SELECT * FROM myTable WHERE number=:number and name=:name',
  parameters: [
    { name: 'name', value: { stringValue: 'Roberto Perez' } },
    { number: 'age', value: { longValue: 10 } },
  ]
).promise()

Specifying all the mapping can be quite tedious and then the answer also comes with an unfriendly format and with its own data type of dataApi. For example:

[[{"longValue":1},{"stringValue":"first row"},{"stringValue":"{\"int_value\": 1, \"float_value\": 1.11, \"string_vale\": \"string1\"}"},{"stringValue":"1.12345"},{"doubleValue":1.11},{"longValue":1},{"stringValue":"1976-11-02 08:45:00"},{"isNull":true},{"isNull":true},{"isNull":true},{"isNull":true},{"stringValue":"2021-03-03 15:51:48.082288"},{"stringValue":"1976-11-02"}]]

So this library is in charge of doing all this mapping work, handling transactions, etc... for you.

How to use this module

dataApiClient = new DataApiClient(secret_arn, databaseArn, databaseName, region)

Running a query ⚙️

Once initialized, running a query is super simple. Use the query() method and pass in your SQL statement:

result = await dataApiClient.query('SELECT * FROM myTable')

By default, this will return the converted elements.

For example, for this database:

CREATE TABLE aurora_data_api_node_test (id SERIAL, a_name TEXT, doc JSONB DEFAULT '{}', num_numeric NUMERIC (10, 5) DEFAULT 0.0, num_float float, num_integer integer, ts TIMESTAMP WITH TIME ZONE, field_string_null TEXT NULL, field_long_null integer NULL, field_doc_null JSONB NULL, field_boolean BOOLEAN NULL, tz_notimezone TIMESTAMP, a_date DATE);
INSERT INTO aurora_data_api_node_test (a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES ('first row', '{"string_vale": "string1", "int_value": 1, "float_value": 1.11}', 1.12345, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');

this query

await dataApiClient.query("select * from aurora_data_api_node_test")

will return this elements:

  • Items:
      [{
        "id": 1,
        "a_name": "first row",
        "doc": {
          "int_value": 1,
          "float_value": 1.11,
          "string_vale": "string1"
        },
        "num_numeric": 1.12345,
        "num_float": 1.11,
        "num_integer": 1,
        "ts": "1976-11-02T11:45:00.000Z",
        "field_string_null": null,
        "field_long_null": null,
        "field_doc_null": null,
        "field_boolean": null,
        "tz_notimezone": "2021-03-03T18:51:48.082Z",
        "a_date": "1976-11-02"
      }]

Running a query with parameters ⚙️

To query with parameters, you can use named parameters in your SQL, and then provide an object containing your parameters as the second argument to the query() method and the client does the conversion for you:

result = await dataApiClient.query(
    'SELECT * FROM myTable WHERE id = :id',
    { 'id': 2 }
)

Transactions

You can initialize a transaction with this method:

const transaction = await dataApiClient.beginTransaction()

and then you can run queries (INSERT,UPDATE,DELETE,SELECT) and finally you can do commit/rollback.

For example:

const transaction = await dataApiClient.beginTransaction()
const thirdInsert = "INSERT INTO aurora_data_api_node_test (a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES ('first row', '{\"string_vale\": \"string1\", \"int_value\": 1, \"float_value\": 1.11}', 1.12345, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');"
const response = await transaction.query(thirdInsert)
console.log(response)
const responseThirdInsert = await transaction.query('SELECT * FROM aurora_data_api_node_test where id=:id', {id: 3})
const resultExpected = [{
    "id": 3,
    "a_name": "first row",
    "doc": {
        "int_value": 1,
        "float_value": 1.11,
        "string_vale": "string1"
    },
    "num_numeric": 1.12345,
    "num_float": 1.11,
    "num_integer": 1,
    "ts": "1976-11-02T11:45:00.000Z",
    "field_string_null": null,
    "field_long_null": null,
    "field_doc_null": null,
    "field_boolean": null,
    "tz_notimezone": "2021-03-03T18:51:48.082Z",
    "a_date": "1976-11-02"
}]
//Do rollback
await transaction.rollbackTransaction() // or await transaction.commitTransaction()
const responseThirdInsertDoesntExists = await dataApiClient.query('SELECT * FROM aurora_data_api_node_test where id=:id', {id: 3})

Running a batchQueries ⚙️

If you want, you can run batchQuery. You create a query with an array with the parameters and then you invoke the batchQuery method. The result is the final number of modified records.

const insert = "INSERT INTO aurora_data_api_node_test (id, a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES (:id, :a_name, '{\"string_vale\": \"string1\", \"int_value\": 1, \"float_value\": 1.11}',:num_numeric, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');"
const params = [{
    id: 9,
    a_name: 'a',
    num_numeric: 1.123
},
{
    id: 10,
    a_name: 'b',
    num_numeric: 2.123
},
{
    id: 11,
    a_name: 'c',
    num_numeric: 3.123
},
{
    id: 12,
    a_name: 'd',
    num_numeric: 4.432
}
]
const inserts = await dataApiClient.batchInsert(insert, params)

or with an update...

const update = "UPDATE aurora_data_api_node_test SET A_NAME = :a_name WHERE ID=:id;"
const params = [{
 id: 1,
 a_name: 'a',
 num_numeric: 1.123
},
{
  id: 2,
  a_name: 'b',
  num_numeric: 2.123
}
]
const inserts = await dataApiClient.batchQuery(update, params)

You can use batchQuery inside a transaction too. For example:

const transaction = await dataApiClient.beginTransaction()
            const insert = "INSERT INTO aurora_data_api_node_test (id, a_name, doc, num_numeric, num_float, num_integer, ts, tz_notimezone, a_date) VALUES (:id, :a_name, '{\"string_vale\": \"string1\", \"int_value\": 1, \"float_value\": 1.11}',:num_numeric, 1.11, 1, '1976-11-02 08:45:00 UTC', '2021-03-03 15:51:48.082288', '1976-11-02');"
const params = [{
    id: 14,
    a_name: 'a',
    num_numeric: 1.123
},
{
    id: 15,
    a_name: 'b',
    num_numeric: 2.123
},
{
    id: 16,
    a_name: 'c',
    num_numeric: 3.123
},
{
    id: 17,
    a_name: 'd',
    num_numeric: 4.312
}
]
const inserts = await transaction.batchInsert(insert, params) // or await transaction.batchQuery(update, params)
await transaction.commitTransaction() // or await transaction.rollbackTransaction()

Running a query with pagination ⚙️

If you want, you can run a pagination query. You only need to use the queryPaginated method. This method receives the sql, the parameters and the pageSize. If you pass a pageSize with the value 50, the query only returns the first 50 elements. The default value for the pageSize is 100. An example of execution:

const response = await dataApiClient.queryPaginated('SELECT * FROM aurora_data_api_node_test', [],50)

Test ⚙️

For runs the test, you only need to configure a .env.json with your values.

DB_SECRET_ARN=<SECRET_ARN>
DB_RESOURCE_ARN=<RESOURCE_ARN>
DB_DATABASE=<DATABASE_NAME>
REGION=<REGION>

Authors ✒️

  • Carefull Engineering Team

Expressions of Gratitude 🎁

  • tell others about the project 📢
  • Invite someone of the team for a 🍺 or a coffee ☕ .
  • etc.
1.0.7

3 years ago

1.0.6

3 years ago

1.0.5

3 years ago

1.0.4

3 years ago

1.0.2

3 years ago

1.0.1

3 years ago

1.0.3

3 years ago

1.0.0

3 years ago