dynaflow

1.1.9 • Public • Published

Dynaflow

🐊
For all your flow needs
A high-level driver for using DynamoDB in NodeJS

Build Status Coverage Status

Table of Contents

Features

  • Perform all standard DynamoDB operations
  • Automatic pagination
  • Batch write streams

Example

npm install dynaflow
const db = new Dynaflow({ region: 'us-east-1' });
    
await db.createTable({
  TableName: 'MyTable',
  AttributeDefinitions: [{ AttributeName: 'id', AttributeType: 'N' }],
  KeySchema: [{ AttributeName: 'id', KeyType: 'HASH' }],
});
    
await db.putItem({
  TableName: 'MyTable',
  Item: {
    id: { N: 1 },
    someValue: { S: 'hello world' },
  },
});
 
const response = await db.getItem({
  TableName: 'MyTable',
  Key: {
    id: { N: 1 },
  },
});
 
assert(response.Item.someValue.S === 'hello world');

Architecture

Dynaflow uses promises heavily. In some cases, it leverages a higher level promise paradigm referred to as a river.

Rivers

In JavaScript, you might be familiar with the use of promises to deal with asynchronous events. A promise represents a single event — but what if we want to represent many events happening over time? This is what Rivers are for. Rivers are composable object streams (similar to ReactiveX Observables) that fit very nicely into the JavaScript ecosystem. Check out the wise-river page to learn more about why they’re more powerful than traditional streams.

Normally, when you query DynamoDB it responds with data broken up into pages. They do this for good reason, but typically in our applications we’d rather deal with the query results as if it were an iterable, or a stream of objects. Using Rivers, you can .map(), .filter(), and .reduce() over the results without dealing with pagination logic. It will automatically fetch new pages until you’re done using the river chain, at which point it will stop automatically.

The most powerful use of Rivers in dynaflow is to provide an abstraction for DynamoDB’s batch write functionality. DynamoDB supports batch requests, allowing you to put and delete items in different tables at the same time, which is useful for saving bandwidth. Normally, this is a complicated process involving “partial errors,” “retries,” and more. But with dynaflow, you can easily create a bandwidth-efficient firehose by just writing objects to a river — the complicated logic is handled for you, so you can easily operate on the results as they occur.

API

new Dynaflow(options)

Creates and returns a new client for DynamoDB. The given options are used directly by the aws-sdk.

Reading

.query(params) -> river

Performs a Query operation by passing the params to ask-sdk, automatically paginating through each page in the result set. The returned river contains each result object of each request made.

const params = {
  TableName: 'MyTable',
  KeyConditionExpression: 'myPrimaryKey = :val',
  ExpressionAttributeValues: { ':val': { B: 'hZn6NqO18x8=' } },
  ItemsOnly: true,
};
 
db.query(params)
  .filter(validateItem)
  .map(transformItem)
  .forEach(logItem)
  .drain()
  .then(() => {
    console.log('all done!');
  });

If the river is rejected or cancelled, iterating will stop and no more requests will be made.

If the ItemsOnly option is true, the river will contain each individual table item, rather than the entire result objects.

Other methods of reading data

.scan(params) -> river

Similar to .query(), but performs a Scan operation instead.

.getItem(params) -> promise

A promisified version of the corresponding aws-sdk method.

Writing

.batchWriteItem(requests, [params]) -> river

Given a river of request objects, this method will group those requests into batches before sending them to DynamoDB. The returned river contains the results of each batch request that is sent.

A request object can be either:

Each result object has (in addition to the fields returned by DynamoDB) a Count field, indicating how many requests were successfully processed. If a request results in an error, the associated Error object will also have a Count field, indicating how many requests failed.

const requests = [
  { TableName: 'MyTable', Item: someData },
  { TableName: 'MyTable', Item: otherData },
];
 
db.batchWriteItem(River.from(requests))
  .consume(({ Count }) => {
    console.log(`Processed a batch of ${Count} items!`);
  });

Each result object will always have an empty UnprocessedItems field, because this method automatically handles retries for you.

If the Timeout option is given, incomming requests will not be buffered for longer than the specified number of milliseconds.

If the Manual option is true, the returned river will output the batch objects without sending them to DynamoDB. Each batch object has a .send() method which you MUST use to send execute the batch request, which returns a promise for the request's result.

Other methods of writing data

.putItem(params) -> promise

A promisified version of the corresponding aws-sdk method.

.updateItem(params) -> promise

A promisified version of the corresponding aws-sdk method.

.deleteItem(params) -> promise

A promisified version of the corresponding aws-sdk method.

Schema

Schema related methods

.createTable(params) -> promise

A promisified version of the corresponding aws-sdk method.

.updateTable(params) -> promise

A promisified version of the corresponding aws-sdk method.

.deleteTable(params) -> promise

A promisified version of the corresponding aws-sdk method.

.describeTable(params) -> promise

A promisified version of the corresponding aws-sdk method.

.listTables(params) -> river

Similar to .query(), but performs a ListTables operation instead.

Metadata

Metadata related methods

.describeLimits(params) -> promise

A promisified version of the corresponding aws-sdk method.

.describeTimeToLive(params) -> promise

A promisified version of the corresponding aws-sdk method.

.listTagsOfResource(params) -> river

Similar to .query(), but performs a ListTagsOfResource operation instead.

.tagResource(params) -> promise

A promisified version of the corresponding aws-sdk method.

.untagResource(params) -> promise

A promisified version of the corresponding aws-sdk method.

.waitFor(params) -> promise

A promisified version of the corresponding aws-sdk method.

.updateTimeToLive(params) -> promise

A promisified version of the corresponding aws-sdk method.

Versioning

We follow the Semantic Versioning convention where a version number is composed of MAJOR.MINOR.PATCH where we increment each with the following rules:

  • MAJOR version when you make incompatible API changes,
  • MINOR version when you add functionality in a backwards-compatible manner, and
  • PATCH version when you make backwards-compatible bug fixes.

Testing

The tests are made to be run against a locally running DynamoDB service. To assist with spinning up a local instance several helper commands have been added to the Makefile using docker.

make docker-dynamo-start - Will spin up an instance of a dynamodb-local image bound to a configurable local port.

make docker-dynamo-stop - Will destroy the previously created docker instance.

make test - Will run the test suite against the configurable local port.

make test-docker - Will create the docker instance, run the tests, and destroy the docker instance.

The make variables CONTAINER_NAME (default dynaflow-testing-dynamo) and BIND_PORT (default 8000) can be used to configure the name of the created container and the local port to use: make test-docker CONTAINER_NAME=testing-dynaflow BIND_PORT=6545

asciicast

Contributing

Contributions, big or small, are welcome. If you have a suggestion for a feature or an issue with the library please feel free to make a Github issue so that we can be made aware of it. If you have written something that you think would be a good addition we would love you to make a PR so that we can work together to see if the changes can be integrated.

Package Sidebar

Install

npm i dynaflow

Weekly Downloads

11

Version

1.1.9

License

none

Unpacked Size

49 kB

Total Files

21

Last publish

Collaborators

  • ssawa
  • nambok