PouchDB Dump CLI
Dump a PouchDB or CouchDB database to a file, or to multiple files. Then you can load it into another PouchDB or CouchDB database.
When you do so, it will be as if you had replicated from one database to the other, but without all the standard HTTP chattiness of the CouchDB protocol.
In other words, this is a very fast way to do initial replication. And it's also safe to do multiple times, since the "load" operation is idempotent.
To load the dump file, use pouchdb-load.
$ npm install -g pouchdb-dump-cli
To dump a CouchDB:
$ pouchdb-dump http://localhost:5984/mydb > dump.txt
To dump a LevelDB-based PouchDB:
$ pouchdb-dump /path/to/my/db > dump.txt
Full usage and examples:
Examples: pouchdb-dump http://localhost:5984/mydb > dump.txt Dump from the "mydb" CouchDB to dump.txt pouchdb-dump /path/to/mydb > dump.txt Dump from the "mydb" LevelDB-based PouchDB to dump.txt pouchdb-dump /path/to/mydb -o dump.txt Dump to the specified file instead of stdout pouchdb-dump /path/to/mydb -o dump.txt -s 100 Dump every 100 documents to dump_00.txt, dump_01.txt, dump_02.txt, etc. pouchdb-dump http://example.com/mydb -u myUsername -p myPassword > dump.txt Specify a CouchDB username and password if it's protected Options: -h, --help this help message -o, --output-file output file (else will dump to stdout) -u, --username username for the CouchDB database (if it's protected) -p, --password password for the CouchDB database (if it's protected) -s, --split split into multiple files, for every n docs
2.0.0 is a breaking change that increases the default batch size for non-split dumps from 1 to 100. This makes the out-of-the-box performance of this tool much beter. You can still get one-doc-per-line if you set the
--split to 1.
I doubt anybody actually wants that, but I'm still incrementing the major version.
Note: This does not break compatibility with
pouchdb-load will still accept dumpfiles created with either 1.x or 2.x.