pull-level

pull-stream interface to levelup

pull-level

pull-stream interface to levelup

read items in database.

var pl = require('pull-level')
var pull = require('pull-stream')
 
var db = require('levelup')('/tmp/pull-level-example')
 
pull(pl.read(db), pull.collect(console.log))

read items in database, plus realtime changes

pull(
  pl.read(db, {tail: true}),
  //log data as it comes, 
  //because tail will keep the connection open 
  //so we'll never see the end otherwise. 
  pull.through(console.log),
  //note, pull-streams will not drain unless something is 
  //pulling the data through, so we have to add drain 
  //even though the data we want is coming from pull.through() 
  pull.drain()
)

If you just want the realtime inserts, use live

pull(
  pl.live(db, {tail: true}),
  pull.through(console.log),
  pull.drain()
)

To write, pipe batch changes into write

pull(
  pull.values([
    {key: 0, value: 'zero', type: 'put'},
    {key: 1, value: 'one',  type: 'put'},
    {key: 2, value: 'two',  type: 'put'},
  ]),
  pl.write(db)
)

If you are lazy/busy, you can leave off type. In that case, if value is non-null, the change is considered a put else, a del.

pull(
  pull.values([
    {key: 0, value: 'zero'},
    {key: 1, value: 'one'},
    {key: 2, value: 'two'},
  ]), 
  pl.write(db)
)

With pull-level it's easy to create indexes. just save a pointer to the key.

like this:

pull(
  pull.values([
    {key: key, value: VALUE, type: 'put'},
    {key: '~INDEX~' + VALUE.prop, value: key,  type: 'put'},
  ]),
  pl.write(db)
)

then, when you want to do a read, use asyncMap

pull(
  pl.read(db, {min: '~INDEX~', max: '~INDEX~~'})
  pull.asyncMap(function (ecb) {
    db.get(e.value, function (value) {
      cb(null, {key: e.value, value: value})
    })
  }),
  pull.collect(console.log)
)

We want to keep a realtime count of everything in the database. When ever something is inserted, we increment. But, we need to check the records that are currently in the database.

Since it takes some time to scan the database, we need to make sure we have done that before giving an answer. We can read it all with one stream, using {sync: true} to be notified of when we have read out all the old records.

First all the old records are read from the non-live stream, then you get one {sync: true} element, then all the new item.

var sum = 0, ready = false, waiting = []
 
//call get count to know s 
function getSum (cb) {
  if(!ready) waiting.push(cb)
  else cb(null, sum)
}
 
pull(
  pl.read(db, {sync: true}),
  pull.drain(function (op) {
    if(op.sync) {
      //if we see a data element with this it means 
      ready = true
      while(waiting.length) waiting.shift()(null, count)
    }
    //increment our counter! 
    if(Number.isFinite(+op.value.amount)) //filter out non numbers & NaN. 
      sum += op.value.amount
  })
)
 

MIT