A node library for Airship's Real Time Data Streaming (formerly Connect)
var connect = require('urban-airship-connect')
var connectStream = connect('appKey', 'authToken')
connectStream.on('data', function (data) {
// will be called with each event
})
// write to the stream to set filters/offset/start
connectStream.write({start: 'LATEST'})connect(appKey, authToken[, options]) -> duplexStream
appKeyis your Airship app keyauthTokenis your RTDS auth tokenoptionsis an optional object accepting parameters:urito specify the URI to a running RTDS instanceparserto specify a custom JSON parser of signature:parse(String) -> Object; defaults toJSON.parse
The returned duplex stream reads and writes plain JavaScript objects. Writes are posted to RTDS (Real Time Data Streaming), and the responding events are emitted as plain JavaScript objects over time.
If an error is encountered during a request, a bad status code is encountered,
or JSON parsing of an event fails, an 'error' event will be emitted with the
error attached. For JSON parse errors, the blob that caused the error will be
available as the blob property of the error.
It is possible for a Custom Event value to be a number that cannot be parsed by JavaScript see docs. If you suspect you might encounter this issue, you may want to provide your own JSON parser, see API.
This module will handle connecting and maintaining a connection to Airship's Real Time Data Streaming service. If its connection is ever severed, it will do its best to resume at the last seen offset.
The resume_offset is not stored externally, so if the process using this
module is killed, crashes, or otherwise exits it will be lost. In general it is
advisable to track this offset yourself.
var fs = require('fs')
var connect = require('urban-airship-connect')
var lookupStream = require('dotpath-stream')
var writeFile = require('file-replace-stream')
var connectStream = connect('appKey', 'authToken')
var filters = {}
// set up our pipeline for saving offsets
connectStream
// pull out just the `offset` property
.pipe(lookupStream('offset'))
// write it to `last_offset.txt`
.pipe(writeFile('last_offset.txt'))
try {
// try to read our offset file
filters.resume_offset = fs.readFileSync('last_offset.txt', {encoding: 'utf8'})
} catch (err) {
// fall back to starting at latest offset
filters.start = 'LATEST'
}
// write the filters to RTDS to start streaming events
connectStream.write(filters)writeFilecould be swapped out for any writable stream for persistence.- If streams aren't your thing, you could alternatively listen to the
dataevent ofconnectStreamand persist theoffsetproperty from the passed event however you please.
A simple command-line utility ua-connect is bundled along with this module. It
should be mostly used for testing your credentials or other aspects of your
RTDS setup, not for anything "mission-critical".
ua-connect <app-key> <token> [options]
Where options are:
--uri <uri>Use a custom uri for RTDS--offset <offset>Start at<offset>--earliestStart at earliest offset--key <app-key>Explicitly set app key--token <token>Explicitly set token
Apache-2.0