33 Commits

Author SHA1 Message Date
Brian M. Carlson
8f2355e454 0.3.0 2014-09-15 20:56:39 -04:00
Brian M. Carlson
d6eab36b66 Make tests a bit more robusto 2014-09-15 20:56:34 -04:00
Brian M. Carlson
33f6ecc11b Add workflow boilerplate files 2014-09-15 20:49:09 -04:00
Brian C
36572a8b7b Merge pull request #16 from drob/fix-docs
Fixes pipe from a file to table example in README.md.
2014-09-15 20:48:23 -04:00
Brian C
0c5d08edae Merge pull request #19 from drob/transform-opts
Accept stream options in constructors, pass to internal transform streams.
2014-09-15 20:47:38 -04:00
Dan
b78a3eb845 Accept stream options in constructors, pass to internal transform streams.
Includes tests.
2014-09-15 15:01:39 -04:00
Dan
107f007249 Fixes pipe from a file to table example in README.md. 2014-08-10 13:33:33 -07:00
Brian C
25b8d6da5f Update README.md
Providing clarity for #6
2014-05-01 11:07:33 -05:00
Dan Robinson
1c0c8871c1 Bump version 2014-04-07 11:56:28 -07:00
Dan Robinson
beb54334e2 Merge pull request #12 from drob/error-handling
Adds handling for errors after initial response.
2014-04-07 11:55:36 -07:00
Dan Robinson
1db9b3ec3d Adds handling for errors after initial response.
Includes a test.
2014-04-06 22:04:20 -07:00
Brian M. Carlson
1822f399d2 Bump version 2014-04-06 11:09:28 -05:00
Brian C
208861d057 Merge pull request #11 from drob/add-test
Adds a smoke test for copying with BINARY.
2014-04-06 11:06:16 -05:00
Brian C
1957554301 Merge pull request #10 from drob/handler-leak
Fixes another event handler leak.
2014-04-06 11:06:09 -05:00
Dan Robinson
bef3ba5fcd Adds a smoke test for copying with BINARY. 2014-04-06 04:50:54 -07:00
Dan Robinson
591a11c955 Fixes another event handler leak.
It turns out 'error' handlers were leaking as well, although more slowly.
2014-04-06 03:49:25 -07:00
Brian M. Carlson
b40918ddb8 Bump version 2014-03-30 00:10:18 -04:00
Brian C
b8946265c3 Merge pull request #8 from drob/handler-leak
Fixes an event handler leak.
2014-03-30 00:06:42 -04:00
Dan Robinson
d649905dbb Fixes an event handler leak.
Includes a test. Also includes async as a devDependency.
2014-03-28 04:19:19 -07:00
Brian C
22da85448e Merge pull request #7 from pensierinmusica/master
Update README.md
2014-03-27 20:43:46 -04:00
Alessandro Zanardi
531f72dcb3 Update README.md
Got rid of duplicate variable
2014-03-27 16:16:33 +01:00
Brian M. Carlson
dc6521b7c1 Fix to work with pg >= 2.8.2 2013-12-09 11:41:17 -05:00
Brian C
16e2001064 Update README.md 2013-11-08 00:49:19 -06:00
Brian C
5bb93308eb Update README.md 2013-11-08 00:47:52 -06:00
Brian C
e5f864bac0 Update README.md 2013-11-08 00:47:15 -06:00
Brian C
bfc0353d2c Update README.md 2013-11-08 00:43:41 -06:00
Brian C
90012d84d7 Update README.md 2013-11-08 00:42:43 -06:00
Brian C
808c284b55 Update README.md 2013-11-08 00:40:45 -06:00
Brian C
b48283f4aa Update README.md 2013-11-08 00:39:13 -06:00
Brian C
5c03015715 Update README.md 2013-11-08 00:38:30 -06:00
Brian C
18fa1fdeb3 Create README.md 2013-11-08 00:15:16 -06:00
Brian M. Carlson
824936176c Bump version 2013-10-28 21:50:54 -05:00
Brian M. Carlson
2f5b37e0ad Do not emit row event 2013-10-28 21:50:45 -05:00
10 changed files with 306 additions and 49 deletions

6
.travis.yml Normal file
View File

@@ -0,0 +1,6 @@
language: node_js
node_js:
- "0.10"
- "0.11"
env:
- PGUSER=postgres

14
Makefile Normal file
View File

@@ -0,0 +1,14 @@
.PHONY: publish-patch test
test:
npm test
patch: test
npm version patch -m "Bump version"
git push origin master --tags
npm publish
minor: test
npm version minor -m "Bump version"
git push origin master --tags
npm publish

94
README.md Normal file
View File

@@ -0,0 +1,94 @@
## pg-copy-streams
COPY FROM / COPY TO for node-postgres. Stream from one database to another, and stuff.
## how? what? huh?
Did you know the _all powerful_ PostgreSQL supports streaming binary data directly into and out of a table?
This means you can take your favorite CSV or TSV or whatever format file and pipe it directly into an existing PostgreSQL table.
You can also take a table and pipe it directly to a file, another database, stdout, even to `/dev/null` if you're crazy!
What this module gives you is a [Readable](http://nodejs.org/api/stream.html#stream_class_stream_readable) or [Writable](http://nodejs.org/api/stream.html#stream_class_stream_writable) stream directly into/out of a table in your database.
This mode of interfacing with your table is _very fast_ and _very brittle_. You are responsible for properly encoding and ordering all your columns. If anything is out of place PostgreSQL will send you back an error. The stream works within a transaction so you wont leave things in a 1/2 borked state, but it's still good to be aware of.
If you're not familiar with the feature (I wasn't either) you can read this for some good helps: http://www.postgresql.org/docs/9.3/static/sql-copy.html
## examples
### pipe from a table to stdout
```js
var pg = require('pg');
var copyTo = require('pg-copy-streams').to;
pg.connect(function(err, client, done) {
var stream = client.query(copyTo('COPY my_table TO STDOUT'));
stream.pipe(process.stdout);
stream.on('end', done);
stream.on('error', done);
});
```
### pipe from a file to table
```js
var fs = require('fs');
var pg = require('pg');
var copyFrom = require('pg-copy-streams').from;
pg.connect(function(err, client, done) {
var stream = client.query(copyFrom('COPY my_table FROM STDIN'));
var fileStream = fs.createReadStream('some_file.tdv')
fileStream.on('error', done);
fileStream.pipe(stream).on('finish', done).on('error', done);
});
```
## install
```sh
$ npm install pg-copy-streams
```
## notice
This module __only__ works with the pure JavaScript bindings. If you're using `require('pg').native` please make sure to use normal `require('pg')` or `require('pg.js')` when you're using copy streams.
Before you set out on this magical piping journey, you _really_ should read this: http://www.postgresql.org/docs/9.3/static/sql-copy.html, and you might want to take a look at the [tests](https://github.com/brianc/node-pg-copy-streams/tree/master/test) to get an idea of how things work.
## contributing
Instead of adding a bunch more code to the already bloated [node-postgres](https://github.com/brianc/node-postgres) I am trying to make the internals extensible and work on adding edge-case features as 3rd party modules.
This is one of those.
Please, if you have any issues with this, open an issue.
Better yet, submit a pull request. I _love_ pull requests.
Generally how I work is if you submit a few pull requests and you're interested I'll make you a contributor and give you full access to everything.
Since this isn't a module with tons of installs and dependent modules I hope we can work together on this to iterate faster here and make something really useful.
## license
The MIT License (MIT)
Copyright (c) 2013 Brian M. Carlson
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.

View File

@@ -1,24 +1,30 @@
module.exports = function(txt) {
return new CopyStreamQuery(txt)
module.exports = function(txt, options) {
return new CopyStreamQuery(txt, options)
}
var Transform = require('stream').Transform
var util = require('util')
var CopyStreamQuery = function(text) {
Transform.call(this)
var CopyStreamQuery = function(text, options) {
Transform.call(this, options)
this.text = text
this._listeners = null
this._listeners = {}
this._copyOutResponse = null
this.rowCount = 0
}
util.inherits(CopyStreamQuery, Transform)
var eventTypes = ['close', 'data', 'end', 'error']
CopyStreamQuery.prototype.submit = function(connection) {
connection.query(this.text)
this.connection = connection
this._listeners = connection.stream.listeners('data')
connection.stream.removeAllListeners('data')
var self = this
eventTypes.forEach(function(type) {
self._listeners[type] = connection.stream.listeners(type)
connection.stream.removeAllListeners(type)
})
connection.stream.pipe(this)
}
@@ -31,10 +37,12 @@ var code = {
CopyStreamQuery.prototype._detach = function() {
this.connection.stream.unpipe()
this.connection.stream.removeAllListeners('data')
var self = this
this._listeners.forEach(function(listener) {
self.connection.stream.on('data', listener)
eventTypes.forEach(function(type) {
self.connection.stream.removeAllListeners(type)
self._listeners[type].forEach(function(listener) {
self.connection.stream.on(type, listener)
})
})
}
@@ -60,10 +68,14 @@ CopyStreamQuery.prototype._transform = function(chunk, enc, cb) {
}
while((chunk.length - offset) > 5) {
var messageCode = chunk[offset]
//complete
if(messageCode == code.c) {
//complete or error
if(messageCode == code.c || messageCode == code.E) {
this._detach()
this.connection.stream.unshift(chunk.slice(offset + 5))
if (messageCode == code.c) {
this.connection.stream.unshift(chunk.slice(offset + 5))
} else {
this.connection.stream.unshift(chunk.slice(offset))
}
this.push(null)
return cb();
}
@@ -78,7 +90,7 @@ CopyStreamQuery.prototype._transform = function(chunk, enc, cb) {
var slice = chunk.slice(offset, offset + length)
offset += length
this.push(slice)
this.emit('row')
this.rowCount++
} else {
break;
}

View File

@@ -1,22 +1,23 @@
var CopyToQueryStream = require('./copy-to')
module.exports = {
to: function(txt) {
return new CopyToQueryStream(txt)
to: function(txt, options) {
return new CopyToQueryStream(txt, options)
},
from: function (txt) {
return new CopyStreamQuery(txt)
from: function (txt, options) {
return new CopyStreamQuery(txt, options)
}
}
var Transform = require('stream').Transform
var util = require('util')
var CopyStreamQuery = function(text) {
Transform.call(this)
var CopyStreamQuery = function(text, options) {
Transform.call(this, options)
this.text = text
this._listeners = null
this._copyOutResponse = null
this.rowCount = 0
}
util.inherits(CopyStreamQuery, Transform)
@@ -39,7 +40,7 @@ CopyStreamQuery.prototype._transform = function(chunk, enc, cb) {
lenBuffer.writeUInt32BE(chunk.length + 4, 0)
this.push(lenBuffer)
this.push(chunk)
this.emit('row')
this.rowCount++
cb()
}
@@ -54,7 +55,7 @@ CopyStreamQuery.prototype.handleError = function(e) {
this.emit('error', e)
}
CopyStreamQuery.prototype.streamData = function(connection) {
CopyStreamQuery.prototype.handleCopyInResponse = function(connection) {
this.pipe(connection.stream)
}

View File

@@ -1,6 +1,6 @@
{
"name": "pg-copy-streams",
"version": "0.1.0",
"version": "0.3.0",
"description": "Low-Level COPY TO and COPY FROM streams for PostgreSQL in JavaScript using",
"main": "index.js",
"scripts": {
@@ -27,6 +27,7 @@
"concat-stream": "~1.1.0",
"gonna": "0.0.0",
"lodash": "~2.2.1",
"heroku-env": "~0.1.1"
"heroku-env": "~0.1.1",
"async": "~0.2.10"
}
}

67
test/binary.js Normal file
View File

@@ -0,0 +1,67 @@
var assert = require('assert')
var gonna = require('gonna')
var async = require('async')
var concat = require('concat-stream')
var _ = require('lodash')
var pg = require('pg.js')
var from = require('../').from
var to = require('../').to
var testBinaryCopy = function() {
var client = function() {
var client = new pg.Client()
client.connect()
return client
}
var fromClient = client()
var toClient = client()
queries = [
'DROP TABLE IF EXISTS data',
'CREATE TABLE IF NOT EXISTS data (num BIGINT, word TEXT)',
'INSERT INTO data (num, word) VALUES (1, \'hello\'), (2, \'other thing\'), (3, \'goodbye\')',
'DROP TABLE IF EXISTS data_copy',
'CREATE TABLE IF NOT EXISTS data_copy (LIKE data INCLUDING ALL)'
]
async.eachSeries(queries, _.bind(fromClient.query, fromClient), function(err) {
assert.ifError(err)
var fromStream = fromClient.query(to('COPY (SELECT * FROM data) TO STDOUT BINARY'))
var toStream = toClient.query(from('COPY data_copy FROM STDIN BINARY'))
runStream = function(callback) {
fromStream.on('error', callback)
toStream.on('error', callback)
toStream.on('finish', callback)
fromStream.pipe(toStream)
}
runStream(function(err) {
assert.ifError(err)
toClient.query('SELECT * FROM data_copy ORDER BY num', function(err, res){
assert.equal(res.rowCount, 3, 'expected 3 rows but got ' + res.rowCount)
assert.equal(res.rows[0].num, 1)
assert.equal(res.rows[0].word, 'hello')
assert.equal(res.rows[1].num, 2)
assert.equal(res.rows[1].word, 'other thing')
assert.equal(res.rows[2].num, 3)
assert.equal(res.rows[2].word, 'goodbye')
queries = [
'DROP TABLE data',
'DROP TABLE data_copy'
]
async.each(queries, _.bind(fromClient.query, fromClient), function(err) {
assert.ifError(err)
fromClient.end()
toClient.end()
})
})
})
})
}
testBinaryCopy()

View File

@@ -5,20 +5,30 @@ var concat = require('concat-stream')
var _ = require('lodash')
var pg = require('pg.js')
var testRange = function(top) {
var client = function() {
var client = new pg.Client()
client.connect()
client.query('CREATE TEMP TABLE numbers(num int, bigger_num int)')
return client
var copy = require('../').from
var client = function() {
var client = new pg.Client()
client.connect()
return client
}
var testConstruction = function() {
var highWaterMark = 10
var stream = copy('COPY numbers FROM STDIN', {highWaterMark: 10, objectMode: true})
for(var i = 0; i < highWaterMark * 1.5; i++) {
stream.write('1\t2\n')
}
assert(!stream.write('1\t2\n'), 'Should correctly set highWaterMark.')
}
testConstruction()
var testRange = function(top) {
var fromClient = client()
var copy = require('../').from
fromClient.query('CREATE TEMP TABLE numbers(num int, bigger_num int)')
var txt = 'COPY numbers FROM STDIN'
var stream = fromClient.query(copy(txt))
var rowEmitCount = 0
stream.on('row', function() {
@@ -33,10 +43,10 @@ var testRange = function(top) {
fromClient.query('SELECT COUNT(*) FROM numbers', function(err, res) {
assert.ifError(err)
assert.equal(res.rows[0].count, top, 'expected ' + top + ' rows but got ' + res.rows[0].count)
console.log('found ', res.rows.length, 'rows')
//console.log('found ', res.rows.length, 'rows')
countDone()
var firstRowDone = gonna('have correct result')
assert.equal(rowEmitCount, top, 'should have emitted "row" event ' + top + ' times')
assert.equal(stream.rowCount, top, 'should have rowCount ' + top + ' ')
fromClient.query('SELECT (max(num)) AS num FROM numbers', function(err, res) {
assert.ifError(err)
assert.equal(res.rows[0].num, top-1)

View File

@@ -1,27 +1,32 @@
var assert = require('assert')
var gonna = require('gonna')
var concat = require('concat-stream')
var _ = require('lodash')
var async = require('async')
var concat = require('concat-stream')
var pg = require('pg.js')
var copy = require('../').to
var client = function() {
var client = new pg.Client()
client.connect()
return client
}
var testConstruction = function() {
var txt = 'COPY (SELECT * FROM generate_series(0, 10)) TO STDOUT'
var stream = copy(txt, {highWaterMark: 10})
assert.equal(stream._readableState.highWaterMark, 10, 'Client should have been set with a correct highWaterMark.')
}
testConstruction()
var testRange = function(top) {
var client = function() {
var client = new pg.Client()
client.connect()
return client
}
var fromClient = client()
var copy = require('../').to
var txt = 'COPY (SELECT * from generate_series(0, ' + (top - 1) + ')) TO STDOUT'
var stream = fromClient.query(copy(txt))
var rowEmitCount = 0
stream.on('row', function() {
rowEmitCount++
})
var done = gonna('finish piping out', 1000, function() {
fromClient.end()
})
@@ -30,9 +35,55 @@ var testRange = function(top) {
var res = buf.toString('utf8')
var expected = _.range(0, top).join('\n') + '\n'
assert.equal(res, expected)
assert.equal(rowEmitCount, top, 'should have emitted "row" ' + top + ' times but got ' + rowEmitCount)
assert.equal(stream.rowCount, top, 'should have rowCount ' + top + ' but got ' + stream.rowCount)
done()
}))
}
testRange(10000)
var testLeak = function(rounds) {
var fromClient = client()
var txt = 'COPY (SELECT 10) TO STDOUT'
var runStream = function(num, callback) {
var stream = fromClient.query(copy(txt))
stream.on('data', function(data) {
// Just throw away the data.
})
stream.on('end', callback)
stream.on('error', callback)
}
async.timesSeries(rounds, runStream, function(err) {
assert.equal(err, null)
assert.equal(fromClient.connection.stream.listeners('close').length, 0)
assert.equal(fromClient.connection.stream.listeners('data').length, 1)
assert.equal(fromClient.connection.stream.listeners('end').length, 2)
assert.equal(fromClient.connection.stream.listeners('error').length, 1)
fromClient.end()
})
}
testLeak(5)
var testInternalPostgresError = function() {
var fromClient = client()
// This attempts to make an array that's too large, and should fail.
var txt = "COPY (SELECT asdlfsdf AS e) t) TO STDOUT"
var runStream = function(callback) {
var stream = fromClient.query(copy(txt))
stream.on('data', function(data) {
// Just throw away the data.
})
stream.on('error', callback)
}
runStream(function(err) {
assert.notEqual(err, null)
fromClient.end()
})
}
testInternalPostgresError()

View File

@@ -1,2 +1,3 @@
require('./copy-from')
require('./copy-to')
require('./binary')