unifiyng to see 'the forest'

This commit is contained in:
Simon Martín 2018-06-08 15:03:21 +02:00
parent 8e8ccc9c67
commit e259a51623
2 changed files with 147 additions and 153 deletions

View File

@ -10,6 +10,13 @@ const rateLimitsMiddleware = require('../middlewares/rate-limit');
const { RATE_LIMIT_ENDPOINTS_GROUPS } = rateLimitsMiddleware;
const errorHandlerFactory = require('../services/error_handler_factory');
const StreamCopy = require('../services/stream_copy');
const StreamCopyMetrics = require('../services/stream_copy_metrics');
const Logger = require('../services/logger');
const { Client } = require('pg');
const zlib = require('zlib');
const PSQL = require('cartodb-psql');
const copyTo = require('pg-copy-streams').to;
const copyFrom = require('pg-copy-streams').from;
function CopyController(metadataBackend, userDatabaseService, userLimitsService, statsClient) {
this.metadataBackend = metadataBackend;
@ -60,52 +67,155 @@ CopyController.prototype.route = function (app) {
function handleCopyTo (streamCopy) {
return function handleCopyToMiddleware (req, res, next) {
const sql = req.query.q;
const { userDbParams, user } = res.locals;
const filename = req.query.filename || 'carto-sql-copyto.dmp';
const logger = new Logger(global.settings.dataIngestionLogPath, 'data-ingestion');
let metrics = new StreamCopyMetrics(logger, 'copyto', sql, user);
res.header("Content-Disposition", `attachment; filename=${encodeURIComponent(filename)}`);
res.header("Content-Type", "application/octet-stream");
streamCopy.to(
res,
req.query.q,
res.locals.userDbParams,
res.locals.user,
function(err) {
if (err) {
return next(err);
}
// this is a especial endpoint
// the data from postgres is streamed to response directly
const pg = new PSQL(userDbParams);
pg.connect(function (err, client, done) {
if (err) {
return next(err);
}
);
let responseEnded = false;
let connectionClosedByClient = false;
const copyToStream = copyTo(sql);
const pgstream = client.query(copyToStream);
res
.on('error', err => {
metrics.end(null, err);
pgstream.unpipe(res);
done();
return next(err);
})
.on('close', () => {
if (!responseEnded) {
connectionClosedByClient = true;
// Cancel the running COPY TO query
// See https://www.postgresql.org/docs/9.5/static/protocol-flow.html#PROTOCOL-COPY
const runningClient = client;
const cancelingClient = new Client(runningClient.connectionParameters);
cancelingClient.cancel(runningClient, pgstream);
const err = new Error('Connection closed by client');
metrics.end(null, err);
pgstream.unpipe(res);
// see https://node-postgres.com/api/pool#releasecallback
done(err);
return next(err);
}
})
.on('end', () => responseEnded = true);
pgstream
.on('error', err => {
if (!connectionClosedByClient) {
metrics.end(null, err);
pgstream.unpipe(res);
done(err);
return next(err);
}
})
.on('data', data => metrics.addSize(data.length))
.on('end', () => {
metrics.end(copyToStream.rowCount);
done();
return next(null, metrics);
})
.pipe(res);
});
};
}
function handleCopyFrom (streamCopy) {
return function handleCopyFromMiddleware (req, res, next) {
streamCopy.from(
req,
req.query.q,
res.locals.userDbParams,
res.locals.user,
req.get('content-encoding') === 'gzip',
function(err, metrics) {
if (err) {
return next(err);
}
const sql = req.query.q;
const { userDbParams, user } = res.locals;
const gzip = req.get('content-encoding') === 'gzip';
const { time, rows } = metrics;
if (!time || !rows) {
return next(new Error("No rows copied"));
}
res.send({
time,
total_rows: rows
});
const logger = new Logger(global.settings.dataIngestionLogPath, 'data-ingestion');
let metrics = new StreamCopyMetrics(logger, 'copyfrom', sql, user, gzip);
const pg = new PSQL(userDbParams);
pg.connect(function (err, client, done) {
if (err) {
next(err);
}
);
let copyFromStream = copyFrom(sql);
const pgstream = client.query(copyFromStream);
pgstream
.on('error', err => {
metrics.end(null, err);
req.unpipe(pgstream);
done();
next(err);
})
.on('end', function () {
metrics.end(copyFromStream.rowCount);
done();
const { time, rows } = metrics;
if (!time || !rows) {
return next(new Error("No rows copied"));
}
res.send({
time,
total_rows: rows
});
});
let requestEnded = false;
req
.on('error', err => {
metrics.end(null, err);
req.unpipe(pgstream);
pgstream.end();
done();
next(err);
})
.on('close', () => {
if (!requestEnded) {
const err = new Error('Connection closed by client');
metrics.end(null, err);
const connection = client.connection;
connection.sendCopyFail('CARTO SQL API: Connection closed by client');
req.unpipe(pgstream);
done();
next(err);
}
})
.on('data', data => {
if (gzip) {
metrics.addGzipSize(data.length);
} else {
metrics.addSize(data.length);
}
})
.on('end', () => requestEnded = true);
if (gzip) {
req
.pipe(zlib.createGunzip())
.on('data', data => metrics.addSize(data.length))
.pipe(pgstream);
} else {
req.pipe(pgstream);
}
});
};
}
@ -114,12 +224,12 @@ function validateCopyQuery () {
const sql = req.query.q;
if (!sql) {
next(new Error("SQL is missing"));
return next(new Error("SQL is missing"));
}
// Only accept SQL that starts with 'COPY'
if (!sql.toUpperCase().startsWith("COPY ")) {
next(new Error("SQL must start with COPY"));
return next(new Error("SQL must start with COPY"));
}
next();

View File

@ -12,126 +12,10 @@ module.exports = class StreamCopy {
}
to(res, sql, userDbParams, user, cb) {
let metrics = new StreamCopyMetrics(this.logger, 'copyto', sql, user);
const pg = new PSQL(userDbParams);
pg.connect(function (err, client, done) {
if (err) {
return cb(err);
}
let responseEnded = false;
let connectionClosedByClient = false;
const copyToStream = copyTo(sql);
const pgstream = client.query(copyToStream);
res
.on('error', err => {
metrics.end(null, err);
pgstream.unpipe(res);
done();
return cb(err);
})
.on('close', () => {
if (!responseEnded) {
connectionClosedByClient = true;
// Cancel the running COPY TO query
// See https://www.postgresql.org/docs/9.5/static/protocol-flow.html#PROTOCOL-COPY
const runningClient = client;
const cancelingClient = new Client(runningClient.connectionParameters);
cancelingClient.cancel(runningClient, pgstream);
const err = new Error('Connection closed by client');
metrics.end(null, err);
pgstream.unpipe(res);
// see https://node-postgres.com/api/pool#releasecallback
done(err);
return cb(err);
}
})
.on('end', () => responseEnded = true);
pgstream
.on('error', err => {
if (!connectionClosedByClient) {
metrics.end(null, err);
pgstream.unpipe(res);
done(err);
return cb(err);
}
})
.on('data', data => metrics.addSize(data.length))
.on('end', () => {
metrics.end(copyToStream.rowCount);
done();
return cb(null, metrics);
})
.pipe(res);
});
}
from(req, sql, userDbParams, user, gzip, cb) {
let metrics = new StreamCopyMetrics(this.logger, 'copyfrom', sql, user, gzip);
const pg = new PSQL(userDbParams);
pg.connect(function (err, client, done) {
if (err) {
return cb(err);
}
let copyFromStream = copyFrom(sql);
const pgstream = client.query(copyFromStream);
pgstream
.on('error', err => {
metrics.end(null, err);
req.unpipe(pgstream);
done();
return cb(err);
})
.on('end', function () {
metrics.end(copyFromStream.rowCount);
done();
return cb(null, metrics);
});
let requestEnded = false;
req
.on('error', err => {
metrics.end(null, err);
req.unpipe(pgstream);
pgstream.end();
done();
return cb(err);
})
.on('close', () => {
if (!requestEnded) {
const err = new Error('Connection closed by client');
metrics.end(null, err);
const connection = client.connection;
connection.sendCopyFail('CARTO SQL API: Connection closed by client');
req.unpipe(pgstream);
done();
return cb(err);
}
})
.on('data', data => {
if (gzip) {
metrics.addGzipSize(data.length);
} else {
metrics.addSize(data.length);
}
})
.on('end', () => requestEnded = true);
if (gzip) {
req
.pipe(zlib.createGunzip())
.on('data', data => metrics.addSize(data.length))
.pipe(pgstream);
} else {
req.pipe(pgstream);
}
});
}
};