7.8 KiB
Copy Queries
Copy queries allow you to use the PostgreSQL copy command for efficient streaming of data to and from CARTO.
The support for copy is split across two API end points:
http://{username}.carto.com/api/v2/sql/copyfrom
for uploading data to CARTOhttp://{username}.carto.com/api/v2/sql/copyto
for exporting data out of CARTO
Copy From
The PostgreSQL COPY
command is extremely fast, but requires very precise inputs:
- A
COPY
command that describes the table and columns of the upload file, and the format of the file. - An upload file that exactly matches the
COPY
command.
If the COPY
command, the supplied file, and the target table do not all match, the upload will fail.
"Copy from" copies data "from" your file, "to" CARTO. "Copy from" uses multipart/form-data to stream an upload file to the server. This avoids limitations around file size and any need for temporary storage: the data travels from your file straight into the database.
api_key
provided in the request URL parameters.sql
provided either in the request URL parameters or in a multipart form variable.file
provided as a multipart form variable; this is the actual file content, not a filename.
Composing a multipart form data upload is moderately complicated, so almost all developers will use a tool or scripting language to upload data to CARTO via "copy from".
Example
For a table to be readable by CARTO, it must have a minimum of three columns with specific data types:
cartodb_id
, aserial
primary keythe_geom
, ageometry
in the ESPG:4326 projection (aka long/lat)the_geom_webmercator
, ageometry
in the ESPG:3857 projection (aka web mercator)
Creating a new CARTO table with all the right triggers and columns can is tricky, so here is an example:
-- create the table using the *required* columns and a
-- couple more
CREATE TABLE upload_example (
the_geom geometry,
name text,
age integer
);
-- adds the 'cartodb_id' and 'the_geom_webmercator'
-- adds the required triggers and indexes
SELECT CDB_CartodbfyTable('upload_example');
Now you are ready to upload your file. Suppose you have a CSV file like this:
the_geom,name,age
SRID=4326;POINT(-126 54),North West,89
SRID=4326;POINT(-96 34),South East,99
SRID=4326;POINT(-6 -25),Souther Easter,124
The COPY
command to upload this file needs to specify the file format (CSV), the fact that there is a header line before the actual data begins, and to enumerate the columns that are in the file so they can be matched to the table columns.
COPY upload_example (the_geom, name, age)
FROM STDIN WITH (FORMAT csv, HEADER true)
The FROM STDIN
option tells the database that the input will come from a data stream, and the SQL API will read our uploaded file and use it to feed the stream.
To actually run upload, you will need a tool or script that can generate a multipart/form-data
POST, so here are a few examples in different languages.
CURL Example
The curl utility makes it easy to run web requests from the command-line, and supports multi-part file upload, so it can feed the copyfrom
end point.
Assuming that you have already created the table, and that the CSV file is named "upload_example.csv":
curl -X POST \
--data-binary @upload_example.csv \
"http://{username}.carto.com/api/v2/sql/copyfrom?api_key={api_key}&sql=COPY+upload_example+(the_geom,+name,+age)+FROM+STDIN+WITH+(FORMAT+csv,+HEADER+true)"
Python Example
The Requests library for HTTP makes doing a file upload relatively terse.
import requests
api_key = {api_key}
username = {api_key}
upload_file = 'upload_example.csv'
sql = "COPY upload_example (the_geom, name, age) FROM STDIN WITH (FORMAT csv, HEADER true)"
url = "http://%s.carto.com/api/v2/sql/copyfrom" % username
with open(upload_file, 'rb') as f:
r = requests.post(url, params={'api_key':api_key, 'sql':sql}, data=f)
if r.status_code != 200:
print r.text
else:
status = r.json()
print "Success: %s rows imported" % status['total_rows']
A slightly more sophisticated script could read the headers from the CSV and compose the COPY
command on the fly.
Response Format
A successful upload will return with status code 200, and a small JSON with information about the upload.
{"time":0.046,"total_rows":10}
A failed upload will return with status code 400 and a larger JSON with the PostgreSQL error string, and a stack trace from the SQL API.
{"error":["Unexpected field"],
"stack":"Error: Unexpected field
at makeError (/repos/CartoDB-SQL-API/node_modules/multer/lib/make-error.js:12:13)
at wrappedFileFilter (/repos/CartoDB-SQL-API/node_modules/multer/index.js:39:19)
...
at emitMany (events.js:127:13)
at SBMH.emit (events.js:201:7)"}
Copy To
"Copy to" copies data "to" your desired output file, "from" CARTO.
Using the copyto
end point to extract data bypasses the usual JSON formatting applied by the SQL API, so it can dump more data, faster. However, it has the restriction that it will only output in a handful of formats:
- PgSQL text format,
- CSV, and
- PgSQL binary format.
"Copy to" is a simple HTTP GET end point, so any tool or language can be easily used to download data, supplying the following parameters in the URL:
sql
, the "COPY" command to extract the data.filename
, the filename to put in the "Content-disposition" HTTP header. Useful for tools that automatically save the download to a file name.api_key
, your API key for reading non-public tables.
CURL Example
The SQL to start a "copy to" can specify
- a table to read,
- a table and subset of columns to read, or
- an arbitrary SQL query to execute and read.
For our example, we'll read back just the three columns we originally loaded:
COPY upload_example (the_geom, name, age) TO stdout WITH (FORMAT csv, HEADER true)
The SQL needs to be URL-encoded before being embedded in the CURL command, so the final result looks like this:
curl \
--output upload_example_dl.csv \
"http://{username}.carto.com/api/v2/sql/copyto?sql=COPY+upload_example+(the_geom,name,age)+TO+stdout+WITH(FORMAT+csv,HEADER+true)&api_key={api_key}"
Python Example
The Python to "copy to" is very simple, because the HTTP call is a simple get. The only complexity in this example is at the end, where the result is streamed back block-by-block, to avoid pulling the entire download into memory before writing to file.
import requests
import re
api_key = {api_key}
username = {api_key}
download_file = 'upload_example_dl.csv'
sql = "COPY upload_example (the_geom, name, age) TO stdout WITH (FORMAT csv, HEADER true)"
# request the download, specifying desired file name
url = "http://%s.carto.com/api/v2/sql/copyto" % username
r = requests.get(url, params={'api_key':api_key, 'sql':sql, 'filename':download_file}, stream=True)
r.raise_for_status()
# read save file name from response headers
d = r.headers['content-disposition']
savefilename = re.findall("filename=(.+)", d)
if len(savefilename) > 0:
with open(savefilename[0], 'wb') as handle:
for block in r.iter_content(1024):
handle.write(block)
print "Downloaded to: %s" % savefilename
else:
print "Error: could not find read file name from headers"