Pipeline Built-in Functions
On this page
SingleStore provides two built-in functions with pipelines to help load data.SET
clause.
pipeline_ source_ file()
Pipelines persist the name of a file by using the pipeline_
function.SET
clause to set a table column to the name of the pipeline data source file.
For example, given the table definition CREATE TABLE b(isbn NUMERIC(13), title VARCHAR(50));
, use the following statement to set the titles of files while ingesting data from AWS S3.
CREATE PIPELINE books ASLOAD DATA S3 's3://<bucket_name>/Books/'CONFIG '{"region":"us-west-2"}'CREDENTIALS '{"aws_access_key_id": "<access_key_id>","aws_secret_access_key": "<secret_access_key>"}'SKIP DUPLICATE KEY ERRORSINTO TABLE b(isbn)SET title = pipeline_source_file();
For more information on using the pipeline_
function to load data from AWS S3, refer to Load Data from Amazon Web Services (AWS) S3.
pipeline_ batch_ id()
Pipelines persist the ID of the batch used to load data with the pipeline_
built-in function.SET
clause to set a table column to the ID of the batch used to load the data.
For example, given the table definition CREATE TABLE t(b_
, use this statement to load the batch ID into the b_
column:
CREATE PIPELINE p AS LOAD DATA ... INTO TABLE t(@b_id,column_2) ...SET b_id = pipeline_batch_id();
Last modified: November 27, 2024