Load Data
Load by Data Source

Data Pipeline Load of Parquet Data from S3

A common setup for loading files in a batch into the  System is to load from a bucket on  S3 with time-partitioned data. Often, you must perform a batch load repeatedly to load new files.

Ocient uses data pipelines to transform each document into rows in one or more different tables. The loading and transformation capabilities use a simple SQL-like syntax for transforming data. This tutorial guides you through a simple example load using a small data set in  format. The data in this example comes from a test set for the  Business Intelligence tool.

Parquet Loading Recommendations

Follow this set of recommendations for an optimal loading experience of Parquet files.



Prerequisites

This tutorial assumes that:

  1. The Ocient System has network access to S3 from the Loader Nodes.
  2. An Ocient System is installed and configured with an active Storage Cluster. For details, see Ocient Application Configuration.

Step 1: Create a New Database.

To begin, load two example tables in a database. First, connect to a SQL Node using the Commands Supported by the Ocient JDBC CLI Program . Then, execute the CREATE DATABASE SQL statement to create the metabase database.

SQL


Step 2: Create a New Table in the Database.

Create the orders table in the new database. First, connect to that database (e.g., connect to jdbc:ocient://sql-node:4050/metabase), and then execute this CREATE TABLE SQL statement that specifies to create a table with these columns and a clustering index based on the user_id and product_id columns:

  • created_at as a  timestamp that is not nullable.
  • id, user_id, and product_id as integers that are not nullable.
  • subtotal, tax, total, and discount as floating point numbers.
  • quantity as an integer.
SQL


The database creates the orders table, and you can begin loading data.

Step 3: Preview and Create a Data Pipeline.

Create data pipelines using the CREATE PIPELINE SQL statement. To load data, you first create a pipeline with the definition of the source, data format, and transformation rules using a SQL-like declarative syntax. Then, you execute the START PIPELINE SQL statement to start the load. You can observe progress and status using system catalog tables and views.

Each Ocient pipeline defines a single data source and the target table or tables into which data loads. A data source includes the location of the source and filters on the source to define the specific data set to load. This tutorial loads data from two data sources, where each source is located in a directory on the same S3 bucket.

First, inspect the data that you plan to load. Each document has a JSON format similar to this example.

JSON


Inspecting the file with Pandas, you can see this schema with details.

Parquet Schema Description


In this case, Ocient automatically transforms the data to the target columns using some sensible conventions. In other cases, loads require some transformation. Most transformations are identical to functions that already exist in the SQL syntax of the Ocient System.

Prior to creating a pipeline, you can use the PREVIEW PIPELINE SQL statement to create your pipeline iteratively. This statement returns a result set that shows the final values that would be loaded but does not load the data into the target table.

Preview the orders_pipeline pipeline for the orders data set from your database connection prompt. Use the S3 data source with endpoint https://s3.us-east-1.amazonaws.com, bucket ocient-examples, and filter metabase_samples/parquet/orders.parquet. Specify the parquet format. Load the data into the public.orders table. The SELECT part of the SQL statement maps the fields in the Parquet file to the target columns in the created table. In this example, limit the result set to the first five records in the data source.

SQL


Output

SQL


With this preview, you can confirm that the results of the pipeline match your requirements. If there is an issue, you can update the statement and run the PREVIEW PIPELINE statement again until it meets your needs.

Next, create the pipeline named orders_pipeline for the orders data set from your database connection prompt. The pipeline has three main sections:

  • SOURCE — Loads data from S3. Set the S3 endpoint, bucket name, and filter for the Parquet files.
  • EXTRACT — Sets the format to Parquet.
  • INTO ... SELECT — Targets the public.orders table and selects the chosen fields from the Parquet records. In this case, all data is available at the top level of the Parquet records, so the example references the fields by the attribute name (e.g., $id, $user_id, etc.). For nested data, reference the nested fields using dot notation (e.g., $order.user.first_name). Each field maps to a target column using the as syntax.
SQL


After you successfully create this pipeline, execute the START PIPELINE SQL statement to start the load.

SQL


Step 4: Observe the Load Progress.

With your pipeline running, data begins to load immediately from the S3 files that you defined. If there are many files in each file group, the load process first sorts the files into batches, partitions them for parallel processing, and assigns them to Loader Nodes.

You can check the pipeline status and progress by querying the information_schema.pipeline_status system catalog table or executing the SHOW PIPELINE_STATUS SQL statement.

SQL


Output

SQL


After the status of the pipeline changes to COMPLETED, all data is available in the target table.

After a few seconds, the data is available for query in the public.orders table.

SQL


Output

SQL


You can drop the pipeline with the DROP PIPELINE orders_pipeline; SQL statement. Execution of this statement leaves the data in your target table, but removes metadata about the pipeline execution from the system.

Related Links