Author: Matillion
Date Posted: Aug 22, 2024
Last Modified: Oct 17, 2024
Pre-built pipelines for loading streamed data
Process the latest files from cloud storage to maintain tables in your cloud data platform.
If you have configured a streaming pipeline with Amazon S3 or Azure Blob as the destination, these Data Productivity Cloud pre-built pipelines can be used to load the Avro files into Snowflake or Databricks.
Requirements
To load the files into Databricks, you must use a Data Productivity Cloud project configured with a Hybrid SaaS agent.
To load the files into Snowflake, you can use a Data Productivity Cloud project configured with either a Full SaaS or Hybrid SaaS agent.
Installation
- Download the latest zip file for your target data platform below.
- Open a branch on your Data Productivity Cloud project.
- If you already have a folder named “Matillion Pre-built Pipelines” in the root of your project, delete it.
- Hover over the root folder in your project, click the three dot menu and select “Import”.
- Browse to and select the zip file.
- You should now have a folder named “Matillion Pre-built Pipelines” containing the latest version of the pipelines.
Usage
Open the orchestration pipeline “Matillion Pre-built Pipelines > Example”.
Follow the instructions on the notes in this pipeline to copy the Run Orchestration component into your own orchestration pipeline, and configure it to load your Avro files into your data platform.
Downloads
Licensed under: Matillion Free Subscription License
- Download pre_built_pipelines_streaming_databricks_20241016.zip
- Target: Databricks
- Download pre_built_pipelines_streaming_snowflake_20241017.zip
- Target: Snowflake