Hi, I'm Ask INFA!
What would you like to know?
ASK INFAPreview
Please to access Ask INFA.

Table of Contents

Search

  1. Introducing Mass Ingestion
  2. Getting Started with Mass Ingestion
  3. Connectors and Connections
  4. Mass Ingestion Applications
  5. Mass Ingestion Databases
  6. Mass Ingestion Files
  7. Mass Ingestion Streaming
  8. Monitoring Mass Ingestion Jobs
  9. Asset Management
  10. Troubleshooting

Mass Ingestion

Mass Ingestion

Databricks Delta target

Databricks Delta target

Use a
streaming ingestion
task to write data to a Databricks Delta target. To create a Databricks Delta target, use the Databricks Delta connection type. The Databricks Delta target requires a Databricks cluster version 6.3 or later.
Databricks Delta is an open source storage layer that provides ACID transactions and works on top of existing data lakes. Databricks uses proprietary Delta software to manage stored data and allow fast access to the data.
You can access Delta Lake tables built on top of the following storage types:
  • Azure Data Lake Storage (ADLS) Gen2
  • Amazon Web Services (AWS) S3
The Databricks Delta target writes data to one or more Delta Lake tables on Databricks. You can use the Databricks Delta target in a
streaming ingestion
task for the following use cases:
  • Ingest bulk data from all streaming sources into Databricks Delta tables
  • Merge change data capture (CDC) from all streaming sources and write to Databricks Delta tables
The Databricks Delta connection uses a JDBC URL to connect to the Databricks cluster. When you configure the target, you specify the JDBC URL and credentials to use to connect to the cluster. You also define the connection information that the target uses to connect to the staging location in Amazon S3 or Azure Data Lake Storage Gen2.
You specify the tables in Delta Lake to which you want to write the data. The target writes data from record fields to table columns based on matching names.

0 COMMENTS

We’d like to hear from you!