Every little thing begins with good knowledge, so ingestion is commonly your first step to unlocking insights. Nonetheless, ingestion presents challenges, like ramping up on the complexities of every knowledge supply, maintaining tabs on these sources as they alter, and governing all of this alongside the way in which.
Lakeflow Join makes environment friendly knowledge ingestion straightforward, with a point-and-click UI, a easy API, and deep integrations with the Information Intelligence Platform. Final yr, greater than 2,000 prospects used Lakeflow Hook up with unlock worth from their knowledge.
On this weblog, we’ll evaluation the fundamentals of Lakeflow Join and recap the most recent bulletins from the 2025 Information + AI Summit.
Ingest all of your knowledge in a single place with Lakeflow Join
Lakeflow Join affords easy ingestion connectors for functions, databases, cloud storage, message buses, and extra. Underneath the hood, ingestion is environment friendly, with incremental updates and optimized API utilization. As your managed pipelines run, we deal with schema evolution, seamless third-party API upgrades, and complete observability with built-in alerts.
Information + AI Summit 2025 Bulletins
At this yr’s Information + AI Summit, Databricks introduced the Basic Availability of Lakeflow, the unified method to knowledge engineering throughout ingestion, transformation, and orchestration. As a part of this, Lakeflow Join introduced Zerobus, a direct write API that simplifies ingestion for IoT, clickstream, telemetry and different related use instances. We additionally expanded the breadth of supported knowledge sources with extra built-in connectors throughout enterprise functions, file sources, databases, and knowledge warehouses, in addition to knowledge from cloud object storage.
Zerobus: a brand new method to push occasion knowledge on to your lakehouse
We made an thrilling announcement introducing Zerobus, a brand new modern method for pushing occasion knowledge on to your lakehouse by bringing you nearer to the info supply. Eliminating knowledge hops and decreasing operational burden allows Zerobus to supply high-throughput direct writes with low latency, delivering close to real-time efficiency at scale.
Beforehand, some organizations used message buses like Kafka as transport layers to the Lakehouse. Kafka affords a sturdy, low-latency approach for knowledge producers to ship knowledge, and it’s a well-liked selection when writing to a number of sinks. Nonetheless, it additionally provides further complexity and prices, in addition to the burden of managing one other knowledge copy—so it’s inefficient when your sole vacation spot is the Lakehouse. Zerobus supplies a easy resolution for these instances.

Joby Aviation is already utilizing Zerobus to instantly push telemetry knowledge into Databricks.
Joby is ready to use our manufacturing brokers with Zerobus to push gigabytes a minute of telemetry knowledge on to our lakehouse, accelerating the time to insights — all with Databricks Lakeflow and the Information Intelligence Platform.”
— Dominik Müller, Manufacturing unit Programs Lead, Joby Aviation, Inc.
As a part of Lakeflow Join, Zerobus can be unified with the Databricks Platform, so you’ll be able to leverage broader analytics and AI capabilities immediately. Zerobus is presently in Personal Preview; attain out to your account group for early entry.
🎥 Watch and be taught extra about Zerobus: Breakout session on the Information + AI Summit, that includes Joby Aviation, “Lakeflow Join: eliminating hops in your streaming structure”
Lakeflow Join expands ingestion capabilities and knowledge sources
New absolutely managed connectors are persevering with to roll out throughout varied launch states (see full record beneath), together with Google Analytics and ServiceNow, in addition to SQL Server – the primary database connector, all presently in Public Preview with Basic Availability coming quickly.
We now have additionally continued innovating for patrons who need extra customization choices and use our present ingestion resolution, Auto Loader. It incrementally and effectively processes new knowledge recordsdata as they arrive in cloud storage. We’ve launched some main price and efficiency enhancements for Auto Loader, together with 3X sooner listing listings and automated cleanup with “CleanSource,” each now usually obtainable, together with smarter and cheaper file discovery utilizing file occasions. We additionally introduced native help for ingesting Excel recordsdata and ingesting knowledge from SFTP servers, each in Personal Preview, obtainable by request for early entry.

Supported knowledge sources:
- Purposes: Salesforce, Workday, ServiceNow, Google Analytics, Microsoft Dynamics 365, Oracle NetSuite
- File sources: S3, ADLS, GCS, SFTP, SharePoint
- Databases: SQL Server, Oracle Database, MySQL, PostgreSQL
- Information warehouses: Snowflake, Amazon Redshift, Google BigQuery
Throughout the expanded connector providing, we’re introducing query-based connectors that simplify knowledge ingestion. These new connectors will let you pull knowledge instantly out of your supply techniques with out database modifications and work with learn replicas the place change knowledge seize (CDC) logs aren’t obtainable. That is presently in Personal Preview; attain out to your account group for early entry.

🎥 Watch and be taught extra about Lakeflow Join: Breakout session on the Information + AI Summit, “Getting Began with Lakeflow Join”
🎥 Watch and be taught extra about ingesting from enterprise SaaS functions: Breakout session on the Information + AI Summit that includes Databricks buyer Porsche Holding, “Lakeflow Join: Seamless Information Ingestion From Enterprise Apps”
🎥 Watch and be taught extra about database connectors: Breakout session on the Information + AI Summit, “Lakeflow Join: Simple, Environment friendly Ingestion From Databases”
Lakeflow Join in Jobs, now usually obtainable
We’re persevering with to develop capabilities to make it simpler so that you can use our ingestion connectors whereas constructing knowledge pipelines, as a part of Lakeflow’s unified knowledge engineering expertise. Databricks just lately introduced Lakeflow Join in Jobs, which lets you create ingestion pipelines inside Lakeflow Jobs. So, in case you have jobs as the middle of your ETL course of, this seamless integration supplies a extra intuitive and unified expertise for managing ingestion.

Prospects can outline and handle their end-to-end workloads—from ingestion to transformation—multi function place. Lakeflow Join in Jobs is now usually obtainable.
🎥 Watch and be taught extra about Lakeflow Jobs: Breakout session on the Information + AI Summit “Orchestration with Lakeflow Jobs”
Lakeflow Join: extra to come back in 2025 and past
Databricks understands the wants of information engineers and organizations who drive innovation with their knowledge utilizing analytics and AI instruments. To that finish, Lakeflow Join has continued to construct out sturdy, environment friendly ingestion capabilities with absolutely managed connectors to extra customizable options and APIs.
We’re simply getting began with Lakeflow Join. Keep tuned for extra bulletins later this yr, or contact your Databricks account group to affix a preview for early entry.
To attempt Lakeflow Join, you’ll be able to evaluation the documentation, or try the Demo Middle.