Get startedGet started for free

Why the Lakehouse?

1. Why the Lakehouse?

Welcome to Introduction to Databricks Lakehouse!

2. What you'll learn

In this course, you'll explore the platform that's becoming the default foundation for modern data teams. By the end, you'll understand how data is organized in a lakehouse, how to manage compute and notebooks, how to govern and share data securely, and how to deploy your work to production.

3. Meet the Instructor

This course was developed in partnership with Gang Wang, a Senior Data Scientist at Origin Energy in Australia with over nine years of post-PhD experience building data pipelines and analytics platforms on Databricks. Let's get started.

4. The data lake promise

For years, organizations relied on data lakes to store massive amounts of raw data cheaply. Data lakes are incredibly flexible - you can dump in structured tables, semi-structured JSON, even images and logs. There's no need to define a schema upfront. But that flexibility came with a serious downside.

5. The lake's dark side

Without structure, data lakes often became data swamps. There was no guarantee that the data you found was clean, current, or correct. Governance was bolted on after the fact. Access control was manual. And if you wanted to run analytics or train a machine learning model, you typically had to copy data into a separate warehouse or a specialized tool. That meant more complexity, more cost, and more room for errors.

6. The warehouse trade-off

Data warehouses solve the quality problem. They offer reliability, strong governance, and fast query performance. But they come with their own trade-offs. Warehouses are expensive to scale, rigid in their schema requirements, and limited to structured data. If your data doesn't fit neatly into rows and columns, the warehouse can't help. Many organizations ended up running both - a lake for flexibility and a warehouse for reliability - duplicating data and effort.

7. Enter the Lakehouse

The lakehouse was designed to end this compromise. It's a single platform that brings together the best of both worlds: the low-cost, flexible storage of a data lake with the reliability, governance, and performance of a data warehouse. You don't have to choose. And you don't have to copy data between systems. One platform, one copy of the data, multiple workloads.

8. What makes it work in Databricks?

What makes the lakehouse possible in Databricks? Four key ingredients. First, open file formats like Delta Lake add reliability to cheap cloud storage. Second, ACID transactions ensure your data stays consistent even with concurrent reads and writes. Third, unified governance through Unity Catalog means one place to manage permissions, lineage, and compliance. And fourth, everything runs on one platform - analytics, machine learning, and real-time applications, all using the same data.

9. Lakehouse improves data quality

Compared to a standalone data lake, the lakehouse dramatically improves data quality. Schema enforcement catches bad data before it enters your tables. Transaction logs track every change, so you always know what happened and when. And time travel lets you query previous versions of your data - perfect for auditing or rolling back mistakes. The result is a lake that you can actually trust.

10. Summary

Let's recap. Data lakes offer flexibility and low cost but lack reliability. Data warehouses deliver performance and governance but at higher cost and less flexibility. The lakehouse combines both into a single platform, powered by Delta Lake for reliability, ACID transactions for consistency, and Unity Catalog for governance. Next, we'll see how the lakehouse organizes data into layers using the medallion architecture.

11. Let's practice!

Now let's test your understanding of the Lakehouse with some exercises.

Create Your Free Account

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.