Тёмный
No video :(

Understanding Delta Lake - The Heart of the Data Lakehouse 

Bryan Cafferky
Подписаться 41 тыс.
Просмотров 6 тыс.
50% 1

Опубликовано:

 

21 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 17   
@mainakdey3893
@mainakdey3893 2 месяца назад
at last somebody is clearing the confusion, Good job Bryan
@amarnadhgunakala2901
@amarnadhgunakala2901 Год назад
Thank you Brother, this helps people.
@stylish37
@stylish37 11 месяцев назад
Top stuff Bryan! Thanks a lot for this playlist
@BryanCafferky
@BryanCafferky 11 месяцев назад
YW
@gatorpika
@gatorpika Год назад
Great explanation! Thanks!
@BryanCafferky
@BryanCafferky Год назад
You're welcome!
@rahulberry5341
@rahulberry5341 Год назад
Thanks for the nice explanation
@BryanCafferky
@BryanCafferky Год назад
YW
@gautamgovinda5140
@gautamgovinda5140 3 месяца назад
Cool👍
@parisaayazi8886
@parisaayazi8886 3 месяца назад
Thanks Bryan! I'm wondering how it's possible to create a CSV table using the CREATE TABLE command, which allows us to write SQL queries against it, but we can't use saveAsTable with format('csv') to achieve the same result
@BryanCafferky
@BryanCafferky 3 месяца назад
Originally Spark could not create updatable tables. Instead it could only create a schema for a flat file like a CSV. The schema describes the data in the file so SQL select statements can be used on it. You can't update the table though and it is not a Managed table meaning if you drop the table for the CSV file, the file remains. Updateable tables (supports CRUD and ACID) was added with Delta tables.
@parisaayazi8886
@parisaayazi8886 3 месяца назад
@@BryanCafferky thanks a lot.
@panzabamboo1901
@panzabamboo1901 Год назад
Hi Brian, would you be able to elaborate more on the file types, currently supporting etl jobs running databricks, still using trial and error to figure out the file type/ how to load em
@BryanCafferky
@BryanCafferky Год назад
Hi Panza, Assuming you mean source files types to be read, most file types supported via Spark, i.e. csv, json, SQL databases, parquet, delta, avro. Are you looking for a specific type?
@user-cj2wt4mi5b
@user-cj2wt4mi5b 7 месяцев назад
Thanks, this is great video and well explained
@BryanCafferky
@BryanCafferky 7 месяцев назад
Thanks. In my experience, it is important to have the original data you loaded into a DW bc 1) troubleshooting issues, 2) recovery if some part of the data fails to load - you reload from the copy, 3) auditability - you can show what you loaded. It's especially critical if you cannot go back at a later date and retrieve that data again from the source.
@sajeershahul8361
@sajeershahul8361 Год назад
How can I not subscribe 👌🏽
Далее
Data Lakehouse: An Introduction
25:00
Просмотров 19 тыс.
Open Data Foundations across Hudi, Iceberg and Delta
34:24
Data Warehouse vs Data Lake vs Data Lakehouse
9:32
Просмотров 43 тыс.
Core Databricks: Understand the Hive Metastore
22:12
Просмотров 15 тыс.
The Hot Technology You Need to Learn!
23:57
Просмотров 6 тыс.
Understanding Data Lakehouse
11:46
Просмотров 7 тыс.
Should You Use Databricks Delta Live Tables?
9:57
Просмотров 6 тыс.
Why a Data Lakehouse Architecture
8:02
Просмотров 57 тыс.