Тёмный

Best Practices for Unit Testing PySpark 

Databricks
Подписаться 116 тыс.
Просмотров 1,9 тыс.
50% 1

This talk shows you best practices for unit testing PySpark code. Unit tests help you reduce production bugs and make your codebase easy to refactor. You will learn how to create PySpark unit tests that run locally and in CI via GitHub actions. You will learn best practices for structuring PySpark code so it’s easy to unit test. You’ll also see how to run integration tests with a cluster for staging datasets. Integration tests provide an additional level of safety.
Talk By: Matthew Powers, Staff Developer Advocate, Databricks
Here’s more to explore:
Big Book of Data Engineering: 2nd Edition: dbricks.co/3Xp...
The Data Team's Guide to the Databricks Lakehouse Platform: dbricks.co/46n...
Connect with us: Website: databricks.com
Twitter: / databricks
LinkedIn: / data…
Instagram: / databricksinc
Facebook: / databricksinc

Опубликовано:

 

18 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 3   
@sezaiburakkantarci
@sezaiburakkantarci 3 дня назад
Matthew is doing incredible work. A lot value packed into just 17 minutes. 💚
@ProjectFinalAudio
@ProjectFinalAudio 6 дней назад
Great talk, will try to normalize these concepts. Can't count how many hours we put into refactoring.
@comptedodoilya
@comptedodoilya Месяц назад
i came specifically for local unit testing which is th part i struggle in
Далее
Inside Out 2: BABY JOY VS SHIN SONIC
00:19
Просмотров 4,8 млн
СМЕЛОЕ РЕШЕНИЕ #shorts
00:14
Просмотров 23 тыс.
ХИТРАЯ БАБУЛЯ #shorts
00:20
Просмотров 1,2 млн
Debugging 101: Replace print() with icecream ic()
12:36
The ONLY REASON To Unit Test
8:26
Просмотров 79 тыс.
What does larger scale software development look like?
24:15
Databricks Workflows
8:20
Просмотров 4,7 тыс.
Inside Out 2: BABY JOY VS SHIN SONIC
00:19
Просмотров 4,8 млн