Тёмный

Lesson 9A 2022 - Stable Diffusion deep dive 

Jeremy Howard
Подписаться 122 тыс.
Просмотров 32 тыс.
50% 1

Опубликовано:

 

7 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 15   
@markhopkins8731
@markhopkins8731 Год назад
Love your simple explanation of a manifold Jonathan. It's the first time it's made sense to me. Looking forward to the coming lectures.
@al3030
@al3030 Год назад
Thank you for this deep dive. The sampling explanation especially was helpful to try to get an intuition for what the model does.
@timandersen8030
@timandersen8030 Год назад
Appreciate this supplemental deep dive into code of stable diffusion!
@saidmoglu
@saidmoglu Год назад
pretty good video to further understand SD!
@spider853
@spider853 Год назад
I finally understand the schedulers! Thank you!
@adityagupta-hm2vs
@adityagupta-hm2vs 2 месяца назад
Also, are we using latent space as gradients here, as we are subtracting gradients from the latent, which we typically do from weights in conventional NN ?
@alexrichmonkey7845
@alexrichmonkey7845 Год назад
Please explain the ancestral samplers.
@adityagupta-hm2vs
@adityagupta-hm2vs 2 месяца назад
How do we decide the scaling factor in VAE part i.e. 0.18215, any hint on how to decide it ? I did try changing and could see the different output, but what's a good way to choose ?
@climez
@climez Год назад
This is useful but I wish you went into more detail here and there. Is some CLIP or similar model included in the stable diffusion implementation? If so, are precomputed weights of the CLIP model used to calculate noise_prediction in each step? I.e. we pass the current noisy image (in a latent space) and the text embedding to CLIP and then calculate the gradient for each voxel of the image so that something (semantic similarity?) is maximized? I wish you would say what happens during training of the mode and what then happens during inference :).
@AM-yk5yd
@AM-yk5yd Год назад
I'm surprised how... complexity(?) raised up. It's second day and I only on 4th minute, spent 30 minutes debugging my coding-along session (I wrote rand_like instead of randn_like and my parrot photo went green instead of grambled)
@offchan
@offchan Год назад
rand is uniform whereas rand is normal (gaussian)
@howardjeremyp
@howardjeremyp Год назад
Feel free to skip over lessons 9A and 9B if you don't feel ready for them just yet - they're optional extras for those looking to dig deeper.
@JohnSmith-he5xg
@JohnSmith-he5xg Год назад
Why do you "sample()" from the latents? Does this mean the latents are not the same between runs?
@jaivalani4609
@jaivalani4609 Год назад
How can it perform the custom action. basically how can we fine tune it for our input and target image we want as per our text action
@DinoFancellu
@DinoFancellu 3 месяца назад
Don't like all this jumping around. Would be much easier to simply go through it, in a linear fashion, explaining as you go. Disappointing
Далее
Lesson 9B - the math of diffusion
51:36
Просмотров 32 тыс.
Modus males sekolah
00:14
Просмотров 11 млн
Lesson 9: Deep Learning Foundations to Stable Diffusion
2:15:16
Why Does Diffusion Work Better than Auto-Regression?
20:18
How AI 'Understands' Images (CLIP) - Computerphile
18:05
Intro to FastHTML
7:29
Просмотров 66 тыс.
I Built a Neural Network from Scratch
9:15
Просмотров 276 тыс.
These are the Worst Construction Mistakes Ever Made
12:52
A Hackers' Guide to Language Models
1:31:13
Просмотров 523 тыс.
Stable Diffusion explained (in less than 10 minutes)
9:56