Тёмный

Lesson 23: Deep Learning Foundations to Stable Diffusion 

Jeremy Howard
Подписаться 122 тыс.
Просмотров 6 тыс.
50% 1

(All lesson resources are available at course.fast.ai.) In this lesson, we work with Tiny Imagenet to create a super-resolution U-Net model, discussing dataset creation, preprocessing, and data augmentation. The goal of super-resolution is to scale up a low-resolution image to a higher resolution. We train the model using AdamW optimizer and mixed precision, achieving an accuracy of nearly 60%. We also explore the potential for improvement by examining the results of other models on Tiny Imagenet from the Papers with Code website.
We discuss the limitations of using a convolutional neural network for image super-resolution and introduce the concept of U-net, a more efficient architecture for this task. We implement perceptual loss, which involves comparing the features of the output image and the target image at an intermediate layer of a pre-trained classifier model. After training the U-net model with the new loss function, the output images are less blurry and more similar to the target images.
Finally, we discuss the challenges of comparing different models and their outputs. We demonstrate how perceptual loss has improved the results significantly, but also note that there isn't a clear metric to use for comparison. We then move on to gradually unfreezing pre-trained networks, a favorite trick at fast.ai. We copy the weights from the pre-trained model into our model and train it for one epoch with frozen weights for the down path. This results in a significant improvement in loss.

Опубликовано:

 

7 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 11   
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
38:00 no need of augmentation calback
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
1:28:00 replacing some layers of unet with pre trained classifier
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
1:06:00 modulelist is just like sequential but it doesnt autofwd we need to define the fwd method
@thehigheststateofsalad
@thehigheststateofsalad Месяц назад
I see you are trying to help. However, can you combine all your comments into one and delete the others? It's just messy.
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
34:00 trivial augmentations
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
57:00 why in superres squeez and again unsquezz why not we use stride1
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
1:09:00 weight initialization of unet
@mattzucca4102
@mattzucca4102 Год назад
😮
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
1:22:00 comb loss
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
49:00 superres
@satirthapaulshyam7769
@satirthapaulshyam7769 9 месяцев назад
59:00
Далее
Men Vs Women Survive The Wilderness For $500,000
31:48
Fixing Plastic with Staples
00:18
Просмотров 1,4 млн
Friends
00:32
Просмотров 958 тыс.
How Stable Diffusion Works (AI Image Generation)
30:21
Просмотров 147 тыс.
How might LLMs store facts | Chapter 7, Deep Learning
22:43
The U-Net (actually) explained in 10 minutes
10:31
Просмотров 101 тыс.
These are the Worst Construction Mistakes Ever Made
12:52
Variational Autoencoders
15:05
Просмотров 496 тыс.
How AI 'Understands' Images (CLIP) - Computerphile
18:05