Тёмный
No video :(

07 - Classification, an energy perspective - PyTorch 5-step training code 

Alfredo Canziani
Подписаться 39 тыс.
Просмотров 2,8 тыс.
50% 1

Course website: bit.ly/DLFL22-...
Playlist: bit.ly/DLFL22-Y...
Speaker: Alfredo Canziani

Опубликовано:

 

28 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 6   
@kasparhidayat9293
@kasparhidayat9293 10 месяцев назад
Hi Alfredo - thanks for your videos. Just as a note on the step to zero gradients, I felt a simpler way to think about it is that pytorch is storing the results of past computations and these past results need to be cleared prior to future training batches. Found it confusing when you argue that zeroing + L.backward() are conceptually linked when I don't think they are.
@alfcnz
@alfcnz 10 месяцев назад
There’s a reason why these previous gradients are stored. I have a entire section about it. To perform backpropagation in PyTorch one needs to execute two commands: zeroing + backward. Backward alone does two things: it computes and accumulate the gradient. So, if it is preceded by zeroing the previous grads, then it just computes the new grads. That’s why I’m insisting that ‘zeroing + backward’ accounts for a single statement, i.e. backpropagation.
@Francis-gg4rn
@Francis-gg4rn Год назад
Thank you for making this!
@alfcnz
@alfcnz Год назад
Anytime 😇😇😇
@user-co6pu8zv3v
@user-co6pu8zv3v Год назад
Hello, Alfredo. Thank you for the video! It's nice to spend a Saturday morning watching a lecture. one question: Will your book be available for public sale?
@alfcnz
@alfcnz Год назад
For sale in print and for free in digital version.
Далее
Reforged | Update 0.30.0 Trailer | Standoff 2
02:05
Просмотров 413 тыс.
Variational Autoencoders
15:05
Просмотров 494 тыс.
I'm Coding an Entire Physics Engine from Scratch
9:19
Why Does Diffusion Work Better than Auto-Regression?
20:18
3 PYTHON AUTOMATION PROJECTS FOR BEGINNERS
17:00
Просмотров 1,6 млн