Тёмный
No video :(

06L - Latent variable EBMs for structured prediction 

Alfredo Canziani
Подписаться 39 тыс.
Просмотров 10 тыс.
50% 1

Опубликовано:

 

28 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 20   
@cambridgebreaths3581
@cambridgebreaths3581 3 года назад
Wonderful, more and more🙃. Many thanks, Alfredo!!!
@alfcnz
@alfcnz 3 года назад
Haha, I don't see the end 😭🤣😭🤣
@geekyrahuliitm
@geekyrahuliitm 3 года назад
Thanks for making these videos publically available. :-)
@alfcnz
@alfcnz 3 года назад
😇😇😇
@ShihgianLee
@ShihgianLee 2 года назад
Thank you for uploading this Alf! This wasn't in the 2020. After watching the 2021 EBM lecture, I feel that everything is about pushing down and up energy. This clarifies it! The interpolation vs extrapolation in high dimension space is interesting. It is like a detective work to deduce result in high dimensional space 😀
@alfcnz
@alfcnz 2 года назад
You're welcome 😊😊😊 Next semester there'll be more material on energy stuff. 🔋🔋🔋
@ShihgianLee
@ShihgianLee 2 года назад
🥳🥳🥳
@user-co6pu8zv3v
@user-co6pu8zv3v 3 года назад
Thank you, Alfredo :)
@alfcnz
@alfcnz 3 года назад
You're welcome 😁
@sebastianpinedaarango8239
@sebastianpinedaarango8239 3 года назад
Thanks for the video. I really like it. Unfortunately, the background does not help to read the equations sometimes. I would suggest to look for another approach to increase the contrast between the font and the background.
@alfcnz
@alfcnz 3 года назад
Usually students follow along with the PDF version of the slides, so I thought it was not a big problem. But yeah, I've got your point. I've been constantly experimenting new techniques, some work better than others.
@siddhantrai7529
@siddhantrai7529 3 года назад
Hi Alfredo, Just a small doubt at 1:11:30 (at end of factor graph), when Yann mentioned that the algo is dp and it is in linear time. But the way he explained the algo, it was more like Dijkstras greedy search, which is O(V log E). As far as I remember, Dp based shortest path that work on network exhaust ively, have O(VE) time complexity, like bellman-ford. Please do correct me if I am wrong. I know this isn't of much concern here, but it bugged me a bit, thus wanted to clarify. Thank you.
@666zhang666
@666zhang666 2 года назад
In a GAN part: How do we know that Gen(z) when we learn it to produce 'y^hat' with lowest possible energy will always produce 'wrong sample'? (so sample for which we want to increase energy). Maybe it can happen that it will produce something correct?
@shrey-jasuja
@shrey-jasuja Год назад
I have a doubt, In previous video Yann explained that while training EBM in contrastive learning methods with joint embedding methods, we take negative samples in such a way that they are very different so that the system learns better, but in Graph transformer networks we took the best possible answer for contrastive learning. So how does it works?
@alfcnz
@alfcnz Год назад
You need to add some time stamps or it’s going to be impossible for me to address your question.
@shrey-jasuja
@shrey-jasuja Год назад
@@alfcnz I am talking about the discussion between the time instants 1:39:00 and 1:43:00
@-mwolf
@-mwolf 2 года назад
What does "averaging the weights over time" mean exactly? at 43:40
@alfcnz
@alfcnz 2 года назад
w_{t} = a₁ w_{t-1} + a₂ w_{t-2} + a₃ w_{t-3} + …
@oguzhanercan4701
@oguzhanercan4701 2 года назад
56:35 ,,,, hi from Turkey :)
@alfcnz
@alfcnz 2 года назад
👀👀👀
Далее
02L - Modules and architectures
1:42:27
Просмотров 22 тыс.
04L - ConvNet in practice
51:41
Просмотров 10 тыс.
03L - Parameter sharing: recurrent and convolutional nets
1:59:48
Why Does Diffusion Work Better than Auto-Regression?
20:18
ICLR 2020: Yann LeCun and Energy-Based Models
2:12:12
Просмотров 21 тыс.
MIT Introduction to Deep Learning | 6.S191
1:09:58
Просмотров 505 тыс.