Тёмный

Orignal transformer paper "Attention is all you need" introduced by a layman | Shawn's ML Notes 

Yuxiang "Shawn" Wang
Подписаться 483
Просмотров 6 тыс.
50% 1

Thank you for checking out my video notes on the orignal transformer paper "Attention is all you need", as introduced by a layman - me! I would love to share my ML learning journey with you.
Paper information:
- Vaswani, Ashish, et al. "Attention is all you need." Advances in neural information processing systems 30 (2017).
Please let me know in the comment section regarding any questions, points of discussion, or anything you would like see next. See you in the next video!

Наука

Опубликовано:

 

5 апр 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 18   
@oo_wais
@oo_wais Месяц назад
one of the very few videos i found on youtube that explains the architecture very well
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
Thank you so much for the recognition!
@tk-og4yk
@tk-og4yk Месяц назад
Another Video! Looking forward to watching.
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
Haha thank you for your support! It was an old deck I made a year ago, so I might as well record it :)
@matthewritter1117
@matthewritter1117 Месяц назад
Incredible content and your style is a perfect mix of confident and relatable. Keep it up!
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
I appreciate the encouragement :)
@OEDzn
@OEDzn Месяц назад
amazing video!
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
Thank you!
@420_gunna
@420_gunna Месяц назад
Seems like a great video, subbed! 🙂
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
Thanks for the sub! Appreciate the recognition ❤️
@s8x.
@s8x. Месяц назад
please do more videos like this
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
Thank you! Will do :)
@aga5979
@aga5979 3 дня назад
Thank you for the very valuable explanation. But in what f ucking world do laymen speak with dot product , cosine and e to the power of time and time prime? 😅😅😂😂.
@isiisorisiaint
@isiisorisiaint Месяц назад
pretty okay until andrew's attention slide, then when it comes to your own explanations things become murky, and when you get "explain" the decoder, and then the full codec, you're swiping everything under the rug in a few short seconds when in fact this is exactly the section you should have spent most of time. all in all, a nice video until adrew's slide, basically worthless afterwards
@yuxiangwang9624
@yuxiangwang9624 Месяц назад
Thanks for the feedback! Will learn to improve :) Would you mind explain in more details on which part I was missing for the encoder details? I can look into those and see if I can add some later!
@isiisorisiaint
@isiisorisiaint 25 дней назад
@@yuxiangwang9624 darn, i got a notification that you responded to my comment, but only the first line of your reply was shown ("Thanks for the feedback! Will learn to improve :)"), and i didn't actually open to see your full reply until now. I will be back to you with the details, sorry for the delay...
@nxlamik1245
@nxlamik1245 День назад
Work on explainging things easily. It seems u have enough knowledw but you made it difficult
@MrMusk-it5nz
@MrMusk-it5nz Месяц назад
You aren't definitely a layman
Далее
The Attention Mechanism in Large Language Models
21:02
The Most Important Algorithm in Machine Learning
40:08
Просмотров 204 тыс.
Why Does Diffusion Work Better than Auto-Regression?
20:18
This is why Deep Learning is really weird.
2:06:38
Просмотров 323 тыс.
Vision Transformer Basics
30:49
Просмотров 16 тыс.