Тёмный

Softmax (with Temperature) | Essentials of ML 

Kapil Sachdeva
Подписаться 9 тыс.
Просмотров 3,7 тыс.
50% 1

Опубликовано:

 

27 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 32   
@ssshukla26
@ssshukla26 2 года назад
Great to see a new video after so many days... Will watch it afterwards... thank you Sir....
@KapilSachdeva
@KapilSachdeva 2 года назад
🙏
@ninobach7456
@ninobach7456 11 месяцев назад
This video was one big aha moment, thanks! A lot of weight readjusting
@KapilSachdeva
@KapilSachdeva 11 месяцев назад
🙏
@lielleman6593
@lielleman6593 Год назад
Awsome explanation ! thanks
@KapilSachdeva
@KapilSachdeva Год назад
🙏
@rbhambriiit
@rbhambriiit 2 года назад
Thanks for making it simple and clear.
@KapilSachdeva
@KapilSachdeva 2 года назад
🙏
@abhishekbasu4892
@abhishekbasu4892 10 месяцев назад
Amazing Explanation!
@KapilSachdeva
@KapilSachdeva 10 месяцев назад
🙏
@oguzhanercan4701
@oguzhanercan4701 2 года назад
Great explanation, thanks a lot
@KapilSachdeva
@KapilSachdeva 2 года назад
🙏
@peterorlovskiy2134
@peterorlovskiy2134 11 месяцев назад
Great video! Thank you Kapil
@KapilSachdeva
@KapilSachdeva 11 месяцев назад
🙏
@victorsilvadossantos2769
@victorsilvadossantos2769 3 месяца назад
Great video!
@kalinduSekara
@kalinduSekara 7 месяцев назад
Greate explanation
@KapilSachdeva
@KapilSachdeva 6 месяцев назад
🙏
@murphp151
@murphp151 2 года назад
This is brilliant
@KapilSachdeva
@KapilSachdeva 2 года назад
🙏
@mrproxj
@mrproxj 2 года назад
Hi, thanks for this video. Now I know why my classifier always predicted with such high confidence, be it correct or incorrect. Could there be something else other than temperature to solve this? I would like to determine how confident the model is in its prediction. Is temperature the way to go?
@KapilSachdeva
@KapilSachdeva 2 года назад
Another technique is called label smoothing. It is related but applied to ground truth labels. See - proceedings.neurips.cc/paper/2019/file/f1748d6b0fd9d439f71450117eba2725-Paper.pdf Also there is something model calibration but I have not yet applied them to neural networks.
@mrproxj
@mrproxj 2 года назад
Thanks a lot. This will come a lot in handy!
@SM-mj5np
@SM-mj5np 25 дней назад
You're awesome.
@krp2834
@krp2834 2 года назад
Instead of using using exp function in softmax to make logits positive what if we shift the logits by least logit value [1, -2, 0] => [3, 0, 2]. This also ensures relativity between logits.
@KapilSachdeva
@KapilSachdeva 2 года назад
Thanks Prasanna; forgot to mention that the transformation should be differentiable.
@Gaetznaa
@Gaetznaa 2 года назад
The operation is differentiable; isn’t it just an ordinary subtraction (by 2 in the example)?
@krp2834
@krp2834 2 года назад
@@Gaetznaa The min operation which is required to find the minimum logit to subtract is not differentiate I guess.
@ssssssstssssssss
@ssssssstssssssss 2 года назад
​@@krp2834 The min isn't differentiable, but it's still a differentiable function at other points. But if you do that, the minimum value will be guaranteed to always have a "probability" of zero. That may not be desirable... It also will prevent you from using loss functions like KL Divergence or Cross Entropy. Also, they will not be "logits". I suggest you review the definition of logit
@behnamyousefimehr8717
@behnamyousefimehr8717 7 месяцев назад
Good
@zhoudan4387
@zhoudan4387 4 месяца назад
I thought temperature was like getting a fewer and saying random things:)
@KapilSachdeva
@KapilSachdeva 4 месяца назад
Depends on the context. Here it is about logits. In LLM apis it is to control the stochasticity/randomness.
@HellDevRisen
@HellDevRisen 6 месяцев назад
Great video; thank you :)
Далее
Variational Autoencoder - VISUALLY EXPLAINED!
35:33
Просмотров 13 тыс.
Softmax - What is the Temperature of an AI??
8:34
Просмотров 14 тыс.
Why Do Neural Networks Love the Softmax?
10:47
Просмотров 66 тыс.
Knowledge Distillation in Deep Learning - Basics
9:51
Detection Head | Essentials of Object Detection
15:32
Softmax Function Explained In Depth with 3D Visuals
17:39
Softmax Regression (C2W3L08)
11:48
Просмотров 164 тыс.
Application of Calculus in Backpropagation
14:45
Просмотров 19 тыс.
The U-Net (actually) explained in 10 minutes
10:31
Просмотров 114 тыс.