Тёмный

Activation Functions - Softmax 

Finn Eggers
Подписаться 2,2 тыс.
Просмотров 36 тыс.
50% 1

Наука

Опубликовано:

 

14 май 2018

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 45   
@finneggers6612
@finneggers6612 3 года назад
so for anyone wondering. there is an issue in the derivative of this one! It is not as simple as I stated. I was a bit younger when I did this so please have mercy with me :) The rest should still be correct.
@earltan739
@earltan739 5 лет назад
Thanks for the most concise and forthright explanation of the softmax activation function I've ever seen!
@dor00012
@dor00012 4 года назад
what are you taking about? i can barely understand what he's saying.
@alinamanmedia
@alinamanmedia 3 года назад
The best explanation I've heard of.
@giovannipizzato6888
@giovannipizzato6888 4 года назад
Amazing explanation, I loved the fact you took some example numbers, did the calculations, and showed how to values are modified by the function. Really got the point home. Keep it up!
@99strenth3
@99strenth3 5 лет назад
Good explanation 10/10
@GamelessAnonim
@GamelessAnonim 4 года назад
Damn, why couldn't everyone explain it like this? I am dumb and I need an explanation like I am a 5-year-old but most of the explanations on the internet assume that we are all smart as fuck. Thank you!
@jays907
@jays907 5 лет назад
Thank you so much for the explanation!
@pankaj_pundir
@pankaj_pundir 5 лет назад
Great, found the difference between softmax and sigmoid. Thanks
@nuriel3
@nuriel3 Год назад
GREAT VIDEO ,Thank you !
@ShermanSitter
@ShermanSitter 4 года назад
At 5:30 the light bulb went on. THANK YOU! :)
@optimusagha3553
@optimusagha3553 2 года назад
Thanks, easy to follow👏🏾👏🏾
@omaral-janabi9186
@omaral-janabi9186 5 лет назад
I Love it :) thanks
@clubgothica
@clubgothica 5 лет назад
Excellent explanation.
@feridakifzade9070
@feridakifzade9070 4 года назад
Perfect ✌🏻✌🏻
@guillemguigocorominas2898
@guillemguigocorominas2898 5 лет назад
I think what you mean in the last part about the difference between using a Sigmoid or a Softmax for classification is that for a binary classification problem you only need the probabilities of the two outcomes and a threshold, let's say if I predict A with over 50% probability then my prediction is A, otherwise my prediction is B. For a multi-classification task however you want to normalize over all possible outcomes to obtain a prediction probability for each class
@finneggers6612
@finneggers6612 5 лет назад
yeah exactly. I might not have pointed that out good enough
@ax5344
@ax5344 5 лет назад
at @2:06, there seems to be a typo. x should be [0,1,2,3,4,5] instead of [1,2,3,4,5,6]. f(0) = 1/1+e(0) = 0.5; f(1) !=0.5
@finneggers6612
@finneggers6612 5 лет назад
Yeah you are right. My bad. Thanks for noticing!
@DamianReloaded
@DamianReloaded 4 года назад
I wonder in which cases it's advantageous to use softmax over using percentages of the total sum? Numerically it seems softmax is good for separating big values from smaller ones: EDIT: **googles** Apparently is exactly that. To make high values more evident.
@swaralipibose9731
@swaralipibose9731 3 года назад
You just got a new like and subscriber
@familypart2617
@familypart2617 4 года назад
I love your videos!!!!! It helped me create my very first AI ever! Ur tutorials are so concise! I was wondering, if you know how to do it, could you make a tutorial on q learning in Java, then deep q learning in Java, the deep q learning being something I have been struggling with implementation
@finneggers6612
@finneggers6612 4 года назад
What part are you struggling with? I have implemented it. Basically you should first implement a version with a table without neural networks. After that, you replace the table with a neural network and add a replay buffer. I have code which works. You can also add me on discord (Luecx@0540) and we can talk about it in detail
@familypart2617
@familypart2617 4 года назад
@@finneggers6612 I have the concept down for the basic q learning, however, I cannot figure out how to even begin to train the AI, like what inputs to give it and how to train the AI with the reward and punishment. I tried to send a friend request to chat with u on discord about it, but it didn't work, I can give u my discord real quick KRYSTOS THE OVERLORD#4864
@idobooks909
@idobooks909 3 года назад
This (3:20) little thing tells a lot about you and is the way to reach more subs. Thanks!
@okonkwo.ify18
@okonkwo.ify18 Год назад
There’s no problem with sigmoid , all activation functions have their uses
@Ip_man22
@Ip_man22 5 лет назад
thanks a lot
@ahmedelsabagh6990
@ahmedelsabagh6990 4 года назад
Excellent explanation
@ahmedelsabagh6990
@ahmedelsabagh6990 3 года назад
Super excellent
@alaashams8137
@alaashams8137 3 года назад
respect
@dogNamedMerlin
@dogNamedMerlin 5 лет назад
Thanks, Finn- helpful! I don't think you mention why you need the exponential functions in the Softmax definition. If you showed some negative example values as components of your a-vector (totally legitimate outputs of e.g. a layer with a tanh activation function) it would be easier to see that without them, you wouldn't be guaranteed probabilities bounded by zero and one.
@finneggers6612
@finneggers6612 5 лет назад
You are correct! I did not think about that when I made the video but your critic is 100% correct. Thank you for pointing this one out. the exponential function has a nice derivative behaviour and the output value is always > 0. Everything else would not make sense in this context.
@ccuuttww
@ccuuttww 5 лет назад
the problem of softmax is about the derivative form here math.stackexchange.com/questions/945871/derivative-of-softmax-loss-function?rq=1 u must consider two case 1. i not equal to k and 2. i equal to k I have calculate it by myself but I m not sure if it is right can u go through it once?
@yusuferoglu9287
@yusuferoglu9287 5 лет назад
thanks for the explanation Finn!! I have a question. whenever ı google "derivative ofsoftmax function ı always find something like this " eli.thegreenplace.net/2016/the-softmax-function-and-its-derivative/ ". I am working on a project that is about pure java implementation of multi Layered NN can you help me how can ı use derivative of softmaxfunction.
@american-professor
@american-professor 4 года назад
why do we use e?
@omarmiah7496
@omarmiah7496 3 года назад
my understanding is that we use e because it doesn't change the probability by much as opposed to multiplying by a constant such as 100. it's a form of normalizing data. check out ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-8ps_JEW42xs.html - he goes deeper into the actual use of euler's constant e
@TheFelipe10848
@TheFelipe10848 4 года назад
Congrats on making this so simple to understand, you actualy know what the function does. I sometimes wonder if people actually even understand the content they reproduce or are just too lazy to try to put things in a way others can understand. Einstein did famously say: "If you can't explain it simply you don't understand it well enough"
@finneggers6612
@finneggers6612 4 года назад
well I hope I explained it right. I was a lot younger and I feel like the derivative might not be that simple.... Hope its still remotely correct :)
@alexkatz9047
@alexkatz9047 4 года назад
why do we use "e"?
@finneggers6612
@finneggers6612 4 года назад
I am not 100% sure but maybe its a combination of: "The derivative is pretty simple" and "We need something exponential". The latter one so that probabilities make a little bit more sense
@RamVelagala
@RamVelagala 4 года назад
thanks for the explanation.
@privacyprivate9330
@privacyprivate9330 4 года назад
what is "e" ? how i can get the "e" value ?..
@privacyprivate9330
@privacyprivate9330 4 года назад
in 4.24 Minute
@finneggers6612
@finneggers6612 4 года назад
privacy private It’s probably Eulers number. It’s about 2.7 but in every programming language it should be defined somewhere.
@privacyprivate9330
@privacyprivate9330 4 года назад
@@finneggers6612 Thank you
Далее
Лекция 2.1 - Softmax
31:08
Просмотров 16 тыс.
Activation Functions - EXPLAINED!
10:05
Просмотров 112 тыс.
Это реально работает?!
00:33
Просмотров 1,3 млн
Китайка Шрек всех Сожрал😂😆
00:20
Наташа Кампуш. 3096 дней в плену.
00:58
Softmax - What is the Temperature of an AI??
8:34
Просмотров 12 тыс.
Why do we use "e" in the Sigmoid?
4:54
Просмотров 9 тыс.
The Softmax : Data Science Basics
13:09
Просмотров 49 тыс.
Why Do Neural Networks Love the Softmax?
10:47
Просмотров 64 тыс.
Categorical Cross - Entropy Loss Softmax
8:15
Просмотров 16 тыс.
NEAT - Introduction
21:27
Просмотров 83 тыс.
Softmax Function Explained In Depth with 3D Visuals
17:39
Which Activation Function Should I Use?
8:59
Просмотров 263 тыс.
What is the derivative of the Softmax Function?
37:26
КРАХ WINDOWS 19 ИЮЛЯ 2024 | ОБЪЯСНЯЕМ
10:04