Тёмный
Learn Statistics with Brian
Learn Statistics with Brian
Learn Statistics with Brian
Подписаться
linktr.ee/briangreco
The Rao-Blackwell Theorem Explained
15:12
14 дней назад
What is a degree of freedom?
12:30
2 месяца назад
The Cramer-Rao Lower Bound ... MADE EASY!!!
10:38
3 месяца назад
Link functions for GLMs... MADE EASY!!!
8:56
5 месяцев назад
Maximum Likelihood Estimation ... MADE EASY!!!
9:12
5 месяцев назад
Skewness... MADE EASY!!!
2:59
5 месяцев назад
Unbiased Estimators ... Made Easy!
6:03
5 месяцев назад
Inverse Transform Sampling ... MADE EASY!!!
7:00
6 месяцев назад
Regularization... Made Easy!!!
4:44
6 месяцев назад
Probability vs. Likelihood ... MADE EASY!!!
7:31
6 месяцев назад
The Method of Moments ... Made Easy!
9:02
6 месяцев назад
Chebyshev's Inequality ... Made Easy!
9:46
7 месяцев назад
Empirical Rule (68-95-99.7 Rule) and Z-scores!
8:39
8 месяцев назад
What is R-Squared (R^2) ... REALLY?
8:53
8 месяцев назад
Комментарии
@jjjokkkerrr
@jjjokkkerrr 4 часа назад
6:34 I get the feeling that this is impossible but great video, thanks!
@statswithbrian
@statswithbrian 3 часа назад
lmao, editing is hard, thanks for watching far enough to notice :)
@GPAgoBRRR
@GPAgoBRRR 19 часов назад
Is z-test better for 2 variables than chi square ? Or is it just simpler
@statswithbrian
@statswithbrian 5 часов назад
They are exactly the same in the sense that your chi-squared statistic will literally be the square of your z statistic (eg, if z = 2, your chi-squared statistic will be 4). The only advantage of the z-test is that it more naturally allows you to do a 1-sided test, whereas chi-squared tests are 1-sided by default. For a two-sided z test and a chi-square test of a 2x2 table, you will get exactly the same results (p-value)
@cangozpinar
@cangozpinar День назад
Loved the way you intuitively derived the Markov Inequality. In my opinion it is much better than the proof given in the Introduction to Probability book by Dimitri and John. Thank you very much.
@kelvinwannnn7594
@kelvinwannnn7594 2 дня назад
You made a great explanation better than my professor! Can you make a Video about completeness statistics ?
@statswithbrian
@statswithbrian 2 дня назад
I'm thinking about it - someone requested a minimal sufficient statistic video today, so I might create a playlist with a short video on completeness.
@TheAlen081
@TheAlen081 3 дня назад
Shoutout to educational yt 📈
@statswithbrian
@statswithbrian 3 дня назад
Thank you for my first-ever Super Thanks! :)
@MarcoBova
@MarcoBova 3 дня назад
Really neat explanation and video, could you explain minimal sufficiency with concrete example as in this video?
@statswithbrian
@statswithbrian 3 дня назад
I might make some videos in the future about minimal sufficiency or completeness, but probably not anytime soon. Most sufficient statistics you would ever find are minimal - minimal just means you're not including an extra information. For example, if the sample mean is sufficient, then the pair (sample mean, sample median) would NOT be minimal sufficient because it contains extra information. Similarly, the original complete data set itself is always sufficient, but it's usually not going to be minimal sufficient, because it contains a whole bunch of extra information.
@statswithbrian
@statswithbrian 2 дня назад
Someone else yesterday requested a video on completeness, so I am going to make a series of short videos on minimal sufficiency, completenes, and ancillary statistics that should be out in the next week or so.
@danijeI1337
@danijeI1337 4 дня назад
Thank you, i just recently passed a half of the oral exam with Chebyshev's Inequality question :)
@brazilfootball
@brazilfootball 7 дней назад
Great work, quick question! Why is it ok to use a normal distribution for response variables like weight if weight can't be negative, or zero? I see it a lot, but don't understand why it's so common.
@statswithbrian
@statswithbrian 7 дней назад
There's pretty much nothing that *really* follows a normal distribution - it's all approximations. Take height for example - and suppose the height follows an approximately normal distribution with mean = 64 inches and sd = 4 inches. Even though a normal distribution has some probability of being less than 0 (which is impossible), because that is 16 standard deviations away from the mean, the probability is basically 0 anyways (less than 1 in a billion billion billion billion billion billion). So yes, you're totally right that it's impossible, but assuming it's normal makes things easy and the probability calculations are often pretty accurate!
@brazilfootball
@brazilfootball 7 дней назад
@@statswithbrian Works for me, thank you!
@cleign195
@cleign195 8 дней назад
Really good explanation, much appreciated, Thanks!
@lted6924
@lted6924 9 дней назад
may I ask the differences with chernoff bound and chebyshev inequality
@statswithbrian
@statswithbrian 9 дней назад
Chernoff bounds are a further extension - in Markov’s we use the mean, in Chebyshev’s we use the variance, but in Chernoff bounds we use the moment generating function which gives us an even better bound. If you look at the Chernoff bound Wikipedia page, you’ll see the proof is very similar to this Chebyshev proof, as an extension of Markov.
@lted6924
@lted6924 9 дней назад
thanks
@Sokrabiades
@Sokrabiades 10 дней назад
Solid.
@QuantNovice
@QuantNovice 13 дней назад
That video is gold for every stats student! Thanks a lot for this amazing content!
@dhruvkhare235
@dhruvkhare235 16 дней назад
wow! mad mad respect!!
@СквозьГоризонт-м7ъ
Best explaination, i finally understand. Do you have video about R and it's relation to the R^2? I have seen the video of Veritasium about IQ where he shows the graph with regression and tell labout getting R^2 from R. I wanted to undrstand both, now i know what is R^2, R left.
@statswithbrian
@statswithbrian 17 дней назад
I don't have a video on the correlation coefficient r. It's just the square root of R^2, but r will be either negative or positive depending on whether the line is going up or down. For example, if R^2 = 0.49, then r will be either 0.7 (for a line that is going up) or -0.7 (for a line that slopes downward). So r gives you a little more information (the direction), but it doesn't have an easy interpretation - 0.7 doesn't really "mean" anything. R^2 is a little more general, because R^2 exists for any type of regression model (multiple regression, or more complicated forms of regression), whereas the correlation coefficient r only applies to simple linear regression where there is 1 predictor variable.
@СквозьГоризонт-м7ъ
@@statswithbrian Wow, thanks for the answer. Now i understand. You are the best.❤
@个人主义-h7t
@个人主义-h7t 17 дней назад
Thank you, this helped alot.
@statswithbrian
@statswithbrian 17 дней назад
Thanks for watching, glad to hear it! :)
@andrashorvath2411
@andrashorvath2411 18 дней назад
Great video, only let me express to you that you jumped a huge step when introducing the Poisson equation into the one that simplifies to ((n-1)/n)^t. Why did you transform the probability formalization into that equation? A huge step and I think many might lost the track here. It would be nice if you added the explanation in a comment. Thanks.
@statswithbrian
@statswithbrian 18 дней назад
The data follows a Poisson distribution, so anything would need to be eventually translated into something involving the Poisson formula. It relies on the fact that X1 follows a Poisson (lambda) distribution, sum of x2…xn follows a Poisson ((n-1)lambda) distribution, sum of x1…xn follows a Poisson (n*lambda) distribution
@jakeaustria5445
@jakeaustria5445 19 дней назад
Thank You
@jakeaustria5445
@jakeaustria5445 19 дней назад
Thank You
@RoyalYoutube_PRO
@RoyalYoutube_PRO 19 дней назад
Thanks a lot for making this... Very very helpful
@rodrigokr1_
@rodrigokr1_ 19 дней назад
coool, thanks!
@gabrielwilliams4444
@gabrielwilliams4444 19 дней назад
Your mastery of the topic is evident when you are able to explain it so easily like that. Provides the intuition in a precise manner and with clear connection to the application. Thank you very much! Excellent teacher, please continue sharing, happily subscribed!
@statswithbrian
@statswithbrian 19 дней назад
Thank you! :)
@aleksandarbrozek717
@aleksandarbrozek717 19 дней назад
So at the beginning you said you couldn't use values of degrees celsius as a random variable but why not? In the video it is said that the values must have a minimum value hence they should be non negative so they always have a minimum of 0, but degrees celsius do have a minimum value (the absolute zero) even though they are real numbers.
@statswithbrian
@statswithbrian 19 дней назад
The requirement is not that the variables have to have *any* minimum - the minimum has to be 0 (the variable has to be non-negative). That's why we could only use Markov's Inequality for degrees Kelvin, because the minimum actually would be 0. For Celsius, since it is possible to be as low as −273.15, Markov's inequality would not work.
@Arriyad1
@Arriyad1 19 дней назад
This has nothing to do with the video, but I tried to buy a biased coin from Amazon - no result -, and searching for other means to get one (like Galton Boards, they would be great in class), I had to find out that biased coins probably do not exist. Help me out if you know more about that.
@statswithbrian
@statswithbrian 19 дней назад
I honestly don’t know if it’s physically possible but I’m guessing not. Would love an update if you find one.
@andrashorvath2411
@andrashorvath2411 18 дней назад
I believe you could make a biased coin by welding a straight wire with specific length to one side that is sticking straight outward from the middle. This way when flipping the coin, the wire stick will operate as a weight trying to balance the coin pointing downward, like a keel for a ship, so more times will the coin point with the wire downward. I've not tried this but I believe this could work.
@rabiasufian8919
@rabiasufian8919 19 дней назад
You are a hidden gem. Keep it up. We r here to support you.
@statswithbrian
@statswithbrian 19 дней назад
Thank you! :)
@user-yq4ts4hp1x
@user-yq4ts4hp1x 20 дней назад
Great video! Thxxx!!!
@sherwinprospero
@sherwinprospero 20 дней назад
Simple yet effective explanation! thank you!
@abrrrik
@abrrrik 20 дней назад
awesome!
@abrrrik
@abrrrik 20 дней назад
Thank you so much for this video!! It is so fantastic! So easy to understand in a very interesting way
@DrewAlexandros
@DrewAlexandros 21 день назад
In your final slide, you say that the link function maps from the original scale to "the parameter of the relevant probability distribution". You also say the parameter is personalised.... Is your final slide saying that in general, the link function maps to the parameter of the data's distribution? e.g. "p" in Bernoulli, "sigma" in Rayleigh? Apologies if i haven't understood this correctly.
@statswithbrian
@statswithbrian 21 день назад
Yes, the link function is just transforming a real number with no restrictions (negative infinity to infinity) to something with the correct possibilities for the parameter of interest. In logistic regression, if we were predicting the probability of having diabetes based on weight, you and me would each get a personalized parameter p based on our weight. The heavier person might have p = 0.7, reflecting the fact that their weight makes it more likely that they may have diabetes. The lighter person might have p=0.3. But they will both be between 0 and 1 no matter eat because the link function transformed the scale to ensure that it’s between 0 and 1, which regular linear regression did not do.
@AlokAsthana1954
@AlokAsthana1954 22 дня назад
The two systems are very well explained. Thanks
@xaitpri7905
@xaitpri7905 22 дня назад
Before this I was going through other RU-vid videos. Everyone saying missing data. But in the first min you said it's missing col of a sheet and not row. This cleared my confusion. Thank you sir
@statswithbrian
@statswithbrian 22 дня назад
I totally agree, I used to be confused by the same thing :)
@samuelkobinagyasi5168
@samuelkobinagyasi5168 23 дня назад
Very explicit explanation and very clear. Thank you
@douglasespindola5185
@douglasespindola5185 24 дня назад
I'm feeling a Statquest vibe here and this is very good! Keep the good work, bro! Subscribed!
@statswithbrian
@statswithbrian 24 дня назад
Thank you!
@celinaflores6480
@celinaflores6480 24 дня назад
Good, thank you, but I need to know analisys methods with outliers
@lexi218
@lexi218 24 дня назад
You explained this so well! I wish my lecturers explained everything this way.
@sachinmohanty4577
@sachinmohanty4577 24 дня назад
nicely explained :)
@jakeaustria5445
@jakeaustria5445 24 дня назад
Thank You
@sarahdonnelly5236
@sarahdonnelly5236 26 дней назад
thank you!
@PotatoMan1491
@PotatoMan1491 26 дней назад
Years after university this finally makes sense to me, bless you sir!
@andrashorvath2411
@andrashorvath2411 27 дней назад
Amazing explanation with great examples.
@sarahdonnelly5236
@sarahdonnelly5236 Месяц назад
Thank you!!
@infinitesum
@infinitesum Месяц назад
FANTASTIC EXPLANATION!!!!!!!! Can't get any better.
@newbie8051
@newbie8051 Месяц назад
Wow, great follow up with your prev video explaining probabilities and likelihood Thanks sir
@amitdhiman6567
@amitdhiman6567 Месяц назад
I was struggling to understand MLE for few days now. Finally came across your video and your video totally rewired my brain to understand this concept. You are awesome man. Keep on making such videos.
@statswithbrian
@statswithbrian Месяц назад
Thank you, really appreciate it!
@snehashishghosh7258
@snehashishghosh7258 Месяц назад
Clear and concise
@TejasNarendraChaudhari
@TejasNarendraChaudhari Месяц назад
Wonderfull...
@richardtvtv
@richardtvtv Месяц назад
Awesome video! This answers my fundamental question about the difference between the two, but as is with all things statistics raises 100s more.
@ecarg007
@ecarg007 Месяц назад
At 5:06, in the yellow box said "It is NOT a likelihood in this context either" confused me. When dealing with a single data point, the likelihood function is simply the probability density function (PDF) evaluated at that point. So, the likelihood of observing a specific value x from a normal distribution with mean μ and variance σ² is given by: L(μ, σ² | x) = f(x; μ, σ²) where f(x; μ, σ²) is the PDF of the normal distribution. Therefore, for a single value, the likelihood and density value are equivalent.
@statswithbrian
@statswithbrian Месяц назад
Yes, they are the same number, but the interpretation is different. A likelihood might be the same number as a pmf or pdf, of course, that’s the entire point of the video.
@ecarg007
@ecarg007 Месяц назад
By the way, with quite different explanations at 3:33 of ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-pYxNSUDSFH4.htmlsi=vI0gQGTKWks68Hhm
@statswithbrian
@statswithbrian Месяц назад
@@ecarg007You seem to understand that likelihood is sometimes the same number as a pdf. But a likelihood and a pdf are never going to have the same interpretation because a likelihood is not a probability. Not sure what you’re trying to say about statquest’s video.
@ecarg007
@ecarg007 Месяц назад
​​@@statswithbrian"It is NOT a likelihood in this context either", here "in this context" does it mean the following? In the density function with x=1, the density value is 0.2419707, it is not a likelihood. But you can say, the likelihood of observing x=1 from a normal distribution with mean 0 and variance 1 is 0.2419707.
@statswithbrian
@statswithbrian Месяц назад
@@ecarg007 when we know the true parameters are mu = 0 and sigma = 1, 0.2419707 is the probability density at x=1, which allows us to compute probabilities around x=1. When we do not know the true parameters and only know our observed data x=1, 0.2419707 is only the likelihood in the universe where mu = 0 and sigma=1. The likelihood would be a completely different with different parameters. For example, when x = 1, the likelihood is 0.3989 when mu =1 and sigma = 1. So 0.3989 and 0.2419707 cannot both be probabilities (or probability) densities in this context. They are likelihoods under different scenarios.
@afai264
@afai264 Месяц назад
Great video, thank you. I was often confused between probability and likelihood and would just move on to the next topic still not fully understanding the difference, the "multiple universes" idea is the key part for me which finally explains it! Thank you.