Тёмный

14. Classification and Statistical Sins 

MIT OpenCourseWare
Подписаться 5 млн
Просмотров 55 тыс.
50% 1

Опубликовано:

 

6 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 40   
@leixun
@leixun 4 года назад
*My takeaways:* 1. L1 and L2 logistic regression 5:22 2. Receiver operating characteristic 16:00 3. Statistical Sins 26:55 3.1 Example1: statistics about the data is not the same as the data, we should plot and visualise data 28:40 3.2 Example2: lying with charts, e.g. Y-axis start point 30:57 3.3 Example3: lying with charts, e.g. Y-axis start point, no Y-axis label, confusing X-axis label 32:45 3.4 GIGO: garbage in garbage out: analysis of bad data is worse than no analysis at all 35:40 3.5 Survivor bias: it's not easy to get random samples in real life 41:35, in such cases, we can't use the empirical rule, central limit theorem and standard error on them 46:38
@user-r1g5i
@user-r1g5i 3 года назад
41:50 - the aircraft is P-47 "Thunderbolt"
@djangoworldwide7925
@djangoworldwide7925 Год назад
I smile when he smiles I feel like he's my beloved professor
@adiflorense1477
@adiflorense1477 4 года назад
I really like your teaching style that is easy to understand even for me who is zero in the field of data science
@JohnCena963852
@JohnCena963852 5 лет назад
The red sox and the cardinal on fox is pure gold.
@jasonbarr5176
@jasonbarr5176 4 года назад
I'm a Cardinals fan and I had almost successfully forgotten that World Series ever happened before watching this 😂😂😂
@nashsok
@nashsok 10 месяцев назад
I was playing around with the Titanic data and noticed another correlation between features - The average ages of the passengers was not evenly spread across the different cabin classes, with the cabins having average ages of 39.16 for first class, 29.51 for second class, and 24.82 for third class. Examining the weights that logistic regression provides when evaluated on just a single cabin class shows that age within a cabin class is strongly associated with the passenger surviving.
@shobhamourya8396
@shobhamourya8396 5 лет назад
@25:00 As sensitivity increases specificity decreases, so plotting sensitivity vs specificity will result in a convex curve. Whereas sensitivity vs 1 - specificity will result in a concave curve and AOC will be easier to visualize. Also, 1 - specificity results in FP/(TP+FN) which is False Positive Rate, so it will be TPR(sensitivity) vs FPR that is the focus will be on positives...
@JCResDoc94
@JCResDoc94 4 года назад
youre getting there
@jftsang
@jftsang Год назад
Doesn't doing sensitivity instead of 1 - sensitivity just flip the curve left-right, and keep the area the same? Or am I missing something?
@studywithjosh5109
@studywithjosh5109 4 года назад
Wow this course has been really great!
@JCResDoc94
@JCResDoc94 4 года назад
31:20 traditionally pink is boys and blue is girls. it is a 19thC switch that saw that occur.
@McAwesomeReaper
@McAwesomeReaper 11 месяцев назад
Mention Calhoun's party affiliation in the future! Its important that people know.
@RodrigoNishino
@RodrigoNishino 3 года назад
Every time when such awesome class ends I wonder to myself: "Why aren't they clapping?" 👏 👏 👏 👏 👏 👏
@JamBear
@JamBear 3 года назад
Start an undergrad at MIT and you'll see why.
@berndczech1554
@berndczech1554 5 лет назад
At 41:40 it's a en.wikipedia.org/wiki/Republic_P-47_Thunderbolt
@Speed001
@Speed001 2 года назад
Ah, I guessed a spitfire. That plane with the teeth painted on the front.
@WhaleTasteGood
@WhaleTasteGood 4 года назад
thank you for this wonderful lecture
@ebateru
@ebateru 3 года назад
Great lecture. Just a quick question : the age coefficient is very small (-0.03), but are all the features normalized before fitting logistic regression? If they are not, then age has a much bigger impact since the difference between a 20 and a 50 year old is (30x-0.03) =-0.9 which is almost twice the impact of being in the third cabin.
@sharan9993
@sharan9993 3 года назад
Features are normalized
@wwmheat
@wwmheat Год назад
I have same question. And I've downloaded the source code and I can see that features are NOT normalized
@stephenadams2397
@stephenadams2397 4 года назад
For further GIGO clarification see climate model
@mmahgoub
@mmahgoub 4 года назад
That bit of how FOX news ignorely plotted a graph is really funny 😂
@UrgeidoitNet
@UrgeidoitNet 7 лет назад
great job!
@batatambor
@batatambor 4 года назад
So professor first makes a model with perfect collinearity and doen't explain the issues related to be doing that. The first model felt in to a dummy variable trap.
@haneulkim4902
@haneulkim4902 4 года назад
Thank you!
@victorcy
@victorcy 3 года назад
when worse than random, just reverse the prediction :-)
@kvnsrinu
@kvnsrinu Год назад
I answered the question at 8:30. I also need a candy:)
@samarhabib2614
@samarhabib2614 Год назад
The comment at 35:18 is hilarious
@videofountain
@videofountain 7 лет назад
Thanks. Video time point .... ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-K2SC-WPdT6k.htmlm33s . I still want to know what quote was [not] shown, so that I can insert my own picture.
@mitocw
@mitocw 7 лет назад
The original slide read: "A Thing of the Past?" and "Insert Photo Here," referring to garbage-in-garbage-out data (see slides 19-20 in the deck found on the OCW site: ocw.mit.edu/6-0002F16).
@MrGarysjwallace
@MrGarysjwallace 2 года назад
He doesn’t know the sensitive of variance to determine the secret weighted formula. Shameful!!
@Speedymisha
@Speedymisha 6 лет назад
Man this guy is quite left winged haha, but a good lecture
@Guinhulol
@Guinhulol 2 года назад
Well, he's smart, therefore is on the right side of history, politics aside, John Guttag is a damn good teacher, too be fair, they all are.
@wwmheat
@wwmheat Год назад
Thanks for the great lecture. For me a bit confusing thing in the ROC curve (blue curve at 18:46) was that it wasn't showing the values of 'p' explicitly. Here is 'p' color-coded drive.google.com/file/d/15-AUzxuzvgFPfUoU3MxyCI1I1-Z19qWO/
Далее
15. Statistical Sins and Wrap Up
44:43
Просмотров 38 тыс.
9. Understanding Experimental Data
47:06
Просмотров 61 тыс.
PUBG MOBILE | Metro Royale: Fun Moments #4
00:16
Просмотров 144 тыс.
8. Sampling and Standard Error
46:45
Просмотров 79 тыс.
13. Classification
49:54
Просмотров 134 тыс.
5. Random Walks
49:21
Просмотров 179 тыс.
20 Most Common Logical Fallacies
52:46
Просмотров 93 тыс.
12. Clustering
50:40
Просмотров 300 тыс.
MIT Introduction to Deep Learning | 6.S191
1:09:58
Просмотров 608 тыс.