Тёмный

Deep Learning Foundations: Misha Belkin's Talk on deep learning through the prism of interpolation 

Soheil Feizi
Подписаться 4,1 тыс.
Просмотров 2,1 тыс.
50% 1

Course webpage: www.cs.umd.edu/...
Abstract: In the past decade the mathematical theory of machine learning has lagged far behind the successes of deep neural networks on practical challenges. In this lecture I will outline why the practice of neural networks precipitated a crisis in the theory of Machine Learning and rethinking of certain basic assumptions. I will discuss how the concept of interpolation (fitting the data exactly) clarifies many of the underlying issues leading to new theoretical analyses. Finally, I will briefly mention some new results showing how interpolated predictors may relate to practical settings where the training loss, while small, is not usually driven to zero.
The main reference paper is the review: arxiv.org/abs/...
I will also briefly discuss some of the results in arxiv.org/abs/... and arxiv.org/abs/...

Опубликовано:

 

8 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 4   
@yaram.bahram6715
@yaram.bahram6715 2 года назад
Very interesting remarks on generalization. Thank you for sharing this great lecture!
@jfjfcjcjchcjcjcj9947
@jfjfcjcjchcjcjcj9947 2 года назад
Thanks for the lecture! I think that there are a couple of points that would make the lecture more digestible if they are explained a bit more in depth. First, is the definition of interpolation, different people imply different things when referring to interpolation. I presume that Misha here by interpolation refers to the ability of estimators to achieve zero training error and still not overfit to the hold-out test set? The second aspect which might require some more explanation is @40:05 when Misha is describing the difference between interpolation accuracy and interpolation in the l2-norm. I presume that what is meant by l2-norm interpolation is that 2 estimators with the same l2-norm on their parameters exhibit the same generalisation capabilities, is that right?
@alexisfernandez8052
@alexisfernandez8052 2 года назад
I can show you how! Let know!
@TheMaP142
@TheMaP142 2 года назад
Michael
Далее
Oops 😅
00:10
Просмотров 2,8 млн
БЕРЕМЕННА В 16 - 5 ВЫПУСК (МАМА В 16)
1:53:50
🍎 Apple Pie Cinnamon Roll Donut #Shorts
00:21
Просмотров 3,5 млн
Deep Learning Foundations by Soheil Feizi : Transformers
1:28:55
Oops 😅
00:10
Просмотров 2,8 млн