Тёмный

Week 10 - Lecture: Self-supervised learning (SSL) in computer vision (CV) 

Alfredo Canziani
Подписаться 39 тыс.
Просмотров 18 тыс.
50% 1

Course website: bit.ly/DLSP20-web
Playlist: bit.ly/pDL-RU-vid
Speaker: Ishan Misra
Week 10: bit.ly/DLSP20-10
0:00:00 - Week 10 - Lecture
LECTURE Part A: bit.ly/DLSP20-10-1
In this section, we understand the motivation behind Self-Supervised Learning (SSL), define what it is and see some of its applications in NLP and Computer Vision. We understand how pretext tasks aid with SSL and see some example pretext tasks in images, videos and videos with sound. Finally, we try to get an intuition behind the representation learned by pretext tasks.
0:01:15 - Challenges of supervised learning and how self-supervised learning differs from supervised and unsupervised, with examples in NLP and Relative positions for vision
0:12:39 - Examples of pretext tasks in images, videos and videos with sound
0:40:26 - Understanding what the "pretext" task learns
LECTURE Part B: bit.ly/DLSP20-10-2
In this section, we discuss the shortcomings of pretext tasks, define characteristics that make a good pre-trained feature, and how we can achieve this using Clustering and Contrastive Learning. We then learn about ClusterFit, its steps and performance. We further dive into a specific simple framework for Contrastive Learning known as PIRL. We discuss its working as well as its evaluation in different contexts.
1:01:50 - Generalization of pretext task and ClusterFit
1:19:08 - Basic idea of PIRL
1:38:09 - Evaluating PIRL on different tasks and questions

Опубликовано:

 

16 июл 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 28   
@swarajshinde3950
@swarajshinde3950 3 года назад
Thankyou Alfredo and Ishan Misra for such wonderful lecture !!
@alfcnz
@alfcnz 3 года назад
You're welcome! 😊
@user-co6pu8zv3v
@user-co6pu8zv3v 3 года назад
Thank you, Alfredo. Great lectures
@alfcnz
@alfcnz 3 года назад
Glad you like them!
@anrilombard1121
@anrilombard1121 Год назад
2 years after the course has been published Ishan Misra is still making breakthroughs and just as fascinating. Thanks Alfredo for getting him on! Really useful for intermediate or even beginner students to see different work and type of people leading ML/AI!
@alfcnz
@alfcnz Год назад
🔥🔥🔥
@akosahbroni8723
@akosahbroni8723 4 года назад
Thank you, Alfredo, for sharing. I've been waiting for it. Your course has been very helpful to me. I hope you'll share the week-11 video soon.
@alfcnz
@alfcnz 4 года назад
Oh, comments copy & paste 😕
@InnovWelt
@InnovWelt 3 года назад
Thanks for sharing the course, Alfredo. You are doing great service to the students all around the world. Hope you are safe as well!
@alfcnz
@alfcnz 3 года назад
You're welcome 😀 Yeah, hopefully someone will benefit from my work 😊
@anrilombard1121
@anrilombard1121 Год назад
@@alfcnz We certainly are!
@alfcnz
@alfcnz Год назад
❤️❤️❤️
@darkmythos4457
@darkmythos4457 4 года назад
Thank you for providing us with such great lectures
@alfcnz
@alfcnz 4 года назад
Our pleasure!
@harriszeng6231
@harriszeng6231 3 года назад
thank you very much! i am willing to listen more self-supervised deep learning from your sharing
@alfcnz
@alfcnz 3 года назад
Pushing the 2021 edition soon! 😇😇😇
@TheSSB007
@TheSSB007 3 года назад
This is amazing! Thank you for sharing
@alfcnz
@alfcnz 3 года назад
You're welcome 😄
@yakupgorur
@yakupgorur 3 года назад
Thanks
@ShihgianLee
@ShihgianLee 3 года назад
I love to have a guest lecture to tell us the journey self-supervised learning and relates them to different papers. I will have to check out a few of the papers. I love the Q&As too! Thank you so much for sharing!
@ekkapricious
@ekkapricious Год назад
Regarding the question at the end about batch normalization. What is the intuition behind why batch normalization may pose a problem? Is the idea that if both positive and negative samples are within the same batch then due to the normalization process the mean between the negative and positive samples would cancel out and result in dilution of the polarity of the samples and hence reduce how positive and negative the samples are?
@tarunnarayanan1513
@tarunnarayanan1513 4 года назад
Hello Dr.Canziani, a small clarification in the content at 1:00:00 of the video. Ishan talks about how they trained a resnet-50 for a pretext task and saw that the mAP of the representations began to drop as they reached the res5 block of the encoder network because it starts to learn high-level signals of the pretext task itself. In another paper by Google termed 'Revisiting Self-Supervised Learning for Visual Representations' (arxiv.org/pdf/1901.09005.pdf), they've mentioned the exact opposite in section 3.1 of the paper, where they say the representation quality remains intact due to the skip-connections making residual units invertible which facilitates preservation of information. Can you please clarify this contradiction for me?
@imisra_
@imisra_ 3 года назад
Great question Tarun. I am aware of this observation from the paper by Kolesnikov et al., 2019. Unfortunately, in multiple versions of our experiments (and those of our collaborators, other work) we have found that res5 does perform worse than res4. Apart from the work of Kolesnikov et al., almost all the prior work I am aware of that uses ResNet-50 with a pretext task makes the same observation. I am not quite sure why this observation is completely opposite to Kolesnikov et al.
@dhathrisomavarapu2089
@dhathrisomavarapu2089 4 года назад
Hello Dr. Canziani, Does the model training for Pretext tasks also need data that has labels already? I am guessing it does not, since the goal of pretext task training is to identify embeddings/features agnostic to labels, right?
@alfcnz
@alfcnz 4 года назад
Exactly, it does not. The pretext task belong to the “self-supervised learning” family, which constructs annotations in an automated manner.
@ijaAI
@ijaAI 3 года назад
it doesn't need a label. it just uses annotations based on the data structure
@jonathansum9084
@jonathansum9084 4 года назад
Do anyone have a notebook that is the pretask in this video, for example rotating and citing them apart and shuffle them in Pytorch? I think I may need them. Thank You.
@alfcnz
@alfcnz 4 года назад
How about torchvision.transforms? pytorch.org/docs/stable/torchvision/transforms.html
Далее
Week 10 - Practicum: The Truck Backer-Upper
1:00:26
Просмотров 4,5 тыс.
What Is Self-Supervised Learning and Why Care?
9:43
Просмотров 17 тыс.
Lecture 1: Introduction to Power Electronics
43:22
Просмотров 11 тыс.