“Masked Autoencoders Are Scalable Vision Learners” paper explained by Ms. Coffee Bean. Say goodbye to contrastive learning and say hello (again) to autoencoders in #ComputerVision! Love the simple, yet elegant idea!
► Check out our sponsor: Weights & Biases 👉 wandb.me/ai-coffee-break
📺 Vision Transformer explained: • Vision Transformers ex...
Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
donor, Dres. Trost GbR, Yannik Schneider
➡️ AI Coffee Break Merch! 🛍️ aicoffeebreak.creator-spring....
Paper 📜: He, Kaiming, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Doll'ar and Ross B. Girshick. “Masked Autoencoders Are Scalable Vision Learners.” (2021). arxiv.org/abs/2111.06377
References:
🔗 blog.keras.io/building-autoen...
🔗 www.deeplearningbook.org/
🔗 / 1462446494766837773
📺 ViT video: • An image is worth 16x1...
📺 DeiT: • Data-efficient Image T...
📺 Swin Transformer: • Swin Transformer paper...
Outline:
00:00 Intro
00:41 Weights & Biases (Sponsor)
02:10 What are autoencoders?
05:03 Differences between vision and language masked autoencoding
07:02 How does masked autoencoding work for images?
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
Patreon: / aicoffeebreak
Ko-fi: ko-fi.com/aicoffeebreak
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
----------------
🔗 Links:
AICoffeeBreakQuiz: / aicoffeebreak
Twitter: / aicoffeebreak
Reddit: / aicoffeebreak
RU-vid: / aicoffeebreak
#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research
5 июл 2024