Тёмный
No video :(

AAAI 20 / AAAI 2020 Keynotes Turing Award Winners Event / Geoff Hinton, Yann Le Cunn, Yoshua Bengio 

ICML IJCAI ECAI 2018 Conference Videos
Подписаться 2,6 тыс.
Просмотров 43 тыс.
50% 1

*Highlighted Topics*
02:52 [Talk: Stacked Capsule Autoencoders by Geoffrey Hinton]
36:04 [Talk: Self-Supervised Learning by Yann LeCun]
1:09:37 [Talk: Deep Learning for System 2 Processing by Yoshua Bengio]
1:41:06 [Panel Discussion]
Auto-chaptering powered by VideoKen (videoken.com/)
For indexed video, conftube.com/v...
*All Topics*
03:09 Two approaches to object recognition
03:53 Problems with CNNs: Dealing with viewpoint changes
04:42 Equivariance vs Invariance
05:25 Problems with CNNs
10:04 Computer vision as inverse computer graphics
11:55 Capsules 2019: Stacked Capsule Auto-Encoders
13:21 What is a capsule?
14:58 Capturing intrinsic geometry
15:37 The generative model of a capsule auto-encoder
20:28 The inference problem: Inferring wholes from parts
21:44 A multi-level capsule auto-encoder
22:30 How the set transformer is trained
23:14 Standard convolutional neural network for refining word representations based on their context
23:41 How transformers work
24:43 Some difficult examples of MNIST digits
25:20 Modelling the parts of MNIST digits
27:03 How some of the individual part capsules contribute to the reconstructions
28:37 Unsupervised clustering of MNIST digits using stacked capsule autoencoders
31:25 The outer loop of vision
31:36 Dealing with real 3-D images
32:51 Conclusion
36:04 [Talk: Self-Supervised Learning by Yann LeCun]
36:25 What is Deep Learning?
38:37 Supervised Learning works but requires many labeled samples
39:25 Supervised DL works amazingly well, when you have data
40:05 Supervised Symbol Manipulation
41:50 Deep Learning Saves Lives
43:40 Reinforcement Learning: works great for games and simulations.
45:12 Three challenges for Deep Learning
47:39 How do humans and animals learn so quickly?
47:43 Babies learn how the world works by observation
48:43 Early Conceptual Acquisition in Infants [from Emmanuel Dupoux]
49:33 Prediction is the essence of Intelligence
50:28 Self-Supervised Learning = Filling in the Blanks
50:53 Natural Language Processing: works great!
51:55 Self-Supervised Learning for Video Prediction
52:09 The world is stochastic
52:43 Solution: latent variable energy-based models
53:55 Self-supervised Adversarial Learning for Video Prediction
54:12 Three Types of Learning
55:30 How Much Information is the Machine Given during Learning?
55:54 The Next Al Revolution
56:23 Energy-Based Models
56:32 Seven Strategies to Shape the Energy Function
57:02 Denoising AE: discrete
58:44 Contrastive Embedding
1:00:39 MoCo on ImageNet
1:00:52 Latent-Variable EBM for inference & multimodal prediction
1:02:07 Learning a (stochastic) Forward Model for Autonomous Driving
1:02:26 A Forward Model of the World
1:04:42 Overhead camera on highway. Vehicles are tracked
1:05:00 Video Prediction: inference
1:05:15 Video Prediction: training
1:05:30 Actual, Deterministic, VAE+Dropout Predictor/encoder
1:05:57 Adding an Uncertainty Cost (doesn't work without it)
1:06:01 Driving an Invisible Car in "Real" Traffic
1:06:51 Conclusions
1:09:37 [Talk: Deep Learning for System 2 Processing by Yoshua Bengio]
1:10:10 No-Free-Lunch Theorem, Inductive Biases Human-Level AI
1:15:03 Missing to Extend Deep Learning to Reach Human-Level AI
1:16:48 Hypotheses for Conscious Processing by Agents, Systematic Generalization
1:22:02 Dealing with Changes in Distribution
1:25:13 Contrast with the Symbolic AI Program
1:28:07 System 2 Basics: Attention and Conscious Processing
1:28:19 Core Ingredient for Conscious Processing: Attention
1:29:16 From Attention to Indirection
1:30:35 From Attention to Consciousness
1:31:59 Why a Consciousness Bottleneck?
1:33:07 Meta-Learning: End-to-End OOD Generalization, Sparse Change Prior
1:33:21 What Causes Changes in Distribution?
1:34:56 Meta-Learning Knowledge Representation for Good OOD Performance
1:35:14 Example: Discovering Cause and Effect
1:36:49 Operating on Sets of Pointable Objects with Dynamically Recombined
1:37:36 RIMS: Modularize Computation and Operate on Sets of Named and Typed Objects
1:39:42 Results with Recurrent Independent Mechanisms
1:40:17 Hypotheses for Conscious Processing by Agents, Systematic Generalization
1:40:46 Conclusions
1:41:06 [Panel Discussion]
1:41:59 Connection between Neural Networks as a Computer Science and a Machine Learning Concept - Natural Competition
1:45:35 Idea of Differentiation: Representation and Listening
1:49:36 Alternate to Gradient Based Learning
1:51:04 What is the role of university when Facebook, Google can manage these enormous projects
1:53:50 What do you think students to read?
1:54:50 Mechanisms for Human Level AI
1:57:59 Where do new ideas come from? How do you decide which one works out?
1:59:54 How should I proceed when people writes me reviews and doesn't like my research?
2:01:53 Publications effect on the field
2:05:36 Can we code during AI doing science
2:06:52 What is not General Intelligence, how to measure? and Neural Architecture
2:08:44 Disagreements

Опубликовано:

 

28 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 29   
@444haluk
@444haluk 4 года назад
Bodyguard: How tough are you? Geoffrey Hinton: I classified MNIST digits with 98.7% accuracy. Bodyguard: Yeah? That's not tough. Geoffrey Hinton: Without using the labels. Bodyguard: Sir, here you go, sir. Sorry, sir.
@gallicios
@gallicios 4 года назад
what a historical moment. and I was there. what a magic moment to be alive in history of AI
@josy26
@josy26 4 года назад
Hinton's explanation of transformer is the best I've seen so far
@ZachDoty0
@ZachDoty0 4 года назад
What a time to be alive. This is inspiring :) I really wish they had answered the question at the end about The Measure of Intelligence paper. We need better ways to measure AGI and better datasets to train it on.
@errorderivative3907
@errorderivative3907 4 года назад
Pure joy to watch
@Marcos10PT
@Marcos10PT 4 года назад
Thank you for posting this. I loved the Q&A session and even took some notes, it's very intimate and inspiring!!
@TheAIEpiphany
@TheAIEpiphany 3 года назад
"Stacked Capsule Autoencoders" paper has got 37 citations so far. I think Geoff might be onto something here hahaha (not being sarcastic!). Nice presentation! I especially liked the Q/A part. Their presentations were way too abstract for a broad audience as you'd have to have some more context about their particular research area to get something super valuable out of it. Coming from a guy who does this for living in Microsoft and also in his free time. So pretty much all the time I'm awake. We need more fireside chats with AI minds on RU-vid. As for the keynotes I got from this talk: attention and self-supervision will play a significant role in the future of ML research. Also, "crazy" ideas like capsule nets or more generally questioning the current methods is super important - we need researchers who are working without overfitting to h-index as their main metric and instead focus on the long-term goals of improving ML for the greater good.
@willd1mindmind639
@willd1mindmind639 3 года назад
The biggest problem in computer vision is not specific to machine learning and neural networks, as opposed to the inherent way images are encoded for computer processing. This starts with the way most camera sensors work to convert light waves into pixels using the bayer filter. And the output of that is simply a file with a bunch of pixels which are nothing more than 3 numeric values representing red, green and blue color values for each pixel. These numbers are not light waves that are converted by the eye into bio chemical electrical changes as they move from the retina into the visual cortex. And when humans see images on a computer they don't see numbers they see the light waves projected by the monitor or display device. All of that is a big problem in computer vision in many ways, from color perception, to pixel processing, segmentation and so on. For colors the primary problem is that when your eye looks at an image it is very easy to pick out specific color regions, such as a red apple in a bowl of fruit, based on light spectrum values received in the retina. A computer cannot do this because each pixel has a unique r,g,b value and there is no simple way to define what groups of pixels represent a specific color and every pixel has "red" in it. Which means that for computer vision most of the processing involves coming up with some way of determining the relationship between each pixel and the pixel next to it. This is generally called a convolution where you go through a region of pixels and apply a mathematical operation in order to "simplify" it for purposes of generating a statistical relationship over a group of pixels. But that still isn't the same as seeing a "red" apple in the bowl of fruit like the human eye. And that is because the human eye can easily group various light signals together as part of coherent shapes and objects. This grouping is called segmentation in computers and is a difficult task that depends on using some kind of convolution algorithm to try and assign mathematical relationships for this kind of grouping based on large volumes of data. And this is where CNNs come into play for computer vision, but these CNNs do not provide the same ability to process visual signals as humans do because the human visual system has no concept of pixels or convolutions.
@ManChoitube
@ManChoitube 4 года назад
Thank you for sharing!!!
@goldilockszone4389
@goldilockszone4389 4 года назад
I had to download the video and watch it in VLC, 'cause I couldn't hear a thing
@pratik245
@pratik245 2 года назад
I think i have bettered the understanding on Yashua Bengio on consciousness and meta learning and why deep learning depedendce on bits for computing and inference search algorithm of gradient descent are its limitations to what it can do.
@zianyang1011
@zianyang1011 4 года назад
Capsule for 2020. Btw for first comment.
@oudarjyasensarma4199
@oudarjyasensarma4199 4 года назад
have the video of them receiving the Turing award come out yet?
@pratik245
@pratik245 2 года назад
Yesterday i had given a comment on twitter on contrastive learning, i would like to correct my statement that there is dimensionlity redcution but it is in the form of a concept. So, basically its just another way of looking at transfer from an entropy perspective. Comparable entropic systems are similar, hence, i don't see many conceptual different between what transformer q, k, v do and energy function do. Its just another good name.
@iwbagiai
@iwbagiai 4 года назад
Africa is a country
@jonatan01i
@jonatan01i 4 года назад
Did they decide that all of them will hit the mic?
@MrNightLifeLover
@MrNightLifeLover 4 года назад
Are the slides available?
@SamA-kb9ug
@SamA-kb9ug 4 года назад
Can someone use machine learning to clear the damn audio
@alexter-sarkisov8321
@alexter-sarkisov8321 3 года назад
10:31 what is it you do with the coordinate frame??
@pratik245
@pratik245 2 года назад
Nice conference
@renegadeMoscow
@renegadeMoscow 4 года назад
To Yann Le Cunn: learning to plan complex action sequences actually are very easy to solve. I have a solution
@marcospereira6034
@marcospereira6034 4 года назад
Where's your working notebook?
@pratik245
@pratik245 2 года назад
Hinton' view of vision as hierarchical is not true. In perception, hierarchy does play a role but hardly in vision. Vision is what a 3D mapping by eyes of an external environment which is used as data for infinite dimensional decision making. World exists in infinite dimesnional space of physical planes. Eyes are evolved to create a 3d physical mapping of these planes and passed to human perception engine which makes sense of this using physical and meta physical spaces where human perception wander, although in a very optimized manner(much much better than computing) to find optimized paths or decision making. Those who excel in finding paths which are best are often people excelling in those dimensions whether its chess, football, occult, porn, movies or whatever. However, enhancing a dimesnion is like writing the best function for this nuance. We live in a world of recursion as well as non-recursion of several functions which are either a composition, association or hierarchical. All at the same time.
@pratik245
@pratik245 2 года назад
Thats why i believe colloborating systems of AI will be the best restricting functional space for each subsystem. Pretty much how humans are. However, computer vision best approach is not thinking of a universal computer vision but area restricted and our expectation of things we highly expect in that area, rare events are rare for humans too and then colloborating computer vision systems to form the universal computer vision. Pretty much like a local driver for each area transferring driving to next local expert.
@jdavid5257
@jdavid5257 3 года назад
Need a better mic...
@pratik245
@pratik245 2 года назад
Somehow glucose is more efficient than electricity. 😂
@jacobi2067
@jacobi2067 4 года назад
CNNs are rubbish! lol
@TheReferrer72
@TheReferrer72 4 года назад
If Hinton says their rubbish, they are rubbish!
Далее
The Fan’s Fang Skin🔥 | Brawl Stars Sneak Peek
00:16
Yann LeCun: "Energy-Based Self-Supervised Learning"
1:00:10
Yoshua Bengio: Deep Learning | Lex Fridman Podcast #4
42:19
AI DEBATE : Yoshua Bengio | Gary Marcus
2:02:25
Просмотров 35 тыс.
The Turing Lectures: The future of generative AI
1:37:37
Просмотров 591 тыс.