This channel needs more viewers. I'm not in love with the content but its very beat to learn how certain everyday things function. Or how certain problems are solved.
+zocke1r if the machine can not properly identify a person, it will set off a flag that goes into danger mode... it will shoot a laser that immediately vaporizes the unidentifiable object. In the future, if the machine can not positively identify someone... that will be unacceptable to Big Brother.
Ilman how many times will the machine rotate the photo to identify the object... how long before the laser is used? We can not predict exactly how AI will work... people believe it will be smarter than us, they say it will learn to make it's own decisions, and what it does could be very unpredictable.
+MAXXHEW Through binary search it will calculate at which direction it should rotate. Once it finds the correct angle, the layers are used in all angles already checked. That's why we shouldn't give a rotated picture to an AI. :D
Nice video, as always. Hey, I know it's not really your style to talk about games (except maybe Tetris and Portal), but have you guys heard of "No Man's Sky". I think it would make a great topic to talk about that game and procedurally generated worlds, either on this channel or on Numberphile. Anyways, keep on, whatever the topic, your videos are always fun to watch :)
Solution: Use makeup to paint additional eyes, mouths, etc. on your face. You might look like some weird alien from some SF horror flick, but face recognition software will not get any meaningful results from you.
nah bro. see, i can map the brightness value of the image of your face to tell me what area has high glossy reflection highlights, and this will tell me where your real eye is as opposed to the ones you painted on.
won't the water in your eye also have glossy reflection? Besides to teach the neural network to deal with fake makeup you would need thousands of examples to train it correctly, so it could try to use specular reflections to find which eye is real. And you wont really find thousands of distinct examples easily.
My father, Woody Bledsoe, PhD, pioneered the first Facial Recognition techniques in the 1960's at his company Panoramic Research, near Sunnyvale CA, while working on contracts for the CIA/NSA. My dad was a pioneer in the fields of AI, Pattern Recognition, Automatic Theorem Proving and Computer Reasoning, and past president of IJCAI. Many people now use variants of his initial algorithms in many Facial Recognition software systems. It would be nice if he was to receive more credit for these innovative and groundbreaking techniques, IMO. -LW Bledsoe, Austin TX
I was reading an article about facial recognition algorithms today, and it was righfully mentioned that Dr. Woody Bledsoe was the first one. However most of his work was not published because he was developing them for government.
I don't know if I'm being stupid so I apologise if it's a silly question. But my understanding is you are looking for a particular position of a feature. You then have an input of pixel coordinates (so some point in a picture). You then put that into a learned regressor (either via deep learning or some other machine learning algorithm) and it comes out with a vector that points to the most likely direction the feature you are looking for is. This vector is learned from all the training data manually set by people. My question is how does the computer decide what each pixel coordinate relates to in the training date? To elaborate, in the video, Dr. Michel Valstar used the eyebrow as an example and said that because it's an eyebrow, the algorithm gives an output vector that points to the most likely direction of the feature (inner eye). How does it know that the pixel value there is an eyebrow in both training and test data?
I'd really like to see a tutorial on which changes one would have to do to a picture of a face (i.e. in Photoshop or GIMP), so that it can't be used in order to create a biometric facial profile of a person any more, while still keeping the picture recognizeable for a human, so that a human that sees the picture can still identify the person, but a computer can not. Is that even possible? If so, what has to be altered in a picture for that?
I think he got a little bit wrong. He said that Gradient Descent is used to determine if a sliding window has detected an eye and is related to deciding that an eye has been found. I don't know about the exact algorithm they're using for deciding when to stop, but Gradient Descent is a training algorithm. The local optima problem for Gradient Descent has only to do with TRAINING the algorithm that takes a patch of the image says "here is an eye", and nothing whatsoever to do with deciding when to stop scanning through a picture. Otherwise; a good video, keep it up.
Do we have any idea of how our brain does facial recognition, and how it compares to our computer vision algorithms? I guess our visual processing is quite brute force since we have massive parallel processing ability (while computers are better at sequential processing), with dedicated hardware (since brain damages to certain areas can destroy the ability; prosopagnosia) and some kind of probability based heuristic since we see faces in literally anything (pareidolia).
+Stein Gauslaa Strindhaug one of the things you need to get rid of is the idea that the brain is a computer. It sort of is in, the way it deals with information, but a computer bases everything of of calculation and numerical values, whereas with brains it's a little more abstract. If you were to tap into the data feed from a CPU and compare to a similar data feed from a brain (however you did this), you would probably see that the CPU data is deterministic, whilst the Brain data would be wildly abstract. It's like comparing machine code to arabic.
***** AI will follow an exponential growth curve once it figures out it can self-modify. As its intelligence grows, it will be able to modify itself even further, to the point that we can no longer understand what it is thinking. Initially, AI will be slow to grow, but it will accelerate as time goes on. In biological systems, a Natural Intelligence (us) does not have the ability to self modify beyond a certain extent.
Evidently there will be physical limits, but answer this for me: If you were capable of modifying your mind and you knew it would make you more intelligent, would you do it?
So how is Valstar's group going to scale from a thousand faces to a million faces? Will they take their existing set of faces, find eigenfaces or some sort of basis between vector maps for facial features, and then decompose any new face into a combination of these bases?
+Allen Law I assume you were asking for identical twins. It is hard -or almost impossible- even for the human beings to recognize identical twins just from their face. What reveals the identical twins is mostly the way they behave rather than just appearance (or sometimes the hair style, outfit preferences etc but still not only face or facial muscle movements.). So probably I would say no it couldn't recognize the difference between identical twins unless it is deeply trained to distinguish the differences in the patterns of two people.
+Cigdem TURAN In fact you may recognize their expressions rather than their face geometry (A smiles all the time, B is sad all the time. If face is smiley, it's probably A). Would two identical twins smile in the same exact way? Maybe two identical twins have two discernible sets of facial expressions. (Just guessing, I don't know anything about twins or how emotions are translated to facial expressions)
+Enrico Piazza are there REALLY "identical" twins, like in "their facial features are 100% the same"? If the answer is no, an extremely good software, paired with extremely high quality photos maybe could make out the differences.
Enrico Piazza I do not know if there is any physiological or computational studies on expression differences of twins but worth to check it. As you said, we may end up finding a unique way of expressing emotions.
Why not use the vertical symmetry of the face? By going to the lowest point of the face the chin. Then Through the curvature of the jawline you find the center of the jaw which is the centerline of the face. Then the mouth is above . Nose is above and the eyes are above and out on both sides of the nose.
+Dave Preston Not a bad thought, and some amount of symmetry is likely used - however, not all faces are going to be looking straight at the cameras. Like in this video, he was never facing the camera right on and so if the face is a bit rotated in 3D space you'll have a hard time using symmetry. The left eye and right eye could be different distances from the lens and thus bigger or smaller, etc.
+Alexander Reynolds Also to take into account they are using deep learning algorithms which I assume they take any picture as input the algorithm will sort out which are the markers before even going to detect the facial features for the eyes etc. So if you would want to go with the chin you would first start to compute where it is, normally you can do that but I think it just consumes even more resources than simply learning a few vectors which tell you the feature you're after. But maybe I'm wrong...
+Alexandru Gheorghe Oh ok . I just figured that the chin is always the lowest point of the face and if they turn their face the curvature of the chin would be different on one side which could be calculated. I figured the positions of the eyes, nose, ears and mouth are some relative constant based on structural distance. If you found one you could calculate the others centers based on a normalizing algorithm.
+Dave Preston I like your idea but if i can play the devils advocate, for example, could you teach a computer to differentiate between a narrow face, to one that is slightly turned from the camera? Probably, but how feasible is that w/ modern tech?
+rentacow That's a good question. Even damage would be a problem but since the face has a slope or angle . You could look for the symmetric slope and if it doesn't match you could look to see if the other slope is closer to vertical then search for the nose outer point. Then you could work a scheme to get results.
What feature type do you guys use when training the classifier ,HAAR or LBP? also which one of the two would you say is faster when using positive samples bigger then 5000 images?
sir am i right in saying geometrical features based algorithms uses less features when compared to appearance based methods for face expression recognition ??? i badly need to understand this.
Is it plausible to have a general learning AI algorithm to create its own initial training data, instead of having to do so manually, like you did for the face recognition?
+· 0xFFF1 there are some preliminary studies on general AI and automatic training based on pure observation and inferential rules, but this is pretty much science fiction (yet).
+· 0xFFF1 Yes and automated feature detection is almost certainly something a lot of these algorithms are going to use. But sometimes it's easier to just tell the thing what eyes look like rather than hope it figures it out on its own.
How difficult would it be to adapt this technology to CCTV and suveillance, with faces being presented at awkward angles, not directly front-on, slightly concealed etc?
+Sam Horler This is in use, through Bayes Theorem plays a role here. Say you have an algorithm that correctly detects if a person is or is not a wanted criminal 99.9% of the time. You put that to work in a CCTV camera in a public street and you get a bunch of warnings about criminals. Unfortunately, since the vast majority of people are not criminals you're just going to wind up with a big list of false positives.
Often such systems rely on motion tracking and will patiently wait for that opportunity to grab one really good frame to perform recognition on, all the while continuing the motion tracking. Though many will have some degree of capability to compensate for a limited range of off axis recognition so long as sufficient features can be identified (which also allows them to deal with imperfections like glasses/sunglasses or facial hair). A lot of those details have to be inferred, though, as the vendor's jealously guard any intellectual property that they feel will give them an edge on the competition.
+Sam Horler This method is robust to faces not directly head on. It's part of the reason this method works so well, in fact. If you are looking for predefined sizes, then it won't work. But for example if you detect what you think is an eye, and it is rotated a bit, then it will point to where the nose should be given the rotation, etc. So this network of self referencing is actually what makes this algorithm work well. Of course, for very extreme angles it probably wouldn't be great, especially on a low-res camera.
+Luke Stutters Kinda how machine learning / AI works. You need a lot of data to parse, but once you have that you can make a lot of inferences automatically with statistics and thus starting from a bit of data yourself, the machine learning algorithm could parse through thousands more of images and create a much larger database to use.
+Chris Baldwin The positions are the same, just rotated. It be possible to check for that. I mean, it could just check the points relative to each other, not relative to the photo's orientation.
that regression sounds a little questionable. How about starting with head location in a low res version, then you have narrowed down eye locations quite a bit
OK, so you start every video with a pair of pseudo HTML tags, i.e. then end the video by closing the inner tag ()… I have not yet observed the closing wrapper tag . Am I blind, or do you not understand how HTML works?
+Ich Nichtdu FACS is the system which names the facial muscle movements which can also be used for facial expression recognition; and mostly be used. So if you find the facial points more accurately, you can observe their movements more accurately and it eventually ends up a better insight of FACS. And when he says the mouth corners will be wider and upper, it is for sure he is talking about the FACS system. AU #12 - Lip Corner Puller to be exact.
Cigdem TURAN I too thought (and still think) it's very likely to be employed here as a tool for analyzing and categorizing of facial expressions. But I guess you're not 100% sure either?
+Jebus Chris You could pretty much deceive it with a photo of the person you want to detect. One time I could deceive a cell phone detector with a smiley face drawn on a paper... Nowadays systems can only detect faces.. recognizing different faces apart from one another is a much more difficult task... even the same person can be completely different depending of the situations...
Eyebrows are above eyeballs...unless maybe you're talking about kids hanging off a jungle gym, or somebody trying to stretch out their back by being in inversion boots, or.... :-) It just depends.
i don't like it...too hand-coded...come up with an algorithm that learns to identify the key features automatically and then we're talkin....i doubt the brain does it your way
+Kevin Durden mmm... don't bring brains into this. They are a completely different beast. The pattern recognition systems in the brain use associations and employ a vast parallel processing matrix. The brain employs techniques for facial recognition which are almost wholly alien to what computers do. The fact is that the brain doesn't quite work mathematically as such. While computers use binary, in the same way that the brain does (eh, kinda) brains also use various neurotransmitters and hormones to appropriate information and disseminate it between neurons. It's a highly complex field, but I can safely say that what works for a brain would not work for a computer. It's a completely different system, with completely separate architecture