I make concise educational videos on artificial intelligence, deep learning, image and video processing, computer vision, computational photography, and various topics in computer science.
Playlists 📌 Deep Learning Crash Course 📌 Image and Video Processing 📌 Hands-on Deep Learning: TensorFlow Coding Sessions 📌 Computer Science in 5 Minutes 📌 Machine Learning in Finance 📌 Cognitive Science Inspirations
Subscribe for more videos!
This is my personal RU-vid channel. Opinions are my own and do not express the opinions of my current or former employers, including Netflix, Apple, Intel, and Motorola.
i wonder whats the point in image compressio... oh wait people usualy store a lot of images on their devices and people usualy use phones for taking pictures which have small storage
I've suspected for a while that the best way to eliminate any lingering vestiges of an accent, should you wish to do so, is probably to record your own voice and listen to yourself from a 3rd person perspective-which this video seems to confirm. Interestingly, I found that as you (partly/mostly) lose your original accent, you also lose the ability to speak authentically/consistently with that (former) accent. Meaning, you can't fake it any more than you current accent permits. In fact, to "regain authenticity," you would probably have to record your own voice and compare it to a recording of someone who's still accustomed to speaking the language that forms the basis of the accent on a daily basis. Of course the same is ultimately true for dialects as well.
At 5:40 (Model Distillation), why does the new model to be trained on the new, specific task receive as an input the same large dataset as the large pre-trained model? I thought that only the mainstream model should receive the large dataset as an input.
you have explained so well, i have seen so many videos , but, the way you explain from start to end is vary relative to what we are learning. very very good explanation about stable diffusion work-flow.
Surprisingly, you have it backwards: LLMs always _extrapolate_ and never _interpolate*._ Essentially due to the curse of dimensionality. Even in compressed latent spaces, neighbors are simply light-years away from each other. There is no convex hull boundary around the training distribution from which the model can sample/interpolate. These things are the real deal. Turns out that Sutskever was right -- if you train on a sufficiently general problem with enough compute/parameters/data, then reasoning pops out. It's obvious in retrospect. If brute force works well enough for mother nature, why not for us? It's maybe a disappointingly prosaic, inefficient, or intellectually humbling discovery; but reality has a long history of popping humanity's ego bubble. Turns out that humanity's intellect is no more special than our place within the solar system. So it goes. Bring on the autoregressive revolution. The skeptic's position will slowly erode away as their arbitrary goal posts are continually scored against and consequently moved further back. AI masters the "impossible" game Go. Then language, textual knowledge. Produces incredible audio, and understands yours. Understands images and produces remarkable works of its own. AI reaches modality after modality, with continuous improvements. As I write this, transformers are revolutionizing robotics. Surely once an AI fetches a glass of water for a skeptic, they'll _finally_ have an epiphany...right? An alien mind can only be intelligent if it supports the exact same modalities as us, right? And only makes mistakes like humans, right? And masters _every_ human skill at _only_ prodigy level or above, right? And reciting past material implies a good memory for humans, but merely being an autocorrect for LLMs, right? And when machines imagine something we enjoy, it's imitating creativity, but if it's something we dislike, it's hallucinating. And... ..Ah, let myself go a little bit, didn't I? I'll stop ranting at vague critics. 😬 Can I make one last observation though? About job anxiety: you know who's not jazzed about AI taking out grunt work? The grunts. They're rarely emphasized, but universally present, collateral damage from technological disruption. The net good of revolutions contain the gross misery of displaced millions. I worry about them, and their chance of adapting to our neat new tools. I'm not sure a laid off career truck driver will be so chipper struggling to understand what these tools are, let alone using them to pay his family's mortgage. I apologize for the negativity. It just felt like you were cheerfully waving away justified anxiety for what is the heaviest, most profound shift in human history. We're _commoditizing an on-demand supply of general intelligence._ We aren't just adding handy buttons to Photoshop or whatever. That's like watching the electrification of America and pointing to the potential of brighter lamps. What we're going through requires gravitas, deep empathy, and steadfast mindfulness. It's the only way we'll make it through this together. Buckle up. * See _Learning in High Dimension Always Amounts to Extrapolation_ (LeCun et al., 2021)
Kinda crazy to think just 3 years ago the idea of an AI making art was a niche and curious idea. Now look at where we are. Mind boggling. I remeber a time when openAI's gpt technology was practically a toy used for auto generated text adventure games, and now it has reshaped our entire civilization. Everyone has heard of chatGPT, my grandmother knows.
Starting the New Year with something new! I've just launched my own podcast: Cognitive Creations! Episodes cover cognitive science, computer science, AI, and more. Join me on this journey into the science and art of the human mind. Spotify: podcasters.spotify.com/pod/show/cognitive-creations/
you are so much great my professor only teach the compression technique but dont even explain when and how and why they are used m thanks to you i understand this now
Wow, that's well explained. The only thing I think is (kinda) wrong: It should not be Megabyte (MB), but instead Mebibyte (MiB). Because Megabytes are base 2 which means 12MB = 12.582.912 Bytes, and Mebibyte are base 10 which means 12MiB = 12.000.000 Bytes. Another fix would be to say that the original image is 4.096x3.072 pixels... However, this mistake is in almost every Literature, so it doesn't actually matter and, moreover, it's not the point of the video... ❤