Тёмный
Aleksa Gordić - The AI Epiphany
Aleksa Gordić - The AI Epiphany
Aleksa Gordić - The AI Epiphany
Подписаться
x-Google DeepMind, x-Microsoft engineer explaining AI. ❤️

▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬

💼 LinkedIn - www.linkedin.com/in/aleksagordic/
🐦 Twitter - twitter.com/gordic_aleksa
👨‍👩‍👧‍👦 Discord - discord.gg/peBrCpheKE

📺 RU-vid - ru-vid.com
📚 Medium - gordicaleksa.medium.com/
💻 GitHub - github.com/gordicaleksa
📢 AI Newsletter - aiepiphany.substack.com/

▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬

💰 BECOME A PATREON OF THE AI EPIPHANY ❤️

If my RU-vid videos, GitHub projects, and blogs help you,
consider helping me out by supporting me on Patreon!

The AI Epiphany - www.patreon.com/theaiepiphany
One-time donation - www.paypal.com/paypalme/theaiepiphany
Best LLM? Qwen 2 LLM w/ author Junyang Lin
1:03:33
Месяц назад
Ishan Misra (Meta) - Emu Video Generation
55:26
5 месяцев назад
InstructPix2Pix (w/ OpenAI's Tim Brooks)
59:19
5 месяцев назад
LLaMA 2 w/ Thomas Scialom (LLaMA 2 lead)
1:00:23
8 месяцев назад
EleutherAI Pythia w/ Hailey Schoelkopf
1:03:33
8 месяцев назад
Tanishq Abraham (PhD at 19) - MedARC
1:10:59
8 месяцев назад
OpenAI DALL-E 3 with James Betker (1st author)
52:43
10 месяцев назад
Комментарии
@ramensusho
@ramensusho День назад
This was some nice explanation
@EatuopLyreqzgj-f5l
@EatuopLyreqzgj-f5l День назад
Robinson Sandra Hall Melissa Lee David
@devanshugupta9749
@devanshugupta9749 2 дня назад
Where do they apply the transformation of input features from D to (D, H_j), I didn't find it in the code.
@viswavsn
@viswavsn 4 дня назад
Thanks for the clear instructions. When I installed and tried running the cuda samples, it errored out. It got fixed when I installed cmake. sudo apt install cmake
@unsaturated8482
@unsaturated8482 5 дней назад
FYI , we are still adding them 27:00 , but at the same time, instead of seperate times , Great diagram regardless thanks.
@kimchi_taco
@kimchi_taco 6 дней назад
Glad to watch! RoPE, emb * sqrt(fan_in), RMSNorm need f32!!! 32:00 Alexa, awesome question!
@present-bk2dh
@present-bk2dh 8 дней назад
crazy to see that this was just 3 years ago
@XnndjehdhkNxndjhds
@XnndjehdhkNxndjhds 9 дней назад
Hernandez Scott Thompson Richard Clark Christopher
@GeorgesSegundo
@GeorgesSegundo 10 дней назад
Your personal story is so Epic that you redefined my course in life. I taught it was impossible to reach Big Tech without Ph.D. and you are the very living proof that it is. Thank you so much for your existence brother, now i believe that i can do that too.
@ignasa007
@ignasa007 13 дней назад
Hey, sorry, 21st minute, you say we need r+1 layers for information to have propagated to a node r-hops away. Why is that? Shouldn't r layers be enough? For example, a 1-layer MPNN will aggregate information from B_1 (nodes 0-hops and 1-hop away).
@user-tg1hz8bl9g
@user-tg1hz8bl9g 13 дней назад
What languages support f?
@World-um5vo
@World-um5vo 14 дней назад
Excellent - Can you do a video on CLIPstyler
@AmirHosseinAlamdar
@AmirHosseinAlamdar 15 дней назад
Great! Also the code section was a very good idea. Saved a lot of time thanks.
@VioFax
@VioFax 16 дней назад
Pretty sure cuDNN installation on linux has made people self-delete before... They really do make it harder than it needs to be. If anything updates you gotta do it over too.
@jimbig3997
@jimbig3997 19 дней назад
In a year, it's going to look really stupid everyone in 2024 paying large chunks of money for video-game video cards so that they can do AI/ML. For one thing, none of these cards have the RAM to run even modestly sized models.
@JeremyWinston-fo5fz
@JeremyWinston-fo5fz 17 дней назад
So what do i do? How does anyone manages to run AI and Machine Learning if modest size models cannot be runned? Should i buy a pc and build it or there is a good laptop for it? Please enlighten me.
@jimbig3997
@jimbig3997 15 дней назад
@@JeremyWinston-fo5fz Seems like the answer is a lot of RAM with an NPU. It runs LLMs faster than a GPU and you are not bottlenecked with the amount of RAM on the GPU card, which is usually not enough to hold a full model. Even just a fast CPU would be fine, like a 7950x or an ARM chip (M4). But even a cheap NPU addition can get things like a Raspberry Pi running LLMs. It seems incorrect to spend $1k+ on a GPU with only 12GB of RAM with the intention of running a 70+GB LLM.
@carlosrondonmoreno9796
@carlosrondonmoreno9796 21 день назад
Wonderful tutorial! Really really good examples. Thank you so much!
@Kutlutr
@Kutlutr 24 дня назад
Wonderful video, to the point, starting with the paper, understanding the necessary background and moving forward with practical exercise, very nice, keep up the good work bro.
@user-ds3xv7li6i
@user-ds3xv7li6i 26 дней назад
So, let's say we have trained all animals with images and texts except wolf and dog images and texts, if we were to ask to a model which has been trained with CLIP and zero-shot classification to draw a picture of a dog and a wolf sitting together, would it still be able to draw them? But it hasn't been trained neither the words of dog and wolf nor images of wolf and dog, if we were to describe as "draw a picture of an animal which howls and looks similar to a fox or canine and lives in the forest" there is a chance it would be able to predict it, but if we say the exact sentence of "draw a picture of wolf" it wouldn't be able to draw it, right? Because it doesn't know the meaning of "wolf", or how it looks like, or it is an animal or a table, in its universe it would be like humans trying to predict how 10th dimension looks like? Am I right?
@mehdizahedi2810
@mehdizahedi2810 28 дней назад
awesome presentation, thanks.
@TheAIEpiphany
@TheAIEpiphany 29 дней назад
Thomas joined me to discuss LLaMA 3 work! Also - if you need some GPUs check out Hyperstack: console.hyperstack.cloud/?Influencers&Aleksa+Gordi%C4%87 who are sponsoring this video! :)
@SocratesMethods
@SocratesMethods 29 дней назад
Informative. Currently training my own quantized models using Llama3.1 open source.
@objectobjectobject4707
@objectobjectobject4707 29 дней назад
Ide gasss !!!!! Bravo Aleksa i hvala :)
@DailySFY
@DailySFY 29 дней назад
Thanks a lot !!! Please continue the awesome work.
@wazmo20
@wazmo20 Месяц назад
Protector of the ocean, bestower of ML content
@roydaboii9925
@roydaboii9925 Месяц назад
Do you still work at deepmind?
@bluetran1
@bluetran1 Месяц назад
incredible!
@abbashaider7165
@abbashaider7165 Месяц назад
That is a great hands-on tutorial which has a perfect mix of theory and practical usage. Thanks
@alejandroreza6593
@alejandroreza6593 Месяц назад
Great video. I'm starting my research for machine learning. I subscribed after looking at the video. Hopefully you do the same for specific needs. I'm interested in trading.
@DynestiGTI
@DynestiGTI Месяц назад
Please do another interview in the future on Groq!
@cutiepets664
@cutiepets664 Месяц назад
You have masters in CS and electronic
@bfc7649
@bfc7649 Месяц назад
nice videos!
@davidro00
@davidro00 Месяц назад
2 years in, still tryin to understand that damn regex
@Viewable11
@Viewable11 Месяц назад
Various machine learning experts - real experts who design and run ML systems as a business, not amateurs on RU-vid - recommend 4 CPU threads per GPU. Considering a dual GPU system, the Ryzen 3600 is the cheapest 12 thread CPU. Alternatives from newer generations are Ryzen 5600, Intel 12400F and Ryzen 7600. All Intel CPUs consume more than double the amount of power per performance than AMD CPUs. From a long term cost perspective, chosing AMD is a no brainer because the Intel CPUs will cost a fortune on electricity bills.
@JeremyWinston-fo5fz
@JeremyWinston-fo5fz 17 дней назад
Source? I do not doubt you but i am struggling, i don't know what to buy to make my first pc for deep learning, or if buying a laptop would be better. My budget is 1500 more or less.
@wwkk4964
@wwkk4964 Месяц назад
Excellent 👌 wish he had more time
@SigmayetB
@SigmayetB Месяц назад
Awesome mate , good on you
@mkamp
@mkamp Месяц назад
Awesome session. Great and relatable speaker. 2h would have been even better. 😂
@MrAmgadHasan
@MrAmgadHasan Месяц назад
Thank you Aleksa and Justin for this interesting talk. Could you please invite someone from the Deepseek team for a talk?
@charliesteiner2334
@charliesteiner2334 Месяц назад
23:00 Interesting take, but I feel like this is a "greedy search" approach to technological development, when our society could really benefit from look-ahead. I would have been interested in what he would have said if you pressured him to talk about what he thinks beneficial *pure* research, not translating quickly into products, would look like.
@AIBrewingWithNik
@AIBrewingWithNik Месяц назад
thankyou aleksa for such interesting guests
@TheAIEpiphany
@TheAIEpiphany Месяц назад
@@AIBrewingWithNik thanks Nik, lmk whom you'd like to see next!
@MrAmgadHasan
@MrAmgadHasan Месяц назад
​@@TheAIEpiphanyCan we have someone from the deepseek team?
@TheAIEpiphany
@TheAIEpiphany Месяц назад
We had Junyang author of Qwen 2 one of the best LLMs with us today! If you need some GPUs check out Hyperstack: www.hyperstack.cloud/free-credit-landing-page-op-2? who are sponsoring this video! :) and you'll also get some free credits ;)
@user-fv7br5it2b
@user-fv7br5it2b Месяц назад
Thanks a lot for your videos on GNNs! After scrolling through multiple papers/articles and not understanding them, I am finally starting to get some intuition on GNNs after watching your videos. And now going back to those papers, I started to actually understand them. Thanks a lot!
@TF2Shows
@TF2Shows 2 месяца назад
The adversarial loss - i think the explanation is wrong You said the discriminator tries to maximize it, however, you have just shown that it tries to minimize is (the term becomes 0 if D(x) is 1 and D(\hatX) is 0). So the discriminator tries to minimize it (and because its a loss function it makes sense), and the generator tries to do the opposite, maximize it, to fool the discriminator. So I think you mis-labeled the objective: L_GAN we try to minimize (minimize loss) in order to train the discriminator.
@karikoga320
@karikoga320 2 месяца назад
How would you store database password and username in Streamlit/Hugging Face?
@atabhatti
@atabhatti 2 месяца назад
Good coverage of the paper. I like how you go into each aspect in depth. I think you struggle--and you acknowledge it yourself--explaining some of the more complicated parts of the paper. If you can nail how to communicate those you would make an awesome teacher.
@aradhyamathur7355
@aradhyamathur7355 2 месяца назад
are these notebooks available on your github as well ?
@aradhyamathur7355
@aradhyamathur7355 2 месяца назад
also the weight file is not longer available any drive link or other location ?
@thomashirtz
@thomashirtz 2 месяца назад
Why wasn't wjk in the autor's list at the end ? It seems like he gave you the initial spark to do that amazing content btw
@context_eidolon_music
@context_eidolon_music 2 месяца назад
Incredibly fast.
@freddiekarlbom
@freddiekarlbom 2 месяца назад
Thanks for this great interview!
@filipstankovic5307
@filipstankovic5307 2 месяца назад
Hej, video sam da si na svojoj stranici napisao da si zavrsio etf, pa me zanima da li si pod onim opisom mislio na to da si zavrsio smer za elektroniku? 😄 Pozdrav iz Srbije!
@TheAIEpiphany
@TheAIEpiphany 2 месяца назад
Aham elektroniku sam zavrsio :)
@navissivan
@navissivan 2 месяца назад
The loss part is still very confusing to me, why add the same loss twice with different weights, and as you said in the end going to zero? Anyone knows?