Тёмный
No video :(

Are GFlowNets the future of AI? 

Edward Hu
Подписаться 6 тыс.
Просмотров 27 тыс.
50% 1

Опубликовано:

 

5 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 155   
@alfellati
@alfellati 5 месяцев назад
Finally a real AI expert with a great track record, inventor and researcher, author of multiple research papers on AI. Great having you brother.
@lobiqpidol818
@lobiqpidol818 5 месяцев назад
I clicked the video thinking.... 😑 Oh look another "AI Expert". I invented LoRA.. Ok sir you have my attention. Hard to find real experts out here. Glad I clicked 😀.
@Fordtruck4sale
@Fordtruck4sale 5 месяцев назад
OpenAI SHOCKS entire industry by allowing employee to run own youtube channel 🤯🤯🤯
@edsonjr6972
@edsonjr6972 5 месяцев назад
Lmao same. Insta subscription
@ds920
@ds920 5 месяцев назад
Same here😂
@jakublala
@jakublala 5 месяцев назад
exactly my thoughts
@shadyworld1
@shadyworld1 5 месяцев назад
😂😂😂😂 ❤
@Nex_Addo
@Nex_Addo 5 месяцев назад
"Huh, what's this guy's deal? Well, GFlowNets sound cool, let's check him out" "I invented LoRA, Prof. Bengio was my PhD Advisor and I'm currently an OpenAI Research Scientist" "Ah, so 'the real' is this guy's deal...gotcha". Easy subscribe.
@pollomarzo
@pollomarzo 5 месяцев назад
Great content! Might want to get a mic for your next one, it'll make all the difference
@edwardjhu
@edwardjhu 5 месяцев назад
Thx for the feedback!
@jamesbrown99991
@jamesbrown99991 5 месяцев назад
I had trouble understanding the speech sometimes
@chaacbze
@chaacbze 5 месяцев назад
Thanks for your awesome content Mr Hu; agree about the sound.
@arjandhaliwal4962
@arjandhaliwal4962 5 месяцев назад
Fantastic video! No hype just the intuition behind current research. I love this type of content because it sets the foundational thinking required to make sense of research papers (which are normally opaque to someone who isn't deep in the field of study)
@nathangonzales-hess6569
@nathangonzales-hess6569 5 месяцев назад
That was great. Thanks for sharing. I really appreciate the simple style, no distracting animated plots or fancy editing. Look forward to more!
@automatalearninglab
@automatalearninglab 5 месяцев назад
Nice! Turn up your volume! Loved the topic! :)
@chitarito100
@chitarito100 5 месяцев назад
Hey, I’m glad you’re on RU-vid! It’s great to have an expert in the field break down complex topics and intuition and also explain real world applications not just theory. I think it would be nice if you could have some summary pointers overlaid in the video or on a summary slide with you overlaid on it. Different people have different learning styles and being able to read helps.
@diga4696
@diga4696 5 месяцев назад
Thank you for taking the time to produce great content!
@definty
@definty 5 месяцев назад
Oh wow. Awesome dude! Thanks for your work with LoRA and crating this video! Subbed. You sir are awesome!
@JL-zl6ot
@JL-zl6ot 5 месяцев назад
super interesting and clear presentation as always. Thank you so much for making this content available!
@alexander73848
@alexander73848 5 месяцев назад
People like you made RU-vid great! Thank you for contributing
@shibohao8930
@shibohao8930 2 месяца назад
Great video! Looking forward to your video explaining the relation between GFN and Max-Entropy RL
@arnoldpalmer-fv7pf
@arnoldpalmer-fv7pf 5 месяцев назад
So much groundbreaking research broken down into an easy to follow 7 minute video, I love it 🙏
@Seekerofknowledges
@Seekerofknowledges 5 месяцев назад
Thank you wholeheartedly.
@giyaseddinbayrak
@giyaseddinbayrak 5 месяцев назад
The kitchen background is awesome. Just a better mic is all you need. Love to hear abou other reasings in addition to your lab's works. Do you accept visiting researchers by any chance 😊
@explorer945
@explorer945 5 месяцев назад
+1
@christopherd.winnan8701
@christopherd.winnan8701 5 месяцев назад
Bill Mollison would be proud - Optimisation rather than maximisation is the best way to go.
@novantha1
@novantha1 5 месяцев назад
So, articulating this in a way that makes sense to me: In traditional machine learning it's pretty common to treat the weights and/or biases (where relevant) as learnable parameters, in the sense that they'll be optimized by the chain rule to produce the lowest loss answer. We are not limited exclusively to modifying weights, though. You can introduce a low rank matrix (I wonder who came up with that? 🤔) instead of modifying the base weights, but you could also treat the prompt, or sampler parameters as learnable parameters, too, leading to interesting and occasionally useful results. In this case, it's taking it a step further, and training a sampler model, and optimizing it for "good pathways", which is clear to see in the example of reasoning; If one "pathway" leads to the model just giving an answer, that's bad reasoning and won't reliably be correct (Maybe you should call up one of your elementary teachers and apologize for not understanding why they made you show your work!), and if the reasoning doesn't follow a logical cadence ("How many rabbits are there in New York in the spring?" "Well, the sides of a Tie fighter have been hypothesized to be Ion thrust surfaces and solar panels...") that could also be recognized as a poor fit by the sampler model. It'd be very interesting to see how this technique compares to something like training a dedicated resolver model to take the best reasoning steps of 1000+ generations (scaling with test-time compute), because this looks very similar to my eyes, though I might be missing something, as I'm not a true "AI researcher" or the like.
@NobleCaveman
@NobleCaveman 5 месяцев назад
Comment for the algorithm 😊 I'm just a layperson in this domain but still found the concept interesting and relatively clear to follow! Makes me wonder whether there is a finite population of reasoning pathways that can solve any logical problem, or if thats even a relevant question to ask. I'm not familiar with the last concept you introduced about maximum entropy.. sounds intresting! Thank you
@edwardjhu
@edwardjhu 5 месяцев назад
I think the intuition is that solving logical problems is basically searching for a valid reasoning chain (the space of reasoning chains is practically infinite IMO, tho). the intuition for max-entropy RL is that we want to explore all kinds of chains before committing to a particular subset.
@bionh
@bionh 5 месяцев назад
Hell Yes! BTW…Sir, please ask your colleagues at OpenAI to delete the customer noncompete clause, it’s dystopian and causes ai safety issues.
@sumitsp01
@sumitsp01 5 месяцев назад
Sir, please make more videos explaining fundamental concepts as well as latest breakthroughs in AI field. Your style of explanation is neat and pin point. Everyone will get benefited out of these videos.
@matnem
@matnem 5 месяцев назад
Thanks for this! I'd be cool if you could improve the audio quality for future videos
@shaunakgalvankar4502
@shaunakgalvankar4502 5 месяцев назад
Awesome!!I am about to start my PhD in cs and stuff like this is way motivating and inspiring…keeping sharing your research through videos…I literally read the Lora paper…then watched your video then went back and read the paper again and then came back and watched the video again
@anatolianlofi
@anatolianlofi 5 месяцев назад
Did you just casually explain what Yann LeCun vaguely hinted at? Where are all the cameras? Internet is a weird place, I'm watching the one-ring-to-rule-them-all being explained in under 10 minutes. So cool.
@Drone256
@Drone256 5 месяцев назад
That was excellent. Thank you for taking the time to share things like this.
@Jandodev
@Jandodev 5 месяцев назад
I found a way to do this at my company with just LLM's! I used LLM's to not only generate the process steps but also make a liminal/expansion map of the ideas themselves that could lead to those individual processing steps! The trick to make it run wasn't to fit to a "reward function" but to use the "reward function" for back propagation as the starting point to then determine what was needed from the user to retrieve and select the correct path!
@james-cucumber
@james-cucumber 5 месяцев назад
Thanks for including subtitles with this video. I’m pretty sure (but not certain given you may have access to better models than regular Joes) that it’s been manually edited or written from scratch. If so, extra thank you!
@RamphyRojas
@RamphyRojas 5 месяцев назад
Thank you for making it open to everyone. ❤❤❤ thank you from Venezuela!
@MaxenceFrenette
@MaxenceFrenette 5 месяцев назад
Love the style of video. Hope you can explain some of these more advances things in the future, it's really interesting.
@christophkogler6220
@christophkogler6220 5 месяцев назад
Great content, just wish the audio quality was a bit better.
@satyamtiwary6220
@satyamtiwary6220 5 месяцев назад
Would love it if you did some implement from scratch videos in pytorch.. teaching us how to innovate, by teaching us how to implement ideas from scratch such as LORA, and Energy based approaches
@ellielikesmath
@ellielikesmath 5 месяцев назад
i was trying to come up with something like this, in that i wanted to train a generator which would be the inverse of a classifier, and the classifier gave a score to how good a solution was drawn from some range. this looks miles and miles more sophisticated than what i was doing with tf and pytorch, but i definitely understand, at least on that level of abstraction, why such a development is necessary. i look forward to trying this, cheers.
@jarno_r
@jarno_r 5 месяцев назад
Such a great video, massively underrated channel! Liked & subbed, keep it up
@oryxchannel
@oryxchannel 5 месяцев назад
what could be a powerful companion to the context of this video is a voice-centric personal assistant. It has enough of a "Context profile" on you to judge inflection, tone synthesize inference and intent with SOTA voice recognition tools just laying around. this would at least morph the zero-shot/few shot dilemma in prompting every query.
@candrewlee14
@candrewlee14 5 месяцев назад
This was fantastic! Thank you, it’s great to hear from a real expert in this AI mega-hype cycle.
@vahanpetrosyan3448
@vahanpetrosyan3448 5 месяцев назад
Very very impressive video Edward! As a startup founder in the Gen AI space, it's hard to find youtube channels that go further into the future than what we see in the market. Such video can certainly change someone's product roadmap :))))
@Bbb78651
@Bbb78651 5 месяцев назад
Thank you so much for the video Edward. Its an inspo seeing you make videos and take off. Im currently a masters of science student in data science and Im always excited about NN architectures and new ML algos. Whenever easy, could you please share 1-2 tips for writing good research papers in ML? I recently started in a lab that does neuroscience-ML, and rlly want to make an impact there
@scapegoat079
@scapegoat079 5 месяцев назад
Very interesting stuff, I'd love to use this algorithm to see how it works. Just a little note, one thing that would help a lot is reducing the reverb/cleaning your audio a bit. I'm sure there are a lot of ai tools online that can do it quick. Thank you for making this educational content!
@kibrutemesgen1759
@kibrutemesgen1759 5 месяцев назад
Amazing explanation with clear explanation of the intuition. It would be great if you give us one more vedio on implementation or detail explanation of how Gflownet is used to enhance reasoning of LLMs (i.e the paper)
@Shaunmcdonogh-shaunsurfing
@Shaunmcdonogh-shaunsurfing 5 месяцев назад
Nice delivery. Thanks for bringing this to us.
@alexeialpaka
@alexeialpaka 5 месяцев назад
Just found your channel, I'm 100% impressed bro. Just a tip for the audio, if you don't have a good microphone, just use a free ai for speech enhancement, it works great. RU-vid ban the comments if you say the name of the AI, because it's also the name of a company apparently :/
@edwardjhu
@edwardjhu 5 месяцев назад
thx for the feedback! i actually use a Rode VideoMic. (i'll fire the guy who processed the audio.) can you DM me the tool you know for audio processing? you can leave a message on my website edwardjhu.com 🙏
@alexeialpaka
@alexeialpaka 5 месяцев назад
@@edwardjhu oh man, now I feel bad for the fired guy 😅 anyway, I'll send u the message
@kc12394
@kc12394 5 месяцев назад
Please keep doing these videos explaining how state of the art developments on AI work. Hell you explained it so well I'll listen to any ai concept you decide to explain. Subscribed!!
@zacharykosove9048
@zacharykosove9048 5 месяцев назад
Awesome video, it really sparks my interest in neural nets. Thanks for the broad overview of what works and what doesn't when training a model.
@ser1ification
@ser1ification 5 месяцев назад
Great video Edward. Just one request: Please use a better mic next time. Cheers!
@miikalewandowski7765
@miikalewandowski7765 5 месяцев назад
Brilliant topic and explanation! Looking forward for your deep dives and if you like some observations or insights you have made on your saga as a researcher 😊
@divingdabbler2035
@divingdabbler2035 5 месяцев назад
Please dive into the connected topics and foundations that have inspired this method in future videos. Can't wait!
@davidw8668
@davidw8668 5 месяцев назад
Good content, audio is hard to follow unfortunately.
@pramey100
@pramey100 5 месяцев назад
Finally someone teaching machine learning recipes from their actual kitchen.😅
@hoteny
@hoteny 5 месяцев назад
It is so cool that you made something so abstract to me that just magically works (lora). No idea how you even begin to understand the concept, let alone the code, of ai stuff.
@user-ut4zh3pw7l
@user-ut4zh3pw7l 4 месяца назад
waiting for those future videos. lets generalize ood
@justinpresent
@justinpresent 5 месяцев назад
thanks edward for the gentle intro!
@jayyoung7577
@jayyoung7577 5 месяцев назад
Wow❤ please post more deep dive videos. Thank you for sharing!
@MrErick1160
@MrErick1160 5 месяцев назад
I'm actually honoured to having ML classes taught by the inventor of Lora lmao. I haven't studied RL yet, would like to know how this works
@peace_and_blessings1111
@peace_and_blessings1111 5 месяцев назад
Sir, with all due respect ,please don't take my question in the wrong way. How can a total newbie get started in AI ? What are the books you would recommend reading and the things Id need to do? Sir I really appreciate the clarity and intensity of your incredible thought processes. Love you sir.
@_XoR_
@_XoR_ 5 месяцев назад
This is exactly what I was thinking about a month ago, the combination of gflownets for active chain of reasoning in llms and I was about to call it out that this is how Q* from openai works and make a paper.. And what do you know, a bit later people from openai start talking about it :)
@mikechung7316
@mikechung7316 5 месяцев назад
Amazing! thank you for making these videos!
@houbenbub
@houbenbub 4 месяца назад
Awesome video, thanks for making it :)
@arian386
@arian386 5 месяцев назад
Sweet RU-vid recommendations
@H1kari_1
@H1kari_1 5 месяцев назад
Hyper fresh AI news and concepts right from one of the best under-the-radar sources? Count me in, subbed.
@user-wr4yl7tx3w
@user-wr4yl7tx3w 5 месяцев назад
this is really informative. thanks for taking time to make such videos.
@anthonybernstein1626
@anthonybernstein1626 5 месяцев назад
Oh, this could be useful in a couple of areas. Subbed, thanks for the video!
@faizanjaved1443
@faizanjaved1443 5 месяцев назад
Here's a "Below are some new topics for you to discuss: - Grok-1 - Claude Opus - Gemini Ultra - Poe - Figure humanoid powered by OpenAI - Hugging Face - Sam Altman 7+ trillion - Google AI in iPhones - NVIDIA Disney robots Let me know if you need any help to choose the best topics for your content."
@BaKimura03
@BaKimura03 5 месяцев назад
Love the fact that you’re organic. Didn’t even lead with the big credential.. unlike some of these goofs who have done far less. 👏
@wevii9043
@wevii9043 5 месяцев назад
I believe you can bring a lot of wisdom to the table as a researcher, maybe you can talk about the future of AI based on your experience, like do you think AGI will get developed anytime soon, or some major problems the researchers and developers might run into that might hinder progress, or how soon will we reach some sort of ceiling?
@marcfruchtman9473
@marcfruchtman9473 5 месяцев назад
Hello Edward, Thank you for this presentation. [Subscribed]. GFlowNet sounds pretty good after listening to your explanation. To be clear, are you saying that it includes discovery to prevent it from getting stuck in a local maxima? Also, If possible, please check your microphone. It sounds like it is partially under water with some echo.
@PharaoTeti
@PharaoTeti 5 месяцев назад
Great scientist and communicator!
@aga5979
@aga5979 5 месяцев назад
The audio can be much clearer. I have a hard time hearing. Thank you for explaining.
@KW-jj9uy
@KW-jj9uy 5 месяцев назад
Hmm for robotics, that may be good for learning to control the robots. You only have a few samples of live data. Most RL relies on big datasets and is a no go
@charliesteiner2334
@charliesteiner2334 5 месяцев назад
Neat! If you've heard of this sampling as "quantilization," you might already know that the safety / "normalness" of the training distribution can get diluted if you do repeated sampling (as in search through a tree of possible sequences). Any plans to tackle this problem?
@ABG1788
@ABG1788 5 месяцев назад
Thanks, for the video!!!!!!!! I appreciate you
@srivatsasrinivas6277
@srivatsasrinivas6277 5 месяцев назад
Is the very high level idea that if you could generate a large number of diverse samples which minimize a loss function you are increasing the number of different local minima you're sampling from, thus reducing the chance that your trained model hit a dumb, or extremely deep minima?
@RonLWilson
@RonLWilson 5 месяцев назад
This looks very promising!
@RonLWilson
@RonLWilson 5 месяцев назад
BTW, I have been working on a graphical way of building Ontology models that I am calling UniML (Universal Modeling Language) that can model the GflowNets process quite readily much like OWL or RDF graphs, but even more graphically. Something like this might help develop these nets in that it is a way to better visualize them. I made a number of introductory videos that I uploaded on my RU-vid channel that describe them in more detail into their what's and why's plus a Padlet virtual corkboard that takes a bit of of a closer look at them. UniML can model NNs as well, or just treat them as functions (Chips).
@AK-ox3mv
@AK-ox3mv 5 месяцев назад
Exciting. Waiting for more videos✌️
@yiannishadjiyianni7737
@yiannishadjiyianni7737 5 месяцев назад
This might be dumb but I wonder how applicable can GFlowNets be in an object detection/instance segmentation context where the distribution created is based on the millions of images the model will train on and the "pathways" of the help the model be even more accurate
@faizanjaved1443
@faizanjaved1443 5 месяцев назад
Hey there! Can we talk about Q*, the AGI developed by Sam Altman? I'm excited to discuss this with you, it's one of the most interesting topics for me after Sora.
@shehrozeshahzad4363
@shehrozeshahzad4363 5 месяцев назад
Thanks a lot!
@jan3477
@jan3477 5 месяцев назад
Great video, but it was a bit too fast for me to keep up with the complex topic.
@saikalyan3966
@saikalyan3966 5 месяцев назад
Thank u brother, just a quick question, how do u get access to hardware? And how can a independent ai researcher access hardware and find interesting people in the domain, thanks🌹 mu pi was awesome too ❤
@explorer945
@explorer945 5 месяцев назад
thanks for the video. audio quality could be better
@minzi7216
@minzi7216 5 месяцев назад
I get lost right at 3:45. If we’re sampling proportional to the reward function do we still not end up with the “hacky” maxing sample occurring the most? Is the idea that at this point humans take over and evaluate the that batch and can easily disregard the “hacky” result while still getting other useful results? In LLM outputs we already get log probabilities. Would this be a superior way of selecting the best choice given the log probabilities?
@edwardjhu
@edwardjhu 5 месяцев назад
yes we will get a lot of the hacky samples but some samples might actually work (versus none will work if we maximize). also one can raise the temperature of the reward distribution to get more diverse samples.
@minzi7216
@minzi7216 5 месяцев назад
Interesting, thank you for answering my question! Looking forward to your future videos!
@Zeddd7
@Zeddd7 5 месяцев назад
Great video, but audio can be improved! I suggest you buy the DJI mics, Rode Wireless Pro, or HollyLark M2... Or find a different room to record in with less reflective surfaces. Those mics are small enough, good quality, and can clip on to your shirt.
@edwardjhu
@edwardjhu 5 месяцев назад
thx for the feedback! believe it or not this was recorded with a Rode VideoMic Go II 😅any tool recs for audio cleaning?
@dr.mikeybee
@dr.mikeybee 5 месяцев назад
Are you minimizing KL Divergence? I'm looking forward to the next video. :)
@edwardjhu
@edwardjhu 5 месяцев назад
not directly. this paper explains the difference and the advantage of GFlowNets: arxiv.org/pdf/2210.00580.pdf
@MrZidanegenius
@MrZidanegenius 5 месяцев назад
Lots of information. I hope this could be applied in robotics which suffers a lot due to magnitude of data it receives
@MrErick1160
@MrErick1160 5 месяцев назад
So if I understand this correctly, this is about optimising the data that we feed into the model right? Taking the right distribution that allows max likelihood for a certain output we want?
@NamTran-cc1ml
@NamTran-cc1ml 5 месяцев назад
GFlowNets + Mamba_ssm architecture, thought?
@AIShipped
@AIShipped 5 месяцев назад
This sounds exciting, however I still didnt really understand the problem this is trying to solve. Could it be used in rl like training a nn to play snake more sample efficient? I guess I am trying to say try to rephrase it like I am 10 or using a subject that I at some knowledge about. Thx!
@farhadkarimi
@farhadkarimi 5 месяцев назад
Lmao so this guys literally a genius😂
@farhadkarimi
@farhadkarimi 5 месяцев назад
Instantly subscribed
@vtrandal
@vtrandal 5 месяцев назад
Fantastic
@JohnRoodAMZ
@JohnRoodAMZ 5 месяцев назад
So the real question, is what haircut will you have in the next video? 🤣 Awesome content.🙏🏻
@seblund
@seblund 5 месяцев назад
Great video as per usual!
@joeyhandles
@joeyhandles 5 месяцев назад
Yeah I knew a few of those words
@GNARGNARHEAD
@GNARGNARHEAD 5 месяцев назад
that's cool, so if I'm understanding this correctly, you're training domain specific knowledge while maintaining the broader perspective of all the training data? that's cool. that's cool.
@GNARGNARHEAD
@GNARGNARHEAD 5 месяцев назад
posterior or potential reasoning chains, that's cool.
@fabiosilva9637
@fabiosilva9637 5 месяцев назад
Under 1M subs gang
@JumpDiffusion
@JumpDiffusion 5 месяцев назад
Good stuff
@legobuildingsrewiew7538
@legobuildingsrewiew7538 5 месяцев назад
I love the hair (d)evolution.
@user-lf4ir3mp2f
@user-lf4ir3mp2f 5 месяцев назад
great man
@cameronlamb153
@cameronlamb153 5 месяцев назад
Is this not just a system using a metropolis Monte Carlo reward function?
@JohnRoodAMZ
@JohnRoodAMZ 5 месяцев назад
Vote for the deeper dive into max entropy RL + path consistency
@catalystzerova
@catalystzerova 5 месяцев назад
oh shit it’s the lora guy!!!
@leonlysak4927
@leonlysak4927 5 месяцев назад
Yep, instantaneous sub.
Далее
Has Generative AI Already Peaked? - Computerphile
12:48
The Most Important Algorithm in Machine Learning
40:08
Просмотров 406 тыс.
ДО ВСТРЕЧИ НА РАЗГОНЕ
52:11
Просмотров 438 тыс.
The moment we stopped understanding AI [AlexNet]
17:38
Mamba Might Just Make LLMs 1000x Cheaper...
14:06
Просмотров 127 тыс.
The True Story of How GPT-2 Became Maximally Lewd
13:54