Тёмный
No video :(

Machine Learning with Flax - From Zero to Hero 

Aleksa Gordić - The AI Epiphany
Подписаться 54 тыс.
Просмотров 18 тыс.
50% 1

Опубликовано:

 

29 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 44   
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Notebook: github.com/gordicaleksa/get-started-with-JAX/blob/main/Tutorial_4_Flax_Zero2Hero_Colab.ipynb My referral link: deepnote.com/referral?token=823d18856ad5 (you get 20h of "Pro machine" for free!). And you support the channel. ❤️ Again, I'll never promote anything I don't find valuable myself - so I strongly suggest you check them out. ❤️
@anuragranjak3829
@anuragranjak3829 2 года назад
this series is really amazing man more jax please
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Thanks man! Yup, I'm going to upload Haiku very soon and then in a couple of months I'll have some more coming. ;)
@sophiawright3421
@sophiawright3421 Год назад
@@TheAIEpiphany extremely useful series on JAX!! when can we expect the next video and more on JAX to drop :)
@Queracus
@Queracus Год назад
@@TheAIEpiphany hey. first of all thank you for all the Jax and Flax videos. Still basically the only good ones on the whole web. Is Haiku video still planned?
@jonathanstreater
@jonathanstreater 4 месяца назад
@@TheAIEpiphany Penzai is out now and it also looks really cool
@mikesmith853
@mikesmith853 2 года назад
Awesome stuff. I can't wait for the Haiku video!
@chenweicui7887
@chenweicui7887 2 года назад
I just couldn't allow myself to enjoy the content without paying my gratitude; Donation made😄
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Oohhh thanks a lot! 😄😄😄 Appreciate it! 🙏 Glad you found it useful!
@Murphyalex
@Murphyalex 2 года назад
At 01:10:58 I thought you said "ingradients" and thought it was a great pun, but then I listened again and it was just "ingredients" (which did make more sense, after all). I think I will save this video to come back to reference as there was a lot of info given, all of it great. I'm in awe and wish I could create the kind of content you do.
@jayrothenberger4354
@jayrothenberger4354 2 года назад
Excellent video. I appreciate all the detail.
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Thanks man 🙏
@adityakane5669
@adityakane5669 2 года назад
Looking forward to the JAX-framework series!
@TheAIEpiphany
@TheAIEpiphany 2 года назад
This one is everything you need to get u started with Flax (+ the 3 JAX videos). I'll cover Haiku next :)
@theneuralmancer
@theneuralmancer 2 года назад
@@TheAIEpiphany Yes, please! I've run into so many problems with Haiku even though I'm more of a fan of its design principles. It doesn't help that the community seems much smaller. Could I request for something along the lines of an ActorCritic class where - the "forward" outputs a tuple of the actor and critics outputs - the actor and critic share a layer I ran into massive problems trying to get that off the ground
@adityakane5669
@adityakane5669 2 года назад
@@TheAIEpiphany Looking forward to that as well. Big fan here!
@vivekpadman5248
@vivekpadman5248 5 месяцев назад
Your haiku video never came out sir 😅, thanks for this series btw
@navubot
@navubot 2 года назад
loved it
@directorans
@directorans 2 года назад
great video!
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Ty 🙏
@tempdeltavalue
@tempdeltavalue 2 года назад
23:05 why dividing by 2 ? shouldn't be where n_samples?
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Let me try and clarify. First of all - It doesn't matter that much, the optimal point in the param space is invariant to loss scaling by a positive const (or in general, in theory, you could apply an arbitrary strictly monotonic function - like log), and the optimization process should be fairly robust to "whatever" positive constant we put in there. Div by 2 is there simply because of a convention (d/dx(x^2) = 2*x and that's why you'll see div by 2 in the literature the analytical expression looks cleaner, you end up with only x instead of 2*x). Even in the multi dim case where inner would give you L = x1^2 + x2^2 + ... when you do dL/dx1 you end up with 2*x1, and similarly for other dimensions (x2, etc.).
@quantumjun
@quantumjun 2 года назад
Thank you so much for the video!! Would you like to share what kind of software you are using to improve your productivity :P
@TheAIEpiphany
@TheAIEpiphany 2 года назад
You're welcome! Well I use OneNote, Keep for notes. I set alarms on my phone not to forget stuff (although I just started doing that recently). Other than that, I just minimize distractions and try to have deep work sessions very often. :)
@quantumjun
@quantumjun 2 года назад
@@TheAIEpiphany Thank you for your reply😊
@demoysegment5488
@demoysegment5488 Год назад
Nice tutorial ! But I still have a little question: what if we have multiple dropouts in a model? I noticed that the params are organized in dicts to be passed into init or apply, then what are the names of the multiple dropouts? I dont feel like they are using the same rngs. How to distinguish them?
@SYBIOTE
@SYBIOTE 2 года назад
New framework? Nice
@TheAIEpiphany
@TheAIEpiphany 2 года назад
It's been around for a while but it's starting to get more traction (same as Haiku! I'm going to cover it next up)
@varunsai9736
@varunsai9736 2 года назад
Awesome
@TheAIEpiphany
@TheAIEpiphany 2 года назад
🙏
@chandanpradhan25
@chandanpradhan25 2 года назад
Does FLAX or HAIKU support complex data? I mean can we give a complex dataset as input to the NN and get complex output?
@TheAIEpiphany
@TheAIEpiphany 2 года назад
Of course! DeepMind and Google Research are building everything in Haiku/Flax - that should answer your question. 😅 (TFDS is usually used for data loading itself)
@chandanpradhan25
@chandanpradhan25 2 года назад
@@TheAIEpiphany Thank you!!
@willbrenton8482
@willbrenton8482 2 года назад
Could anyone help out/point to some help on using JAX for timeseries modeling. Very difficult to use the common sliding window method when generating features. Is there a fix or does it need rethought entirely?
@firqaaqilanoorarasyi2570
@firqaaqilanoorarasyi2570 2 года назад
Can you provide the code for CNN that have nn.Dropout( )? I have follow your suggestions in the video but still not work
@oleksiygrechnyev7120
@oleksiygrechnyev7120 2 года назад
Currently (19 Mar 2022) your notebook fails at "import flax" stage (on colab, didn't try other options). I could only run it by replacing stuff with the CPU-only !pip install jax jaxlib flax
@oleksiygrechnyev7120
@oleksiygrechnyev7120 2 года назад
Update: Deepnote indeed works. Interesting.
@mariolinovalencia7776
@mariolinovalencia7776 2 года назад
Can you do one haiku please?
@MrEnyecz
@MrEnyecz Год назад
Thanks a lot. I must say, at the first glance, this FLAX/JAX thing seems to be even a bigger crap than pytorch is (which is a big thing to me). It's hard to understand why people cannot do things such simply like with keras. Why they need to do all these dance with a simple training? With keras, you just throw the layers on the top of each other, do a model.compile() + model.fit(), and ready. If you want to do tricks, you can, but you don't need it. With FLAX, generate this variable, add that param. Param has one syntax, variable has the other. Keep in mind that this layer needs this variable that needs another. Just doing a synchronized batch normalization seems to be rocket science... Anyway thx. again, hopefully, I will get used to this thing too.
@ScottLeGrand
@ScottLeGrand Год назад
Because the people who design these frameworks are purist Computer Scientists and Mathematicians with little to no consumer product experience. So they design what they want and your job is to praise them for doing so because they are very very very easily triggered and really ought to resolve their academic PTSD, but, I digress. That said, Jax is the least crappy just for treating accelerators like GPUs and TPUs as first class citizens instead of a sad afterthought like everything else. Both Andrej Karpathy and Francois Chollet have demonstrated how to simplify as you desire. And yet they don't get the backing to build a new framework that way. Maybe John Carmack can do a thing or two about this.
@shubhampatel6908
@shubhampatel6908 7 месяцев назад
Keras is pretty good for rapid prototyping, but its not that good when it comes to configuring you model. Tbh, I have seen people don't even know how 3d matrix multiplication is done and using keras and building models. It works fine til you get nans, or get stuck at a local minima loss value or get other errors. That time you don't know what to do in Keras cause you never visualized your model or cared about hyper-parameters and just created a model with any random architecture you liked. pytorch on the other hands is very close to mathematical side so you realize what is actually happening, compare to keras model which are just `feed the data and get the output whatever it is`. Also another part, in research papers every tiny bit of edge matters in performance, this is where things like JAX shines.
@MrEnyecz
@MrEnyecz 7 месяцев назад
​@@shubhampatel6908 "Also another part, in research papers every tiny bit of edge matters in performance, this is where things like JAX shines." you said. That is quite a bold statement, more papers use torch than JAX.
@shubhampatel6908
@shubhampatel6908 7 месяцев назад
@@MrEnyecz your srgument doesn't make much sense. When torch was new, more papers used tensorflow than torch lol. Also I didn't say every research, I said the research where every bit of edge matters.
Далее
ConvNeXt: A ConvNet for the 2020s | Paper Explained
40:08
Useful gadget for styling hair 💖🤩
00:20
Просмотров 1,8 млн
НЕ ИГРАЙ В ЭТУ ИГРУ! 😂 #Shorts
00:28
Просмотров 303 тыс.
Demo: JAX, Flax and Gemma
8:12
Просмотров 3,5 тыс.
Simon Pressler: Getting started with JAX
29:49
Просмотров 1,9 тыс.
Why Does Diffusion Work Better than Auto-Regression?
20:18
This is why Deep Learning is really weird.
2:06:38
Просмотров 384 тыс.
NeurIPS 2020: JAX Ecosystem Meetup
1:02:15
Просмотров 27 тыс.