Тёмный

AI Inference is ABOUT to CHANGE!!! 

1littlecoder
Подписаться 74 тыс.
Просмотров 4,7 тыс.
50% 1

🔗 Links 🔗
Apple M4 Chip
www.apple.com/newsroom/2024/0...
MLX Community on Hugging Face
huggingface.co/mlx-community
Apple MLX Framework - github.com/ml-explore/mlx
❤️ If you want to support the channel ❤️
Support here:
Patreon - / 1littlecoder
Ko-Fi - ko-fi.com/1littlecoder
🧭 Follow me on 🧭
Twitter - / 1littlecoder
Linkedin - / amrrs

Наука

Опубликовано:

 

6 май 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 51   
@rodvik
@rodvik Месяц назад
Good luck to folks who are trapped in the Apple closed garden model. But they are not for me. I will stick with PC's I can rely on them not deliberately obsoleting them. Apples vision of AI is closed and censored. Thats up to them. Its not for me though.
@1littlecoder
@1littlecoder Месяц назад
Their MLX thingy is quite open, but don't expect that from Hardwares. so you are right there!
@mshonle
@mshonle Месяц назад
The Nvidia CUDA ecosystem is equally closed. There are no real open source projects that could be a drop in replacement for CUDA.
@1littlecoder
@1littlecoder Месяц назад
@@mshonle absolutely it's one of the reasons why I love seeing someone bold for accelerated computing
@fontende
@fontende Месяц назад
i agree, it's like car where you must use only Apple controlled and sold gasoline/benzine
@rodvik
@rodvik Месяц назад
@@mshonle good point.
@simplemanideas4719
@simplemanideas4719 Месяц назад
Which llms can be run on the most powerful M4 tablet?
@dissolvesoftware
@dissolvesoftware Месяц назад
Amazing
@1littlecoder
@1littlecoder Месяц назад
Thank you! Cheers!
@prashantsolanki007
@prashantsolanki007 Месяц назад
I really doubt the future of on-device LLM. I mean it nice to chat and stuff but if you think from quality and usage. The larger models are always going to be better. Unless the gap in performance between large and small models get reduced, which is highly unlikely. Also technologies like GROQs LPU are just too fast right now and most likely may get better. For sure there are going to be use cases for on-device LLMs but it's just you can always go on internet and search with almost minimal latency.
@1littlecoder
@1littlecoder Месяц назад
I'm definitely looking forward to see how much Groq can scale up! On-device LLM has its own place I guess. it's not going to replace API calls in building SaaS but for things like grammar correction, local RAG and even controlling the computer, I think it's got its thing!
@StuartJ
@StuartJ Месяц назад
I think we are going to see a hybrid, where you have small LLMs running locally, which draw on agent AIs hosted in the cloud.
@Raphy_Afk
@Raphy_Afk Месяц назад
You don't need a super smart LLM for everything. For voice command or video games a small and fast on device LLM will be preferable.
@Serifinity
@Serifinity Месяц назад
I just don't get the need for an on-device LLM, unless it's better than cloud based I don't think I would use it. Besides an on-device LLM would most likely be out of date and dependant on Apple/Google updating the LLM when they want. Compare that to something like Pi which is up to date on anything I ask about. Even something like Perplexity and the rumoured OpenAI search engine released on Thursday.
@Trevor0713
@Trevor0713 Месяц назад
it depends on the situsation. in my case, a fine-tuned llama 3 8B is good enough. Why people are looking at local LLM? because data privacy in some cases is a mandated requirement for business use by law, not just a nice to have.
@jmirodg7094
@jmirodg7094 Месяц назад
Could similar things be done with the latest Ryzen 8700G and sufficient system ram?
@serkhetreo2489
@serkhetreo2489 Месяц назад
Would m4 be placed in apple mini?
@GetzAI
@GetzAI Месяц назад
1:41 is this just marketing added or is it really a physical difference in the architecture?
@supercurioTube
@supercurioTube Месяц назад
Interesting take but fairly speculative since we don't know yet if the new NPU is helpful in any way. mlx and llama.cpp are running llms on the GPU or CPU as fallback, and in both cases the performance is mostly dependent on the memory bandwidth available. Typically, despite sharing the same RAM, the GPU memory bandwidth is a lot higher than the CPU cores combined. What about the NPU memory bandwidth on the M4? It wasn't mentioned in the keynote, but it's what matters most for llms inference. BTW they're comparing to Intel Core Ultra's NPU, since Qualcomm's platform is not released yet.
@ps3guy22
@ps3guy22 Месяц назад
So the Base M4 chip has 38 TOPs (Neural engine only). Nice
@Custodian123
@Custodian123 Месяц назад
Just not interested in being trapped with Apple 😂.
@TheGalacticIndian
@TheGalacticIndian Месяц назад
Whether on the device or in the cloud, these models are capable of compressing data so intensely that they are the perfect spy tool for many applications. Just think about it - if the average size of an LLM model is a few gigabytes, and it contains most of the knowledge humanity has created in the known past, what is it for the model to compress one person's data. These will be bits in size that can be indistinctly attached to any file, or encoded in a file so that they are invisible. And sent over an electrical network that has long been connected to the Internet. And all this still without the IoT running 24/7. No place to hide😉
@mvasa2582
@mvasa2582 Месяц назад
Certainly will dominate the Edge and could be a challenge for inferencing. This is a bigger challenge for AMD as opposed to Nvidia
@mohamedhatem325
@mohamedhatem325 Месяц назад
very good video bro
@RickySupriyadi
@RickySupriyadi Месяц назад
wait until apple said they invent LLM i once read that somewhere in RU-vid comment
@mvasa2582
@mvasa2582 Месяц назад
You hit the nail about Intel and AMD not capturing the market potential. Unfortunate but true!
@vickeythegamer7527
@vickeythegamer7527 Месяц назад
still snapdrgon x elite with 45 tops on base with higher effeciency of less than 4 wattt
@RunForPeace-hk1cu
@RunForPeace-hk1cu 29 дней назад
it's not even out yet LOL
@vickeythegamer7527
@vickeythegamer7527 29 дней назад
@@RunForPeace-hk1cu same for m4 ( you are just ignoring the fact that apple launched it hurry because they were scared )
@adriangpuiu
@adriangpuiu Месяц назад
Apple has had AI chips before all this AI thing came out , but, apple did not say anything about it , they just minded their business. now is coming out and vocal cause they know they might be left behind the curtain and they need to earn money
@kostik
@kostik Месяц назад
WWDC is in June
@fouziaanjums6475
@fouziaanjums6475 Месяц назад
Hey @1littlecoder could you please make a video on Idefics2 model from Hugging Face , also its fine tuning for custom dataset, I would be really glad😊
@IdPreferNot1
@IdPreferNot1 Месяц назад
Have you never seen apple PR before?????
@Spreadshotstudios
@Spreadshotstudios Месяц назад
Am I about to live in a world where I actually have to buy Apple products to be on bleeding edge of graphics? I'm not convinced yet. Companies like groq etc can make accelerated cards but the reason nvidia is on top of every ML engineers mind is due to CUDA unless I am missing something?
@Macorelppa
@Macorelppa Месяц назад
Ain't No competition for SIRI.
@paulmuriithi9195
@paulmuriithi9195 Месяц назад
M4 is nothing compared to the upcoming TSMC'S A16 N2 chip which is based on GAA tech-what intel calls ribbon 20A and 18A tech. the A16 TSMC chip will head strait to apples devices and shall be capable of running gpt-4 old version on device. the next apple imac and desktop will be insane.
@Raphy_Afk
@Raphy_Afk Месяц назад
I will never buy an apple product, but it could be good for the industry indeed. As for the market I think, on a product level, they are competing or rather preventing the rise of AMD who has invested a lot in its APUs, because M chips are only that, CPUs with integrated graphics. There is no chip dedicated to AI on consumer hardware yet. If AMD manages to make a chip that reach the same performance as apple's it could be a better option as the user could be able to buy and upgrade the ram as he pleases.
@EternalKernel
@EternalKernel Месяц назад
i mean they are apple. they will do everything they can to make mlx incompatible with M1 hardware within the next month. Apple is preditory in the worst ways. I mean it kind of puzzles me because apples whole entire thing is "lets make something expensive, and white." Are they going to try to make things even more expensive then Nvidia?.. no whiter and smoother I guess.. or no more of an "exclusive club" Listen Nvidia is NOT great but apple is way worse.
@metobabba
@metobabba Месяц назад
you are mispronouncing "only"
@1littlecoder
@1littlecoder Месяц назад
How should I say?
@metobabba
@metobabba Месяц назад
​@@1littlecoder "own-lee". dont get me wrong. i just noticed it. english is not my native language too.
Далее
Apple's Silicon Magic Is Over!
17:33
Просмотров 943 тыс.
Artificial Einstein: Did AI just do the impossible?
19:40
Новый дом для пернатого
00:59
Просмотров 638 тыс.
host ALL your AI locally
24:20
Просмотров 709 тыс.
RISC-V 2024 Update: RISE, AI Accelerators & More
14:03
What Is an AI Anyway? | Mustafa Suleyman | TED
22:02
Programming a Guessing Game in Rust!
13:22
Просмотров 104 тыс.
CrewAI Tutorial: Complete Crash Course for Beginners
1:05:43
Калькулятор в iPadOS 18 ➕
0:38
Просмотров 130 тыс.
ПЕРЕХОЖУ НА НОВЫЙ ТЕЛЕФОН!
10:34