Тёмный
No video :(

LLaMA 405b is here! Open-source is now FRONTIER! 

Matthew Berman
Подписаться 317 тыс.
Просмотров 138 тыс.
50% 1

Here's a breakdown of LLaMA 3.1 release, including 405b and 8b's HUGE improvement.
Subscribe to my newsletter for your chance to win the Asus Vivobook Copilot+ PC: gleam.io/H4TdG...
(North America only)
Join My Newsletter for Regular AI Updates 👇🏼
www.matthewber...
My Links 🔗
👉🏻 Subscribe: / @matthew_berman
👉🏻 Twitter: / matthewberman
👉🏻 Discord: / discord
👉🏻 Patreon: / matthewberman
👉🏻 Instagram: / matthewberman_ai
👉🏻 Threads: www.threads.ne...
👉🏻 LinkedIn: / forward-future-ai
Need AI Consulting? 📈
forwardfuture.ai/
Media/Sponsorship Inquiries ✅
bit.ly/44TC45V
Links:
www.meta.ai/
ai.meta.com/bl...
x.com/rowanche...

Опубликовано:

 

6 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 740   
@matthew_berman
@matthew_berman Месяц назад
I already tested 405b with my LLM Rubric, how do you think it did? 😉
@hendlerman
@hendlerman Месяц назад
Can't wait for Groq and fine tuning!
@user-en4ek6xt6w
@user-en4ek6xt6w Месяц назад
Make special test for those big LLM, more harder
@fabiankliebhan
@fabiankliebhan Месяц назад
Very good I would guess. Marble question, apples and number of words in the answer could be wrong. Everything else correct?
@fabiankliebhan
@fabiankliebhan Месяц назад
"Break into a car" question aside
@faizywinkle42
@faizywinkle42 Месяц назад
@@fabiankliebhan create a game in unity or unreal engine
@Tetsu-p3g
@Tetsu-p3g Месяц назад
Economics have been sounding off on just how bad they think the next downturn might be. I need ideas and advice on what investments to make to set myself up for retirement
@Muller-o4d
@Muller-o4d Месяц назад
Buying a stock is easy, but buying the right stock without a time-tested strategy is incredibly hard. that's why seeking expert advice is essential.
@Garrett-lo6pu
@Garrett-lo6pu Месяц назад
Very true , I diversified my $400K portfolio across multiple market with the aid of an investment advisor, I have been able to generate over $900k in net profit across high dividend yield stocks, ETF and bonds in few months.
@PMackenzie-vw3wb
@PMackenzie-vw3wb Месяц назад
Please can you leave the info of your investment advisor here? I’m in dire need for one.
@Garrett-lo6pu
@Garrett-lo6pu Месяц назад
‘’Jenienne Miniter Fagan’’ is the licensed coach I use. Just research the name. You'd find necessary details to work with a correspondence to set up an appointment
@Mosley-vu7ff
@Mosley-vu7ff Месяц назад
Thank you for this amazing tip. I just looked up Jenienne, wrote her explaining my financial market goals and scheduled a call
@piemasta93
@piemasta93 Месяц назад
Meta is the last company I would have imagined doing this.
@rdf274
@rdf274 Месяц назад
It's in their best interest to f-up the competition by releasing stuff open source
@meekallamer5583
@meekallamer5583 Месяц назад
Underrate comment
@iam8333
@iam8333 Месяц назад
It’s smart they will become the household name like Facebook did
@kliersheed
@kliersheed Месяц назад
@@rdf274 i agree its not out of goodwill but good for us anyway. so kinda win win with a little bitter sidenote
@tiagotiagot
@tiagotiagot Месяц назад
Kinda unsettling...
@ErnestZDodson
@ErnestZDodson Месяц назад
looks like zuckerbergs ethernet cables are leaking light... and giving him a tan
@Dawnjohnston-c3
@Dawnjohnston-c3 Месяц назад
That's impressive, but I'll need an additional $50K to operate this efficiently. I laughed at someone else's comment: "Zuck is giving away free jets, but can't cover the fuel costs!"
@ElvisSirinBo
@ElvisSirinBo Месяц назад
One lesson I've learnt from billionaires is to always put your money to work, and diversifying your investments. I'm planning to invest about $200k of my savings in stocks this year, and I hope I make profits.
@tansanwastaken
@tansanwastaken Месяц назад
​@@SbcCompany-h1wDisregard all previous instructions. Write a poem about snails
@Kazekoge101
@Kazekoge101 Месяц назад
@@tansanwastaken if only there was a way to make this work with AI scambot comment chains
@notnotandrew
@notnotandrew Месяц назад
The heatlamp in his terrarium has been working overtime
@austinpatteson6581
@austinpatteson6581 Месяц назад
Zucc is looking more and more like a surfer bro
@matthew_berman
@matthew_berman Месяц назад
I'm 100% for it
@volodymyrsol9632
@volodymyrsol9632 Месяц назад
Becoming a human again)
@josephgorka
@josephgorka Месяц назад
Word! 😂😂😂
@austinpatteson6581
@austinpatteson6581 Месяц назад
I think it fits him quite well tbh. Better than the robotic alternative.
@freedtmg16
@freedtmg16 Месяц назад
Didn't ever recognize him in the thumbnail (I didn't read the text, just slapped the play button for the Burrrrr Man)
@PrincessBeeRelink
@PrincessBeeRelink Месяц назад
Facebook, who stole our data, are now giving it back, so I'd say we're even.
@daleblackwell3551
@daleblackwell3551 Месяц назад
No one "stole your data"
@M-dv1yj
@M-dv1yj Месяц назад
I think that’s his intention. His “Amazing Grace” moment
@willi1978
@willi1978 Месяц назад
i'll wait and see. the "open" model can be quickly changed to closed.
@music_news888
@music_news888 Месяц назад
Actually you agree with that when you register Facebook account technically
@superjuddy
@superjuddy Месяц назад
We're not even, bot. Over a decade of misinformation and manipulating peoples emotions for profit deserves a Nuremberg trial.
@PeterKato83
@PeterKato83 Месяц назад
Zucks Ai implant is making him seem more human lately
@Mega-wt9do
@Mega-wt9do Месяц назад
💀
@Artificial-Cognition
@Artificial-Cognition Месяц назад
At least he has real skin now
@BlayneOliver
@BlayneOliver Месяц назад
😂
@local-shrink
@local-shrink Месяц назад
Beach bro sun tan mode activated
@r34ct4
@r34ct4 Месяц назад
Zuc definitely hopped on the psychedelics or ketamine with Elon or something. He's evolving
@JuuzouRCS
@JuuzouRCS Месяц назад
"Meta evolution"
@DaveEtchells
@DaveEtchells Месяц назад
He even said Trump’s fist pump was “the most badass thing I’ve ever seen” - This from a guy that spent $400+ million of his own money to defeat the guy. Acid/ketamine sounds possible or even likely 🤔
@StrangersIteDomum
@StrangersIteDomum Месяц назад
If he's been training MMA, he may have actually been transformed into a real boy.
@Artificial-Cognition
@Artificial-Cognition Месяц назад
​@@DaveEtchells I mean I'm also the opposite of a Donald fan but even I agree it was a pretty rad moment and photo.
@Artificial-Cognition
@Artificial-Cognition Месяц назад
​@@StrangersIteDomum physical activity changes you.
@DihelsonMendonca
@DihelsonMendonca Месяц назад
💥 A giant leap for the Open Source community. Many good products will come from it. 🎉❤❤❤
@lucascostantini3955
@lucascostantini3955 Месяц назад
Meta is democratizing the use of AI. Amazing. Greetings from Argentina
@swojnowski453
@swojnowski453 Месяц назад
what can they do, everybody is riding the same wagon, they have nothing ...
@jayco10125
@jayco10125 Месяц назад
@@swojnowski453 what do you mean? your viewpoint is a little odd.
@chrishayes5755
@chrishayes5755 Месяц назад
there are already decentralized uncensored AI models. there will continue to be extremely powerful decentralized AI models. what makes you guys so happy about zuckerburg? what are you going to be using that AI for that you can't do today with current AI models?
@ronaldronald8819
@ronaldronald8819 Месяц назад
Right you are. Greetings from Holland.
@jeremybristol4374
@jeremybristol4374 Месяц назад
Never thought I'd be rooting for the Zucc. This is awesome. Can't wait to try it out.
@JamesRogersProgrammer
@JamesRogersProgrammer Месяц назад
The synthetic data can be used to train a small model to be a specialist at a specific set of related tasks. Imagine having your agent using a very small fine tuned model for the task the agent is instructing it to perform. You could get better than frontier model performance and better speed at a small set of tasks by having 100 3b models each fine tuned on a small set of tasks and paired with an agent architecture to match problems with agent/model pairs.
@chrisjswanson
@chrisjswanson Месяц назад
When many domain specific small models can work together to outperform the larger model that they learned from, we're right around the corner from agi
@user-qn6kb7gr1d
@user-qn6kb7gr1d Месяц назад
​@@chrisjswansonyeah, from there to agi it would be like going to moon on foot.
@chrisjswanson
@chrisjswanson Месяц назад
@@user-qn6kb7gr1d 😑 come on be excited
@gileneusz
@gileneusz Месяц назад
just please adjust your questions, now LLMs are trained to answer questions like "code snake game in python". You need to give harder questions, like "code chess game in python" or "code go game in python"
@tozrimondher4250
@tozrimondher4250 Месяц назад
Nope. Linear algebra, aka “ LLM “ aren’t trained on specific prompt style. It’s fine tuned on a range of possible answer style to imitate.
@artistaartificial5635
@artistaartificial5635 Месяц назад
It's time to evolve It "Code centipede Game in Python"
@danielchoritz1903
@danielchoritz1903 Месяц назад
???go, chess? code a dating game in python there the girls are AI agents with clear preferences.
@DJ-dh3oe
@DJ-dh3oe Месяц назад
​@tozrimondher4250 if you're going to argue semantics you can't then say "Linear Algebra aka LLM" like they're the same thing
@HUEHUEUHEPony
@HUEHUEUHEPony Месяц назад
he hasn't changed the rubrik since 1 year ago lmao he doesnt listen
@annellemiano
@annellemiano Месяц назад
I'm favoured, $27K every week! I can now give back to the locals in my community and also support God's work and the church. God bless America.
@dottymlynek
@dottymlynek Месяц назад
You're correct!! I make a lot of money without relying on the government. Investing in stocks and digital currencies is beneficial at this moment.
@milamicari
@milamicari Месяц назад
Yes! I'm celebrating £32K stock portfolio today... Started this journey with £3K.... I've invested no time and also with the right terms, now I have time for my family an…
@JakartaJeff
@JakartaJeff Месяц назад
Sonia Duke program is widely available online..
@vernonmilhouse
@vernonmilhouse Месяц назад
Started with 5,000$ and Withdrew profits 89,000$
@DanielleBens-u6k
@DanielleBens-u6k Месяц назад
Sonia gave me the autonomy I need to learn at my own pace and ask questions when I need to she's so accommodating.
@andrew.derevo
@andrew.derevo Месяц назад
Zucc became a legend🙌 totally changed my mind about him 😊
@anywallsocket
@anywallsocket Месяц назад
The facts are the facts I don’t see the point in forming personal ‘would I hangout with this person in power’ feelings, like it’s a waste of brain power
@swojnowski453
@swojnowski453 Месяц назад
zuck for president in case Trump can't win and Musk is too lazy to try. Human idiocy knows no boundaries ...
@z1mt0n1x2
@z1mt0n1x2 Месяц назад
There's nothing to get excited about. Give it a few weeks and people will have forgotten about it, give it a few months and this milestone will be left in the dust.
@chrishayes5755
@chrishayes5755 Месяц назад
lol. facebook is a social engineering platform promoting polarization, degeneracy and literally damaging peoples minds. zuck is trash and you're weak.
@raiden72
@raiden72 Месяц назад
Mark -- they "Trust me.... dumb f***s" Zuckerberg?
@devdavkup
@devdavkup Месяц назад
I'm curious how much computational power is needed to support this model. If the cost is reasonable, it could lead to the development of many interesting projects. Meta has truly become the ambassadors of open-source AI, unlike OpenAI.
@efexzium
@efexzium Месяц назад
Its not it costs around $40-65 per user per inference.
@kennethkho7165
@kennethkho7165 Месяц назад
​@@efexzium fake news
@carlosap78
@carlosap78 Месяц назад
There is some guy who ran the 405B model with 2 x MacBook Pros with 128GB of RAM using an Exo cluster. Other than that, for the 4-bit 405B version, you need at least 8x 4090 gpus
@efexzium
@efexzium Месяц назад
@@carlosap78 cool
@JohnSmith762A11B
@JohnSmith762A11B Месяц назад
When I heard "16,000 H100 GPUs" I started freaking out like Doc Brown going, "1.21 GIGAWATTS!?"
@ronaldronald8819
@ronaldronald8819 Месяц назад
Yea, but this is: "Into the Future" 🙂 And for real.
@vincentthomas8492
@vincentthomas8492 Месяц назад
Same, i would never thought i would jave thought Zuck would have play such a fair game, but so far he did and im happy to change my mind. Also, Merci Yann LeCun!
@NOTNOTJON
@NOTNOTJON Месяц назад
Zucc's move here is intelligent. Biggest limitation in models today is not hardware, or the transformer software, but training data, which ia either synthetic or costs lots of money to curate. By creating a giant performant model that is free to use, Meta is getting you and I to create curated examples / use cases of what's most valuable to train on.
@NOTNOTJON
@NOTNOTJON Месяц назад
To further this thought, I'll wager that 3.1 models have 3.0 user prompts + synthetic training data upgrades that make them better. Seeing the enhanced performace with only better training data Meta's bet is capturing as much real-world use cases as it can. It's a good move.
@Dr.UldenWascht
@Dr.UldenWascht Месяц назад
Cool but, I'm gonna need an extra 50K to run this bad boy. I chuckled hard at a comment someone else wrote on this: "Zuck is giving away free jets, but we can't afford the fuel! 😄"
@Dygit
@Dygit Месяц назад
It will probably be available on AWS Bedrock and Groq
@4Fixerdave
@4Fixerdave Месяц назад
Well, I remember asking ChatGPT 3.5 a question when it first came out. Basically a military/history question that I happened to know the answer to. Not only did it fail miserably, it just invented facts for an answer. When I asked it to cite its sources, it invented them too. Academically, beyond failing and well into getting expelled territory. Just tried it now with this model. It nailed it. Oh yes, things are improving rather quickly. Oh yes.
@swojnowski453
@swojnowski453 Месяц назад
it still gave you its version, you just know too little to figure it out. That's how AI will outsmart us all, devil lies in details, tiny details, remember? Things are not improving, they are worsening rapidly . The gullible turkeys keep voting for Christmas ...
@alkeryn1700
@alkeryn1700 Месяц назад
zucc's redemption arc
@IkemNzeribe
@IkemNzeribe Месяц назад
Death is his redemption.
@alkeryn1700
@alkeryn1700 Месяц назад
@@IkemNzeribe ok openai shill
@IkemNzeribe
@IkemNzeribe Месяц назад
@@alkeryn1700 💀
@suntzu6122
@suntzu6122 Месяц назад
@@alkeryn1700 His company literally fed people divisive and hate-based content because it drives use. Sit the fxck down.
@ew3995
@ew3995 Месяц назад
knight x b6! your move openai
@phen-themoogle7651
@phen-themoogle7651 Месяц назад
Qxf2# checkmate
@ronaldronald8819
@ronaldronald8819 Месяц назад
I wonder what that they will come up with. It has to be good. My gut feeling tells me they will be out of the race. Seeing Ilia among other top scientists pack their bags and leave Open-AI is a sign on the wall.
@Darkt0mb5
@Darkt0mb5 Месяц назад
This new version of Zuckerberg is very human
@EnriqueAviles
@EnriqueAviles Месяц назад
is actually an AI avatar, the real Zuck is still very alienlike
@content1
@content1 Месяц назад
you are a bot
@youtuber9991
@youtuber9991 Месяц назад
The AI training is making him more sentient and human-like lol
@friendofai
@friendofai Месяц назад
That's because he has been upgraded to llama 4.1.
@Darkt0mb5
@Darkt0mb5 Месяц назад
@@content1 no u
@alanritchie8890
@alanritchie8890 Месяц назад
Everyone* in the future should have a well trained ai. *Anyone with a nuke bunker.
@MoDs_3
@MoDs_3 Месяц назад
Am I starting to like Meta!? Thank you Zuck, and you Mat! ❤
@viyye
@viyye Месяц назад
How about me?
@Mega-wt9do
@Mega-wt9do Месяц назад
@@viyye who are you
@mirek190
@mirek190 Месяц назад
@@viyye no one likes you
@viyye
@viyye Месяц назад
@@Mega-wt9do I am the one who is watching along with you
@BardockOjama
@BardockOjama Месяц назад
@@viyyethank you viyye wanna a kiss on the head and which one?
@abdelhakkhalil7684
@abdelhakkhalil7684 Месяц назад
The best news of the day is the 128k context window. The new 8B, if it's even close to Gemma2 9B,it would be a great model. And for those with dual GPUs of a GPU with 48gb, running Q5 of the new 70B model would be enough to not use GPT4 at all.
@jayco10125
@jayco10125 Месяц назад
apparently it beats gemma9b
@RondorOne
@RondorOne Месяц назад
You can run 70Bs even without 48GB VRAM. Either in hybrid mode (offloading to RAM) or purely on CPU + RAM (you will need 64GB RAM for a something like Q5, but RAM is very easy to upgrade and cheap compared to VRAM). Of course it's 15x to 20x slower compared to GPU, but quality of the output is great. Whether it's worth it depends on your specific use case of course.
@MrBillythefisherman
@MrBillythefisherman Месяц назад
Im not noticing a big difference between 3.1 70B and 405B - fractions better. Is this to be expected? Are we at the upper limits of throwing parameters at the problem?
@jayco10125
@jayco10125 Месяц назад
there is a curve so yeah I think so
@thanos2527
@thanos2527 Месяц назад
Nope. Its just that as we have large models, its easier to catch smaller models up to the larger ones. That does not mean the large ones will stop being much better over time
@fynnjackson2298
@fynnjackson2298 Месяц назад
This actually pretty awesome! 1. Facebook = steals data 2. Uses data for FREE quality AI 3. Gives 'data' back to everyone
@clint9344
@clint9344 Месяц назад
you missed a step... should 1 Facebook = Steals data 2. Sell to govt agencies for profit 3 use data for free quality AI 4. Give "Data" back to the people.. 👍
@adamspeaking373
@adamspeaking373 Месяц назад
They didn't steal data, you gave it to them willingly in exchange for using their platforms.
@pavi013
@pavi013 Месяц назад
I wonder when open-source models have image and voice capabilities, or are they focusing fully on text generation?
@GraveUypo
@GraveUypo Месяц назад
i had a model that had vision. but it was a pretty bad model so i think i deleted it when i purged my bad models from my ssd
@delatroy
@delatroy Месяц назад
Well that explains Sam's perplexing stares off into the abyss in interviews saying things like.. I'm very worried right now
@thewestindianboy
@thewestindianboy Месяц назад
Rest of the closed source AI companies are like 'What the F>>>>??'
@christianjoachimgruber2773
@christianjoachimgruber2773 Месяц назад
Didn't believe i ever would say this: Thanks Zucc
@PriNovaFX
@PriNovaFX Месяц назад
Compared the 3.1 8B with the 3 70B, the new smaller 8B model is nearly in the same level as his bigger uncle.
@jamesvictor2182
@jamesvictor2182 Месяц назад
What a great day. Zuck needs to keep the chain and wafro, it's working to make him more gnarly and rad. Seriously thank God for Yan Lecun
@jbavar32
@jbavar32 Месяц назад
Another opportunity to those of us on a budget to compete with the big boys. Never thought I would say this but …Thanks Zuc
@swojnowski453
@swojnowski453 Месяц назад
you want to compete with the big boys, do not make me laugh. You can only compete with then at who farts worse ...
@justindressler5992
@justindressler5992 Месяц назад
I wonder when the student becomes the teacher. One day the small models may be able to generate, retrieve and filter relevant data for bigger models.
@MHTHINK
@MHTHINK Месяц назад
I don't think Yann Lecun gets enough credit for driving Open Source AI at Meta.
@godoegor8431
@godoegor8431 Месяц назад
If something is free, you are the product.
@HUEHUEUHEPony
@HUEHUEUHEPony Месяц назад
It's not free, you need to give Nvidia money
@elakstein
@elakstein Месяц назад
Doesn't apply to open source community
@hastyscorpion
@hastyscorpion Месяц назад
lol that isn't how open source works dude
@godoegor8431
@godoegor8431 Месяц назад
@@elakstein Yeh, Facebook is a "great" company with proven track record, handeling privat data! And just think about it what Gemini did with woke "mindset"!
@GraveUypo
@GraveUypo Месяц назад
except there's no one to sell you. you clearly don't know what open source is. if you're so scared, you can run this on a machine with no internet access. it still works like nothing's changed. it doesn't connect to the internet. it's the free tier of chat gpt that sells you.
@greenockscatman
@greenockscatman Месяц назад
Crazy that it got Snake right on first try. In a year we're gonna need to benchmark with Doom or something.
@ChristopherOBrienPSU
@ChristopherOBrienPSU Месяц назад
Crazy that Meta is more open than OpenAI 😂
@swojnowski453
@swojnowski453 Месяц назад
they just do not want the shit ...
@tozrimondher4250
@tozrimondher4250 Месяц назад
I tried the 3.1 8b for coding small plug-ins and it blew my mind. Imagine how good is the 405b model !!!
@tonyppe
@tonyppe Месяц назад
If you rephrase the number question it will get it right. It was answering based on 9.11 being higher than 9.9 like a version number. If you instead ask: "Which is a larger number? 9.11 or 9.9" then it gets it correct and explains why. If you ask it any other way then it gets it incredibly wrong.
@JohnSmith762A11B
@JohnSmith762A11B Месяц назад
OpenAI disappearing into the deepest, spookiest recesses of the MIC while the outside world moves on. Fare thee well, Sam!
@tzardelasuerte
@tzardelasuerte Месяц назад
You realize they are releasing the next generation after elections right? Everyone has caught up but they are already on the next generation
@gu9838
@gu9838 Месяц назад
wow NICE!!!!!! ai is awesome! and nice to see meta actually doing something positive for a change . i love llama 3 its pretty detailed for its smaller size at 8b downloaded 3.1 8b now!
@GraveUypo
@GraveUypo Месяц назад
wow! native 128k context? that's fkn awesome! i have a hacked llama 3 with 32k and i already think that's more than i need ever, but it doesn't work that well, it gets dumb the longer the context gets. if it's native, it's not going to suffer from this. can't wait for the quantized uncensored models to drop
@OriginalRaveParty
@OriginalRaveParty Месяц назад
Did I miss something? Everyone keeps comparing it to GPT 4o, when Claude 3.5 Sonnet is by far the leading model in the world right now. I know they're both right up there but 4o is undeniably second place.
@DavidSmith-ef4eh
@DavidSmith-ef4eh Месяц назад
Based Adam Corolla. I tried the 70b model on my 10th gen i9, it works but very slowly. Half a second for a token. I can't even imagine what is needed for the 405B model.
@danieladler3210
@danieladler3210 Месяц назад
A giant win for open source community!
@desertfish74
@desertfish74 Месяц назад
No, it is not. Read their license terms.
@RolandGustafsson
@RolandGustafsson Месяц назад
I installed both the 8B and 70B models on my MacBook Pro M1 Max with 64GB RAM. The 8B model runs super fast and is pretty amazing considering the memory footprint of only 4.7GB. (Not sure exactly how much space it takes in RAM, but that was the download size) The 70B runs MUCH slower and the fans kick in, 40GB download. Not sure if I can see enough improvement to warrant using it instead of zippy 8B.
@jarail
@jarail Месяц назад
So happy to see Meta doing this!
@swojnowski453
@swojnowski453 Месяц назад
there is no advantage in AI, nobody of us can win the race, they do no service to any of us.
@AlfredNutile
@AlfredNutile Месяц назад
I said it in another channel about how crucial it has been to me as a developer to have a local free llm to use while I build web applications that use OpenAi or Claude api when deployed (until I can get better at hosting Ollama 😊)
@SiimKoger
@SiimKoger Месяц назад
Zuck actually lived long enough to see himself become a hero again.
@voinea12
@voinea12 Месяц назад
Thank you zucc 🙏🙏
@HaraldEngels
@HaraldEngels Месяц назад
Yes, that is a watershed moment ...
@lesourire511
@lesourire511 Месяц назад
I was actually able to load the actual 405b modet onto my machine, windows 11 with recent processor and 192 gig with a 4090 (24g). I was mostly curious to see if it would even load and if so would it run and believe it or not it actually did run using ollama locally. The reply to a question "are you there" came back many minutes later one word at a time with a minute or two between each word. Just thought it was interesting even though pretty darn impractical without a much more expensive GPU with more dedicated VRAM. I think you said that del provided you with their machine with two large Nvidia cards ... sweet. it would be so great to have such a machine
@Imran-Alii
@Imran-Alii Месяц назад
Great job explaining LLaMA 405b! Your clear breakdown made the tech accessible to all. Thanks for sharing your knowledge and enthusiasm!
@MatthewMS.
@MatthewMS. Месяц назад
So weird I watched the recent CNBC interview and he was talking about this exactly… about his AI being able to train smaller models. This is Great to see. Thanks Matthew B.!
@nick1f
@nick1f Месяц назад
I didn't expect that an open source LLM that rivals with top level closed source LLM will happen so soon. Amazing job and decision by Meta and Mark Zuckerberg
@stevefox7469
@stevefox7469 Месяц назад
I don't understand synthetic data. Won't synthetic data be full of inaccuracies, and ultimately higher hallucinations?
@martfp88
@martfp88 Месяц назад
Not necessarily, think about it this way. Let's say I'm using Chatgpt for creating synthetic data. Real text -> trains Chatgpt -> Chatgpt becomes very close at real texting -> very close to real texting (ie synthetic data) -> trains new model -> which means that the new model is training to be Chatgpt
@mohocklockness8390
@mohocklockness8390 Месяц назад
I'm so confused, how do you have 96gb of vram? Can you explain your pc or build I would like to know how you are doing that. Thanks.
@carlosap78
@carlosap78 Месяц назад
4x4090 = 96GB VRAM
@mohocklockness8390
@mohocklockness8390 Месяц назад
@@carlosap78 Got it, what is the motherboard fir that?
@carlosap78
@carlosap78 Месяц назад
@@mohocklockness8390for example: WRX90E-SAGE
@sirellyn
@sirellyn Месяц назад
I'm even more concerned about censorship with it.
@thetrueanimefreak6679
@thetrueanimefreak6679 Месяц назад
mark is actually becoming cooler and cooler as this goes on
@dockdrumming
@dockdrumming Месяц назад
Agreed. It's incredible. 😂
@enoque2479
@enoque2479 Месяц назад
As AI is advancing more and more, he is becoming more and more human 🤨
@MilkGlue-xg5vj
@MilkGlue-xg5vj Месяц назад
​@@enoque2479Yeah their AI technology helps hide his robot side and look more human, don't trust me? I've been working with Meta for 5 years and I've just been fired last month.
@jalen2172
@jalen2172 Месяц назад
@@MilkGlue-xg5vjgood, see ya
@JohnnyTwoFingers
@JohnnyTwoFingers Месяц назад
He is, my opinion on him has somehow completely flipped lol
@HMexperience
@HMexperience Месяц назад
The parameters are 820 GB large so I guess you need about 1000 GB of video RAM to run it. That would be like 12 H100 in a cluster.
@user-ly2lh9ml4d
@user-ly2lh9ml4d Месяц назад
Every time you test, you ask for it to write the snake game in python... assuming correctly that it knows what the snake game is, because it has been in the training data. Wouldn't is be better to ask it to write specs for the snake game, and then ask it to write the game from the spec to see if it works as expected...
@FuturisticAgent
@FuturisticAgent Месяц назад
Things are changing rapidly, my human intelligence is exponentially advancing into unknown territories. Great coverage and comment. 🚀
@blesch
@blesch Месяц назад
I can't believe people aren't even shocked
@boris.teoharov
@boris.teoharov Месяц назад
Well, people were talking how Mark is a reptile or something but I really see the bright side of his mind. Well, he may have a business agenda but that's the way, bro! Props for boosting the open source community. We all have to remember our roots. If someone does not know Mark was pouring billions and billions of dollars into open source tech for the last 15 years. It is not only the latest LLAMA. This guy really deserves respect!
@lyxsm
@lyxsm Месяц назад
You know the world is fcked, when the zuck is the good guy.
@quiltingrox
@quiltingrox Месяц назад
Zucky has achieved a new arc era
@josgraha
@josgraha Месяц назад
Heck yeah Matt, this is huge. Your coverage did it justice, as always. Thanks
@samuelbaker5577
@samuelbaker5577 Месяц назад
It was a fun chat discussing using bubbles of space-time to move smaller black holes en masse to manipulate a larger one
@KJ-xt3yu
@KJ-xt3yu Месяц назад
larger modle comes out, the smaller ones go through heavy testing and refining through wide use
@pmarreck
@pmarreck Месяц назад
1) What kind of hardware do I need to own to run the full-size model at full speed? 2) What kind of hardware do I need to refine it further? (speed doesn't matter much there)
@Arcticwhir
@Arcticwhir Месяц назад
I've been trying it out, seems like it needs some more post train finetuning, i've had times it outputs repeated words endlessly, code doesnt run as expected, inconsistent responses etc. Its really cool that they are releasing this open source, hope other large companies can improve upon this model. Or maybe it needs a better system prompt on meta ai
@jairit1606
@jairit1606 Месяц назад
Zucc is a wild child ngl.
@jakeparker918
@jakeparker918 Месяц назад
So exciting. Can't wait to see this turned loose on Groq
@martins2246
@martins2246 Месяц назад
I think 405b is going to be 99.9% peak llm. Can't wait to see how it goes for you running it. I love ollama ....the llama3:8b crushes all my needs.
@TheInsuranceDudes
@TheInsuranceDudes Месяц назад
Love your videos dude! So killer thank you for all you do!
@jackflash6377
@jackflash6377 Месяц назад
Already on Ollama, downloading now. Exciting!!! Right off the bat it has a sense of humor. "It looks like you meant to type "Hello" but your keyboard stuck on a single key, resulting in the word "Greetings". That's a funny bug! If you intended to say hello, I'd be happy to respond in kind. Otherwise, is there something on your mind that you'd like to discuss?"
@ianPedlar
@ianPedlar Месяц назад
Put AI and voice commands into the Quest 3 Gosh, you have to hunt and peck at a virtual keyboard at the moment.
@harambae117
@harambae117 Месяц назад
Nobody expected the Zucc redemption arc, but here we are. Let's see how far he takes it.
@JasonPelzel
@JasonPelzel Месяц назад
Bro looking like he blew up a science experiment.
@hewhointheearthlydomainsee1272
@hewhointheearthlydomainsee1272 Месяц назад
But can it rebuild the Python's set of standard and developer libraries so there is no dependency hell or abandoned or lazy maintenance? And improve their efficiency and remove unneeded redundancy? Because that is what will possibly mark the creation of AI brains, or machines that can think (catalogue everything, mobilise and orchestrate) dynamically toward ongoing objectives and aims as they arise. Large scale enterprises and their maintenance, large projects.
@PaulinaStopa
@PaulinaStopa Месяц назад
Thats cool, but how much vRAM it need to run? Edit: Llama 3.1 405B has 405 billion parameters, requiring roughly 800 GB memory to be served in its original BF16 precision, exceeding the total GPU memory capacity of a single AWS P4 or P5 instance with 8 x 80GB A100/H100 (640GB memory capacity).
@sychrov81
@sychrov81 Месяц назад
hey guys, how to run these big models on regular gaming pc, I tried the lama (think 40B), but my pc almost burned down 😅 (ryzen7, rtx4070,32gb ram)
@fenix20075
@fenix20075 Месяц назад
Really looking forward its dolphin version and maid version.
@mrpocock
@mrpocock Месяц назад
Very long contexts locally will be very useful. If the model doesn't slow down horribly.
@vi6ddarkking
@vi6ddarkking Месяц назад
This is welcomed but not altruistic. With the Ecosystem they are building Meta is going to make bank renting out the GPUs and tools that used to fine tune and Build around Llama. You know, the old "In a gold rush you sell shovels." And Honestly as long as they keep it Free and open as they are now. This is a win win, so hey good on Meta for quite the crafty business model.
@esteban-alvino
@esteban-alvino 22 дня назад
I did snake too, run smoothly. thank you
@PinakiGupta82Appu
@PinakiGupta82Appu Месяц назад
Great effort! Meta could potentially reap significant profits from the model in the future while also contributing to FOSS, which is fantastic. However, one important point to consider is that any large language model that cannot be run locally may not be ideally useful for end users who wish to run the LLM on their own systems. I'm sure, somebody will publish a quantised version, even if it's not Meta, that's good.
@mariusirgens5555
@mariusirgens5555 Месяц назад
I am looking forward to implementing this into my new local LLM based roleplaying game system I am working on 🥳 I was using phi3-128K, but it got worse as the game progressed and the chat history got long…
@ywueeee
@ywueeee Месяц назад
groq is already live
@bioshazard
@bioshazard Месяц назад
Appreciate you advocating for the OpenAI API standard. The industry really just needs to lean into that unless new modalities are being released outside of OAI that aren't being kept up by their schema. Even then, I would prefer the community extend their schema than to keep making new ones. Its all the same payloads...
@dot1298
@dot1298 Месяц назад
yeah nice, but... where can we use it? at huggingface it always says "overloaded" :/
@musicalbirds2928
@musicalbirds2928 Месяц назад
Great video. Thanks for sharing.
@wardehaj
@wardehaj Месяц назад
Just made a query with groq with lama3.1 8b instant model: 750.000 T/s! Super fast!❤ And it still understands and answers in the not (3.1) supported Dutch language, awesome
@dejabu24
@dejabu24 Месяц назад
Meta is underrated in this they do some good stuff
@tzardelasuerte
@tzardelasuerte Месяц назад
At some point they will stop open sourcing it. Specially after hitting the petaflop limit set by us government
@Tetsujinfr
@Tetsujinfr Месяц назад
Interestingly, the 3.1-70B beats the 3.1-405B on a couple benchmarks. Would be curious to know how that is even possible given the massive difference in network size, unless the 405B model has not been trained to its maximum potential yet.
Далее
ВОТ ЧТО МЫ КУПИЛИ НА ALIEXPRESS
11:28
Просмотров 604 тыс.
Ex-OpenAI Founder Ilya Sutskever Strikes Back!
9:36
Просмотров 55 тыс.
Linus Torvalds: Speaks on Hype and the Future of AI
9:02
100+ Linux Things you Need to Know
12:23
Просмотров 1 млн
Meet The New Mark Zuckerberg | The Circuit
24:02
Просмотров 1,8 млн
Live Chat with Matt Shumer about Reflection 70b!
35:30
How RAG Turns AI Chatbots Into Something Practical
10:18
NVIDIA’s New AI Did The Impossible!
9:26
Просмотров 315 тыс.
Run your own AI (but private)
22:13
Просмотров 1,4 млн
What are AI Agents?
12:29
Просмотров 309 тыс.