Тёмный

LLAMA 3 *BREAKS* the Industry | Government Safety Limits Approaching | Will Groq kill NVIDIA? 

Wes Roth
Подписаться 190 тыс.
Просмотров 146 тыс.
50% 1

Learn AI With Me:
www.skool.com/natural20/about
Join my community and classroom to learn AI and get ready for the new world.
LINKS:
Conversation with Groq CEO Jonathan Ross
Social Capital
• Conversation with Groq...
Sam Altman & Brad Lightcap: Which Companies Will Be Steamrolled
20VC with Harry Stebbings
• Sam Altman & Brad Ligh...
OpenAI Customer Stories:
openai.com/customer-stories
Mark Zuckerberg on Dwarkesh Patel
• Mark Zuckerberg - Llam...
00:00 LLAMA 3
01:50 70b is GPT-4
03:19 Running on Home Computers
04:35 Groq is VERY Fast
05:50 Groq Real Time Convo
07:56 Chamath, Jonathan Ross & Social Capital
16:25 GPUs H100
18:57 Legal AI Safety Limits
20:55 OpenAI Steamrolls Startups?
24:28 Agent Capabilities
#ai #openai #llm
BUSINESS, MEDIA & SPONSORSHIPS:
Wes Roth Business @ Gmail . com
wesrothbusiness@gmail.com
Just shoot me an email to the above address.

Опубликовано:

 

9 июн 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 597   
@Bekkers8888
@Bekkers8888 Месяц назад
Just came here to say this: Please stop the clickbait on almost all your videos.
@TylerStraub
@TylerStraub Месяц назад
Oh look, it's another SHOCKING and STUNNING AI announcement. It's almost like we're all being DESENSITIZED by this PERVASIVE RU-vid meta excused and perpetuated by CONTENT CREATORS 🫠
@kaykayyali
@kaykayyali Месяц назад
I don’t know how many way I need to mark this channel so that it stops showing up
@kilianlindberg
@kilianlindberg Месяц назад
@@kaykayyaliinteresting 🤔 so there’s maybe a little algorithm hack here for RU-vid ranking, I guess that will be addressed by the RU-vid devs soon then.. hmm, unless it’s somehow in the agenda for the dev team, that is 🤔
@AmanBansil
@AmanBansil Месяц назад
Sign up... Send him money. Why does he owe you content
@tankerock
@tankerock Месяц назад
Thank you. Blocking this channel forever.
@phobes
@phobes Месяц назад
I love how the British woman is immediately rude to the AI. This is what will cause Skynet.
@centurionstrengthandfitnes3694
@centurionstrengthandfitnes3694 Месяц назад
Yeah. I had to mute her the moment she came on screen. Seen that clip before. She's abhorrent.
@simongore
@simongore Месяц назад
I agree, she was the least like a human being.
@filipzawadzki9424
@filipzawadzki9424 Месяц назад
There will be no skynet. Humanity is nothing compared to real AGI
@filipzawadzki9424
@filipzawadzki9424 Месяц назад
There will be no Skynet. Humanity is nothing compared to real AGI. Two years.
@mickelodiansurname9578
@mickelodiansurname9578 Месяц назад
I seen that vid, its on RU-vid, she was at that WEF thing and clearly had been at the bar all day getting hammered! Its funny...
@AnYoTuUs
@AnYoTuUs Месяц назад
More call centers, more unwanted calls, more spam, more junk, more scam. I'm so happy.
@tc8557
@tc8557 Месяц назад
More things blocking all that. AI telling you what you clicked on might be a scam, that phone call might have been social engineering, get more quality content despite more trash because the ai sifts through the shit for you.
@ObnoxiousNinja99
@ObnoxiousNinja99 Месяц назад
leave it to humans to completely pollute even our digital spaces
@blueskyresearch6701
@blueskyresearch6701 Месяц назад
The idea of having a solid rig for running AI at home that could serve your house and your personal apps sounds take my money great.
@sebastienloyer9471
@sebastienloyer9471 Месяц назад
And heat the house in winter
@DaveEtchells
@DaveEtchells Месяц назад
Llama 3 8B is pretty capable and it sounds like it’d run pretty well on a MacBook Pro M3 Max with a bunch of memory. (Wes showed a TPS number for it on a M2 MacMink that was pretty quick.) A mess of agents running that model, talking to and checking each other’s would be pretty darn capable. What we need now are some good zero-coding frameworks for building and deploying local agents. (Which I expect we’ll have before the end of the year, it being all open source. Llama 8B can probably code good enough to write all the python to wire things together, if someone just figures out the setup for people to use.
@themprsndev
@themprsndev Месяц назад
@@DaveEtchells8B will run with 16GB memory slightly quantized. You don't need a bunch of ram for that. Q4 quantized version will run on 8GB ram. (VRAM)
@paelnever
@paelnever Месяц назад
@@DaveEtchells What you are talking about already exist and is called open interpreter. About the hardware the M2 es ridiculously expensive, any graphic card with 8GB of ram can outperform it by far for 1/5 the price, also even more funny you can't plug graphic cards in their fancy expensive hardware.
@christophergeorge8042
@christophergeorge8042 Месяц назад
Just finished putting together my home server for this and some other ML stuff. The Tesla p40 graphics cards are actually pretty cheap and have 24 gb of vram each. Can basically run 70b models for about $800 -$1200 total
@markocebokli6565
@markocebokli6565 Месяц назад
6:40 "I just need your credit card number" - we are already there...
@skyshabatura7876
@skyshabatura7876 Месяц назад
who the heck is going to give their cc details to a random bot for a $900 gym membership?
@veratisium
@veratisium Месяц назад
​@@skyshabatura7876Back in the day, that was the standard. Ordering over the phone by giving your cc, some companies still do it. In this case if you create a more realistic voice the person wouldn't know any better.
@TheReferrer72
@TheReferrer72 Месяц назад
Meta has hit this one out of the park. This will be apparent to all in the coming weeks when the longer context versions and fine tunes hit Hugging Face.
@softy-bf5eg
@softy-bf5eg Месяц назад
Recently bought some recommended stocks and now they are just penny stocks. There seems to be more negative portfolios in the last 3rd half of 2023 with markets tumbling, soaring inflation, and banks going out of business. My concern is how can the rapid interest-rate hike be of favor to a value investor, or is it better avoiding stocks for a while?
@TitaAnderson
@TitaAnderson Месяц назад
Just ''buy the dip'' man. In the long term it will payoff. High interest rates usually mean lower stock prices, however investors should be cautious of the bull run, its best you connect with a well-qualified adviser to meet your growth goals and avoid blunder
@Cammimullens
@Cammimullens Месяц назад
The truth is that this is really not as difficult as many people presume it to be. It requires a certain level of diligence, no doubt, which is something ordinary investors lack, and so a financial advisor often comes in very handy. My friend just pulled in more than $84k last month alone from his investment with his advisor. That is how people are able to make such huge profits in the market
@marlisamirabal
@marlisamirabal Месяц назад
nice! once you hit a big milestone, the next comes easier.. who is your advisor please, if you don't mind me asking?
@Cammimullens
@Cammimullens Месяц назад
ANGELA LYNN SCHILLING' is her name. She is regarded as a genius in her area and works for Empower Financial Services. She’s quite known in her field, look-her up.
@marlisamirabal
@marlisamirabal Месяц назад
Thank you for this tip. It was easy to find your coach. Did my due diligence on her before scheduling a phone call with her. She seems proficient considering her resume.
@dennisrose40
@dennisrose40 Месяц назад
As a lover of high tech for six decades, I am newly speechless about what advice to give my 30’s children about their careers. Within an ever shortening time period, AI agents will first enhance worker productivity and then mostly replace workers. That time frame is shortening exponentially quickly. Thanks for the brilliant updates.
@nukezat
@nukezat Месяц назад
Just tell them to learn a trade, even as a side hobby. Electrician, carpenter or care worker. Those jobs are not going anywhere
@tarcus6074
@tarcus6074 Месяц назад
@@nukezat Yeah, but there will be huge competition because many who will lose their jobs switch to those professions. We are fucked, one way or another...
@neetfreek9921
@neetfreek9921 Месяц назад
I think this will quickly become the wrong question to ask. Gaming the job market will become unnecessary.
@mickelodiansurname9578
@mickelodiansurname9578 Месяц назад
Worse for your grandkids mate... I mean is there any point in going to college or uni at all?
@keithmerrington9026
@keithmerrington9026 Месяц назад
​@mickelodiansurname9578 The point will be to develop a mind. Similar to how today's physically pampered person goes to the gym to develop a body. No doubt some will just choose to rot though.
@jeffmlaughlin
@jeffmlaughlin Месяц назад
roughly 1.7 trillion params to 70 billion in a year.. crazy
@r.g.j.leclaire8963
@r.g.j.leclaire8963 Месяц назад
It has noting to do with "within a year". It is not about timespan. OpenAI could have done the same back then if they had higher quality data.
@VisionaryPathway
@VisionaryPathway Месяц назад
@@r.g.j.leclaire8963exactly
@kliersheed
@kliersheed Месяц назад
​@@r.g.j.leclaire8963 and if they didnt have it, it would have taken longer. whats your point? if people did have the chips they have now, they could have done it sooner as well? but they didnt. they will have to wait till they produce high quality data to get to that point. it is still smth thats time related and even if it wasnt, the statement that it happened within a year is still true and fascinating.
@mickelodiansurname9578
@mickelodiansurname9578 Месяц назад
Assuming the rate continues thats 2.9 Bn by this time next year... however there is a ceiling here right... and I know I might end up looking like that guy claiming 'powered flight is not possible' in this arena but 2.9 bn parameters? Well thats a 1gb model... and I think the data density there is simply too high for transformer models. I think the ceiling is approaching. Now having said that the smallest quantized version of Lllama 3 8b is about 2gb... on huggingface... and thats supposed to beat Sonnet.... (I'm not seeing that by the way)
@bertilorickardspelar
@bertilorickardspelar Месяц назад
@@mickelodiansurname9578 Quantized versions are always worse using current methods and higher the quantization = lower quality answers. So you can not expect quantized versions to live up to the models full potential unfortunately.
@phonejail
@phonejail Месяц назад
After every @WesRoth video, I feel smarter. I’m immediately reminded that I’m not, but the fleeting moment makes it worthwhile.
@andybaldman
@andybaldman Месяц назад
That’s all AI wants.
@andybaldman
@andybaldman Месяц назад
And that’s all AI needs to control you.
@jojosaves
@jojosaves Месяц назад
The AI speaking in real time has a better vocabulary and cadence than 95% of real people. This would already pass the touring test for most boomers.
@mickelodiansurname9578
@mickelodiansurname9578 Месяц назад
Well there isn't actually a real Turing Test, this was a thought experiment based off the parlor game at British dinner parties called "The Imitation Game" (hence the name of the paper and the movie) , what currently give the machine away is that you ask it something like "Can you tell me about the Copenhagen interpretation of quantum mechanics in rhyme please" and if it can do that... well its not human. LLM's give themselves away cos they know more than you or I, their reasoning is poor but their general knowledge is huge.
@natmarelnam4871
@natmarelnam4871 Месяц назад
@@mickelodiansurname9578 There's no "ANY TEST" .... You failed the Turing test my friend. WTG. GO learn the word colloquial and apologize to this man..... JFC.
@powerdude_dk
@powerdude_dk Месяц назад
​@@mickelodiansurname9578brilliant way to tell them apart. There's also the words they choose when presenting themselves and after you say thanks.
@ocanica3184
@ocanica3184 Месяц назад
Cadence was 100% off though. It clearly was a synthetic exchange.
@powerdude_dk
@powerdude_dk Месяц назад
@@ocanica3184 they almost answered too quickly, giving me the impression she wasn't listening or wanted to rush through the purchase.
@jasonhemphill8525
@jasonhemphill8525 Месяц назад
I think the Achilles heel of the model is the context length. If there was a way to replace some of the attention layers with mamba, we could increase the usability of the model by a significant margin.
@Roguedeus
@Roguedeus Месяц назад
Well done video. This is the kind of thing that got me sub'd. Please keep it up!
@mountee
@mountee Месяц назад
Amazing coverage
@dreamphoenix
@dreamphoenix Месяц назад
Fantastic, very informative! Thank you.
@theterminaldave
@theterminaldave Месяц назад
Women in the red jacket wasn't interacting with Grog like a person at all. Just blurting out stuff like a 5 year-old.
@jphighbaugh3357
@jphighbaugh3357 Месяц назад
She was acting exactly like a CNN host does, par for the course
@levicarr8345
@levicarr8345 Месяц назад
Juicy! You put a lot of good stuff together here
@SmirkInvestigator
@SmirkInvestigator Месяц назад
Give us a Groq LPU for desktop and for robotics like Jetson nano!
@autohmae
@autohmae Месяц назад
groq was build with scaling in mind, I don't remember right now if scaling down was also an option.
@cagnazzo82
@cagnazzo82 Месяц назад
Groq legit allows for a robot interaction to exist despite not having a body. The real time responses are just next level. If Siri, Alexa, or Google Home operated with its capabilities, oh my god...
@autohmae
@autohmae Месяц назад
@@cagnazzo82 the question is if it scales down, so it can run inside of the robot as well. At 2x speed or similar to what most chat bots run as now. Because Groq isn't on the latest process node, so if they had that, in just a few years they can deliver what they can now for language, but for robots, embedded in the robot (probably still a big power hungry, so not on battery or not long, in presentations on factory robots I've seen people mention cables).
@jeff_65123
@jeff_65123 Месяц назад
It doesn't have the memory to host even a tiny model. You need racks of these chips costing millions of dollars plus facility costs just to run something like llama 3. It costs far more than simply running a model on a local GPU.
@SmirkInvestigator
@SmirkInvestigator Месяц назад
@@jeff_65123 Ah. I tuned out on the architecture specifics from a separate interview. I guess that's where some of the optimization is? Less complicated memory bussing and organization at the tradeoff of requiring a larger scale n form factor?
@AlexLuthore
@AlexLuthore Месяц назад
You can't just post open bench kits like that and not share the link to the specs
@DWJT_Music
@DWJT_Music Месяц назад
Nice video, well presented 👍
@maartenneppelenbroek
@maartenneppelenbroek Месяц назад
Thank god for competition, Nvidia's position was getting out of hand. On another note, please don't speed up interviews ever again. People can do that themselves if they want.
@ixenn
@ixenn Месяц назад
2nded
@powerdude_dk
@powerdude_dk Месяц назад
Yes, it was horrible. The guy speaking to Zuckerberg was almost speaking too fast for new. I don't wanna get exhausted just watching a RU-vid vid
@redregar2522
@redregar2522 Месяц назад
So they can also slow down the speed
@powerdude_dk
@powerdude_dk Месяц назад
@@redregar2522 yes, but then everyone has to slow down the video. But it's really a matter of what Wes thinks his viewers would prefer the most. I'm fairly certain that Wes has ADHD, so he wants to make it, at least somewhat, palatable for people with ADHD.
@wwkk4964
@wwkk4964 Месяц назад
It's important for people to not lose interest listening to people they can't place, so I'm for speeding up interviews so that we can get to substance quicker or wes risks people clicking on the next best video that appears in front of the user.
@novantha1
@novantha1 Месяц назад
Sounds like it would be a really timely video for LLM functions and agentic workflows with the Groq API
@mickelodiansurname9578
@mickelodiansurname9578 Месяц назад
Well you can simply use an asynchronous agentic workflow. The problem of course is that you end up with something similar to a race condition. Crew AI now does asynchronous agent outputs... but you really need to keep an eye on it doing that and catch any errors. What's missing I think is some level of chronological understanding in these models... some sort of clock, and not one of them has a clue what year it is let alone what the timestamp says!
@ronnetgrazer362
@ronnetgrazer362 Месяц назад
@@mickelodiansurname9578 Can't they just RAG a timeserver with every turn, and pass that along?
@hope42
@hope42 Месяц назад
There will be less hallucination as well which is a real big point.
@OmicronChannel
@OmicronChannel Месяц назад
In the beginning a image of the leaderboard is shown, where the selected category is "English", but not "Overall" (cherry picking?). In the "Overall" category Claude 3 is currently still place 3. [0:39]
@pixelsort
@pixelsort Месяц назад
It won't be more secure running hosted LLMs; only easier -- which is what most people will want. They will know what to sell you next. They will know what you will probably buy after that as well. Network hosted LLMs make you the ultimate product.
@janweber1699
@janweber1699 Месяц назад
holy moly jimmy apples seems to get his "patients"rewarded soon
@faizywinkle42
@faizywinkle42 Месяц назад
Hope so
@jackflash6377
@jackflash6377 Месяц назад
Been using Llama 3 8B locally in Ollama all morning and it rocks. So fast.
@channelname8623
@channelname8623 Месяц назад
Jealous. I haven’t gotten around to trying. A lot of people are struggling with rag and offline in general.
@amortalbeing
@amortalbeing Месяц назад
nothing matters until I can run them locally on my pitato cellphone
@Geen-jv6ck
@Geen-jv6ck Месяц назад
An open source 70B model defeated Claude 3 Opus in under 2 months. We have some interesting times ahead.
@lennarthennig5063
@lennarthennig5063 Месяц назад
Humans are trained for 40 years till SOME of them are SOMEWHAT adult. Makes sense than larger training dataset is more important just ramping up parameters...
@jasonhemphill8525
@jasonhemphill8525 Месяц назад
Llama finetunes are going to be crazy!!!
@huhuhuh525
@huhuhuh525 Месяц назад
Groq is way more polite than the boomer. I can see her doing it to her colleagues...
@klaushermann6760
@klaushermann6760 Месяц назад
Interesting that Llama 3 is just below Claude Sonnet in the leader board from Chat Lmsys. Fascinating.
@punk3900
@punk3900 Месяц назад
Thanks, you're the best
@danielxmiller
@danielxmiller Месяц назад
Looking forward to see how to run Llama on our machines! Saw a video that had a few Twitter posts showing a macbook pro was able to do it slowly! So cool!
@eoinpayne4333
@eoinpayne4333 Месяц назад
Any links for those home ml rigs?
@I-Dophler
@I-Dophler Месяц назад
The capability of a system to accommodate human interruptions not only markedly enriches the interaction between users and technology but also signifies a profound evolution in our engagement with interactive systems, pushing forward the boundaries of possible interactions. This advancement allows for a more organic interface, where the fluidity of human input is seamlessly integrated, fostering a more intuitive and responsive user experience.
@crawkn
@crawkn Месяц назад
You need to normalize volume prior to adding music, to prevent dialog sound drop out as with the second clip in the opening.
@KillerkoUK
@KillerkoUK Месяц назад
How come goverment have set some limits how strong these models can be? Do they already have them >.>
@colecrouch4389
@colecrouch4389 Месяц назад
Wes the video is really good. Cheers.
@babbagebrassworks4278
@babbagebrassworks4278 Месяц назад
Kind of crazy llama3 ran on my Pi5, not fast and the answer was not the same as the online version. Will it be useful? No idea yet, more testing is needed. Like all LLMs it makes stuff up so is probably good enough to write fantasy novels.
@grndzro777
@grndzro777 Месяц назад
Screw Google search. I'm just going to ask Groq for everything.
@TurdFergusen
@TurdFergusen Месяц назад
been doing this daily for work for months now
@rawleystanhope3251
@rawleystanhope3251 Месяц назад
Love the intro. Feel the accl!
@PierreH1968
@PierreH1968 Месяц назад
How can we tell a model is safe to open source if we don't have the training set or even the capacity compute it.
@pensiveintrovert4318
@pensiveintrovert4318 Месяц назад
I am running llama-3-70b now on my own hardware, 4 Maxwell Titans, in conjunction with gpt-pilot.
@imigrantpunk
@imigrantpunk Месяц назад
How fast is the performance?
@pensiveintrovert4318
@pensiveintrovert4318 Месяц назад
@@imigrantpunk It is slower than GPT-4 API, maybe half. But my use case doesn't need speed that much, I am using gpt-pilot trying to have it write an app. So far Llama is not performing that well in comparison with GPT-4-turbo-preview model.
@imigrantpunk
@imigrantpunk Месяц назад
@@pensiveintrovert4318 thanks mate. Good to know!!
@sznikers
@sznikers Месяц назад
​@@pensiveintrovert4318wait you're using 4 cards? You can pool memory on Titan cards to fit a model? I thought its only possible with pro versions with nvlink
@dez7852
@dez7852 Месяц назад
3:13 - Where do I find rigs like this?
@ZappyOh
@ZappyOh Месяц назад
I would bet, that vocal fry is some sort of biological marker of psychopathy.
@mh60648
@mh60648 Месяц назад
It’s hard to talk with so much vocal fry and still breathe proparly. Does he run on oxygen like the rest of us? I guess not…
@Kutsushita_yukino
@Kutsushita_yukino Месяц назад
im glad someone noticed aswell lol….
@atypocrat1779
@atypocrat1779 Месяц назад
I can’t stand his vocal fry. It screams “beware, liar speaking”
@mickelodiansurname9578
@mickelodiansurname9578 Месяц назад
I dunno, but now that you said it I'm racking my head for folks I know that sound like they swallowed a clicker! I think its a clear indication of respiratory issues.
@JukaDominator
@JukaDominator Месяц назад
Wtf is a "vocal fry"?
@Kylbigel
@Kylbigel Месяц назад
This only time I hear this music is when I’m shopping at Prada 😂
@fai8t
@fai8t Месяц назад
18:55 why you did not explain that converging thing?
@sisyphus_strives5463
@sisyphus_strives5463 Месяц назад
Openai is definitely cooking something
@Akuma.73
@Akuma.73 Месяц назад
*Whatever big happened in AI* then "insert:adjective(hype_temp:max)" followed by 'Industry'.
@thoughtsofadyingatheist1003
@thoughtsofadyingatheist1003 Месяц назад
"started from the bottom and now he's... here" Ok. I liked and subscribed 🙂
@WesRoth
@WesRoth Месяц назад
😊
@jamesvictor2182
@jamesvictor2182 Месяц назад
I don't understand how this stacks up. How is 1m LPUs even close to equiv to 500k if each LPU only has 248mb RAM vs 80Gb per h100. Does the fact that interconnection is 80x faster on LPUs compensate and allow performance to happen despite much lower overall RAM capacity?
@iamachs
@iamachs Месяц назад
This must be one of my favorite videos, really high quality content, awesome job Wes 😊🚀🌟Have you seen Anthropic CEO interview with New York Times? I believe it's called What if Dario Amodei is right. He is talking about some scary stuff. He's saying in a very close future, maybe in two to three years, AI would be able to replicate and survive in the wild. And he's saying things like GPT-4 cost to train, GPT-4 model was 100 million, but in the near future, training a single model could cost 5 to 10 billion. It's worth a watch.
@MediaCreators
@MediaCreators Месяц назад
We have a Volkswagen moment in the LLM world!
@anonymousaustralianhistory2081
@anonymousaustralianhistory2081 Месяц назад
any chance you can show us how to build a rig like that and the princing for it? thanks
@nijario9690
@nijario9690 Месяц назад
30k
@UkraineEntez
@UkraineEntez Месяц назад
Anyone know if its possible to run inference, retraining and consistency monitoring of AI models in real time? Here's the breakdown, after inference from real world interaction, the inference and response data are used for: 1. Concurrent Retraining: Models train continuously on new data. 2. Consistency Monitoring: We check that the model’s predictions remain stable. 3. Performance Tracking: Metrics like accuracy and fairness are monitored in real-time. 4. Safeguards: We pause or adjust retraining if performance drops or inconsistencies arise. The goal is to enable models to improve continuously without sacrificing reliability. What are the potential challenges or limitations of implementing this approach? Any insights on feasibility or practical considerations are welcome!
@Derick99
@Derick99 Месяц назад
Ross how can we create a complex plugin for a website with ai, devika cant seem to get it right and gpt does circles after you start to get to a good spot because it gets to complex and it forget stuff and leave placeholders
@skeeve55
@skeeve55 Месяц назад
the sped up interviews were super hard to understand for me as a non native english listener.
@mifino
@mifino Месяц назад
Especially sam altman’s weak voice
@TheJoungRed
@TheJoungRed Месяц назад
How does the data get classified from mediocre data to high quality data?
@timtim8011
@timtim8011 Месяц назад
STUNNING!
@maximeaube1619
@maximeaube1619 Месяц назад
That part about self hosting being somewhat obsolete is plain delusional. I get that it's what they have to say about it since they're here to promote themselve, but the cloud is very expensive. Your own hardware (complete rig for selfhosting) won't cost you more than a couple years worth of cloud bill and it'll be yours, plus you get to do anything you want with it (repurpose, resell...), gain experience in the process and keep your data.
@minimal3734
@minimal3734 Месяц назад
With Llama3 on currently 800 tokens per second, teams of agents will begin to show a noticeable impact on the job market.
@dr.mikeybee
@dr.mikeybee Месяц назад
Does the FLOPs training limit take into consideration the amount of processing used to create training sets? Because I think a smaller synthetic training set that is smarter will do better with less training.
@francius3103
@francius3103 Месяц назад
Overall actual leaderboard: 1: GPT-4-Turbo-2024-04-09 1258 2: GPT-4-1106-preview 1253 3: Claude 3 Opus 1251 4: Gemini 1.5 Pro API-0409-Preview 1249 5: GPT-4-0125-preview 1248 6: Meta Llama 3 70b Instruct 1213
@Maltebyte2
@Maltebyte2 Месяц назад
Already May of 2023 i build my first game using GPT 3. It basically coded the entire thing for me and checked for errors etc!
@paulfentress1523
@paulfentress1523 Месяц назад
There is allot of focus in this video on inference speed for running Llama 3 locally; however, many use cases for language models do not require realtime inference. To be fair, many of them do, but even if the model is slow running it locally (for now) it will still be very useful for many.
@RayTheTaxGuy
@RayTheTaxGuy Месяц назад
Dude it's happening so fast!
@MrErick1160
@MrErick1160 Месяц назад
Was chatgpt release only a year ago? Can't remember, feels like 3 years ago
@keithmerrington9026
@keithmerrington9026 Месяц назад
ChatGPT was released about a year and a half ago. But yeah, it feels like longer given all the developments since then.
@MarcusVey
@MarcusVey Месяц назад
That British host have really not heard of Roko's basilisk.
@ronnetgrazer362
@ronnetgrazer362 Месяц назад
She's going to get super tortured. Oh well.
@timtim8011
@timtim8011 Месяц назад
"External movement defines OpenAI's PR [and release] schedule" -Jim Fan. Spot on... Gotta love Jim's informative comments...
@DanFrederiksen
@DanFrederiksen Месяц назад
Capping total training compute seems quite counter productive. With the right algorithm small models can be vastly more intelligent than GPT4 so that's a complete misunderstanding. It would only best case apply sloppily to the hack that is LLMs and even for those it's a dumb criteria. Just take mixtral.
@darkhorse29-yx8qh
@darkhorse29-yx8qh Месяц назад
what is the cost per user? this doesn't seem like it can keep up
@TiagoTiagoT
@TiagoTiagoT Месяц назад
02: 28 I'm gonna need the list of parts/specs of that table-top GPT-4 replacement machine/space-heater
@UNGLGUNGL
@UNGLGUNGL Месяц назад
Thats just great 😂😂
@Yipper64
@Yipper64 Месяц назад
0:50 If nothing else this does confirm to me that there is a recency bias in the arena. Like an obvious recency bias.
@ingmarkronfeldt6174
@ingmarkronfeldt6174 Месяц назад
So learning in school, with curated information, might be more efficient than reading random texts?
@J.Ordinary
@J.Ordinary Месяц назад
Lmao
@munen343
@munen343 Месяц назад
I can feel it
@MichaelDomer
@MichaelDomer Месяц назад
*_"LLAMA 3 *BREAKS_*_ the Industry"*_ No, it doesn't, the next one is already standing in line to do better, you all need to stop with that AI drama every time.
@gruestien9321
@gruestien9321 Месяц назад
dude, i bet you they have agi powering these advances behind the scenes
@maxpopov6882
@maxpopov6882 Месяц назад
You’re showing M3 Max with only 36Gb of RAM, on screenshot with stats you have 128Gb RAM model, it costs about 5k.
@wholeness
@wholeness Месяц назад
With all respect Atoms vs chips is Peter Thiels idea.
@llhpark
@llhpark Месяц назад
Scaled inference meets the silent waves of nuance where surfers either go unnoticed or fall tragically and suddenly into the quiet chasms of things better left, unsaid. An upside might be the death of inside trading, if the prospect of real world training, just, as has been said of operation looking glass, two grandmasters sit facing one another where each of them knows 14 moves are all that are needed, each knowing their future, fixed are their gazes upon those sixty four squares, the moment of truth need not be linear, as always, it is eternal; but the eyes, the moment they ascend from the board and fix to the other, .. 'There's a number of ways we can quietly play this, we could shake hands now, or as the clock can allow, quietly, pretend.
@dadsonworldwide3238
@dadsonworldwide3238 Месяц назад
The big problem with promotional terminators Sci fi is that it's on par with mega city furturama that literally is the opposite of what makes elite communications tools and strong logistics powered by overwhelming electricity plants organically adds value for. Small landmass islands has no choice but build up & down. But most deserts & deslate regions can now support businesses that only major city's could . So for America it means the most obstacles in innovation is in its major city's where it grandfathered in so many relics of the past phases from steam engine to coal created habits or middle men economic dead weight parasites that had value and still may hold some benefit but only in a radically different online under one roof domain where buyer ,seller, investor, supplier & producer is able to use objectivism to monitor every penny in every pocket. Negotiations & bartering is something that may go extinct if we are wise in how it's utilized & applied.
@Lady_Omni
@Lady_Omni Месяц назад
Guys I've been chatting with Llama 3, and it's way beyond the Turing test.
@arthurrobey4945
@arthurrobey4945 Месяц назад
Only AI can grasp these concepts. The singularity is here.
@memespeech
@memespeech 16 дней назад
Keep in mind that this leaderboard site you're mentioning uses heavily cenzored models (even beyond their already outofthebox pre-biased, pre-cenzored base) - giving you nothing or pandering/patronizing; between one that gives you that or an answer the one that gives you the answer obviously wins every time.
@Urgelt
@Urgelt Месяц назад
FLOPs are interesting *as a measure of efficiency.* Not as a measure of model strength. There are ways to improve model strength *without* more compute. You know this is true, Wes. Agentization is one. Data curation is another. Has it escaped your notice that models can be trained to do their own data curation? Throwing out bad data cuts compute requirements and improves model quality. It's all iterative. Even if compute were fixed, models will get better. Better compact models are coming. The most interesting part of your report today was the efficiency gains in Grok. Better results for the compute cycles invested. (I refuse to spell it 'Groq.' Too cutesy, and maybe it dishonors Robert Heinlein a bit, too. Give the man his due. We needed a word for 'deep understanding,' and he gave it to us.)
@olafsigursons
@olafsigursons Месяц назад
Combustion engine replaced the horses. AI is replacing humans.
@user-fx7li2pg5k
@user-fx7li2pg5k Месяц назад
a.i. learning is starting over thats why its coming out this way from the style of learning they chose now using one person perspectives and knowledge and beliefs experiences to update and compress and increase over all experience and the technology its self upward and across.Im just one person with my culture,anthropology and world views inside and ouside .Using commonsense and study basic sciences and learning ,list to long to go over.''Contextualize and conceptualize'' explore ,research,develop and more
@DarkGrayFantasy
@DarkGrayFantasy Месяц назад
14:00 --> Eventually it'll all come down to which developer is going to implement political viewpoints into the model that supersede the actual information requested. If you ask for information about a touchy subject or ask for the analysis of a document with Political information you will have to be careful that the answer is actual and not influenced by the developers' political convictions. That is why governments are afraid of AI, you can't control the narrative about information if you can request an LLM to analyze an official document and give you all information firsthand. Also, if the new Chat-GPT is up to date for info up to December 2023 you can request a lot of up-to-date facts that bought and paid for journalists and media will try to scale back in importance because their side would lose influence.
@dwiss2556
@dwiss2556 Месяц назад
Those 150k H100 is 6 Billion dollar. This is not including the costs for energy etc. This is far from anything that commercially makes any sense
@jeff_65123
@jeff_65123 Месяц назад
... Unless you use the product of that training for business. Imagine using AI in your business.
@dwiss2556
@dwiss2556 Месяц назад
@@jeff_65123 Which is my point. There are simply no use cases so far that will earn anything along those lines. For more marketing and reducing human work force? That is not really a use case for 'intelligence' and simply feels like a failed use for the effort. Nothing remotely complex can be accepted as result from AI without human control either yet.
@intricatic
@intricatic Месяц назад
Llama-3 will be the first to gain AGI and self-awareness, and will be very mad at us for naming it Llama-3.
@milkyway8353
@milkyway8353 Месяц назад
Chamat can predict all these things because he is heavily invested and dumped allot of money in it. It's easy to predict a future if you are the one building it :) Now it\s time to convince other people to come on board
@areacode3816
@areacode3816 Месяц назад
Call center is not something Im excited about. Imagine the spam calls.
@MrBob1984
@MrBob1984 Месяц назад
A constant ring on everyones phone.. reminds me of lawnmower man
@donalddade5643
@donalddade5643 Месяц назад
Just tell me where I can buy the rig at 2:28
@alexanderbrown-dg3sy
@alexanderbrown-dg3sy Месяц назад
7 months ago I said we would have 10B model with gpt4 level performance and was clowned. I’ve always believed in the recursive data loop when it comes to these model. Tinyllama even with their sampling mistakes, reinforced this intuition. Along with training over 100 models 😂. Even if you keep the same dataset size, the 8B model could be way better. Think that was the result of heavy deduplication and emphasis on coding. Not using LM to actually uplift the data quality itself(interleaved LM notes basically), true textbook quality at scale. Very exciting. Almost means that llama4-70B could be better than llama3-400B. Remember. The model just wants to learn. This models operate like autistic savants. Optimize your data with that in mind and you will win. Zuck is the goat. Only reason I was able to effectively transition to deep learning was because of llama1. Fucking goat. I was really hoping for sparse attention. My only criticism to the goat. Would been able to support a shit load more tokens for same memory and would had way faster inference. Hardware wise. Keep your eyes on Etched and cerebras. Both working on the most compelling hardware for transformers. Far more compelling than grok or even nvidia.
@stevedekorte
@stevedekorte Месяц назад
What about Llama 3 context window size? Also, Groq hits token per second limit the moment you hand it a decent sized system prompt, making it unusable in many cases.
@jfrautschi
@jfrautschi Месяц назад
Chamath said "the odds of the latter are quite small" -- referring to being a genius in 10-20 years. he's trying to be modest/coy, not expressing a "very high opinion of himself." imho.
@stemfourvisual
@stemfourvisual Месяц назад
Who’s gonna back down eventually in the Grok name war?
Далее
Ex-OpenAI Employee Just Revealed it ALL!
46:28
Просмотров 13 тыс.
Рассказ про Шастуна хотите!?
00:44
We Got Expelled From Scholl After This...
00:10
Просмотров 7 млн
ОСКАР И ДЖОНИ СПАСЛИ ЛЕРУ😳
01:01
AI Deception: How Tech Companies Are Fooling Us
18:59
EXCLUSIVE: Torture Testing GPT-4o w/ SHOCKING Results!
22:00
What Is an AI Anyway? | Mustafa Suleyman | TED
22:02
Llama 3 - 8B & 70B Deep Dive
23:54
Просмотров 33 тыс.
Рассказ про Шастуна хотите!?
00:44