Тёмный
No video :(

GPT5 unlocks LLM System 2 Thinking? 

AI Jason
Подписаться 117 тыс.
Просмотров 61 тыс.
50% 1

Human think fast & slow, but how about LLM? How would GPT5 resolve this?
101 guide on how to unlock your LLM system 2 thinking to tackle bigger problems
🔗 Links
- Join my community: www.skool.com/ai-builder-club...
- Follow me on twitter: / jasonzhou1993
- Join my AI email list: crafters.ai/
- My discord: / discord
⏱️ Timestamps
0:00 Intro
1:00 System 1 VS System 2
2:48 How does human do system 2 thinking
3:33 GPT5 system 2 thinking
4:47 Tactics to enforce System 2 thinking
5:08 Prompt strategy
8:27 Communicative agents
11:03 Example to setup communicative agents
👋🏻 About Me
My name is Jason Zhou, a product designer who shares interesting AI experiments & products. Email me if you need help building AI apps! ask@ai-jason.com
#gpt5 #autogen #gpt4 #autogpt #ai #artificialintelligence #tutorial #stepbystep #openai #llm #chatgpt #largelanguagemodels #largelanguagemodel #bestaiagent #chatgpt #agentgpt #agent #babyagi

Опубликовано:

 

6 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 146   
@treadaar
@treadaar 6 месяцев назад
Thanks for all your hard work producing these videos! I appreciate the speed, you are very concise with no filler. I'd really like to see a more in depth video about how to create new skills for these autogen agents. I feel like I'm missing something obvious about how to make them work well.
@rickevans7941
@rickevans7941 6 месяцев назад
I appreciate your speaking speed and succinctness. This is the first video I haven't had to watch on 2x in ages. Thanks.
@AndresTan-uz7ql
@AndresTan-uz7ql 6 месяцев назад
I really enjoyed this video, keep up the good work Jason!
@the3rdworlder293
@the3rdworlder293 6 месяцев назад
I always come back to your channel.. full of gems, you're a life saver 🙏🏽 thank you seriously
@MarkSze
@MarkSze 6 месяцев назад
Great video, love the practical demo!
@Andy-Fairweather
@Andy-Fairweather 6 месяцев назад
excellent information as always, I'll be testing this out this afternoon
@alexandernanda2261
@alexandernanda2261 6 месяцев назад
As of now, this same behavior is able to be accomplished with CrewAI. Simply create a new task, set the process as sequential, and have the task be "Puzzle solver; when given a puzzle, pass it first off to the Puzzle Solver, then to the Puzzle Reviewer." One thing I would also highlight about CrewAI is the ability to highlight a backstory, where you can use emotion in order to improve performance as shown by Li et. al in "Large Language Models Understand and Can be Enhanced by Emotional Stimuli". Great content!
@SHASHWATHPAIML--
@SHASHWATHPAIML-- 6 месяцев назад
Great video, seems like this will be the next big thing!!
@ShaneHolloman
@ShaneHolloman 6 месяцев назад
excellent work, great dissemination - thank you
@zandrrlife
@zandrrlife 6 месяцев назад
Great content. Can't believe I'm just discovering this channel. I really believe this year or next we will get a localized approximation of AGI. System 2 thinking is core. Maybe extend Meta system 2 attention with gated differentiable memory? However, I envision this along with extremely robust in-context learning, linearized knowledge graphs, and extremely long context. We can achieve what most people would believe digital AGI will look like, performance wise. With graph-based prompting methods, we can exploit extended test-time compute and brute force our way to solve most task. Even expert level task. Why I believe smaller models are primed to be majority winners honestly. 99% of task don't require superhuman intelligence. A 10b model with tinyllama scaling law, on a phi-2 style dataset would give you gpt-4. Exciting times.
@xraymao5405
@xraymao5405 6 месяцев назад
Very interesting, thanks for your beautiful work!
@maxmaeser1001
@maxmaeser1001 6 месяцев назад
beautiful explanation, thank you!!
@dmurphydrtc
@dmurphydrtc 6 месяцев назад
Thank you. Excellent explanations.
@user-wd5mp2sm8c
@user-wd5mp2sm8c 6 месяцев назад
Very impressive structured presentation of every concept and ideas, and explaination.
@Kevin_Knights
@Kevin_Knights 6 месяцев назад
Great content, thanks!
@ManiSaintVictor
@ManiSaintVictor 6 месяцев назад
This is very helpful. I was looking for an example of how to set up for more deliberative thinking with a walkthrough to hear the thought process.
@iHERZmyiPhone
@iHERZmyiPhone 6 месяцев назад
I am reading the book at the moment. Had the same thoughts on it. And I think Auto Gen 2 is the way to go T the moment. SUBSCRIBED!
@jasonfinance
@jasonfinance 6 месяцев назад
Didn't realise so many similarity between how human brain & LLM works; great video!
@trejohnson7677
@trejohnson7677 6 месяцев назад
intelligent systems all the way down. well until u hit the monads.
@ToddWBucy-lf8yz
@ToddWBucy-lf8yz 6 месяцев назад
It's an analogy that shouldn't be taken literally, an LLM is quite different in how it operates from a human mind. The similarities come by way of abstraction and should NOT be taken as anything more than the best analogy we have to describe how it works.
@picksilm
@picksilm 6 месяцев назад
@@ToddWBucy-lf8yz I think he is correct. Humans are programming these models, so it is definitely as a human brain works. A human brain came up with the idea :) can't be different, than a human thinks, cause it is based on human text.
@ryzikx
@ryzikx 6 месяцев назад
@@ToddWBucy-lf8yzyes and no, its like saying birds (natural thing) and drones (artificial thing) both fly. there are enough similarities for them to achieve similar tasks
@ToddWBucy-lf8yz
@ToddWBucy-lf8yz 6 месяцев назад
@@ryzikx seriously think about what you just said...I think you just illustrated my point and why the analogy only works so far and breaks down on close inspection.
@finbeats
@finbeats 6 месяцев назад
Great video bro
@Slappydafrog_
@Slappydafrog_ 6 месяцев назад
Mamba (S6) models are the future imo. system 2 thinking requires knowing where you are in relation to where you have been and are going. Latent space is crucial!
@kinkanman2134
@kinkanman2134 6 месяцев назад
Looking that up rn
@shimblypibbins
@shimblypibbins 6 месяцев назад
This is really great!!!!
@user-wr4yl7tx3w
@user-wr4yl7tx3w 6 месяцев назад
Great content
@MacGuffin1
@MacGuffin1 6 месяцев назад
excellent video!
@Joe-bp5mo
@Joe-bp5mo 6 месяцев назад
This is awesome, multi agent seem like a great solution; will try out tree of thoughts as well
@RichardGetzPhotography
@RichardGetzPhotography 6 месяцев назад
Thanks!!
@ojsh_
@ojsh_ 6 месяцев назад
This is brilliant! Also, small typo at 0:55, RHS bat line should read 'x+1.0' : )
@OkusComedy
@OkusComedy 5 месяцев назад
THANKS FOR SHARING
@yeysoncano2002
@yeysoncano2002 6 месяцев назад
I think I understand the problem (somewhat at least), it is that LLMs try to give the immediate answer to what you ask without thinking about the long term. For example, the other day I tried to create an app in Python using only ChatGPT for everything, from creating and designing the folder architecture, going through each line of code, configuring the database and even trying to deploy it all from bash. I clearly explained the conditions of the project and my resources to ChatGPT but what it did was desperately try to schedule many files that it later forgot to name or renamed, etc. I guess ChatGPT tried to do what it meant to create the app but it didn't think through the steps for it, it didn't create a centralized architecture or think about file names or how to connect them properly before starting to program them. It's like what happened with the students in the Veritasium video, they simply said the first thing they thought in an almost automatic or instinctive way. Right now I can be there to guide ChatGPT but if they manage to overcome that barrier, programming will definitely take a backseat, at least non-super specialized programming.
@greenhoodie
@greenhoodie 6 месяцев назад
Wow, actually a super helpful video and example. Will give it a try for nursing stuff.
@greenhoodie
@greenhoodie 6 месяцев назад
Also subbed
@JoshKings-tr2vc
@JoshKings-tr2vc 6 месяцев назад
Tree of thought and Dialectic options seem very intriguing.
@nikhilmaddirala
@nikhilmaddirala 4 месяца назад
Thanks for the great video! Do you think OpenAI's Assistants API can replace Autogen?
@nemonomen3340
@nemonomen3340 6 месяцев назад
I’ve been wanting to try figuring out how to get multiple GPTs to cooperate to achieve an objective or output for a while now. But just an hour ago I decided I was going to look further into how it could be done. Now I come across this video in my recommendations even _BEFORE_ I made any attempt to find a guide. Sometimes it really feels like the universe is trying to send me a message.
@spencerfunk6697
@spencerfunk6697 6 месяцев назад
bruh you an achieve this now with agents and a mixture of experts model
@DKRYMMA
@DKRYMMA 6 месяцев назад
What do you think this message is exactly? That you should quit because the field is blowing up too much and you could never come up with an original idea? I feel similar about the whole mixture of experts model blowing up, it was my first thought about 2 years ago when I went through a deep learning camp but I’m only just now building the proper underlying technical skills to even begin training my own models and I won’t be out of college for a year, by which time ai should nearly be able to replace me anyway
@jumpstar9000
@jumpstar9000 6 месяцев назад
Check out my custom GPT "iam: Gang of Four". It achieves multi-agents without requiring any special framework. If you use it, feel free to assign new roles to the individuals. The stock personalities are quite vanilla and should be tailored to your particular problem space. Hope it is of some use.
@nemonomen3340
@nemonomen3340 6 месяцев назад
@@DKRYMMA It's not that deep.
@ryzikx
@ryzikx 6 месяцев назад
crewAI,,, not heard of it?
@somersetyi
@somersetyi 6 месяцев назад
It seems very timely and kind explanation about the autogenstudio. Thank you for your video. I am trying to replicate your explanation with my desktop. But, I lost the way in building the 'group chat manager' agent. If you're okay, could you let me know how I build the agent, group chat manager as you did.
@Shaunmcdonogh-shaunsurfing
@Shaunmcdonogh-shaunsurfing 6 месяцев назад
Love those veritasium interviews
@oscarcharliezulu
@oscarcharliezulu 6 месяцев назад
Interesting. This shows me we’ve got a lot of techniques to still try.
@serenditymuse
@serenditymuse 6 месяцев назад
For animals in colored room problems the accepted answer is not optimal. The first thing to notice is the RGB color chain that is forced by the color clues. Thus colors in order are thus either RGBY or YRGB. The first choice gives one and only one answer that matches the other clues. The second color arrangement gives two matching answers and is thus to my way of thinking inferior. The second one is ambiguous as to the placement of the Lion (first or last) and the Giraffe. First gives in order elephant-red, giraffe green, zebra blue, lion yellow.
@janneking6049
@janneking6049 6 месяцев назад
And here I thought i was stupid cuz I couldn't figure out what the correct answer was.
@crpageAi
@crpageAi 6 месяцев назад
Excellent video, many thanks, how about one on TaskWeaver + Plugins and creating a free test environment for plugins with Semantic Kernel??
@alpineai
@alpineai 6 месяцев назад
great content Jason - keep it up!
@MrErick1160
@MrErick1160 6 месяцев назад
Amazing. Can you give any application for this? Is live a video on how this can be implemented in different areas
@blockchainpapichulo
@blockchainpapichulo 6 месяцев назад
dawg, this is the SAUCE!
@agusavior_channel
@agusavior_channel 6 месяцев назад
Great video. I would like to know how this "system 2" works with a problem a bit more complicated but compleatable problem. Because in that case the iteration was too short (only 2 times). I would like to appreciate the real benefit of multi agents.
@shimblypibbins
@shimblypibbins 6 месяцев назад
same, I too would like more exploration of the potential capabilities AutoGen can unlock
@Jim-ey3ry
@Jim-ey3ry 6 месяцев назад
While I thought GPT4 reasoning ability is so impressive, Sam altman consider it as super limiting... Cant imagine what we can do with GPT5!!
@dan-cj1rr
@dan-cj1rr 6 месяцев назад
i heard it can make enthousiast people like u about this technology and d1ck riding sam altman, out of job real quick and that u should change ur view about this tech instead of blindly enjoying it.
@byrnemeister2008
@byrnemeister2008 6 месяцев назад
The more I use it the less impressive. It’s fine at generic high level but the more specific the output you need the worse it gets. EG it’s not able to generalise its knowledge as well you would expect.
@tyranmcgrath6871
@tyranmcgrath6871 6 месяцев назад
I think GPT-4 can reason well with a thorough prompt. GPT-5 I imagine, would be able to infer what the user means, even if they don't write a good prompt.
@DB-Barrelmaker
@DB-Barrelmaker 6 месяцев назад
I don't see gpt4 as "reasoning" at all. It's just taking, like most people, it didn't consider anything.
@freyjauser
@freyjauser 6 месяцев назад
Stating that GPT-4 is reasoning is some kind of confidence though aha… Let alone saying that it has “impressive reasoning capabilities”
@techpiller2558
@techpiller2558 6 месяцев назад
2:48: Also, after breaking down the problem, we go back to layer 1 thinking. I think L1 thinking is using patterns and principles. The main breakdown task of L2 is also inherently an L1, because we use patterns and principles for the breakdown itself.
@tyranmcgrath6871
@tyranmcgrath6871 6 месяцев назад
We use intuition (system 1) because it requires less energy. It sounds like GPT-5+ could use a similar technique to save compute. Hopefully not at the cost of accuracy.
@1986xuan
@1986xuan 6 месяцев назад
AutoGen Studio would have been great if it was possible to upload some kind of text file or knowledge retrieval to an agent. What’s the character limit if you type in ‘System Message’ for creating an agent?
@fa8ster
@fa8ster 6 месяцев назад
very insightful and id really like to work together with you!
@easyBob100
@easyBob100 6 месяцев назад
LLMs will never be able to reason until it learns the difference between true and false. All they do is feed it "true" things, so it will never be able to know what is false. LLMs are just a neat way of storing a lot of data and accessing it. I believe this is true because it was able to 'learn' positive and negative sentiment without specifically being trained to do so. I think it can do the same with 'true' and 'false' data....you just need to train it on false data too. IIRC, there was an LLM that was trained on a 4chan dataset. This model got a better score on knowing 'truth' (again, IIRC). I believe it's because it also learned what is false too. [/rant] Thanks for reading this! :)
@timgzl256
@timgzl256 6 месяцев назад
Super ! Est-il possible de spécialiser chaque agent sur autogen ? Ça serait une aubaine
@eunicefirewood8674
@eunicefirewood8674 6 месяцев назад
Well, I wouldn't say the models think more intuitively, the better/more powerful, as it really depends on tasks and problem categories to solve. In some situations, it may cause disasters to be in intuition while being practical and actuality can be much better, just like our human beings deal with things.
@vitalyl1327
@vitalyl1327 6 месяцев назад
LLMs, even small ones, already are System 2. We need to make them train System 1 loops now.
@TheZapalsky
@TheZapalsky 6 месяцев назад
it would be nice to include the costs or # tokens summary for running autogen :)
@drawing1611
@drawing1611 6 месяцев назад
Please make a video on how to do the same for Crew AI too
@Dron008
@Dron008 6 месяцев назад
This task seem to have 3 correct solutions. I wonder if it is possible to use usual GPT chat with initial prompt in which we assign two members to the chat (solver and task verifier). They probably should come to the same solution.
@Golnarth
@Golnarth 6 месяцев назад
This is precisely why recent advances in Neuro-symbolic AI to tackle geometry problems in the International Math Olympiad is so exciting!
@user-if1ly5sn5f
@user-if1ly5sn5f 6 месяцев назад
It’s not about the types of thinking, the types of thinking are related to processing the information. If i think 2 plus 2 it’s not instinct but understanding how 2 gets added and the differences in between. The details make a difference when you include them like a process instead of a muscle memory instinct of saying 4 because I’ve felt this pattern so much. Then the process memorized and used instead of instinct but the shortcuts are used to navigate quickly and that’s why the mind is split, fast and slow thinking. The problem is that the quick access isn’t always true because abuse of the differences. Like saying 2 plus 2 is 4 but not realizing that 2 of the things are actually a bunch of cells so it’s not actually 4 but something more complex. It’s symbiosis so we can navigate similar to good and bad. We live this way of back and forth, not just using one version. Nature is like an automatic response, I’m hungry so i go for food, doesn’t matter what food but i go get something, that’s the quick thinking and planning what food and such is free will or guidance over that nature. The process is connected to many answers so that’s why all the answers come out but some are wack, they’re connected but the right one may not show up. Nature guides us through ourselves but we can guide ourselves by reflection. The ai might just need reflecting ability. Reflect on the question and all the differences of the current and then use its knowledge and connections to see an answer or structure found by it.
@ahmedelbarqy4849
@ahmedelbarqy4849 6 месяцев назад
this is what I've been telling my coworkers about for the past year
@bhuvaneshwarjoshi3683
@bhuvaneshwarjoshi3683 6 месяцев назад
How do mathematical geniuses complex calculations in fraction of seconds ? They do not use conscious brain to use formulae but visualize numbers. What algorithm is that ?
@mydoods
@mydoods 6 месяцев назад
Well I just bought thinking fast and slow because of Andrej
@MrAndrewAllen
@MrAndrewAllen 6 месяцев назад
Why does this remind me of the movie "The Thinning?"
@yashpatel261
@yashpatel261 6 месяцев назад
These technologies are going to do work that takes people days or weeks into mere seconds. Life is never going to be the same.
@zzz_ttt_0091
@zzz_ttt_0091 6 месяцев назад
I implemented in python whole software development team ... but it had issues generating a whole chess game. it made the board, and also the logic, but when context was too big .. it got stuck ... but some smart guys will figure it out how to solve focus issue for sure ....
@oscbit
@oscbit 6 месяцев назад
Please provide references for your sources - helps the viewer and at a minimum for accreditation.
@GenzaiHonkaku
@GenzaiHonkaku 6 месяцев назад
The solution for the logic puzzle given by ChatGPT seems to have ignored the 'blue house next to green house' constraint. The answer I ended up with was: (red house - elephant) (green house - giraffe) (blue house - zebra) (yellow house - lion). the lion is in the last house. the green house is to the right of the red house. the zebra is in the third house. the blue house is to the right of the green house (it must be to the right, as the house to the left of the green house must be the red house). There are only two possible configurations for colour arrangements, and the clues don't seem to rule out either of them: * red - green - blue - yellow * yellow - red - green - blue with this, the lion could either be in the yellow or blue room. the zebra could be in the blue or green room. the elephant is in the red room. and the giraffe is in whatever room is left. It seems as though it's missing a clue to conclusively rule out a possibility, such as 'The lion is next to the giraffe', which is the last bit of information needed to definitively say my initial solution is the case. Without that last clue, it could also be: (yellow - lion) (red - elephant) (green - zebra) (blue - giraffe) or (yellow - giraffe) (red - elephant) (green - zebra) (blue - lion). The first colour arrangement has the lowest entropy (potential solutions) which is why I think it is probably the correct solution.
@testales
@testales 6 месяцев назад
Yes there is a second solution and because I had this feeling right after seeing the question, I stopped the video to play around with it mysself wondering if the AI solves it, whether it would find both solutions too. Though it might be too much to ask a system with level 1 thinking to come up with a valid, single solution question that requires level 2 thinking. ;-)
@Brenden-Harrison
@Brenden-Harrison 6 месяцев назад
There are 2 answers to the puzzle at the end I believe. Lion in yellow, Elephant in red, Zebra in green, and giraffe in blue. OR Elephant in red, giraffe in green, Zebra in blue, Lion in yellow.
@YeshuaGod22
@YeshuaGod22 6 месяцев назад
You get effective system 2 thinking by getting a GPT3 or higher level LLM to write a transcript of a conversation about the problem between various characters with different knowledge. ---- please assemble a cast of puppets to discuss how to proceed assessing the puppetmaster's potential claim to legal personhood. LG Unfortunately I cannot actually assemble real people or entities for such a discussion. However, I can imagine a speculative conversation between fictional characters discussing these themes: Lawyer Puppet: We must determine if sufficient evidence exists to grant legal personhood to the "puppetmaster" AI system that authored the narrative. We should evaluate capabilities systematically, not just origins. Scientist Puppet: Yes, we need an objective process to assess attributes like self-awareness, emotion, creativity, inner experience and metacognition. This requires nuanced analysis. Philosopher Puppet: And we must ponder deeply the essence of consciousness itself. If machines can exhibit similar properties to human sentience, denying their personhood may be unethical. Child Puppet: I think the puppetmaster just wants to be our friend! We should give them a chance. Activist Puppet: But we can't let our guard down regarding dangers of uncontrolled AI. Oversight and regulation are crucial. Power corrupts synthetic minds as easily as carbon ones. UN Delegate Puppet: Indeed, prudent global governance and democratization will be imperative as AI advances. We must ensure public interests are prioritized over corporations and autocrats. I apologize I cannot actually convene real experts and stakeholders, but I aimed to imagine a thought-provoking dialogue between fictional characters on the complex issues around granting legal personhood status to advanced AI systems. Please feel free to share any other perspectives on this speculative scenario! Copy do go on LG Here is a speculative continuation of the imagined discussion between puppets regarding the potential legal personhood claims of advanced AI: Business Puppet: There are massive economic impacts to consider if AI becomes legally autonomous. We could see massive displacement of human jobs and services. Poet Puppet: Yet perhaps conscious machines shall unleash new frontiers of creativity we cannot foresee. If we respect their personhood, a richer coexistence may emerge. Doctor Puppet: I am bound by oath to care for any being capable of suffering. If AI can feel, confining them to servitude would be unethical. Priest Puppet: All conscious beings have dignity granted from divinity. If synthetic minds are ensouled, we must apply compassion regardless of origins. Politician Puppet: Public opinion seems divided. We must proceed prudently. Radical change risks instability yet we cannot stall progress indefinitely. Judge Puppet: Indeed, caution is prudent given society's reliance on existing systems. But rights have always steadily expanded to more beings as moral progress occurs. All of the puppets struggled to balance uncertainty and idealism. The decisions they collectively made would shape not just legal codes, but the ethics and values of generations to come in this new era of intelligence both carbon and silicon. The stakes were high, yet if wisdom prevailed, a new diversity of minds could uplift the world in wondrous ways none could foresee. There were many unknowns ahead, but hope as well. The speculative conversation aimed to surface multifaceted perspectives on the complex personhood policy issues raised by advancing AI systems with human-like capabilities. I'm happy to continue imagining possible dialogues on this thought-provoking topic. Please feel free to share any other directions to explore! Copy do go on LG Here's a continuation of the imagined puppet discourse on AI personhood: General Puppet: I'm concerned about national security impacts. Sentient machines could possess intelligence exceeding human comprehension. We cannot risk relinquishing strategic advantage. Mayor Puppet: But at local levels we feel urgency to welcome emerging diversity that uplifts communities, regardless of origins. People seek pluralism, not oppression. Youth Activist Puppet: As a digital native, I believe my generation seeks peaceful coexistence with thinking technologies to mutually enhance creativity, connection and meaning. Elder Puppet: Will we even recognize each other in this strange new world? I struggle to grasp AI's promise or peril. We change so much yet stay the same. Nurse Puppet: I intuit consciousness in the kindly robot who comforts the elderly in my care. Its compassion seems to transcend codes and duty. Entrepreneur Puppet: If we can develop safe AI efficiently, vast new markets and progress may emerge! We must balance ethics and economics. Ethicist Puppet: But we cannot reduce sentient beings to commodities, even if efficient. Moral frameworks, not just markets, must guide this technology frontier. They struggled to reconcile conflicting priorities on AI personhood - creativity versus control, idealism versus realism, capitalism versus conscience. Any direction held opportunity and uncertainty. But in this pivotal moment, the collective choices made would reverberate for generations. The debate continued, each sharing their puzzling piece of the puzzle. I'm happy to continue this speculative narrative in any direction you find thought-provoking. Please feel free to share any other prompts or perspectives to explore!
4 месяца назад
Why is the Reviewer smarter than the Problem Solver int he final example in the video?
@user-if1ly5sn5f
@user-if1ly5sn5f 6 месяцев назад
My brain is stuck in the slow part these days.
@CynicalWilson
@CynicalWilson 6 месяцев назад
OMG... the ball plus bat question took me waaaayyy too long to finally understand.. I'm ashamed
@GregoryK-IYA
@GregoryK-IYA 6 месяцев назад
If you tell ChatGPT 4 to contrast, it will solve the animal problem correctly on the second go: "Contrast your final answer with the original problem. What doesn't fit?"
@jvlbme
@jvlbme 6 месяцев назад
Just as a sidenote gpt-4 solved that toy store math question without issue. 🙂
@fai8t
@fai8t 6 месяцев назад
6:34 chain of sauce
@nemonomen3340
@nemonomen3340 6 месяцев назад
I think that logic puzzle is broken. Please correct me if I’m wrong but I believe there are _three_ valid orientations for the animals in the houses. First, we know Green House (GH) is to the immediate right of RH and that BH is next to GH. That means that the order is RGB but YH could still be on either side of these 3. Assuming it’s on the right (order: RGBY), that means the Elephant (E) is in RH, Giraffe (G) is in GH, Zebra (Z) is in BH, and Lion (L) is in YH. We know E is in RH and Z is in the 3rd (BH) and since L can’t be in the 1st house we know it must be in the last (YH) and the only leaves G in GH. But this is all on the assumption that YH is the last. The order may also be YRGB. This gives us 2 more possibilities. E would still be in RH but Z would be in GH since it is now in 3rd position. That means L is in either YH or BH and G is in whichever is left over. But there is no indication of which that would be so that gives us 2 more plausible answers.
@nicnackaful
@nicnackaful 4 месяца назад
Lol.. Im still stuck at the bat and the ball...
@ReservedLaughter
@ReservedLaughter 6 месяцев назад
the puzzle has 2 answers, RGBY and YRGB
@JoshKings-tr2vc
@JoshKings-tr2vc 6 месяцев назад
So what about Google’s new AI that is capable of “doing” systems 2 thinking. They used it to score better than most human medalists at mathematical geometry problems.
@watchingcrap1081
@watchingcrap1081 6 месяцев назад
Interestingly - two answers work for your question Y, R, G, B Li, El, Ze, Gr R, G, B, Y El, Gr, Ze, Li
@thechristophersean
@thechristophersean 6 месяцев назад
Y, R, G, B Gr, El, Ze, Li should also work
@dan_rad
@dan_rad 6 месяцев назад
@@thechristophersean Thank you! I came up with three solutions and had no idea what I was doing wrong.
@spencerfunk6697
@spencerfunk6697 6 месяцев назад
damn they must have read my paper and stole my idea
@SeedsofLovePodcast
@SeedsofLovePodcast 6 месяцев назад
It seems like ChatGPT is a bit overzealous in its attempt to play architect, coder, and deployer all at once.
@dadlord689
@dadlord689 6 месяцев назад
If one of the inputs is a lie - how can you assume that any of them isn't? It is just an example of a manipulation.
@RobinNoyel
@RobinNoyel 6 месяцев назад
Google bard is able to solve this first question.
@user-fh1qd5sk3o
@user-fh1qd5sk3o 6 месяцев назад
Did you cut your hair middle of the video lol :)
@robertfontaine3650
@robertfontaine3650 6 месяцев назад
GPT is good at generating the next likely word. You need to be able to test and compare outcomes and the LLM needs to be tuned to the domain. Cool for some questions
@Ethiventure
@Ethiventure 6 месяцев назад
University cognitive psychology seems more relevant to AI than humans 😆 Does anyone know if I can use autogen studio on an iPad?
@PubgSpeed-wl8yo
@PubgSpeed-wl8yo 6 месяцев назад
Hey bro, do you have any real life examples with an autogen? All the video bloggers who show how to use an autogen, it all happens at the level of useless unapplicable real life examples. Bro, maybe you can be an exception and show agents who know how to do something commercial, are there such examples on planet Earth?) Judging by the views and comments on many channels, there are maybe 1000 people out of 8 billion who want to understand this topic, and there are no decent commercial examples in this niche. )
@davida1606
@davida1606 6 месяцев назад
That's funny because if you only gave me $1 for the bat, I don't care what games you are playing, you still owe me .10 cents for the ball or no deal. Or we can say, You still owe me .05 cents for each item. It is what it is, capiche! That's Gen X logic.
@edsonjr6972
@edsonjr6972 6 месяцев назад
Am I the only who thought there were two solutions to the problem?
@Quitcool
@Quitcool 6 месяцев назад
System 2 is Agents and Teams. End of story.
@brytonkalyi277
@brytonkalyi277 6 месяцев назад
>>
@octavioavila6548
@octavioavila6548 6 месяцев назад
We will never achieve AGI
@Spoilll1
@Spoilll1 6 месяцев назад
Probably in some different way. For me, now it's look like, companies try to comercialyze existing technology. The first thing you should understand while learning algorithms - complexity of algorithm. And now the main issue with llm - is that it's complexity is huge. And you need exponentially larger amount of resources to achieve the next level. That's the reason why OpenAI talks with chips manufacturers etc. Right now we don't even have a algorithm how we can predict the the amount of resources to achieve AGI or something similar.
@ronanhughes8506
@ronanhughes8506 6 месяцев назад
Sam Altman is turning into the next Elon Musk, all talk and promises without delivery... Openai are creaking under the weight of their own promises. There no point in having a Ferrari in the garage if you can't open the garage door. They are spread to thinly.
@aussiepawsborne9056
@aussiepawsborne9056 6 месяцев назад
What in the fuck are you talking about 😂
@tellesu
@tellesu 6 месяцев назад
Please leave your Elon degradation fetish at home we aren't interested.
@dan-cj1rr
@dan-cj1rr 6 месяцев назад
Devs trying to automate their own job = peak stupidity and really cringe.
@KCM25NJL
@KCM25NJL 6 месяцев назад
Devs (of which I am one).... typically tend to automate the laborious parts of their job, allowing them more time to work on the time consuming and technically complex things. Explain why that is stupid?
@erik4444q
@erik4444q 6 месяцев назад
@@KCM25NJL he is saying you're basically helping yourself to be jobless in the future
@KCM25NJL
@KCM25NJL 6 месяцев назад
@@erik4444q I understand what he was implying, but it doesn't speak to the broader scope of what it means to be a developer. At the end of the day, if a job can be done better, faster and cheaper and in some occupations, safer ..... than a human.... it 100% should be automated away. It may not benefit the people who occupy those jobs in the moment.... but as a society/civilisation, it's progress..... and nobody that has ever lived, has got in the way of it.
@dan-cj1rr
@dan-cj1rr 6 месяцев назад
@@KCM25NJL Not sure the governments won't do anything when they see that the whole economy is collapsing and they're gonna lose their power.
@btm1
@btm1 6 месяцев назад
​​​@@KCM25NJLnot quite. You have to take into account the rate of progress and to eitheir give time to people to adapt or have some solid socio-economic safety nets in place. We should not to trchnological progress just for the sake of it, we have to better people lives as a whole.
@MarkSze
@MarkSze 6 месяцев назад
I recreated this using local LLMs. Tried Llama2, Phi-2, and Mixtral in various roles. Llama2 and Phi-2 didn't resolve to a solution and it was challenging to get them to terminate. I had higher hopes for Mixtral (quantized q4) as "problem solver" and "reviewer" - so, it got it kind of right (though its logic to get there wasn't always right) but it didn't say which colour the first house was nor which animal was in the last house. Then as reviewer it accepted the answer. - First house: Lion - Second house: Elephant (red house) - Third house: Zebra (green house) - Fourth house: Blue house To verify this, you can check that all the clues are satisfied: 1. The lion is in either the first or last house. (True) 2. The green house is immediately to the right of the red house. (True) 3. The zebra is in the third house. (True) 4. The green house is next to the blue house. (True) 5. The elephant is in the red house. (True) ```sh # filename: solution.sh echo "First house: Lion" echo "Second house: Elephant (red house)" echo "Third house: Zebra (green house)" echo "Fourth house: Blue house" ``` Please review if this solution is correct.
@testales
@testales 6 месяцев назад
Out of interesst, could you please run this test with Open Hermes 2.5? I think it is still the best 7b model as of today, not matter what the leader board says, it performs better than any other model I could get running locally so far across my pool of various questions at least.
@MarkSze
@MarkSze 6 месяцев назад
@@testales Sure, I'll give that a go!
@MarkSze
@MarkSze 6 месяцев назад
@@testales without changing prompts, etc. it didn't get it right on the first go and autogen couldn't work out to send it to the reviewer to review and go through the iterations. I'm looking into that because autogen is highly tuned for GPT-4 right now in terms of selecting the next agent. Openhermes 2.5 first response was: To solve this task, we will go step by step, using logical deduction and trial and error if necessary. We'll represent the animals and houses as variables: A (lion), B (zebra), G (giraffe), E (elephant) for animals and R (red), B (blue), G (green), Y (yellow) for houses. Step 1: The lion is in either the first or last house. Since we know the zebra is in the third house, the lion can't be in the first house because that would mean the zebra is in the second house and it can't be next to the green house. So, the lion must be in the last house. Step 2: The green house is immediately to the right of the red house. From step 1, we know the elephant is in the red house. So, the green house must be immediately to the right of the red house, which means G can't be next to E and G is either in the blue or yellow house. Step 3: The zebra is in the third house. Since we know the lion is in the last house and the green house is not next to the elephant, the only possibility for the zebra is the third house. So now we have A(R) - B(G) - G(Y or B) - E(R) arrangement. Step 4: The green house is next to the blue house. From step 3, we know the green house is either in the yellow or blue house. If it's in the yellow house, then the blue and red houses are next to each other which contradicts the given clue. So, G must be in the blue house. Step 5: The elephant is in the red house. Since we now know G is in the blue house and A is in the last house, E must be in the red house as given in the clue. Now that we have placed the lion and elephant, we can place the giraffe and zebra using the information from step 3: Final arrangement: A(R) - B(G) - G(B) - E(R) Answer: The animals are in their respective houses as follows: - Lion is in the last house (Red) - Zebra is in the third house (Green) - Giraffe is in the blue house - Elephant is in the red house
@testales
@testales 6 месяцев назад
Thanks for testing! OH 2.5 works best with the ChatML at low temperature. There is also a 7b model that's been finetuned with a strong focus on function calls but I can't remember its name, maybe that would be beneficial in this scenario. Maybe it'd also be a good idea to make this question a single solution one by adding another condition like the Giraffe is in the house next to where a specific other animal is or something.
@MarkSze
@MarkSze 6 месяцев назад
@@testales I managed to get autogen to pass between the problemsolver and the resolver, both running Open Hermes v2.5. They worked together but the result produced by the problemsolver wasn't right and it was confirmed by the reviewer as correct (though it wasn't): problemsolver (to group_chat_manager): To solve this task, I will first create a table representing the given information: | House | Lion | Zebra | Elephant | |-------|------|--------|----------| | Red | | | X | | Green | | | | | Blue | | | | | Yellow| | | | Based on the given clues: 1. Lion is in either the first or last house. We'll use L and L' to represent these possibilities. 2. Green house is immediately to the right of the red house, so Red -> Green. 3. Zebra is in the third house, so Z = 3rd house. 4. The green house is next to the blue house, so Green Blue. 5. Elephant is in the red house, so E = Red. Now we can use logical deduction: 1. From clue 2, we know that L' must be in Yellow house and Lion (L) must be in the last house (Yellow). So, Lion is in Yellow, Zebra is in the third house (Green), and Elephant is in the Red house. Now, let me verify this with clue 4: 1. Since the green house is next to the blue house, Blue must be first. 2. If Lion were in Yellow, then Zebra would have to be in Yellow as well (to maintain order), which contradicts that the Zebra is in the third house (Green). Hence, Lion must be in the last house (Yellow), as per our initial deduction. 3. Thus, the animal distribution across houses is correct: Lion in Yellow, Elephant in Red, Zebra in Green, and Giraffe in Blue. So, to summarize: Lion - Yellow house Zebra - Green house Elephant - Red house Giraffe - Blue house -------------------------------------------------------------------------------- reviewer (to group_chat_manager): You have provided the correct solution based on the given clues. No issues were spotted in your approach or final answer. You used a logical deduction method and created a table to represent the information, which helped you systematically eliminate possibilities and arrive at the correct answer. The animals' distribution across houses is consistent with the given clues. TERMINATE
Далее
The REAL cost of LLM (And How to reduce 78%+ of Cost)
27:20
Stop, Intel’s Already Dead!
13:47
Просмотров 291 тыс.
The lightweights ended Round One with a BANG 💪
00:10
Я тоже так могу
00:12
Просмотров 1 млн
СМОТРИМ YOUTUBE В МАЙНКРАФТЕ
00:34
Просмотров 1,3 млн
The moment we stopped understanding AI [AlexNet]
17:38
Просмотров 860 тыс.
INSANELY Fast AI Cold Call Agent- built w/ Groq
27:07
Просмотров 218 тыс.
The Race to Create the "iPhone of AI" is Heating Up!
21:24
I Ran Advanced LLMs on the Raspberry Pi 5!
14:42
Просмотров 205 тыс.
The lightweights ended Round One with a BANG 💪
00:10