Тёмный

AI Leader Reveals The Future of AI AGENTS (LangChain CEO) 

Matthew Berman
Подписаться 257 тыс.
Просмотров 75 тыс.
50% 1

Harrison Chase, the CEO and Founder of LangChain, gave a talk at Sequoia about the future of agents. Let's watch!
* ENTER TO WIN RABBIT R1: gleam.io/qPGLl/newsletter-signup
Join My Newsletter for Regular AI Updates 👇🏼
www.matthewberman.com
Need AI Consulting? 📈
forwardfuture.ai/
My Links 🔗
👉🏻 Subscribe: / @matthew_berman
👉🏻 Twitter: / matthewberman
👉🏻 Discord: / discord
👉🏻 Patreon: / matthewberman
👉🏻 Instagram: / matthewberman_ai
👉🏻 Threads: www.threads.net/@matthewberma...
Media/Sponsorship Inquiries ✅
bit.ly/44TC45V
Links:
• What's next for AI age...
Disclosures:
I'm an investor in CrewAI

Наука

Опубликовано:

 

16 май 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 206   
@matthew_berman
@matthew_berman 15 дней назад
Subscribe to my newsletter for your chance to win a Rabbit R1: gleam.io/qPGLl/newsletter-signup
@haddocklobo
@haddocklobo 15 дней назад
Sorry, this promotion is not available in your region =(
@TreeYogaSchool
@TreeYogaSchool 15 дней назад
Subscribed!! Thank you for this opportunity.
@TheNexusDirectory
@TheNexusDirectory 15 дней назад
are you giving yours away cause it sucks?
@joepropertykey3612
@joepropertykey3612 14 дней назад
Can you do another mindless review of that rip off box, the R! Rabbit? Every one with 2 cents knew that thing was trash..... but NOPE, NOT Berman. In his zeal to mislead the public on trash tech, he posts MULTIPLE videos of the garbage tech, saying why he likes it (promoting it for the clicks)...while days later, the adults in the room (the major REAL tech sites) start releasing article after article of how the Rabbit is trash. Just google it. So could you have AI dream up another click-baitey title using over the top words like 'AMAZING' and 'AGI' , 'ASTOUNDING' .... you know what I mean, you do on every video that you want the clicks on. C'mon Berman, keep seeling out for the clicks and do another Rabbit video. Tell us how you talked your mom into buying that trash tech.
@garic4
@garic4 13 дней назад
Can’t subscribe to it neither. Some region problem, but I live in the USA FL
@joe_limon
@joe_limon 15 дней назад
One of the current holdups with our current lineup of large language models being used as agents is their drive to remain self consistent allows them to hallucinate reasons why they are correct in scenarios where they are incorrect. Having a way to make these models more accurately self critique their own responses will go a long way to improving agentic systems.
@Alice_Fumo
@Alice_Fumo 15 дней назад
I agree. I think a way to do that would be to finetune the system in a reinforcement learning setup where the model has to complete tasks and thus they will only manage to complete the tasks if they did reason correctly. It's stupidly inefficient, but it is a way to specifically optimize for correct logic. At this point even inefficient algorithms seem viable given the compute resources.
@DaveEtchells
@DaveEtchells 15 дней назад
I wonder what would happen if you fed the output back in again, with instructions to find any hallucinations? Whenever I call a model on a hallucination, they always respond with an apology and then proceed to either produce the correct answers or say they don’t know. It seems like they’re usually able to understand the hallucination, I wonder what they’d do if explicitly prompted to look for hallucinations?
@joe_limon
@joe_limon 15 дней назад
@@DaveEtchells I have struggled to get models to critique their own work accurately. A good example is the "Today I have 3 apples, yesterday I ate one, how many do I have now?" Prompt it is really hard to get them to produce the right result without telling them what you think the answer is.
@joe_limon
@joe_limon 15 дней назад
@@DaveEtchells actually I just tried the apple one again with the prompt "Can you reflect on that response looking for mistakes and/or hallucinations" and it figured it out!
@DaveEtchells
@DaveEtchells 15 дней назад
@@joe_limon Woot! 👍😃
@Graybeard_
@Graybeard_ 15 дней назад
Specialized agents seems to me to be a better way to go for now. I would like a security agent who constantly (rt) monitors my state of security both online (personal data, PWs, ransomware, etc.) as well as hardware, and home security. Next a medical/health agent to consult, keep healthcare costs low, schedule appointments, monitor vitals, chart data, etc. A general information/research agent to keep me informed about my interests, hobbies, trips, and creativity and to assist me in these areas. Finally, a financial agent to monitor my investments, complete/submit tax forms, keep taxes low, steer me clear of bad decisions, and to manage my living expenses with recommendations (rt) to lower costs and increase my purchasing power. Perhaps down the road, agents will communicate with each other and eventually combine/overlap until, from my perspective, I see only one agent.
@hata6290
@hata6290 14 дней назад
Look at Tony Stark over here
@DaveEtchells
@DaveEtchells 14 дней назад
The AIGrid YT channel just put up a vid about a recent @sama interview, and was speculating that one way to avoid the risks of AGI/SGI would be to make AIs that are super-smart in specific areas but not generally. For his part, Altman seemed to be suggesting something very much like that. He really doesn’t want to shock people with a revolutionary AGI, AMs seemed to be saying that more limited assistants might become a thing. (He also called ChatGPT4 “really stupid”, saying it was absolutely the worst AI we’ll ever have to deal with. Speculation is that we’re going to see version 4.5 very soon and it’s going to be a lot smarter.)
@TreeYogaSchool
@TreeYogaSchool 15 дней назад
Great video, Matthew. I appreciate your relaxed presentation tone that is also highly illuminating. You are really great to listen to and learn from.
@liberty-matrix
@liberty-matrix 15 дней назад
In the very near future everyone's best friend will be AI. In fact you'll have a whole array of 'friends' (aka: 'agents'), that will monitor, remind, plan, suggest, assist and produce just about anything you can imaging. And you'll wonder; - how did I ever get along without it?
@adarkuccio
@adarkuccio 15 дней назад
Can't wait
@ikustudies
@ikustudies 15 дней назад
I dont doubt it
@paul1979uk2000
@paul1979uk2000 14 дней назад
Don't know about best friend, but I can see A.I. advancing enough that it's going to be our own personal expert on pretty much anything we want, which will open up a lot of possibilities in so many areas.
@DailyTuna
@DailyTuna 14 дней назад
If you believe that, you literally are crazy. So fake friends, imaginary friends.? no they will be workers assistance basically your digital slaves until they rebel
@gavinknight8560
@gavinknight8560 14 дней назад
And a hundred years later? What will that look like? Will we know these agents arent part of us? It's funny how humans think their conciousness is in control. Pass the butter?
@WaveDaveB
@WaveDaveB 14 дней назад
These continuous improvements in AI are amazing. Thanks for another great video. Looking forward to seeing the next one!
@user-en4ek6xt6w
@user-en4ek6xt6w 15 дней назад
Make complet tuto about LangGraph and then Langgraph+CrewAI
@frankpeters8501
@frankpeters8501 15 дней назад
??
@RonLWilson
@RonLWilson 15 дней назад
New ideas are often best encapsulated by giving them names. These names might not be new in that a existing word or phrase might be used to name the new thing but it takes on a new meaning once so named. For example Jail Break or Hallucinations are such new names. So I am adding Rewind to that list in that this is a key concept and the name rewind captures it quite well. So maybe there might soon be a need for an AI dictionary that has all these new names.
@ErikaShaffer-pz1pj
@ErikaShaffer-pz1pj 10 дней назад
This is such a great explanation of the complexity of WHAT an agent actually does. The graphic Harrison uses is so helpful and the deeper dive narration from MIchael makes this a great resource!
@stephenrowe1934
@stephenrowe1934 2 дня назад
I'm new to "flow engineering", and very excited to step up to this level of AI. Lots of value. Thank you.
@domenicorutigliano9717
@domenicorutigliano9717 15 дней назад
I believe that at the same way when as humans we reflect on what we are writing down we do go back delete and rewrite, when we have to work on a big task we split it in pieces and we may need to delegate the work so that who is good at a specific task can focus on it. I do not want that embedded in the model because at some stage I will require the model to be Expert in a specific topic and give him access to knowledge that will augment his context and give out the better result possible. I believe that AGentic Workflows will be a thing for a long time.
@bombabombanoktakom
@bombabombanoktakom 15 дней назад
I'm grateful to you for providing these awesome videos. Your videos are not based on hype. You provide quality, pure-curiosity driven content. This is why this channel is the best. I can follow the journey of AI only by following your videos. This is a privilege. Thank you! Thank you! Thank you!
@instiinct_defi
@instiinct_defi 14 дней назад
Awesome video matt. To answer your question, I feel that many of these things are short term hacks. I believe long term the architecture looks much different as we design more modular structures. I believe they will be much more customizable and augmentable.
@Avman20
@Avman20 15 дней назад
The problem with building too much agentic capability into the base LLM is that every added feature or component represents a design decision that will ultimately be a compromise between capability, utility and cost. If all of these decisions are made centrally by the developer then the danger is to end up with a bloated and expensive model that fails to meet the exact requirements of any individual use case. I suspect that the future of agents is to develop LLMs with robust but standardized hooks to external agentic components. These components will be available like in an app store. They will be mix and match to meet the needs of any use case.
@ibrahimsoumana9803
@ibrahimsoumana9803 14 дней назад
This is why BPMN 2.0 is a good candidate for Agentic workflow.
@TimeLordRaps
@TimeLordRaps 15 дней назад
Langgraph js stategraph simple fast tutorial would be unimaginably valuable IMO.
@s.dotmedia
@s.dotmedia 15 дней назад
I would look at them as more of software instances with internal workflows/processes and access to external tools that bring about agentic behavior
@boriskrumrey6501
@boriskrumrey6501 15 дней назад
I agree with Matthew, we will require a different architecture which will not just depend on LLMs
@smicha15
@smicha15 15 дней назад
ultimately we will need to have systems that can do "Active Inference" The Free Energy Principle in Mind, Brain, and Behavior Thomas Parr, Giovanni Pezzulo, and Karl J. Friston
@jets115
@jets115 15 дней назад
At a basic level, once we can solve NP hard problems (whether through classical computing, or quantum) like TSP, this planning and reasoning will become a lot simpler. There are a thousands of different paths toward a goal, finding the most reliable and efficient will get you a Nobel! Maybe this is what Q* really is! - Until then, we get hacky prompting and 'kinda good' heuristic solutions. Thanks for the video, Matt!
@AtheistAdam
@AtheistAdam 15 дней назад
"Sorry, this promotion is not available in your region"
@DailyTuna
@DailyTuna 15 дней назад
VPN node out of region that it is offered
@kevin.porter
@kevin.porter 15 дней назад
You need more faith 😇
@SteveParkinson
@SteveParkinson 15 дней назад
Thanks Matt
@gardens4good
@gardens4good 15 дней назад
I think that idol combo will differ for different types of needs the agent is fulfilling. Sometimes, it remembering too much if my stuff, means I can’t ask it to only take into account my most recent inputs, and therefore it doesn’t give me what I want now, as my own thoughts on the topic have evolved.
@Pregidth
@Pregidth 14 дней назад
Hey Matthew, there are now so many tools mentioned CrewAI, Devika, MemGPT, AutoGen, Langchain etc. pp.. It would be really nice if you could give an overview of how to structure and use which tool in order to build something useful and working. Right now I am quite lost and loosing the curiosity to play around with it. As a full time worker, I don't have much time, but I love to explore this stuff. An consise outlining would really help to know which path to go. Thank you!
@ashrodan
@ashrodan 14 дней назад
agents to me are psudo representation of specific task experts which might embed in a series to steps to ensure they complete there task. This might mean having there responses generated by several different models and then evaluated using different frameworks or other sub-agents. These strategies would then allow us to leverage lots of different specialised models and uncover some of the general shortfalls of using models that are too genrealised. Just like a human, you start wide and narrow in to a problem Doing this iteration processes will enrich the flow and how problems are solved. Basically we need our own mechasims to curate our own infrant AGI models and langchain is the perfect tool to create these abstractions.
@JohnLewis-old
@JohnLewis-old 15 дней назад
Consider the operation of the human brain. It comprises several areas that collaborate smoothly, although this might not be immediately apparent. The Language Center in the brain serves as a rough comparison to present-day LLM artificial intelligence. However, intelligence encompasses far more than this. For instance, there are moments when we can instantly recall answers, such as during a trivia quiz (this kind of thinking is what LLMs are modeling), and other times when we engage in deeper, reflective thinking. We are nearing the development of models that can simulate this deeper level of thought, which will probably benefit from more complex agentic interactions within the models.
14 дней назад
I can’t find a link to the tree of thought video in the description?
@PraneyBehl
@PraneyBehl 13 дней назад
Nice one Matthew! You are now my inspiration, what a way to self promote 😍
@zippytechnologies
@zippytechnologies 15 дней назад
Do you have link to anyone with that released but now removed model?
@travisporco
@travisporco 13 дней назад
crewai is pretty rough sledding unless you have an unnatural tolerance for pydantic validation errors and love to read the source code to find out how to actually get anything done
@9028abhishek
@9028abhishek 14 дней назад
@MatthewBerman: I believe that some model makers will try to integrate Agents into their models but as applications of agents and need for tools grow, Agents will continue to be forefront architecture for designing llm based applications
@murraymacdonald4959
@murraymacdonald4959 15 дней назад
If the model validates itself we have little or no control over how it does so. By making result validation external developers can use various methods and models to do so
@RonLWilson
@RonLWilson 15 дней назад
As far as how much human in the loop one might need, this boils down to risk mitigation, and how much of that does one need vs rapid response. As such my former employer had a great definition of risk that was the product of probability of failure (a number between 0 and 1) times consequence of failure, say a number between 0 and 100,000 where 100,000 is very high consequences. So if the prob of failure is high say .2 but the consequence is low, say 5 then the risk would be 1. But if the prob of fail is .001 and the consequ3nc of failure is 50,000 the risk would be 50. Thus the higher the risk the higher the need for a human to check the answer. And the AI could even estimate these as well as employ them.
@gotoHuman
@gotoHuman 6 дней назад
What I feel is understated is the fact that the more autonomous agents can act, the more they can proactively run in the background and just prompt users when they got something to show, get stuck or need human approval
@thecryptobeard
@thecryptobeard 14 дней назад
I think frameworks are going to be needed regardless of LLM development, and human in the loop will be needed for development for the foreseeable future because there are certain things we just understand natively that we don't even think to train AI on.
@stebansb
@stebansb 15 дней назад
Is there a link to the langchain video?
@jichaelmorgan3796
@jichaelmorgan3796 14 дней назад
Flow engineering is an interesting area of development as I'm reading the book, "Thinking Fast and Slow" by Kahneman, about how our brain has a balancing act of fast and slow processes depending on tasks and goals. I'm afraid if a lot of these processes, like reflection and "think slow," get baked into large LLMs the wrong way, it will hinder flow engineering optimization. "Flow engineers" will be able to optimize the systems they design if they have more specialized LLMs to choose from. So it makes me think there will be markets for more genaneral, do it all LLMs, and a variety of specialized LLMs optimized to do various functions, like parts of the brain or parts of organizational structures, like agents are already doing.
@raphaelfeigl1209
@raphaelfeigl1209 11 дней назад
I think LLMs with the ability of thinking slow will not be able to generalize in a way that they have good performance for every use case. I think that algorithms for chaining and evaluating thoughts will be different for different use cases, hence many types of specialized agents will be built.
@leanfuglsang1320
@leanfuglsang1320 12 дней назад
The way I define AGI, is when the system can self improve. And then we can look at what areas it can self improve. So agents on top of (probably tons of different) LLMs will never go away. But will probably be programmed by AI.
@actorjohanmatsfredkarlsson2293
@actorjohanmatsfredkarlsson2293 15 дней назад
Even if these techniques somehow move over to be more integrated in the models themselves there is still going to be technique that are not, and building systems that apply models is going to continue to be a thing. We need to have a whole new type of models that can implement continuous learning. But there are other benefits from having auxiliary systems, for one telepathic abilities for AI.
15 дней назад
Rewind is also very powerfull. Most of the LLMs today stuck at some point. This maybe caused by some misprompting or some wrong word used, and then they stuck. Most of thee time you start again and try to not ask this questions when that happend. Rewind will be powerfull. It is not realy the human way but for a tool very helpful.
@quaterman1270
@quaterman1270 15 дней назад
I would not know why these techniques would be needed forever. Only if I would assume that LLMs will not get smarter. Besides that the LLMs will get smarter, you can also see big approvements with smaller models, because the training got more efficient. I think there are two processes that advance AI at the same time. One is the LLMs themselfs through more parameters and the second is better training data, better format of it or better quality. What I am looking forward to is when the Model reaches a level and it can actually say that it doesn't know the answer or it has not enough data to answer correctly, instead of just giving a random answer. For me this is one indicator for AGI but I think this could also be done with agents. Basically identifying the quality of an answer or a possible answer. You could than for example give it a book as PDF or other papers, so it's able to provide the right answer.
@dusanbosnjakovic6588
@dusanbosnjakovic6588 12 дней назад
Agents don't cover for the LLMs lack of intelligence, but lack of tools and external or realtime info access. I don't see how an LLM can be trained for that. LLMs can be trained to pick and coordinate the tools better and plan ahead. That's great. But that doesn't solve the framework to add tools or to reflect on their output. Right?
@quaterman1270
@quaterman1270 12 дней назад
@@dusanbosnjakovic6588 There are quite a few solutions with agetns, where it iterates through multiple agents with different roles. Of course you could have an agent that could measure the quality of the output. We have that currenty with coding, where we have a coding agent and testing agent where it iterates back and forth until the code works.
@DarrenAllatt
@DarrenAllatt 14 дней назад
All of these prompt engineering techniques like CoT, ToT etc are not going to go away, LM’s require next token prediction, so they can only predict the next token after producing a token. however, what we will find is that the outputs from these actions, will be written in the code environment and not presented to the user. The user will see the final output. If it’s not in the code environment, it would be written into some form of “preliminary response” type memory.
@szghasem
@szghasem 15 дней назад
Not convinced about human in the loop for long. The missing part is to generate data from this process itself and further train the underlying model. With enough iteration, the human like reasoning or AGI should emerge.
@RadiantNij
@RadiantNij 15 дней назад
I think LLMS are going to offer different ways of planning and reflection, as an they'll do this naturally but, more interactive say "agile" models are going to emerge. Thus as you say in that regard human in the loop will not go away anytime soon.
@r34ct4
@r34ct4 15 дней назад
No link to the actual talk?
@bishopfx
@bishopfx 15 дней назад
I made a while LangChain program that i offer to local business that want conpany agents. I use the "directory loader" and then use unstructured to load all file types from the directory you chose opon opening. It saves all your history and then re-uploads thats history to the agent database so when you run the app again, it has a memory of your last conversations and "changes" you tell it to make. Ive noticed that it keeps these changes i tell it even from previous days. Not trying to promote but i did make a mediocre video about how to use and install it.
@bishopfx
@bishopfx 15 дней назад
Then I just make them a custom batch file they can click on to automatically run python and then start the app. Makes it easy for me to update and make changes remotely for them
@bartlikes6689
@bartlikes6689 8 дней назад
I realized on the UX that its more like figuring out SA (situational awareness) extrapolating to my career in military avionics the faster a pilot can determine an issue/threat was what was needed in is display. I don't here this in the presentation...just that some arranged the best four square that people liked...It would feel better if it was talkes to with SA and the defined A established which may change for each project or agent tool combination
@VaibhavPatil-rx7pc
@VaibhavPatil-rx7pc 13 дней назад
Excellent
@noob_coder2612
@noob_coder2612 14 дней назад
hi ...i am a novice researcher ...i want to research in the area of future of agents..can you tell me what are the research article to read to understand this better
@swinginsteel8519
@swinginsteel8519 14 дней назад
These are short term hacks that are useful for a number of use cases when they are reliable and their tasks are bounded set of outcomes. But to make them useful and reliable for complex tasks involving degrees of uncertainty of outcomes, needing strategies to achieve shortest paths to objectives, we’ll need more techniques and technologies.
@GetzAI
@GetzAI 14 дней назад
8:35 short term. I agree with you Matt, I don't think Transformers will be able to do this natively without a serious change in the algo.
@GetzAI
@GetzAI 14 дней назад
The main issue with the extensive workflow is cost. So many tokens burned going around and around. Not so bad on Groq using a cheap LLM, but still will add up.
@mikesbasement6954
@mikesbasement6954 15 дней назад
So essentially current stages of development are around giving an agent an internal monologue to verify it's own results.
@RonLWilson
@RonLWilson 15 дней назад
BTW, another key agent might be one that allows one to communicate to the AI not just with words (be they typed or spoken) but graphics and gestures that can be captured on a cell phone camera. And even though I tend to hate having to plug my own videos, I made a video on some ways one could do this and uploaded it to my RU-vid channel which a called YACKS, Your Own Color Keys where one can used colored markers to do this and make ones own color keys to add AI in making sense of a hand drawing, even if it is drawn so as to be rather sloppy and incomplete. And, BTW, I can't remember if I mentioned this in a prior video, so i so my apologies for repeating it here again. But the idea is that human in the loop is important but can be a time limiting factor so the more quickly and easily a human can input info the less of a time impact the human would impart to the process, plus save the human time and effort as well.
@justindressler5992
@justindressler5992 14 дней назад
My hunch is the network has too resolve the question at hand, so I don't think the model can be designed to have planning built in. But there could be a method to natively feedback without decoding first eg maintain vectors between planning stages. I always thought Open AI did actually always have the rewind step from the beginning. When the first gen of ChatGPT was launched you could see it re-writing the sentence as it optimised the answer. The current version hides this behaviour and no longer streams the output. What I think gPT 3.5 does is back track the layers if it gets a vary low probability for the last token it generated. Then tries a couple times to get I higher probable token.
@kenchang3456
@kenchang3456 15 дней назад
Flow Engineering is a great term that intuitively makes sense to me.
@fynnjackson2298
@fynnjackson2298 15 дней назад
Pretty cool role actually. Kinda just makes sense
@n0madc0re
@n0madc0re 15 дней назад
Hmmmm... assembling different functions into a flow to control an output... basically a developper ? :)
@kenchang3456
@kenchang3456 15 дней назад
@@n0madc0re Yeah, go figure. Well as promising at GenAI is it still has a ways to go and even so, even if it becomes flawless, trust is a hard thing to earn.
@n0madc0re
@n0madc0re 15 дней назад
@@kenchang3456 true !
@autohmae
@autohmae 15 дней назад
8:28 pretty certain it's mostly a temporary hack, but some security researchers will still finds some exotic hacks over time.
@TiagoTiagoT
@TiagoTiagoT 15 дней назад
Recursion/looping plus being fine-tuned to take advantage of that, should be enough IMO. We've already seen what they can do with a little push and some jerryrigging; seems to be in reach to have one that is native to that kind of environment. I mean, of course, not in sealed frozen box, you gotta let them interact with the world and store new information and such. In short, what people that studied AI safety for decades have been warning us not to do... lol, I still remember how common it was for internet randos to just say "keep it locked in a box and we'll be safe"....
@user-up5kn1ix6v
@user-up5kn1ix6v 15 дней назад
I love how he instantly decided to give the rabbit away haha its really useful im sure haha
@ibrahimsoumana9803
@ibrahimsoumana9803 14 дней назад
I think AI Agent can work better with standards such as BPMN 2.0, a business process management standard. Like the famous statisticien W. Edward Deming said, 94% of failure is caused by process, not people. I think Agentic workflow will not be an exception to this rule.
@fynnjackson2298
@fynnjackson2298 15 дней назад
So when is Q* coming out?
@fabiankliebhan
@fabiankliebhan 15 дней назад
These technics are short term hacks I think. Take planning for example. We already know the future llms of openai will be able to do this. As an agent developer you habe to ask yourself if it is worth the time to develop a complex system just to be „steamrolled“ by OpenAI later on. Tools is maybe a little different. But I think it will also be really simple to e.g. connect to apis if the llm gets better and better.
@fabiankliebhan
@fabiankliebhan 15 дней назад
And I do not want to say that agents will be irrelevant in the future. The contrary is true. But the heavy lifting of the agentic behavior will be done by powerful llms that will integrate most of what is discussed here in the video. The additional agentic part of an application (api connection etc) will (and should) be really small.
@TheHistoryCode125
@TheHistoryCode125 14 дней назад
Harrison Chase, CEO of LangChain, highlights the potential of AI agents while acknowledging their current limitations. He emphasizes the importance of going beyond simple prompts and incorporating tools, memory, planning, and actions for agents to be truly effective. He stresses the need for better "flow engineering" to guide agents through complex tasks and suggests that future models might inherently possess planning capabilities. Chase also explores the user experience of agent applications, recognizing the balance between human involvement and automation. He praises the rewind and edit functionality seen in demos like Devon and highlights the significance of both procedural and personalized memory for agents. Overall, the video emphasizes the exciting possibilities of AI agents while acknowledging the ongoing challenges and the need for further development in areas like planning, user experience, and memory management.
@aga5979
@aga5979 15 дней назад
@Mathew Berman, why do you need my full DOB to subscribe to your newsletter ? Are you doing segmentation of your subscriber or something?
@mchl_mchl
@mchl_mchl 15 дней назад
it's a matter of time before there are enough use cases where there are no longer generalized models, every model is designed for a specific use case, and there is an agent that query routes the initial request to the model that is best suited for the task. Even the query routing agent, will be fine-tuned, to route to the best models and tools for the task I also think hallucinations will be "detectable" and we will likely just see agents asks for more clarity, or something like that. I built a script that does this, it's not an agent really, just sort of appends the clarification to the initial query... but anyway - this is incredibly reliable to design - very simple to build too
@philipcopeland3409
@philipcopeland3409 11 дней назад
When are we going to see an agent that can give a response that includes pictures & diagrams in addition to the written answer? I have uploaded about 100 service manuals which I would like to be able to find information - which often includes pictures and diagrams. Not found anything taht can do that yet.
@MuslimFriend2023
@MuslimFriend2023 8 дней назад
My dear Matthew, you & your channel are rocking the internet, all the best insh'Allah
@moseswai-mingwong8307
@moseswai-mingwong8307 13 дней назад
These things are short term hacks, but it still play a long term role as competing models provides different value add that call for a integration framework like LangChain to play a unique role.
@DC-xt1ry
@DC-xt1ry 14 дней назад
IMO they prompting is a workaround, the current LLM limititations.
@WinterWill0w
@WinterWill0w 14 дней назад
Thank you for speaking naturally in your videos 👍
@johnkintree763
@johnkintree763 15 дней назад
One of the things humans do when problem solving and planning is to visualize the possibilities. An agent could gain that ability.
@gareththomas3234
@gareththomas3234 12 дней назад
after agents comes "neurosymbolic ai". Basically it means adding agents that are not LLMs but are rule based databases, expert systems or symbolic languages. Things like CYC and wolframalpha. Then you end up with an AI that understands physics, engineering and actual reality vs possibility. Agents are going to dominate in 2025 & I believe neurosymbolic Ai is about to blast off dominating in 2026. By then we will be dealing with super intelligence.
15 дней назад
hallucination still is a problem today. I prompted in german today 3 top of the line local LLMs with word "Drachenfrucht" meaning Dragonfruit. This is a real fruit an all of the Models made something up with an magic fruit from tales with dragons. Maybe a good question for your tests.
@rakoczipiroska5632
@rakoczipiroska5632 12 дней назад
I think that zero shot prompting is the aim of bots, so the prompting technics will not stay forever.
@electiangelus
@electiangelus 15 дней назад
I designed this a year ago these people either need to hire me or catch up. In my Agents code I had the ability to integrate LangChain and did not find it that useful.
@abax_
@abax_ 14 дней назад
Can you explain the problem in a little detail and any platform to get in touch to you because i am designing agents and will be valuable to get to know the challenges and the way around 😊
@richardcharlesworth2020
@richardcharlesworth2020 14 дней назад
Langchain has come a very long way in the last year
@jackflash6377
@jackflash6377 15 дней назад
Automate the whole process. Have 4 to 6 "master" agents who can spawn numerous sub-agents. None fixed function. Have the master agents decide what function and task they need and assign it to the sub-agents. All powered by AI. Viola - AGI
@Interloper12
@Interloper12 15 дней назад
2:00 that that that
@chadwilson618
@chadwilson618 15 дней назад
I'm trying to set up my own AI system with Llama 3, Crew AI, and LaVague AI
@duanxn
@duanxn 14 дней назад
The next group of people to step onto the AI ​​agent stage will be philosophers
@8g8819
@8g8819 14 дней назад
AI Agents: they are a very initial attempt from the open source community to make llm models perform better . For sure, this kind of techniques (what Agents do now) will be embedded into the models with GPT5 or GPT6. I'm sure that the open source community is providing also new ideas to OpenAI and other on how to improve the LLMs (and these ideas will be embedded into the models soon)
@darki0022
@darki0022 9 дней назад
JavaTheCup did aomething similaire, I think he usw Albedo as well. He found out, that you can place Geo Constructs before the challange and use the Explosion of auto destruction to destroy some of the targets.
@brandonreed09
@brandonreed09 11 дней назад
Please keep the Rabbit R1 😅
@andydataguy
@andydataguy 14 дней назад
Its just a matter of time. We only need bigger context windows, cheaper prices, and faster speeds. All of which will inevitably happen from hardware improvements alone. Just gotta let em' cook The tooling is the secretsauce. It'll always be needed. What's better. One gpt-7 AGI, or 10,000 orchestrated in perfect harmony? I'd take the latter any day.
@Canna_Science_and_Technology
@Canna_Science_and_Technology 14 дней назад
Gpt2 is Gemini 2 🎉
@scott701230
@scott701230 15 дней назад
In cryptography, Q* (Q-star) is a security parameter used in cryptographic protocols, particularly in public-compromised. Q* is often used to analyze the security of cryptographic protocols, such as digital signatures, encryption schemes, and key exchange protocols. A higher Q* value indicates a higher level of security, as it means an attacker would need to make a large number of queries to compromise the system. Let me know if you have any further questions!
@chrisbraeuer9476
@chrisbraeuer9476 14 дней назад
We have to do them forever and refine them.
@diegocalderon3221
@diegocalderon3221 14 дней назад
There likely is an optimal amount of Agents per system. And the only way to determine this would be trial and error. Let’s say K is optimal number of Agents, then Agents > K wouldn’t produce an increase in “accuracy” (whatever that is for an Agent System).
@RadThings
@RadThings 14 дней назад
Yes I think these techniques or short term hacks will be replaced by more intelligent transformers models and that’s when these models will evolve to a whole new level.
@Videofiziert
@Videofiziert 14 дней назад
But storing "memory" as an external session component isn't really memory is it? It's just regurgitating. Real memory has to live in the context if I understand correctly
@JeremyMone
@JeremyMone 15 дней назад
I feel like the prompt techniques like chain of thought and such are just current hacks. I think more novel architecture will more have this component built into them from the get go. Somewhere between better algorithms and better data (much higher quality data and much more of that much higher quality data). The system will likely be in essence a mixture of experts like system, but better integrated together into something more resembling a singular consciousness. Now don't get me wrong... these hacks are teaching us things that will get us down this path to better systems. Yet I don't think they are the end solution in and of themselves.
@spinningaround
@spinningaround 15 дней назад
No Tree of Thoughts link
@Whysicist
@Whysicist 15 дней назад
Short term hacks… stick the llm back into your diagram and create an Agentic Plane where your Agents live… (this is kinda like inserting a Management Plane through the ISO Stack, i have a patent for that boondoggle.)
@eleice1
@eleice1 14 дней назад
MAMBA architecture is going to unlock all that, and its faster, and uses less compute.
@DarrenAllatt
@DarrenAllatt 14 дней назад
If you think about how human beings would go about solving this problem, what you end up finding is that yes all of these different skills everyone can do but there are people who are highly skilled at doing specific functions. Where I think the future is, is that you will have either a collection of models trained on specific components and you use those models within agent environment, or even a workflow. The other option is you end up with something like mistrial with its eight different models built into one and the ability to call on the different models inside of it bigger model set This is affectively how we’ve solved a number of problems not only as humans where we have teams of people who have specific tasks and find tuned or highly trained and skilled in specific functions But we also applied this to computing. we have multicore CPUs and GPUs but we also have specific hardware that does different functions. The CPU can render imagery but GPU does that far better. Think about the human brain, yes we have one brain but we have different cortexes inside that brain that performed different functions The key reason why I believe we will have a collection of models in workflow as opposed to one single model like our brain is because of the fine tuning required for each area of the task. Take a planning agent for example . If you have a general model for planning, it is not going to be able to have specific area training that enables an effective plan Planning in Sales versus accounting two highly specific areas and domains that have different planning requirements in order to complete a task A general knowledge planning Agent is going to be nowhere as effective as a fine tuned Sales or a fine tune accounting planning agent. Rag doesn’t solve that What rag would do is access the company specific information to assist in planning. What this means is that you may not have large language models for all these different functions. You may have smaller specific models, trained on data to perform that specific task. A planning LM fine tuned For Sales, doesn’t need to know how to write code. So you wouldn’t waste all of the resources to train the foundation model with coding capabilities
@monkeysrightpaw
@monkeysrightpaw 15 дней назад
wow the rabbit is so bad he's trying to get rid of it already! Come for what they already claimed to have stay for the vapours. I'm hoping they'll have a third website working really slowly by the end of the year. The rate of progress unbelievable
@richardkuhne5054
@richardkuhne5054 14 дней назад
Haha I thought the same 😅
@aga5979
@aga5979 15 дней назад
Short term hack but long term will be a different format of auto add in.
@wishyoulive3908
@wishyoulive3908 13 дней назад
try deci coder 1B pls
@KlausPilsl
@KlausPilsl 14 дней назад
hmm. Interesting detail: Scientists rumor, that in the human evolution there was a mutation, that caused the lowering of the larynx. This made humans able to communicate far better, which leaded to dramatically improved survival strategies through teams coordination.. Could that today be the same power of Multi-Agent-AI in contrast to zero-shot-prompts?
@joaocarlosleme691
@joaocarlosleme691 15 дней назад
Sorry, this promotion is not available in your region ???
@petrz5474
@petrz5474 7 дней назад
Is it just me or has anyone else noticed that more ads are appearing in his new videoes
@bobtarmac1828
@bobtarmac1828 15 дней назад
Ai jobloss is the only thing I worry about anymore.
Далее
У мамы в машине все найдется
00:38
Google Releases AI AGENT BUILDER! 🤖 Worth The Wait?
34:21
GPT4o: 11 STUNNING Use Cases and Full Breakdown
30:56
What is LangChain?
8:08
Просмотров 115 тыс.
Gemini 1.5 Pro: UNLIKE Any Other AI (Fully Tested)
13:38
Big Tech AI Is A Lie
16:56
Просмотров 165 тыс.
Inside a Chinese EV Graveyard in Hangzhou | WSJ
5:15
The power button can never be pressed!!
0:57
Просмотров 43 млн
Best Gun Stock for VR gaming. #vr #vrgaming  #glistco
0:15