@davidondreji with you program can I build my own private agents And what’s the kick off price I asked Chat gpt and it said 5.75 usd to have 7 agents with a over seer
So the first 60% of this video built up the expectation that we are gonna use offline llama3 py agents, but then at the very end you switch it to using the llama3 available through groq's api. Although you do get the agent working with llama3, its a bit misleading, and it would have been better to straight up say: I havent got Llama3 working offline*, but here is how I got it working through groq's API. *Edit for clarity: llama3 working offline with crewai in the context of this tutorial. *Edit 2: Others have tested the offline ollama3 model recently, and state it now is working. At the time this video was recorded, CrewAI wasn’t collaborating properly with Ollama3 offline-and that was the issue, which should now be fixed.
@@ardagunay4699I think it has more to do with crewai not being correctly configured with the new llama model. It’s possible to use llama2 offline to do the same project, but if you repeat the same steps for llama3 there is a clear breakdown in the crewai step of the process
I also noticed that as of 2024-04-27 the Llama3 (local LLM) does not work with CrewAI. However, you can replace Lllama3 with Erik Hartford's excellent model "dolphin-llama3" and you get the expected result. Dolphin-llama has the additional advantage of being uncensored. Cheers! Keep up the good work!
more then anything i appreciate your showing when and where these processes dont work . the trouble-shooting is a critical part of process and the overhype of these systems is most deceitful when the user actually tries to integrate them and runs into all sorts of issues that were hidden by showmen . really excited for llama3 finetunes and more powerful agentic systems , thinking recursive self-debug and finetuning for the generation of the most understandable debugable code , with proofs and tests , could build a solid foundation .
I have no coding experience and i copied every code you wrote and my crew worked without groq, which is really something! Thanks for the tutorial a lot
@@Instant_Nerf @Instant_Nerf With any big model, you're not going to be able to make much use of any consumer GPU like the 3090. He can run the 8B parameter model with it, but the most sensible route is cloud computing for stuff that's big, which he is doing with Groq. If you're going to run LLM's an absurd amount of time, sure. Get a rack of GPU's or get a high end server processor with large amounts of fast server memory. But for most people, this is not a good use of money.
Bro I just want to say thank you for making this content. It's always super informative in an easy-to-understand format. Out of the 50 (exaggeration, but there's alot) I always find myself looking for your videos first. I've always had an interest in programming and hacking but didn't do much with machine learning. But know I'm a man obsessed. Mainly because of how critical it is for normal civilians to learn how create and train these things. I truly belive the future of humanity depends on it. If corporations kill us all building agi and it escapes especially Gemini we are so screwed because the odds are against us that it finds any value I'm something(humans)that's killing the thing it exists on or believing we will attempt to shut her down. Or government with runaway military ai because they couldn't wait until all the bugs were out when they deploy it.
I appreciate the clean simple video. I didn't run into the same issues as you with Llama 3 though. Still it gave me enough to get my head around this though.
Yea, if you still have to pay for some damn service... I'm having issues getting autogen working with llama3:8b-instruct-fp16 and the teachability module(runs at 42+ t/s though!) It almost never decides to flag things as important/worthy of remembering!... But just started messing with that today. If you have a solution to using agents with only localLLM, no api keys, please let us know! TL;DR - it fails to understand it's asked to basically form a question about what it's supposed to store in the db, so that it could be found that way, and the analyzer just keeps asking this same question every time. Probably need a better analyzer?..hm -------------------------------------------------------------------------------- teachable_agent (to analyzer): Imagine that the user forgot this information in the TEXT. How would they ask you for this information? Include no other text in your response. -------------------------------------------------------------------------------- analyzer (to teachable_agent): What is the context or background information mentioned in the provided text that I should be aware of? Can you remind me what important details are missing from the passage and need to be recalled? --------------------------------------------------------------------------------
@@PrinzMegahertz I'm interested in your setup. I'm using it with LMStudio and getting the same result ... the Executor kicks off, and the GPU ramps up, but nothing happens.
@@richardchinnis I tried with crewai and llama3:8b locally on my computer, 15 minutes later, it still is stuck at > Entering new CrewAgentExecutor chain...
So this started as a good tutorial, ran into some issues and kinda just ended. I did manage to get the groq to work in the end. I do have Llama running on a Docker container. Now i would like to combine both of those . Thanks for the tutorial
I would like to know your thoughts on this: With a limitation that restricts the potential of Llama materials to enhance other major language models. Researchers and developers often want to compare or fine-tune different models to improve their performance or tailor them to specific task? However, due to the restrictions in the licensing terms, they cannot freely utilize the Llama materials to do so unless they specifically use Llama 3.
Introduction to Building AI Agents - 00:00:00 Performance Comparison of Llama Models - 00:00:35 Getting Started: Required Tools and Downloads - 00:01:06 Downloading and Setting Up Llama Models - 00:01:36 Basic Chat with Llama Locally - 00:02:28 Setting Up VS Code and Writing Initial Code - 00:02:56 Installing Required Packages - 00:03:02 Defining and Importing Models and Packages - 00:04:04 Creating the Email Classifier Agent - 00:04:37 Creating the Email Responder Agent - 00:06:25 Defining Tasks for the Agents - 00:07:00 Defining and Running the Crew - 00:07:37 Initial Run and Troubleshooting - 00:08:13 Adding the Gro API for Better Performance - 00:09:26 Final Setup and Testing with Gro API - 00:10:04 Conclusion and Call to Join the Community - 00:12:07
GREAT VID! FINALLY THE INSTRUCT!! ...may I just ask a sidebar question Re. Your VS Code editor window behavior? PLEASE?!! How have you set your VS Code preferences so that the longer length strings you've written for the classifier and responder classes (specifically, the strings stored in 'goal' and 'backstory') when they reach the edge of the editor window they wrap to the next line down, WITH THE NEXT WORD CONTINUING FROM THE CORRECT INDENTATION POSITION (Directly beneath the declaration like: To demonstrate/explain here: || = indicated the edge of the editors window Your editor looks like this: responder = Agent( (\t) goal = "qwer|| tyabcdefghijklmnop", ) My editor looks like this: responder = Agent( (\t) goal = "qwer|| tyabcdefghijklmnop", #
would really be cool how you could get agents to understand it and create a working sample.. its weird as its knows what it is, but can calculate simple.. please give it a try or anyone else ai interested
Hey guys! I'm immersed in the study of AI agents and I'm curious: would it be viable to build an agent that prospects customers for freelance professionals? I envision a system capable of exploring Instagram in an automated way, identifying potential customers and even starting conversations to schedule sales meetings. Is it possible to develop such AI agents? If so, do you know of any videos on RU-vid or any mentors that explain how to create AI agents to automatically prospect customers through Instagram? Is there something like this in development or is this an idea for the future of AI?"
thanks for this! currently trying to figure out what's missing with my groq connection since I'm encountering this error: openai.error.InvalidRequestError: The model `gpt-4` does not exist or you do not have access to it.
I want something to experiment with agents and get the handle of it and experiment on how much better it can help me at work, without spending any money or having particular premium keys. I've watched a lot of videos but i still don't understand what agent builders allow free to use agents, even if it's on a daily token limit.
The issue with these third parties is that there is lot under the hood and you will end with a lot of api calls and billing aspect is to be considered.
Thanks for the video! One question: any clue why do you set OPENAI_API_KEY with groq api key? I found it a bit confusing. Especially when using openai's API_KEY for authentication. Is OPENAI_API_KEY a placeholder in crewai for groq api key? I know a bit nonsense, so what do I missing? Thanks!
I'm a 3d artist and not a programmer by any stretch of the imagination. Is there a change we could have CrewAI with a nice user interface and an installer?
Man, you’re hilarious 😂😂😂 After 10 minutes of video - we can delete all of this and follow official crewai instructions ) But anyway, thanks, video have some gems, but probably not for those who are looking for entry level video.
Hey brother, thank you very much for your channel. I’m a single father, I’ve been following this AI stuff closely. I’m also in school, and have so many coals in the fire it’s not even funny lol. Thank you for your posts, because when I get a decent computer I’ll be able to quickly jump on board. I grew up very poor, my son will have a better life. I need to be on top of this. My next pay check I’ll be joining your community. Any advice on how to get my hands on a decent computer? To run this stuff? What should it have? I don’t want to miss the opportunity to provide for my son
When I use llama 3 8B on ollama or LM Studio, it is much dumber than on OpenRouter. Even after resetting all parameters to factory and loading the llama 3 preset. Even with the full non-quantized 8-bit version on LM studio.
He stored the Groq API key as an OpenAI API key variable with os.environ["OPENAI_API_KEY"], so when llm = ChatOpenAI(model = "some model") is called, it will automatically switch out "some model" against the variable defined in os.environ["OPENAI_MODEL_NAME"], which he set it to be "llama3-70-b-8192". Finally, he had to specify the url from which the model is accessed, so he set os.environ["OPENAI_API_BASE"] to some Groq related url.
Yes... I think almost all modern computers have more than 40GB of free storage. This isn't your RAM, this is how much space you need on your hard drive.
Can You make some better example of this agents. Something that is really helpfull, you always say thay for questions of time you do something basic, But it will be really fascinating if you spend more time doing something that have some realistic value. Thanks
TLDR: Don't bother with this video if you need to run locally He gets 9:38 in, can't get it working with a local Ollama model so just gives up and switches to a remote model. Really annoying if you're coding along with the video then realise it's useless for your purposes. I hope his premium content is better than this otherwise a bunch of people are getting taken for a ride.
I have made state of the art automation scripts for me work and i also added some stealth web scraping methods, how can i train lama model to use my coding methods??
I have ran it with a nvidia 3070 with like just a few tokens a second at float 16 i believe not sure but using recent branch text generation webui I was able to do it. I was also able to run the 8b model in colab for free using some code I got/came up with on huggingface face model card discussions you can see it. I would say 3090 would run it faster though
Does anyone know if you can run this on in iPad locally and upload documents in order to answer queries? For example if you made an app for allergies on a food menu would you be able to upload ingredients of the food menu into the LLM and have it RAG answers similar to "I have a gluten allergy, can i have the Ceasar salad?"
The easiest way would be to deploy all this stuff on a server or a home PC, expose an end point, then write an ipad app to upload docs and chat with doc via your app.
Anyone who is not ultra-rich will never pay 77 USD just to be in your community. It is simply insane. I suggest you take a different approach. Because it wont work.
I appreciate the value you offer with your community, but I want to be honest about my perspective. The current membership fee of 77 USD is simply too high for many, including myself. I understand that there are costs associated with maintaining the community and providing value to the members, but I wonder if there is room for a more accessible membership fee. A fee that is feasible for more people and enables them to participate and expand their knowledge.
We always have a choice - we can either stick our nose into other people's business and give unsolicited criticism, or we can start with ourselves, like earning more and not making ourselves look like a victim.
I'd pay it to be in with this crew if I could. Maybe someday but not today. Education is expensive. Life is rough, it's even rougher when your stupid. I appreciate your videos sir.
Is it possible to use agents to build an AI powered question answering system for pdf documented for academic research purposes? I'd like to build my own so I can avoid the costs of those currently available commercially.
quantize it at 4bit to 1bit so that you are running 13b param model with not much accuracy life then buy a 4090 or maybe even a a100 in your pc build from nvidia, honestly don't better to not buy that much hardware so many 1000s of $