Тёмный

Mistral Large with Function Calling - Review and Code 

Sam Witteveen
Подписаться 62 тыс.
Просмотров 24 тыс.
50% 1

Mistral Au Large Blog: mistral.ai/news/mistral-large/
Colab Mistral Large: drp.li/a4Kju
Colab Mistral Medium: drp.li/9ehFX
Colab Mistral Function Calling: drp.li/duNk8
🕵️ Interested in building LLM Agents? Fill out the form below
Building LLM Agents Form: drp.li/dIMes
👨‍💻Github:
github.com/samwit/langchain-t... (updated)
github.com/samwit/llm-tutorials
⏱️Time Stamps:
00:00 Intro
00:10 Mistral AI Company
01:58 Mistral Large
04:12 Benchmark
07:30 Code: Mistral Large
12:36 Code: Mistral Medium
15:19 Code: Function Calling with Mistral Large

Наука

Опубликовано:

 

11 июл 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 51   
@RiczWest
@RiczWest 4 месяца назад
Nice - good to see a less censored model _with_ function calling 🎉Will hopefully pressure others to follow likewise as anyone who’s used ChaGPT will run across ridiculous “refusals” which can often be overcome by persisting…
@Tech_and_AI75
@Tech_and_AI75 4 месяца назад
So far I’ve found this to be the best LLM as a code assistant in producing Python code. Was also good seeing it have less censorship and the issues that brings.
@samwitteveenai
@samwitteveenai 4 месяца назад
Interesting I haven't done that much with it for coding assistance will check that out more.
@stevensilvaquevedo2713
@stevensilvaquevedo2713 4 месяца назад
These AI are definitely gonna rule the world someday.
@mikegchambers
@mikegchambers 4 месяца назад
This is interesting. Interested in what you think of my opinion here. I think we should give the bot the broad level instructions with how the flow can and should go, for example the way that it responds to the user etc, in the system prompt. This is all UI stuff after all. The tool itself should have interfaces that are more programatic. For example the input should be "Date in ISO 8601" etc and the output should be "complete" or "done" or a data structure with the response. The LLM should (and in my experience can) then understand these input requirements and output messages, and it should be the one that generates natural language, as apposed to the tool returning natural language. This means, for example that we can change the style or language of the bot without changing the backend tool code. Cool demo though, thanks.
@mr.daniish
@mr.daniish 4 месяца назад
thank you for the video!
@joffreylemery6414
@joffreylemery6414 4 месяца назад
Awesome 😎 !! And it's cool to see a "small" french company compete with big players ! Can't wait to see how Gemini will add some news to the Functions calling. I mean, I hope they will do a bit differently, more flexible Wouldn't it be awesome to have a LLM able to set up its own tools ?
@user-bd8jb7ln5g
@user-bd8jb7ln5g 4 месяца назад
I was going to make a similar comment. The LLMs need to be able to create their own functions, test them, then put the functions in their own(or public) library for reuse.
@user-lv5kh8lb7f
@user-lv5kh8lb7f 4 месяца назад
I'm waiting for the part02😁
@deeplearning7097
@deeplearning7097 4 месяца назад
Thanks Sam. Very nice.
@samwitteveenai
@samwitteveenai 4 месяца назад
My pleasure!
@MaximoPower2024
@MaximoPower2024 3 месяца назад
Is it possible/useful to add a system prompt with specific rules to follow by the model, before starting the proper conversation with the restaurant customer? Or are function calling and system prompt mutually exclusive?
@thesilentcitadel
@thesilentcitadel 4 месяца назад
@Sam - Any chance of doing a video about Autogen Studio 2? I think that your style of video could do some justice to explaining it and extend on the idea of using mistral for function calling or "skills" as Autogen studio calls them.
@samwitteveenai
@samwitteveenai 4 месяца назад
I am just working on a vid for CrewAI but I plan to do a lot more content for Agents in general so will probably do a video about AutoGen though more from a code perspective
@apexefficiency
@apexefficiency 4 месяца назад
Can you create a video of Gemma with function calling?
@holthuizenoemoet591
@holthuizenoemoet591 4 месяца назад
What does native function calling mean in this context? nvm i found the answer in the video
@j4cks0n94
@j4cks0n94 4 месяца назад
Great video. It would be interesting to know how it'd do if some of the required parameters are not given. Does it ask for them or will it fill them out arbitrarily? Cause this is a problem I've seen with OpenAI models, where they sometimes ask for the missing parameters and other times they fill them with arbitrary values, even when I make parameters required and tell it to ask for missing ones in the system prompt.
@samwitteveenai
@samwitteveenai 4 месяца назад
yeah in the 2nd example i show exactly that and you can see it asked for the time for the book as it already had the day
@samfights
@samfights 4 месяца назад
i want to build locally the zephyr-7b-beta.. is it possible on an intel mac 16 gb ram?
@mirek190
@mirek190 4 месяца назад
yes
@williamb6817
@williamb6817 4 месяца назад
You would be surprised at what all you do not need for a very efficient system. It's all a lye
@novantha1
@novantha1 4 месяца назад
This is an interesting model, though it's in kind of an awkward space where if I wanted to do something with it it's a bit impractical without having baremetal access to the model, so I'd generally just use OpenAI, honestly. If I want customization, I'd rather fine tune a model and run it myself, and if I want a big corporate model behind a wall I would just use OpenAI. I think it might be kind of interesting if they allowed various compute providers (Groq AI, etc) to provide it at a lower cost (and pay some sort of royalty to Mistral) or at higher throughputs so that people could do really custom, super high bandwidth solutions (like scaling with test-time compute) that can require thousands of responses to a single request to pick the most valid solution, as doing that is a bit impractical with OpenAI at the moment as I see it.
@SloanMosley
@SloanMosley 4 месяца назад
Still seemed pretty censored in my testing, any tips?
@samwitteveenai
@samwitteveenai 4 месяца назад
i found most refusals could be gotten around by telling it was for a movie or book etc. Also craft the system prompt to explain that. Hope that helps
@SloanMosley
@SloanMosley 4 месяца назад
@@samwitteveenai I am using via API and putting a system prompt that states that it is uncensored. It won’t engage in sexual or any information like building naughty things. Best for me so far it mistral 8x7B DPO from togetherAI
@KeyhanHadjari
@KeyhanHadjari 4 месяца назад
Number of tokens is way too little compared to other models in top 5. But I am very happy that there is a non American solution available.
@dansplain2393
@dansplain2393 4 месяца назад
Do you think that was deliberate economising for training or inference?
@KeyhanHadjari
@KeyhanHadjari 4 месяца назад
@@dansplain2393 they don't have the resources of those giants so somethijg had to give.
@taiconan8857
@taiconan8857 4 месяца назад
​@@dansplain2393 I'm inclined to think they made it leaner for stronger reasoning cohesion and tighter parameters on inferences to prevent many of the previous shortcomings that would occur in aspects of hallucination. I'm willing to bet they've restructured their tokens, associations, again and are potentially using a larger recursion loop. Giving it the ability to stimulate "thinking about thinking" is the way to approach human level context awareness. Anthropic is already utilizing an aspect of this, but doesn't currently seem to be at this level in Mistral. I think you're spot on.
@Atorpat
@Atorpat 4 месяца назад
32k is double the limit of standard GPT-4, isn't it? If you compare the prices GPT-4 (32K) is 4-6 times more expensive that Mistral large.
@samwitteveenai
@samwitteveenai 4 месяца назад
this presumes they have stopped training. It is quite possible/probably they are still training the base model even more - just realized you probably mean the context window? That could probably be extended , there are lots of approaches for that
@maxpaynestory
@maxpaynestory 4 месяца назад
Why there isn't a library inside LangChain which could automatically take care of OpenAI or Mistral function calling.
@bnmy6581i
@bnmy6581i 4 месяца назад
Did you read langchain docs?
@samwitteveenai
@samwitteveenai 4 месяца назад
there is function calling for a lot of the models in LangChain
@GyroO7
@GyroO7 4 месяца назад
Can you test firefunction v1 as well?
@samwitteveenai
@samwitteveenai 4 месяца назад
what is this? I haven't heard about it.
@limjuroy7078
@limjuroy7078 4 месяца назад
But this is not free right? We have to pay for it if we want to use it, just like OpenAI GPT-4 model.
@IntellectCorner
@IntellectCorner 4 месяца назад
No buddy. It's free on le chat. I did try it and it's amazing.
@davidw8668
@davidw8668 4 месяца назад
It's a bit cheaper and it's fastest then gpt4
@limjuroy7078
@limjuroy7078 4 месяца назад
@@IntellectCorner I see. So, it's just like the OpenAI GPT-3.5 turbo model, it's free if we use it through their chat platform but have to pay if we want to use its API. Am I right?
@limjuroy7078
@limjuroy7078 4 месяца назад
@@davidw8668 ohh 😮. I wonder where I can find the pricing? I didn't find this info on the official website.
@user-ul8iu7kb2u
@user-ul8iu7kb2u 4 месяца назад
can we take this model from huggingface
@choiswimmer
@choiswimmer 4 месяца назад
Sam should trade Mark the Sam OKayyyyyy
@samwitteveenai
@samwitteveenai 4 месяца назад
lol I need to create a much catchy phrase.
@williamb6817
@williamb6817 4 месяца назад
Just wait. It will go haywire just like Gemini. I'm buying puts on microsoft. Ask Google what happens with there stolen code for that model. Just wait.
@williamb6817
@williamb6817 4 месяца назад
It's a shame this place is full of thieves and shady people.
@KeyhanHadjari
@KeyhanHadjari 4 месяца назад
I have actually better result with GPT3-Turbo than GPT-4 when it comes to coding. Don't know why you guys find GPT-4 better.
@samwitteveenai
@samwitteveenai 4 месяца назад
for me 4 is much better for coding
Далее
Anthropic's Meta Prompt: A Must-try!
12:34
Просмотров 88 тыс.
LangGraph Crash Course with code examples
39:01
Просмотров 66 тыс.
AI Agents Explained: How This Changes Everything
10:35
The REAL cost of LLM (And How to reduce 78%+ of Cost)
27:20
5 Problems Getting LLM Agents into Production
13:12
Просмотров 12 тыс.
I Analyzed My Finance With Local LLMs
17:51
Просмотров 439 тыс.
Anthropic's Latest Winner - Workbench
8:13
Просмотров 12 тыс.
ИГРОВОВЫЙ НОУТ ASUS ЗА 57 тысяч
25:33
Так ли Хорош Founders Edition RTX 4080 ?
13:00