Тёмный

Long prompts vs fine tuning LLMs 

Steve (Builder.io)
Подписаться 97 тыс.
Просмотров 18 тыс.
50% 1

More tips in my full article: www.builder.io...
#ai #developer #javascript

Опубликовано:

 

11 сен 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 8   
@stonedoubt
@stonedoubt 21 день назад
Multi-shot examples mixed in with chain of thought (an example for appropriate thinking) is just unmatched in my opinion
@clamhammer2463
@clamhammer2463 18 дней назад
fine tuning saves you on the token cost. you dont have to provide the examples with every request for a very specific format. with a larger system with many more users, this is preferable over larger instruction prompts.
@ricko13
@ricko13 21 день назад
thanks for your all great tips 🎉
@threepe0
@threepe0 21 день назад
Interesting. I think longterm cost savings for input tokens can be a big factor. If you’re spending 200k a month on llm calls because and your prompt is several thousand tokens, it really starts to make sense to fine-tune. Benchmarking standards are also kinda tricky. Saying it’s “as good or better” I’ve found leaves a lot of room for it being absolute crap in certain scenarios. 4o is “as good or better” than 4 legacy, which is laughable when you’re asking 4o a complex coding question or stuck in a loop where it provides the same bad suggestion over and over. Thank you for sharing
@N7Tonik
@N7Tonik 19 дней назад
but input tokens are veryyy cheap compared to output tokens and for fine tuning you would need to consider the cost of collecting data, fine tuning itself and when you wanna change smth that will also cost you
@threepe0
@threepe0 6 дней назад
@@N7Tonik very cheap… at what scale? I’ve found that very cheap can turn into $200k pretty quickly. Collecting data is free if you’re already doing it for analysis. And fine tuning jobs are pretty darn cheap relatively too. There are considerations to be made. Saying it’s cheap and moving on only gets you so far. Even at small scales, fine tuning starts to make sense if you have the data needed and aren’t getting the results you need with prompt engineering
@kartikm7
@kartikm7 21 день назад
If you want the llm to output JSON strictly it's a great idea to actually enable format JSON mode while making the API call to the server to get the llm response. Format as JSON is supported by Ollama, Groq, Open AI and probably even other inference api providers. So clubbing the prompt shown in the video + format as JSON should give the best outputs
@gjsxnobody7534
@gjsxnobody7534 20 дней назад
that's interesting and good info
Далее
RAG vs. Fine Tuning
8:57
Просмотров 7 тыс.
To mahh too🫰🍅 #abirzkitchen #tomato
01:00
Просмотров 3,4 млн
Cute
00:16
Просмотров 6 млн
Good refactoring vs bad refactoring
6:19
Просмотров 17 тыс.
LoRA & QLoRA Fine-tuning Explained In-Depth
14:39
Просмотров 39 тыс.
The # 1 Mistake Websites Make with Lazy Loading Images
24:28
Cursor Is Beating VS Code (...by forking it)
18:00
Просмотров 57 тыс.
Essential AI prompts for developers
8:31
Просмотров 63 тыс.
When RESTful architecture isn't enough...
21:02
Просмотров 275 тыс.
How RAG Turns AI Chatbots Into Something Practical
10:18