Тёмный

Why Fine Tuning is Dead w/Emmanuel Ameisen 

Hamel Husain
Подписаться 1,4 тыс.
Просмотров 3,3 тыс.
50% 1

Arguments for why fine-tuning has become less useful over time, as well as some opinions as to where the field is going with Emmanuel Ameisen.
This is a talk from Mastering LLMs: A survey course on applied topics for Large Language Models.
More resources are available here:
bit.ly/applied-llms
00:00: Introduction and Background
01:23: Disclaimers and Opinions
01:53: Main Themes: Trends, Performance, and Difficulty
02:53: Trends in Machine Learning
03:16: Evolution of Machine Learning Practices
06:03: The Rise of Large Language Models (LLMs)
08:18: Embedding Models and Fine-Tuning
11:17: Benchmarking Prompts vs. Fine-Tuning
12:23: Fine-Tuning vs. RAG: A Comparative Analysis
25:03: Adding Knowledge to Models
33:14: Moving Targets: The Challenge of Fine-Tuning
38:10: Essential ML Practices: Data and Engineering
44:43: Trends in Model Prices and Context Sizes
47:22: Future Prospects of Fine-Tuning

Хобби

Опубликовано:

 

30 июн 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 6   
@alaad1009
@alaad1009 3 дня назад
Excellent conversation!!!
@darkmatter9583
@darkmatter9583 День назад
RAG,quantize data? favorite LLM? HELP
@mrwhitecc
@mrwhitecc 3 дня назад
I do not think he understand what happened to the model after fine tuning. Just give one example here, if you have a unique reasoning pattern that there is no chance public pretraining dataset can contain the correlated data , then the SFT is the only way that you can let the model simulate the "reasoning" ability that you want the model to behave , prompt engineering do not help at all , RAG either.
@agenticmark
@agenticmark 3 дня назад
you fine tune for BEHAVIOR, you use RAG for DATA. it fine tuning is how the model interacts with the user, rag is how the model gets factual information. that does not equal prompt engineering....
@agenticmark
@agenticmark 3 дня назад
strange, prompt engineering over fine tuning? if you dont want control-ability sure... prompt engineering will disappear. fine tuning will not. i train voice and chat models (fine tuning) and I have trained dozens of agent foundational models that play nintendo and atari games and a bunch of classifiers. training from scratch (foundational, pretraining) is very very costly. fine tuning is not.
@agenticmark
@agenticmark 3 дня назад
_very_ unscientific claim about the lines on that chart. try trading stocks with that mentality of guessing it will just keep going up!
Далее
🤯 #funny
00:20
Просмотров 761 тыс.
Solving Chollet's ARC-AGI with GPT4o
2:18:02
Просмотров 17 тыс.
Back to Basics for RAG w/ Jo Bergum
44:47
Просмотров 1,2 тыс.
This is why Deep Learning is really weird.
2:06:38
Просмотров 364 тыс.
What is an LLM Router?
9:16
Просмотров 22 тыс.