Тёмный

Adding your own models to Ollama 

Fly․io
Подписаться 7 тыс.
Просмотров 773
50% 1

Опубликовано:

 

25 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 4   
@TheloniousBird
@TheloniousBird 2 месяца назад
Great video! Have your tired if the llama3-40 b runs?
@flydotio
@flydotio 2 месяца назад
Assuming you mean llama3.1 405B: I have tried so much and all my attempts have ended in failure. It's just too big. 405 billion parameters is like 1.6 trillion bits at Q4. This is something ludicrous like 203 GB of vram. The biggest GPU we have is 80 GB of vram. In theory you can use multiple GPUs, but I haven't gotten a config that works reliably. Maybe some time in the future 405B will work (or there will be a dedicated inference service or something), but for now 70B is realistically the limit.
@TheloniousBird
@TheloniousBird 2 месяца назад
@@flydotio Yes, oops I meant the 405B, all of the names are sticking in my lizard brain. I just wondered how people were even using a model that big! Thanks for your response. P.S. I just started deploying my apps on fly - you guys rock!
2 месяца назад
Great video, do you know how to load an encrypted model and load into memory?
Далее
Run your own AI (but private)
22:13
Просмотров 1,6 млн
The intro to Docker I wish I had when I started
18:27
Просмотров 183 тыс.
На самом деле, все не просто 😂
00:45
А вы знали что металл тонет?
00:32
Supercharge your Laravel app with AI
11:36
Просмотров 861
Adding Custom Models to Ollama
10:12
Просмотров 31 тыс.
Anthropic’s New AI Can Control Your Computer!
18:29
host ALL your AI locally
24:20
Просмотров 1,2 млн
Essential AI prompts for developers
8:31
Просмотров 71 тыс.
Let's use Ollama's Embeddings to Build an App
8:21
Просмотров 19 тыс.