Тёмный

Deploy ANY Open-Source LLM with Ollama on an AWS EC2 + GPU in 10 Min (Llama-3.1, Gemma-2 etc.) 

Developers Digest
Подписаться 24 тыс.
Просмотров 3,6 тыс.
50% 1

Опубликовано:

 

21 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 19   
@DevelopersDigest
@DevelopersDigest 2 месяца назад
The best way to support this channel? Comment, like, and subscribe!
@hpongpong
@hpongpong 2 месяца назад
Great concise presentation. Thank you so much!
@DevelopersDigest
@DevelopersDigest 2 месяца назад
Thank you! 🙏
@ryanroman6589
@ryanroman6589 2 месяца назад
this is super valuable. awesome vid!
@DevelopersDigest
@DevelopersDigest 2 месяца назад
Thank you! 🙏
@rembautimes8808
@rembautimes8808 2 месяца назад
Thanks very nice tutorial
@DevelopersDigest
@DevelopersDigest 2 месяца назад
Thank you
@alejandrogallardo1414
@alejandrogallardo1414 2 месяца назад
for models at ~70b, i am getting timeout issues using vanilla ollama. It works with the first pull/run, but times out when i need to reload model. Do you have any recommendations for persistently keeping the same model running?
@DevelopersDigest
@DevelopersDigest 2 месяца назад
github.com/ollama/ollama/pull/2146
@danielgannage8109
@danielgannage8109 2 месяца назад
This is very informative! Thanks :) Curious why you used a g4dn.xlarge GPU ($300/month) instead of a t3.medium CPU ($30/month)? I assumed the 8 Billion parameter model was out of reach with regular hardware. What max model size works with the g4dn.xlarge GPU? To put into perspective, I have a $4K macbook (16gb ram) that can really only run the large (150 million) or medium (100 million parameter) sized model, which i think the t3.medium CPU on AWS can only run the 50 million param (small model).
@rehanshaikh2708
@rehanshaikh2708 12 дней назад
how can i use this endpoint in langchain chatollama?
@dylanv3044
@dylanv3044 2 месяца назад
maybe a dumb question. how do you turn the stream data you received into readable sentences
@DevelopersDigest
@DevelopersDigest 2 месяца назад
You could accumulate tokens and split by the end of sentences . ! ? Etc and then send resp after grouping function like that
@nexuslux
@nexuslux 2 месяца назад
Can you use open web ui?
@ConAim
@ConAim Месяц назад
Stay away from AWS, it will cost you arms and legs in a long run..
@DevelopersDigest
@DevelopersDigest Месяц назад
Which vendors do you prefer? 🙂
@BeCodeless-dot-net
@BeCodeless-dot-net 2 месяца назад
nice explaination
@DevelopersDigest
@DevelopersDigest 2 месяца назад
Thank you!
Далее
She Couldn’t Believe I Did This! 😭 #shorts
00:12
1 Subscriber = 1 Penny
00:17
Просмотров 50 млн
host ALL your AI locally
24:20
Просмотров 1,2 млн
The cloud is over-engineered and overpriced (no music)
14:39
Cursor Composer: Is This How We'll Code Now? 🤔
18:23
The Future of Knowledge Assistants: Jerry Liu
16:55
Просмотров 110 тыс.
I Analyzed My Finance With Local LLMs
17:51
Просмотров 488 тыс.
She Couldn’t Believe I Did This! 😭 #shorts
00:12