Тёмный

Enchanted LLM and the bright path for open language models 

Code to the Moon
Подписаться 71 тыс.
Просмотров 6 тыс.
50% 1

Опубликовано:

 

2 окт 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 32   
@codetothemoon
@codetothemoon 8 месяцев назад
ERRATA: In the video I mention that setting OLLAMA_HOST is an alternative to using ngrok, but that's the case only in scenarios where you only need access on your local network. ngrok apparently lets you leverage your Ollama instance from anywhere, which sounds awesome (thanks to @havokgames8297 for pointing this out)
@melongrasp
@melongrasp 8 месяцев назад
Oh, that's really nice. Thanks for sharing!
@newtonchutney
@newtonchutney 7 месяцев назад
Yepp ngrok can be used to forward your RPi! 😂 But I'd suggest people start looking into tailscale instead.. As it has a lot more security and privacy
@newtonchutney
@newtonchutney 7 месяцев назад
Tailscale is a mesh VPN system.. Not a reverse proxy like ngrok.. BTW..
@Kabodanki
@Kabodanki 8 месяцев назад
LLM not biased by the bay area mentality is the future. I'm glad Mistral is french, there's some hope to get away from censorship
@codetothemoon
@codetothemoon 8 месяцев назад
not sure, it might have a bias towards crepes and baguettes, but I think I'm ok with that!
@fooblahblah
@fooblahblah 8 месяцев назад
You can use ngrok to proxy to your internal machine but via an external hostname or ip
@codetothemoon
@codetothemoon 8 месяцев назад
Thanks - I’ve added this as a pinned errata comment
@devopstoolbox
@devopstoolbox 8 месяцев назад
That is SO COOL!!!
@codetothemoon
@codetothemoon 8 месяцев назад
agree - I know you've been on the Ollama train too 🚂
@undefined24
@undefined24 8 месяцев назад
Looks promising, thanks for sharing.
@codetothemoon
@codetothemoon 8 месяцев назад
thanks for watching!
@lenninlc
@lenninlc 8 месяцев назад
So cool!
@codetothemoon
@codetothemoon 8 месяцев назад
😎
@dpi3981
@dpi3981 8 месяцев назад
What gpu do you use for your setup?
@codetothemoon
@codetothemoon 8 месяцев назад
I have an M1 Max which has an integrated GPU
@vimaximus1360
@vimaximus1360 8 месяцев назад
I would love to see some hardware comparisons, between mac with 32+ GB ram and some Nvidia GPU.
@codetothemoon
@codetothemoon 8 месяцев назад
this video might be what you're looking for! ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-jaM02mb6JFM.html
@vimaximus1360
@vimaximus1360 8 месяцев назад
perfect! thank you @@codetothemoon !
@youpapai
@youpapai 8 месяцев назад
why does `ollama run something` pull/download the model every time ? Is there a setting to cache it or use the cached downloaded model?
@codetothemoon
@codetothemoon 8 месяцев назад
it doesn't, at least for me. everything that appears in the list of language models to choose from is already downloaded and ready to go. That said, they might take a few seconds to load into memory, especially if they are on the larger side. Mistral 7B only takes ~10 seconds or so to load into memory for me. Are you seeing an isue where the model is downloaded on every run?
@youpapai
@youpapai 7 месяцев назад
@@codetothemoonyes. being downloaded every run
@havokgames8297
@havokgames8297 8 месяцев назад
Ngrok would let you access your Ollama without being on the same WIFI as your computer
@codetothemoon
@codetothemoon 8 месяцев назад
ahh got it - thanks for clarifying this! I should have looked into it a bit more. I'll post this in an errata comment.
@havokgames8297
@havokgames8297 8 месяцев назад
@@codetothemoonno worries. No one would expect you to be an expert at everything. I've used Ngrok for example when developing a web app locally that has webhooks and I want an external service to be able to access my local development server. It is perfect for this. The issue is that on the free tier it won't keep the same host name, so when you configure your Enchanted LLM app - if you restart NGrok then the URL will be different. Either you can pay for the service and get static URLs (I believe), or use another static DNS service with a hostname pointing to your machine.
@rnp0728
@rnp0728 8 месяцев назад
Great
@codetothemoon
@codetothemoon 8 месяцев назад
thanks!
@TommiNiemi-hu8pb
@TommiNiemi-hu8pb 8 месяцев назад
ngrok is for nat traversal
@codetothemoon
@codetothemoon 8 месяцев назад
Thanks yeah I made a pinned errata comment about this 😎
@CrazyLuke11
@CrazyLuke11 8 месяцев назад
First 🎉🎉😂😂
@codetothemoon
@codetothemoon 8 месяцев назад
you got it! 🥇
Далее
Language Models For Software Developers in 17 Minutes
16:59
HA-HA-HA-HA 👫 #countryhumans
00:15
Просмотров 565 тыс.
Larger Scale Software Development (and a Big Trap)
17:17
Big Tech AI Is A Lie
16:56
Просмотров 265 тыс.
Rust Demystified 🪄 Simplifying The Toughest Parts
14:05
I love small and awesome models
11:43
Просмотров 20 тыс.
The Uiua Programming Language Caught Me By Surprise
12:24
Run your own AI (but private)
22:13
Просмотров 1,5 млн
Modern All Rust Stack - Dioxus, Axum, Warp, SurrealDB
24:02
How I Made AI Assistants Do My Work For Me: CrewAI
19:21