Тёмный
No video :(

Level Up Your Typescript Skills: Adding Ollama To Your Apps! 

Matt Williams
Подписаться 30 тыс.
Просмотров 25 тыс.
50% 1

The Javascript library for Ollama makes it so much easier to build cools applications with AI. This video will get you up to speed on everything you need to know.
Find the code here: github.com/tec... then videoprojects then find the folder for this video.
Be sure to sign up to my monthly newsletter at technovangelis...
And if interested in supporting me, sign up for my patreon at / technovangelist

Опубликовано:

 

29 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 44   
@userou-ig1ze
@userou-ig1ze 5 месяцев назад
This channel is top on my list lol, I'm addicted to the casual seeming, yet perfect presentations (besides the interesting content and impressive successive iterations of code...). It's clear, concise, intuitive flow and so calm I find myself not always using 2x speed
@technovangelist
@technovangelist 5 месяцев назад
Thanks so much. Nice to know it’s resonating with folks
@DevynDowler-Lewis
@DevynDowler-Lewis 5 месяцев назад
This is remarkably well presented
@technovangelist
@technovangelist 5 месяцев назад
Thanks so much
@brogeby
@brogeby 4 месяца назад
This is exactly the channel I was searching for. Your presentations are very well presented, great examples and you talk about very relevant topics. Thanks!
@yoemasey
@yoemasey 5 месяцев назад
Thank you for explaining the details so carefully and easy to understand. Can't wait to try out Ollama :)
@photorealm
@photorealm 5 месяцев назад
Best video I have seen on this subject by far. Great usable information.
@technovangelist
@technovangelist 5 месяцев назад
Glad it was helpful!
@lucioussmoothy
@lucioussmoothy 5 месяцев назад
nice over view -I'd love to see a view on how we can use JS to train and fine tune an Ollama model (if you haven't posted one yet)
@moonchildeverlasting9904
@moonchildeverlasting9904 5 месяцев назад
this is great matt, you probably forgot that someone made a demo showing mario kart in JS. I think you come abit longer with knowledge, then ai, but we all think differently right?
@jacknimble8331
@jacknimble8331 5 месяцев назад
Thank you very much for these helpful videos! If you have the bandwidth could you go over incorporating embedding for RAG in this app?
@technovangelist
@technovangelist 5 месяцев назад
I'll have that in a video coming next week
@user-he8qc4mr4i
@user-he8qc4mr4i 5 месяцев назад
Thank you for sharing, Matt! I did not realize that JS was an option :-). Any plans to add a video showing Ollama + Lanchain ?
@technovangelist
@technovangelist 5 месяцев назад
Yes. But for most projects it doesn’t add anything. I have to think of a good use case to leverage it.
@MuhammadAzhar-eq3fi
@MuhammadAzhar-eq3fi 5 месяцев назад
Thanks a lot, Sir. ❤
@Pdizzle-ic5sk
@Pdizzle-ic5sk 5 месяцев назад
I downloaded the webUi docker container (Chat GPT like interface) about 3 months ago. Everything worked great. Downloading models was super convenient and easy. Tried to download the latest image on a new machine and the interface got a lot more complex. I dont even know how to download a model anymore. Any chance you can make a video on the web UI docker container? Easy docker compose startup. Downloading the diferenent models, and and prompting them? This tutorial was great! Thanks,
@technovangelist
@technovangelist 5 месяцев назад
Personally I’m not a big fan of most of the web interfaces I have seen. But I should try again.
@joshuandala7669
@joshuandala7669 2 месяца назад
The link isn't working, but I was able to find it. Hopefully you see this comment so that the error can be changed
@technovangelist
@technovangelist 2 месяца назад
Interesting. I think I fixed it, though.
@adityasingh017
@adityasingh017 Месяц назад
Hey anybody knows, how could i use this library in my BROWSER ?
@technovangelist
@technovangelist Месяц назад
What do you want to do? What have you tried? Best to ask on the discord
@grigorikochanov3244
@grigorikochanov3244 5 месяцев назад
Sorry, at 7:00 you say that in chat endpoint parameters are replaced with messages, and show a code sample with /api/generate endpoint. The chat endpoint for Ollama is /api/chat. Is it a mistyping?
@technovangelist
@technovangelist 5 месяцев назад
Doh. Good catch. Yes a typo. It’s right in the GitHub repo.
@CookerSingh
@CookerSingh 5 месяцев назад
Which models currently support function calling, and is there a function calling proxy available for any of these models, or does Ollama provide one?
@technovangelist
@technovangelist 5 месяцев назад
All of them. Well except the embedding models. But it’s not a feature the model itself needs to support.
@truehighs7845
@truehighs7845 5 месяцев назад
Great video as usual, but did you say we can infer with fine-tuned models from ollama?
@technovangelist
@technovangelist 5 месяцев назад
Yes, absolutely. Ollama doesn’t do the fine tuning but it can use the adapters.
@truehighs7845
@truehighs7845 5 месяцев назад
@@technovangelist How do you load them, with 'custom models' ?
@technovangelist
@technovangelist 5 месяцев назад
The docs show how. Create a modelfile and use the adapter instruction
@truehighs7845
@truehighs7845 5 месяцев назад
@@technovangelist ok thanks!
@LucianoTonet
@LucianoTonet 5 месяцев назад
I'm learning so much with this channel. @technovangelist it could be better to see if you add some syntax highlighting on your code I think. Thanks
@abdulrahimannaufal5678
@abdulrahimannaufal5678 5 месяцев назад
Ollama doesn't stream the output, instead prints all at once when I tried with ngrook over https. Is it the same with tailscale ?
@technovangelist
@technovangelist 5 месяцев назад
Haven’t used ngrok before so don’t know how to configure it correctly. You should work on fixing that configuration. Tailscale definitely doesn’t screw with the output of applications
@mshonle
@mshonle 5 месяцев назад
This is random and only tangentially related, but: do you know how to do that thing where your browser sees the localhost as usernames dot machine and there’s also some local CA cert configuration so you can test using https? (Random, yes, but I comment because I care.)
@technovangelist
@technovangelist 5 месяцев назад
Not sure but I think you are talking about wildcard certs. You can do this with certs from letsencrypt using traefik. Technotim I think did a video about this a couple years ago as did Christian L something. Last name starts with l. Lempa? There are other alternatives I think with nginx instead of traefik.
@AINMEisONE
@AINMEisONE 5 месяцев назад
Hi Matt, here are some questions: 1. I have Ollama running but it will not see my Egpu, it's an RTX 4090. I use LM Studio but it's always full of bugs and creates strange replies or does not even load the model. But when it does it sees the CUDA GPU and uses 100% and wow it runs the 70b model perfectly...2. I want to be able to also move from the Intel Macbook Pro Ollama models I downloaded so I do not need to re-download them again and again. Can you make a video for this? You know that most apps now are not on Intel unless you do bootcamp Windows then it runs and the best thing is the RTX 4090 eGPU works incredibly! Help, please.
@technovangelist
@technovangelist 5 месяцев назад
I don’t have a good answer for you. I don’t think the team has put any effort into supporting Intel Macs. When we were first building it Apple silicon is what inspired us to create ollama. But supporting egpu may be a bit further out. If you wanted to load up Linux on there it may have a chance.
@AINMEisONE
@AINMEisONE 5 месяцев назад
@@technovangelist I have an M2 MAX 96GB RAM, the intel is faster, with the RTX 4090... by a lot. with LM studio for now.. but too many bugs in their stuff. It has gotten a lot better..but still not as good as Ollama, big models do not work at all 34b on up.. they get stuck or have very low-level responses.. I will send maybe some other questions. in the meantime to fix some things.
@AINMEisONE
@AINMEisONE 5 месяцев назад
@@technovangelist ok which Linux version to install? I tried Ubuntu 20.04/22.04 and it did not support Ollama it said.. I used the command prompt on the Ollama site and it never worked. Thanks!
@technovangelist
@technovangelist 5 месяцев назад
Asking in the discord will have better results. I know nothing about that.
@florentflote
@florentflote 5 месяцев назад
@DoNotKillThePresiden
@DoNotKillThePresiden 5 месяцев назад
Thanks Matt 😁
@technovangelist
@technovangelist 5 месяцев назад
You bet!
@RickySupriyadi
@RickySupriyadi 5 месяцев назад
Matt Williams Channel just got EXP! 🆙 💫 +100 EXP 💫 +100 EXP 💫 +100 EXP 💫 +100 EXP 💫 +100 EXP LVL 2 (100/1000) EXP Matt Williams Unlocked Titles: 👐 **The Caring One** Lv. 1 (Cares for all types of audiences, taking care of newbies) 🧑‍🏫 **The Guru** Lv. 1 ("Clears Makes complex topics easy to understand) My likes for this channel have skyrocketed!
Далее
Getting Started on Ollama
11:26
Просмотров 50 тыс.
Learn TypeScript Generics In 13 Minutes
12:52
Просмотров 258 тыс.
what will you choose? #tiktok
00:14
Просмотров 7 млн
Fine Tune a model with MLX for Ollama
8:40
Просмотров 1,6 тыс.
I Analyzed My Finance With Local LLMs
17:51
Просмотров 469 тыс.
The Story of Next.js
12:13
Просмотров 569 тыс.
Ollama 0.1.26 Makes Embedding 100x Better
8:17
Просмотров 45 тыс.
PROOF JavaScript is a Multi-Threaded language
8:21
Просмотров 276 тыс.
The Secret Behind Ollama's Magic: Revealed!
8:27
Просмотров 31 тыс.
Is Dify the easiest way to build AI Applications?
13:50