Тёмный
No video :(

Let's use Ollama's Embeddings to Build an App 

Matt Williams
Подписаться 30 тыс.
Просмотров 19 тыс.
50% 1

In this video, we are going to build an app that searches RU-vid transcripts for semantic matches, making it possible to find the right videos for questions folks have.
Find the repo at github.com/technovangelist/videoprojects
Be sure to sign up to my monthly newsletter at technovangelis...
And if interested in supporting me, sign up for my patreon at / technovangelist

Опубликовано:

 

29 авг 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 52   
@southVpaw
@southVpaw 5 месяцев назад
Your content is excellent, thoughtful, and easy to follow. I appreciate you being here. Please keep making embedding apps. I'm finding so much functionality with an LLM + embedding model beyond just straightforward RAG.
@hand-eye4517
@hand-eye4517 5 месяцев назад
you are really nice to listen to thanks sir, and the subject matter and walkthrogh is everything , much appreciated... and you actually provide the repo.. you are a winner
@vikrantkhedkar6451
@vikrantkhedkar6451 5 месяцев назад
The way tou created your own store and used a matching algorithm directly is really helpful to understand what vector databases work under the hood rather than jusr using them directly. Great video will ne using this concepts in a project will update here if anyone is interested. Great video 😀
@piero957
@piero957 5 месяцев назад
Thank you Matt, learning a lot from every video!
@voncolborn9437
@voncolborn9437 5 месяцев назад
Interesting project. I've been mulling something like this for some time. I'm an ML newbe so I really had no idea where to start. I do now. I will be following along, listening intently. Thanks Matt.
@AutisticThinker
@AutisticThinker 5 месяцев назад
I literally had this same idea a couple weeks ago, but as an enhancement to RU-vid so people can find the best content.
@solutionhacker
@solutionhacker 5 месяцев назад
Wonderful video and explanation, as always!
@AINMEisONE
@AINMEisONE 5 месяцев назад
You really are a good teacher! Thanks! So to be clear, you will make an app, that will do the following: 1. take a group of transcripts that deal with same genre, 2. create a index, or database, that can be searched, for a specific question, 3. ollama will be trained for example you talked about making a certain funtion that performs something in 1 video, and in the other you extended this knowledge to include that snippet mentioned in video 1, and. combined it with video 3 to formulate a task, that resolves the question asked?
@mbottambotta
@mbottambotta 5 месяцев назад
love it, thank you Matt! how did you end up choosing `nomic-embed-text`? is that model appropriate for all LLMs? or is embedding orthogonal to the LLM you end up using and can you use `nomic-embed-text` whether you work with llama-2 or mistral?
@sinayagubi8805
@sinayagubi8805 5 месяцев назад
Can you put this series im a playlist? I am really interested in using embeddings to enable a language model to act as a support agent for a company knowing all the information about the company and being able to pull up links.
@CraigRussill-Roy
@CraigRussill-Roy 5 месяцев назад
Supabase for the win !!! I use it for vectors and the api is so easy
@TalkingWithBots
@TalkingWithBots 5 месяцев назад
Man it's brilliant, I need to show it to my friend ❤‍🔥
@petrapetra7551
@petrapetra7551 5 месяцев назад
Hi, Matt. Thanks for your Video. I would like to suggest if in your Video Series, you could evaluate the LanceDb database is a DB Vector. Developed in Rust von packages for Python and Javascript. Thank you and I continue to enjoy your Videos.
@EarlZMoade
@EarlZMoade 5 месяцев назад
This is essentially semantic search right? Like RAG but without generation. Really cool use case.
@DonutMongo
@DonutMongo 5 месяцев назад
Thanks! Great timing, was just thinking about that how to get control over my news consumption with local llms :-)
@chrisBruner
@chrisBruner 5 месяцев назад
Interesting use case.
@Meyu-Sys
@Meyu-Sys 5 месяцев назад
Matt I was using an AI model and saved a session I accidentally deleted that session. I checked the ollama directory and the manifest file for the session as well as the blobs were still there. Is there any way to recover the session?
@technovangelist
@technovangelist 5 месяцев назад
Ollama doesn't really have a concept of session that you can recover. If you are using a GUI frontend for Ollama, some of them may have a way of saving sessions.
@Rikearon
@Rikearon 5 месяцев назад
LangChain would help do some heavy lifting regarding text manipulation and splitting
@technovangelist
@technovangelist 5 месяцев назад
That’s a helluva lot of extra cruft for a tiny bit of convenience
@thaiscvo
@thaiscvo 5 месяцев назад
please make a step by step video using nomic embed in ollama and a open source llm to build a chat with large docs app
@RedWhiteBlue209
@RedWhiteBlue209 4 месяца назад
This is a great video with excellent content. Using TypeScript is a bad choice. Python is a lot more widely used, and you want more people to understand the video fully when they first view it and spread the news. Doing it again with Python?
@technovangelist
@technovangelist 4 месяца назад
Typescript/Javascript tends to get more views. It’s used by far more folks than python as shown by every survey for years. I find it easier to work with and understand. And faster.
@RedWhiteBlue209
@RedWhiteBlue209 4 месяца назад
@@technovangelist Isn't right that Python is the mostly popular programming language for AI, esp. in learning and research?
@technovangelist
@technovangelist 4 месяца назад
Maybe but for no good reason. Python has no benefit here over any other language.
@technovangelist
@technovangelist 4 месяца назад
But thankyou for strengthening the argument for doing more JavaScript/typescript content.
@RedWhiteBlue209
@RedWhiteBlue209 4 месяца назад
I don't know. I watched many AI-related youtube videos. They include school classes and videos from businesses and individuals. Most of the time, they offer implementations in Python or even code working in Colab. If you do a search, you can find out that Python is widely taught in colleges, but not TypeScript/Javascript. Maybe I am wrong. Your video has excellent content, and I feel that I didn't get the most out of it.
@Edward_ZS
@Edward_ZS 5 месяцев назад
I was in one of your videos. There was a github that allows to translate the model files, so I know what each one is. But I can't find it.
@technovangelist
@technovangelist 5 месяцев назад
This one? Sync Ollama Models with Other Tools ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-UfhXbwA5thQ.html
@fearnworks
@fearnworks 5 месяцев назад
Great content!
@squiddymute
@squiddymute 5 месяцев назад
can you create embeddings from images using something like llava ?
@SlykeThePhoxenix
@SlykeThePhoxenix 5 месяцев назад
Are you going to use Docker Compose so it's easy to setup?
@technovangelist
@technovangelist 5 месяцев назад
This video is just about embedding
@technovangelist
@technovangelist 5 месяцев назад
Docker on Mac doesn’t use the gpu so I’d doesn’t make sense for me to include that. And nowhere near as easy as a straight native install
@tanvayvaidya4285
@tanvayvaidya4285 5 месяцев назад
Hi how we can send multiple request ollama. Does this support Multithreading
@HistoryIsAbsurd
@HistoryIsAbsurd 5 месяцев назад
Its like you read my mind lol Was just thinking "How can I do this"
@technovangelist
@technovangelist 5 месяцев назад
Oh no, it was your mind I was taking over.....
@HistoryIsAbsurd
@HistoryIsAbsurd 5 месяцев назад
This...this make so much sense...this explains everything!@@technovangelist​
@truehighs7845
@truehighs7845 5 месяцев назад
And why not just embedding the text transcript?
@technovangelist
@technovangelist 5 месяцев назад
As one thing? Then stuff wouldn’t come up in a match
@truehighs7845
@truehighs7845 5 месяцев назад
​@@technovangelist What you did is very advanced, and probably more effective, but sometimes I wonder if we are not reinventing the wheel, models are able of quite good needle in a haystack with plain text, as long as you remain in a reasonable length of text per promtp. What I did notice is when creating the jasonl it can be inconsistent, but then it takes some trial and error the get the prompt right, or an extra call to repair it, which with Openai becomes quite expensive, so yeah, I am cheating... :) But today I am mourning: bitsandbytes update installed a new triton 2.2.0, that broke flash attention, xformers and, because why not, torch, so I am looking at error logs today, thinking wtf am I doing under the rain wasting my time while I could jump in the car and drive to the south of Spain as they have 35 degrees. But I might isolate the beginning of the flow, for just for dataset creation, it's already a decent tool. It was working quite well before I try to lump half of the finetuning library of Github in my venv. Only I, like you, last time I programmed something, was an online casino in Basic with the 'rnd' function, on a 286. So even the most basic development takes me on a cathartic journey contemplating the abysmal ignorance I bring to the table in terms of development, so I always try to kiss. Peace! 😄 *Edit actually the casino was the opposite of "online" ...
@technovangelist
@technovangelist 5 месяцев назад
Actually in a lot of cases they are very bad at plucking out the right concepts from a long text. You don't have to spend much time with chatgpt or any of the local models to experience that. But this isn't doing that anyway. This one is all about embeddings which is different from asking a model a question
@truehighs7845
@truehighs7845 5 месяцев назад
@@technovangelist Yes I agree, but both solutions, IMO give back 'lumps' of the original text and rag from what I have tested, embedding with various dbs and models is always inferior to finetuning that returns a more organic answer, yet less precise. I guess it's going to depend on the application, probably a mix of both is going to be the best, even if I am only half impressed with OpeanAI's Agents that keep forgetting they have a context. Let us know when you have something we can test, whatever works my friend!
@JeffMcJunkin
@JeffMcJunkin 5 месяцев назад
@@technovangelist I've wondered about taking an LLM-generated summary of the raw transcript and using that summary as the embedding, though. That way you get closer-to-deterministic length so it fits nicely in one embedding. With the length of your videos, I'm betting 32K context length would be enough for the transcript + instructions prompt, plus the output (probably even with chain of summaries).
@iham1313
@iham1313 5 месяцев назад
Why and how is „a list of videos“ not an answer? which would be exactly what a RAG is. but as you point out: yours is different. i just don‘t see how. no offense; just trying to understand.
@technovangelist
@technovangelist 5 месяцев назад
It is. That’s the answer I provide. Normally rag would give the results of a vector db search to the model to generate an answer
@florentflote
@florentflote 5 месяцев назад
@allurbase
@allurbase 5 месяцев назад
Hey Matt, which model can I use for embedding text locally? If you have a code sample lying somewhere it would be awesome. THNX
@technovangelist
@technovangelist 5 месяцев назад
for code samples, look at the ones associated with the video...my github is technovangelist and the repo is videoprojects. the model to use is the one in the video...nomic text embed
@allurbase
@allurbase 5 месяцев назад
@@technovangelist found it 🎊🎊
@joxxen
@joxxen 5 месяцев назад
Amazing
Далее
OpenAI Embeddings and Vector Databases Crash Course
18:41
Whats the best Chunk Size for LLM Embeddings
10:46
Просмотров 11 тыс.
娜美这是在浪费食物 #路飞#海贼王
00:20
Run your own AI (but private)
22:13
Просмотров 1,4 млн
host ALL your AI locally
24:20
Просмотров 1 млн
Optimize Your AI Models
11:43
Просмотров 8 тыс.
The cloud is over-engineered and overpriced (no music)
14:39
The Most Legendary Programmers Of All Time
11:49
Просмотров 553 тыс.
I Analyzed My Finance With Local LLMs
17:51
Просмотров 469 тыс.
Python RAG Tutorial (with Local LLMs): AI For Your PDFs
21:33
Fine Tune a model with MLX for Ollama
8:40
Просмотров 1,1 тыс.
娜美这是在浪费食物 #路飞#海贼王
00:20