Тёмный
Prompt Engineering
Prompt Engineering
Prompt Engineering
Подписаться
Ph.D., Artificial Intelligence & Coding.

Building cool stuff!

▶️ Subscribe: www.youtube.com/@engineerprompt?sub_confirmation=1

Want to discuss your next AI project with me? BOOK NOW:
calendly.com/engineerprompt/consulting-call

For business inquiries email: engineerprompt@gmail.com



Gemini Agents are Surprisingly Good
18:31
Месяц назад
Just Large Enough - Mistral released a BEAST
17:38
2 месяца назад
GPT-4o mini - Can it be used for Agentic RAG?
12:51
2 месяца назад
Agentic RAG: Make Chatting with Docs Smarter
16:11
2 месяца назад
Комментарии
@jackgaleras
@jackgaleras 10 часов назад
A simple question, what time is it... as a language model I have no idea... where are you? I have no idea either, another sum 0+0 and explain how you got the result
@tapikoBlends
@tapikoBlends 11 часов назад
i have 4gb vram and 8gb ram , can i run it?
@mtpscout
@mtpscout День назад
Adding Vision and Voice Activity Detector (VAD) would take it to the next level
@mtpscout
@mtpscout День назад
Adding Vision and Voice Activity Detector (VAD) would take it to the next level
@loganhallucinates
@loganhallucinates День назад
Thanks for the video! So the demo retrieved "pages", if we want the actual paragraph or sentence-level sources we have to do an additional retrieval on the retrieved pages, right? I saw your Gemini PDF video and was wondering how ColPali performs compared to that.
@AmericaGermany
@AmericaGermany День назад
The quality of the text and storyline for Mixtral is significantly better, too. Does the speed of inference change the quality of the output at all?
@lakergreat1
@lakergreat1 День назад
Did verbi get a UI yet?
@HawkFranklinResearch
@HawkFranklinResearch День назад
Contextual retrieval just seems equivalent to GraphRag (by Microsoft) that indexes knowlegde context wise
@brian10508
@brian10508 День назад
In my exprience, it's much better in coding and debugging. GPT-o will constantly gaves me answer that don't really matters to the problems, o1 can always find the problem at first feedback.
@fuba44
@fuba44 День назад
Are you sure the serverless method worked, i find that if the model does not have a config.json file provided on huggingface, it simply wont work. infact i had a lot of problems running quantized models using the serverless endpoint. If you take requests, i would loke to see a more detailed walt through of how to truely use any model (with vLLM support ofc) from huggingface, in the models different quantized states. But thank you for sharing, you truely make some good content.
@andrew-does-marketing
@andrew-does-marketing 2 дня назад
Do you do contract work? I’m looking to get something like this created.
@engineerprompt
@engineerprompt 2 дня назад
Yes, you can contact me. Email is in the video description.
@mrchongnoi
@mrchongnoi 2 дня назад
The issue where the model isn’t generating all of the code likely stems from maintaining a large context. I refer to this as "LLM dementia," which I have also encountered when working with Anthropic. What you need to do is request complete documentation of the code, then start a new chat with the documentation and code included. From there, you can proceed. The challenge you're facing is that you're unable to attach files for preview.
@BalaramakrishnaKamma
@BalaramakrishnaKamma 2 дня назад
If I ask questions about graphs, tables, or images present in the PDF, will it provide an answer?
@pakipower
@pakipower 2 дня назад
I am sure this is a very interesting topic, but i have no idea what are you talking about 😐. Is there anyway somebody can summarize it in simple words. I really want to know.
@nealdalton4696
@nealdalton4696 3 дня назад
Are you adding this to localGPT?
@engineerprompt
@engineerprompt 2 дня назад
Yes, big upgrade is coming :)
@DevtalTalks
@DevtalTalks 3 дня назад
Excellent explanation!!
@crashandersen602
@crashandersen602 3 дня назад
So easy a baby could do it. Don't believe us? We have one following along in this lesson!
@SonGoku-pc7jl
@SonGoku-pc7jl 3 дня назад
thanks!
@SonGoku-pc7jl
@SonGoku-pc7jl 3 дня назад
thanks!
@stanTrX
@stanTrX 4 дня назад
Tabula-py or this? Which is better when it gets to extracting tables?
@yurijmikhassiak7342
@yurijmikhassiak7342 4 дня назад
WHY NOT TO DO SMARK CHANKING ON CONTENT. LIKE WHEN NEW TOPIC STARTS? NEW SENTENCE, ETC? YOU WILL USE FAST LLM TO GENERATE CHANKS. THERE WILL BE LESS NEED FOR OVERLAP.
@Autovetus
@Autovetus 3 дня назад
Chill , dude... Sheesh 🙄
@LatifAmars
@LatifAmars 4 дня назад
What tool did you use to record the video?
@ghulamaliamiri7484
@ghulamaliamiri7484 4 дня назад
is good available? kindly share the link
@souvickdas5564
@souvickdas5564 4 дня назад
How to generate those context for chunks without having the sufficient information to the LLM regarding the chunk? How they are getting the information about the revenue numbers in that example? If it is extracted from the whole document then it will be painful for llm cost.
@zachmccormick5116
@zachmccormick5116 2 дня назад
They put the entire document in the prompt for every single chunk. It’s very inefficient indeed.
@loudcloud1499
@loudcloud1499 4 дня назад
very informational visualizations!
@i2c_jason
@i2c_jason 4 дня назад
Hasn't structured graphRAG already solved this? Find the structured data using a graph, then navigate it to pull the exact example?
@remusomega
@remusomega День назад
How do you think the Graph gets structured in the first place
@faiqkhan7545
@faiqkhan7545 11 часов назад
@@remusomega Any Links to read ?
@forcebrew
@forcebrew 4 дня назад
Thank you for your expertise! Could you recommend a stable and efficient large language model for coding that I can run on my machine without it becoming unresponsive?
@LukePuplett
@LukePuplett 4 дня назад
I was so astonished by how obviously terrible the original "dumb chunking" approach is that I couldn't watch the video.
@DayLearningIT-hz5kj
@DayLearningIT-hz5kj 4 дня назад
Love the Baby ❤️ good father !
@snapman218
@snapman218 5 дней назад
Good information, but having a child crying in the background is unprofessional. Of course now everyone will say I hate children, but I don’t care. I’m sick of unprofessional behavior.
@kerbberbs
@kerbberbs 4 дня назад
Its youtube dawg, nobody cares. Just watch the overlengthed vid and move on. Most people here only came for 2 mins of what's actually important
@ogoldberg
@ogoldberg 4 дня назад
Rude thing to say, and ridiculous. You are the one who is unprofessional.
@tombelfort1618
@tombelfort1618 3 дня назад
Entitled much? How much did you pay him for his time again?
@vikramn2190
@vikramn2190 5 дней назад
Thanks for the easy to understand explanation
@seanwood
@seanwood 5 дней назад
Working with this now and didn’t use the new caching method 😫. Nice to have someone else run through this 🎉😆
@udaym4204
@udaym4204 5 дней назад
does Multi-Vector Retriever Worth It?
@isaacking4555
@isaacking4555 5 дней назад
The baby in the background 🤣
@wwkk4964
@wwkk4964 5 дней назад
🎉baby voices were cute!
@DRMEDAHMED
@DRMEDAHMED 5 дней назад
I want to add this as a the default way the rag is handled in open webUI but its conflicting with other stuff, I tried to make a custom pipeline for it but i'm struggling to make it work is it out of the scope of open web UI or am I just not understanding the documentation properly
@tomwawer5714
@tomwawer5714 5 дней назад
Thanks very interesting. Many ideas came to my head for improving RAG with enhancing chunk
@stonedizzleful
@stonedizzleful 5 дней назад
Great video man. Thank you!
@PhilipAlkire-c3j
@PhilipAlkire-c3j 5 дней назад
Moore Brenda Moore Margaret Gonzalez John
@marc-io
@marc-io 5 дней назад
so nothing new really
@turbo2ltr
@turbo2ltr 5 дней назад
Would this work for asking gemini to write code using a private COM interface by passing the COM documentation via context caching? I've been trying to do this with a custom GPT and have not been able to get it working very well, mostly because of limits on the knowledge files for GPTs.
@tetraocean
@tetraocean 5 дней назад
can chat bot send images with this data? normally embedding only text, but how about with images ?
@steve-g3j6b
@steve-g3j6b 5 дней назад
@Prompt Engineering didnt find a clear answer for my question, so I ask you. as a screenplay writer what do you think is the best model for me? gpt has very short memory. (not enough token memory)
@kees6
@kees6 5 дней назад
Gemini?
@steve-g3j6b
@steve-g3j6b 4 дня назад
@@kees6 why?
@lollots82
@lollots82 3 дня назад
​@@steve-g3j6bhas had 1M token window for a while
@MrGnolem
@MrGnolem 5 дней назад
Isn't this what llama index has been doing for over a year now?
@finalfan321
@finalfan321 5 дней назад
you sound tired but i thin i know why ;)
@limjuroy7078
@limjuroy7078 5 дней назад
What happened if the document contains a lot of images like tables, charts, and so on? Can we still chunk the document in a normal way like setting a chunk size?
@kai_s1985
@kai_s1985 5 дней назад
You can use vision based rag, he described in his previous video.
@limjuroy7078
@limjuroy7078 5 дней назад
@@kai_s1985, so we don't need to chunk our documents if we use vision based RAG? My problem is how are we going to chunk our documents even though the LLM has vision capabilities
@kairatsabyrov2031
@kairatsabyrov2031 5 дней назад
@@limjuroy7078 it is very different from the text based rag. But, I think you need to embed images page by page. Look at his video or read the ColPali paper.
@MrAhsan99
@MrAhsan99 5 дней назад
You can name the little one "Ahsan" just in case, if you are looking for the names.
@jackbauer322
@jackbauer322 5 дней назад
I think the baby in the background disagrees :p
@janalgos
@janalgos 5 дней назад
how does it compare to hybridRAG?
@publicsectordirect982
@publicsectordirect982 4 дня назад
I'd like to know the same 👍
@martinsherry
@martinsherry 5 дней назад
V helpful explanation.