Тёмный
llmware
llmware
llmware
Подписаться
LLMWare by Ai Bloks innovates the latest in cutting edge solutions in AI for enterprises. Our open source AI Platform delivers the most integrated and easy to use out of the box Retrieval Augmented Generation. We also have over 60 models in Hugging Face that are small specialized language models for private cloud or on prem for enterprise use cases. Check us out!
RAG with Text Query (Ex. 4): Fast Start to RAG
16:02
6 месяцев назад
Prompt Models (Ex. 3): Fast Start to RAG (2024)
10:55
6 месяцев назад
Комментарии
@kushis4ever
@kushis4ever 11 дней назад
Hey there, this is all interesting and I tried running your fast start RAG notebooks on google colab. This 5th notebook in the RAG series (Semantic query ) when run on colab throws an error. In fact, you can find the error in the original notebook itself which was uploaded. It seems this notebook never ran well as a notebook but was still uploaded on Github. I have opened an issue there. Hope somebody can have a look.
@HjayejMohamed
@HjayejMohamed 21 день назад
Nice , Is there a way to download and import llmware model from local folder isteade of loading it at the first run of app, for RAG app?
@llmware
@llmware 21 день назад
Hi we have our models in Huggingface as well - please check out our repo: huggingface.co/llmware
@AliAlias
@AliAlias Месяц назад
Hi, can run slim-sql.gguf model using ollama server?
@renatogoncalves8491
@renatogoncalves8491 Месяц назад
Amazing!!!!
@bernardhugueney7804
@bernardhugueney7804 Месяц назад
Excellent! Do you plan to release an updated fine tuned model now that Microsoft released an updated Phi 3 pretrained model?
@jdmusic4188
@jdmusic4188 Месяц назад
I followed this completely but Its not giving the csv. Its only giving the jsonl file
@mikhailkalashnik0v
@mikhailkalashnik0v 2 месяца назад
aarch64 linux support would be great so we can run/test this on linux vms wo having to install it on main apple metal host os or via docker.
@llmware
@llmware 2 месяца назад
Hi Thank you for your feedback. We used to support it for months then we deprecated it about. a month ago because very few people were using it...😅 We will reconsider this in the future if we get more requests!
@llmware
@llmware 2 месяца назад
Example in Repo: github.com/llmware-ai/llmware/blob/a58c2dc7ea94c1a8eef87bc0fd1cc34fb616c743/examples/SLIM-Agents/using-slim-q-gen.py#L4
@dileshtanna5202
@dileshtanna5202 2 месяца назад
Very well explained! Thank you
@TheBialbino
@TheBialbino 2 месяца назад
Thank you for putting a smile on my face
@ezdeezytube
@ezdeezytube 2 месяца назад
Is there a way to combine this with your "example 5" program that runs a semantic query with RAG? This would be such a better interface. Also, it is desirable to avoid the chatbot drop down menu that asks which specific pdf you want analyzed for the semantic query, and instead have it search all of the pdfs in your library OR provide an answer for each pdf just like the example 5 program.
@ezdeezytube
@ezdeezytube 2 месяца назад
This is absolutely brilliant! The utility of this is off the charts! I will have to review all your vids and find out how to run this, and what model suits it best.
@llmware
@llmware 2 месяца назад
Thank you so much for your kind feedback! Please check out our videos and our repo with over 100 examples!
@Psychopatz
@Psychopatz 2 месяца назад
YES, I WANT THIS TO MY GAME, SUPERTHANKS SIR!
@llmware
@llmware 2 месяца назад
LLMWare Re-Ranker Example link: github.com/llmware-ai/llmware/blob/main/examples/Embedding/using_semantic_reranker_with_rag.py
@eugenetapang
@eugenetapang 2 месяца назад
Very Very NIce, Darren is it possible, to have LLMWare automagically boz bpt tale a URL address of a webpage and output it to a format: audio, text, pdf or csv then automatically load it into the biz bot?
@xspydazx
@xspydazx 2 месяца назад
hi great stuff by the way it seems though your not getting greatly noticed !!! perhaps make a few gradio spaces ? Personally i feel that you have nice method with building the slim models (in fact real agent experts) ... but also the wrapper performs a task: which could basically be overfit to the slim model : therby fixing the model in this specific mode: SO: i Think that creating a model from all these models or subsets of these experts as mixture of experts models : with the expert slim models as the experts : (trying to add up to a 7b(total parameters) moe: ie a slim moe? this would actually be very usefull model as you could use the creation of the model prompts internally for the merge directing the specific querys to the correct expert: the concept of slim models is great but the wrappers them selves are also greater .... as thier own sidepeices: hence geven the wraper we could essentially patch and openAI component or a Transformers(pretrained), Or a llamaCpp(gguf) model... hence the customized wrappers that perform the tasks would only need a model and tokenizer to plug in (utilize the embeddings from the model please).... here you can see the product produced .... now the service is designing these overfit experts , and wrapers : hence an object model which i think you have ... the object model is the Free Library... for open source community to build thier own wrappers .... and your comercial wrappers either as example (to show case your departments skillls hence business models would come direct to have thier custom models produced!) ... i can see why you have not been popular and even skipped by many !
@gnosisdg8497
@gnosisdg8497 3 месяца назад
very useful but where exactly is the code for all this ? specific all this !!! Some people are not good with coding and want to test these examples, but this specific has everything one needs....Will you be able to produce some ready examples just like this one ?
@llmware
@llmware 3 месяца назад
github.com/llmware-ai/llmware/blob/main/examples/Use_Cases/biz_bot.py
@llmware
@llmware 3 месяца назад
Thank you for the question - here is our example code in our repo. We will also produce a notebook of this shortly and will post here as well.
@gnosisdg8497
@gnosisdg8497 3 месяца назад
@@llmware thank you for the answer.This is good stuff ! thanks again
@gnosisdg8497
@gnosisdg8497 3 месяца назад
@@llmware please also make it generic meaning give the option to upload the files also the option to upload maybe the small lm? for each usage? just ideas !
@llmware
@llmware 2 месяца назад
@@gnosisdg8497 Hi we have uploaded previous videos that perform these individual tasks (this is representing some of our capabilities in one bot)... Please check out our repo or join our discord and we can answer more questions there as well!
@QorQar
@QorQar 3 месяца назад
Download the repo, open the example, and just run it and it will work, because I tried in Kolab and the example did not work for me
@llmware
@llmware 3 месяца назад
We are working on turning many of our YT videos into Colab notebooks as well and will post these notebooks as we make them.
@QorQar
@QorQar 3 месяца назад
I tried to run a chat in Colab several times and it downloaded the model and it did not complete. Should I run the example from the same folder without modifying its real path? I don’t know the problem.
@llmware
@llmware 3 месяца назад
Hi we can help you trouble shoot in discord: discord.gg/95V8hdS2RG Please join and let us know what you are seeing and we can help you.
@llmware
@llmware 3 месяца назад
Here are also some common problems we see (if this applies to anything you're seeing): llmware-ai.github.io/llmware/troubleshooting
@amrohendawi6007
@amrohendawi6007 3 месяца назад
this is beautiful content! exactly what I was looking for but summarizing this work in a blog with evaluation table and some diagram of the workflow of the benchmarking would have been great.
@llmware
@llmware 3 месяца назад
Hi Thank you so much for your kind feedback! While it is not exactly a perfect match for this particular video, we do have a blog post on this topic. Please take a look! 😃medium.com/@darrenoberst/how-accurate-is-rag-8f0706281fd9
@ROKKor-hs8tg
@ROKKor-hs8tg 3 месяца назад
He stops at the first question and does not answer for 11 minutes
@JebliMohamed
@JebliMohamed 3 месяца назад
🎯 Key points for quick navigation: 00:18 *📄 Introduction to document parsing, chunking, and data extraction.* 00:33 *🛠️ Advanced techniques for extracting images, tables, and automating workflows.* 01:17 *📚 Preparing datasets for self-supervised learning and fine-tuning.* 01:31 *💡 Focus on data wrangling and Microsoft Office documents.* 02:14 *🗂️ Accessing public Microsoft Word, PowerPoint, and Excel documents.* 03:22 *📂 Downloading and preparing Microsoft Office documents.* 04:03 *🛠️ Setting up the environment to parse and chunk documents.* 05:12 *🔍 Smart chunking strategies and their configurations.* 06:22 *📑 Parsing tables and images from documents.* 07:32 *🗃️ Exporting tables into CSV files.* 08:28 *🖼️ Running OCR on extracted images.* 09:54 *📄 Creating a consolidated JSONL file.* 10:35 *📊 Building a dataset for unsupervised testing.* 11:14 *⚡ Parsing 152 files in 6 seconds using a local Mac M1.* 12:37 *🔍 Running OCR and storing text in the library.* 13:17 *⏱️ Comparing the speed of digital parsing versus OCR.* 14:23 *📁 Exploring file artifacts created during parsing.* 16:29 *📄 Reviewing the created dataset.* 19:44 *🎥 Closing remarks and upcoming example videos.* Made with HARPA AI
@llmware
@llmware 3 месяца назад
This is so helpful - thank you!!
@JebliMohamed
@JebliMohamed 3 месяца назад
Loved the video! The step-by-step guide on parsing docs and data was super helpful. I was really impressed by how you used OCR to pull text from images in Microsoft Office files - that was cool. The smart chunking strategy explanation was also 👌.
@llmware
@llmware 3 месяца назад
Thank you so much for your kind feedback! ☺
@user-kk1li5mk7q
@user-kk1li5mk7q 3 месяца назад
This is really a nice way of extracting data and converting the unstructured data into structured form. I believe the data after extraction can be used as a data source for the RAG pipeline and probably LLMs can give more accurate answers.
@llmware
@llmware 3 месяца назад
Thank you so much for your observation - we also believe that documents parsed in this manner will enhance accuracy of LLMs in a RAG workflow!
@Army_76-g4s
@Army_76-g4s 3 месяца назад
wow!!!!! I'm wowed!
@llmware
@llmware 3 месяца назад
Thank you so much! 🥰
@kushis4ever
@kushis4ever 3 месяца назад
That's the best video and simplest program I have come across ingesting pdfs at large scale. I immediately looked into your playlist but couldn't identify the next video to digest the extracted data. Can you kindly guide me to the right video?
@llmware
@llmware 3 месяца назад
Hi @kushis4ever thank you so much for the kind comment! Yes we actually have a playlist for getting started with LLMWare: ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-uW3fElxcri4.html&pp=gAQBiAQB
@llmware
@llmware 3 месяца назад
Our Fast Start to RAG playlist in our channel will help you get started with RAG using LLMWare - please also join our discord community to interact with us and to learn more tips and tricks! discord.gg/4pMEYHxR2K
@MarxOrx
@MarxOrx 3 месяца назад
First 🎉
@llmware
@llmware 3 месяца назад
@bijucyborg
@bijucyborg 3 месяца назад
Congratulations for being part of GitHub accelerator.
@JehovahsaysNetworth
@JehovahsaysNetworth 3 месяца назад
Exactly 👍 start simple are be creative
@mrpeeker
@mrpeeker 3 месяца назад
i run into an issue when the model generates multiple answers, everything is by default, i just said hi to it and here is the answer i get. Hello! How can I assist you today? Whether it's answering questions, providing information, or helping with tasks, feel free to let me know what you need help with. <|assistant|> Hi there! I'm here to help you with any inquiries or assistance you may require. Just ask away! <|assistant|> Greetings! I'm ready and eager to provide any assistance you may need. What can I do for you today? <|assistant|> Hey! I'm your virtual helper. Just say the word and I'll be at your service for any assistance you may need. What can I do for you today? <|assistant|> Good day! As your virtual assistant, I'm ready and able to provide any support you may require. Please go ahead and ask your question or describe the task at hand. I'll do my best to assist you.<|end|><|assistant|> Hello! It's great to have you here. As your virtual assistant, I'm ready and available to help answer your questions and provide any assistance you may need. Just let me know how I can be of service!<|end|><|assistant|> Hey! Welcome! As your virtual assistant, I'm here to provide any support and assistance you may require. Just tell me how I can help and we'll get started right away!<|end|><|assistant|> Hey there! It's great to have you here. As your virtual assistant, I'm ready and eager to provide any assistance you may need. Just ask away and I'll do my best to help!<|end|><|assistant|> Good day! As your virtual assistant, I'm ready and able to provide any support and assistance you may require. Please go ahead and ask your question or describe the task at hand. I'll do my best to assist you in any way I can.<|end|><|assistant|> Welcome! It's great to have you here. As your virtual assistant, I'm ready and able to provide any support and assistance you may require. Just tell me how I can be of service and we'll get started right away!<|end|><|assistant|> Hey! It's great to have you here. As your virtual assistant, I'm ready and eager to provide any assistance and support you it is running on my nvidia 4060ti. ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: CUDA_USE_TENSOR_CORES: yes ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 4060 Ti, compute capability 8.9, VMM: yes
@llmware
@llmware 3 месяца назад
link to the example used: github.com/llmware-ai/llmware/blob/main/examples/UI/gguf_streaming_chatbot.py
@kom_senapati
@kom_senapati 3 месяца назад
I also have submitted similar app but with flask
@llmware
@llmware 3 месяца назад
Sounds awesome - can't wait to check it out!
@adhivp5594
@adhivp5594 3 месяца назад
Any ways to speed it up in Mac m1 air?
@llmware
@llmware 3 месяца назад
Hi It is the amount of memory you have in the Mac M1 Air. If you have 8GB it is suboptimal but with 16GB it should be fine.
@karthikb.s.k.4486
@karthikb.s.k.4486 3 месяца назад
Nice . May I know the theme used for pycharm IDE
@llmware
@llmware 3 месяца назад
Hi The theme we use is just basic "dark" theme and nothing else special... I hope this answers your question but please let me know if there are others!
@aqvayli
@aqvayli 3 месяца назад
What is the best small multilinguage model for sentiment and ton classification?
@andrewdang3401
@andrewdang3401 3 месяца назад
Keep up the work !
@umahatokula9586
@umahatokula9586 3 месяца назад
I love your videos but again, font size too tiny... Can hardly see anything
@llmware
@llmware 3 месяца назад
Hi This video was actually filmed before your feedback but in the later videos we do try to zoom in on the code more. Thank you for your patience! ❤
@figs3284
@figs3284 3 месяца назад
I think you turned the sound down too much 😂 I don't think there's any audio.
@llmware
@llmware 3 месяца назад
🤣We will try to get it right next time lol
@arunprasad8704
@arunprasad8704 4 месяца назад
I tried to pass bank statement which is in pdf format. but the tables within the pdf is not getting extracted. any change I need to make to improve parsing?
@manishadinesh2797
@manishadinesh2797 2 месяца назад
Hi even i have the same problem statement. Did u get any idea?
@llmware
@llmware 4 месяца назад
Link to the BLING PHI-3-GGUF: huggingface.co/llmware/bling-phi-3-gguf
@chrisbo3493
@chrisbo3493 4 месяца назад
Would like to see a tutorial to use it on an Android smartphone: local FOSS LLM, speech and image recognition with sufficient performance IMO are crucial
@RobynLeSueur
@RobynLeSueur 4 месяца назад
I'm a big fan of phi3, very impressive for something that fits on a smart phone. Will have to try out Bling.
@figs3284
@figs3284 4 месяца назад
Amazing stuff as always! Just turn down intro outro music a bit difficult to hear what you were saying.
@llmware
@llmware 4 месяца назад
Thank you so much for your great feedback! Will turn the music down for the next videos.
@johnkintree763
@johnkintree763 4 месяца назад
IMHO, the game changer will be a local language model that is an interface to a graph database that is local for personal and private data, and merges personal and public data with digital agents running on millions of other devices into a decentralized graph structure world model for planning collective actions.
@aravindchakrahari8966
@aravindchakrahari8966 4 месяца назад
I am having issues with this line: source = prompter.add_source_document(contracts_path, contract, query=key) The documents are not loaded and facing below error: source = prompter.add_source_document(contracts_path, contract, query=key) WARNING:root:No source materials attached to the Prompt. Running prompt_with_source inference without source may lead to unexpected results. Note: Made sure the docs are in the directory Please help!
@llmware
@llmware 4 месяца назад
Hi can you pls check if the sample files downloaded in the path /llmware_data/sample_files/ If not there then you can force download refresh as option in the pull sample files... If you are still having problems, please find us in discord so we can help you more there! discord.gg/pUvKzYujdM
@raptorate2872
@raptorate2872 4 месяца назад
You guys are doing God's work in the local LLM space. Wish you guys the best in the days to come
@llmware
@llmware 4 месяца назад
Thank you so much for your awesome feedback and for your support! We truly appreciate it!!💝
@johnkintree763
@johnkintree763 4 месяца назад
I would love to see performance measures for llmware models running on a Oneplus 11 with 16 GB of RAM. It has as good of hardware as any Android smartphone sold in 2023.
@johnkintree763
@johnkintree763 4 месяца назад
Maybe Oneplus or Qualcomm would pay for creating a package for llmware to run on the 16 GB Oneplus 11.
@johnkintree763
@johnkintree763 4 месяца назад
Brilliant. It's lovely seeing the pieces coming together for digital agents running on personal devices.
@llmware
@llmware 4 месяца назад
Thank you @johnkintree763! The rate of innovation in AI is truly amazing!
@adrianasolano1848
@adrianasolano1848 4 месяца назад
I get error: ModuleNotFoundError No Module named 'llmware'. Eventhough I cloned the repo as is from github. Any ideas?
@llmware
@llmware 4 месяца назад
Hi can you pls join us in discord in our technical support section and we will be happy to help you there. discord.gg/xnT8k4mCsh
@llmware
@llmware 4 месяца назад
Yes, if you cloned the repo, then please copy the examples out of their folder tree into the main path as a peer to the /llmware source - and you should be good to go!
@williamsokol0
@williamsokol0 4 месяца назад
What is the IDE used to look at your python code? I am not sure what is good for working with LLM
@llmware
@llmware 4 месяца назад
We use Pycharm - the free community edition. I hope this helps! Please also feel free to join us in our discord server with any questions.