Тёмный

RAG From Scratch: Part 2 (Indexing) 

LangChain
Подписаться 42 тыс.
Просмотров 22 тыс.
50% 1

This is the second video in our series on RAG. The aim of this series is to build up an understanding of RAG from scratch, starting with the basics of indexing, retrieval, and generation. This video focuses on indexing, covering the process of document loading, splitting, and embedding.
Code:
github.com/langchain-ai/rag-f...
Slides:
docs.google.com/presentation/...

Опубликовано:

 

1 июн 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 11   
@user-wj7ys7mk2d
@user-wj7ys7mk2d 3 месяца назад
For the very first time in my mechanical engineering life, I think i learned something in detail about software engineering! Thank you!
@alexenax1109
@alexenax1109 3 месяца назад
Amazing series! Thank you Lance!
@hamzafarouk1538
@hamzafarouk1538 3 месяца назад
Thanks. Nice video short and clear. But why do you need to store the embedding in a vector db
@Orcrambo
@Orcrambo 3 месяца назад
For efficient similarity searches
@horyekhunley
@horyekhunley 3 месяца назад
When performing RAG, you need to encode your input data into embeddings because that is what the LLM understands, it is from these embeddings that the model will perform decoding of the output and subsequently give you the result you've asked for. These embeddings need to be stored somewhere, they can be small or very large, some vector dbs can be free, opensource and be in memory like faiss and chroma, they can also be paid and hosted like pinecone
@clivejefferies
@clivejefferies 2 месяца назад
​@horyekhunley thanks for the insight
@sepia_tone
@sepia_tone 2 месяца назад
@@horyekhunley the process is a little different The LLM doesn't touch the embeddings. The embeddings are used to convert the documents to a form that can more quickly and accurately be compared to the question (which is also converted to an embedding). This is done by an embeddings model (in this example an embeddings model from OpenAI (referred to as OpenAIEmbeddings() in the code) is used. These embeddings and their associated documents need to be stored somewhere (in this example Chroma is used). This is the indexing phase. After the comparison between the embedding of the question and the stored documents, a subset of the documents which have a high similarity (in embedding space) with the question is given to the LLM. This is the retrieval phase. Finally, the LLM uses the returned documents and it's own knowledge to reason and given an answer to the user. This is the generation phase..
@tongweiwang9775
@tongweiwang9775 Месяц назад
@@sepia_tone Still didn't quite get what's the role of indexing here. Relevant documents are retrieved based on the embeddings of the split documents and the embedding of the question. So what's indexing doing here?
Далее
RAG From Scratch: Part 3 (Retrieval)
5:14
Просмотров 15 тыс.
She’s Giving Birth in Class…?
00:21
Просмотров 4,7 млн
UZmir & Mira - Qani qani (Snippet)
00:26
Просмотров 358 тыс.
What is Retrieval-Augmented Generation (RAG)?
6:36
Просмотров 515 тыс.
RAG From Scratch: Part 1 (Overview)
5:13
Просмотров 57 тыс.
LangGraph 101: it's better than LangChain
32:26
Просмотров 42 тыс.
RAG From Scratch: Part 4 (Generation)
6:25
Просмотров 12 тыс.