Quick 2 minute Tutorial to understand LLM (RAG)s
What are ๐ฅ๐ฒ๐๐ฟ๐ถ๐ฒ๐๐ฎ๐น ๐๐๐ด๐บ๐ฒ๐ป๐๐ฒ๐ฑ ๐๐ฒ๐ป๐ฒ๐ฟ๐ฎ๐๐ถ๐ผ๐ป (๐ฅ๐๐) ๐ฆ๐๐๐๐ฒ๐บ๐?
Here is an example of a simple RAG based Chatbot to query your Private Knowledge Base.
First step is to store the knowledge of your internal documents in a format that is suitable for querying. We do so by embedding it using an embedding model:
๐ญ: Split text corpus of the entire knowledge base into chunks โ a chunk will represent a single piece of context available to be queried. Data of interest can be from multiple sources, e.g. Documentation in Confluence supplemented by PDF reports.
๐ฎ: Use the Embedding Model to transform each of the chunks into a vector embedding.
๐ฏ: Store all vector embeddings in a Vector Database.
๐ฐ: Save text that represents each of the embeddings separately together with the pointer to the embedding (we will need this later).
Next we can start constructing the answer to a question/query of interest:
๐ฑ: Embed a question/query you want to ask using the same Embedding Model that was used to embed the knowledge base itself.
๐ฒ: Use the resulting Vector Embedding to run a query against the index in the Vector Database. Choose how many vectorsโฆ