What's RaaG?
Ever wanted to have a chat with a LLM (Large Language Model) like GPT-4o, but with the ability to ask it questions about your data? RaaG is the answer to your needs!
RaaG is the acronyme for "RAG as a GPT"
Why do we need dRAGon RaaG?
You can already use LangChain or LlamaIndex to query your data, so why do you need RaaG?
As you know, a typical RAG application is composed of two main components with each several sub-tasks :
Indexing: This involves the offline process of ingesting data from a source and indexing it.
Load: The data is loaded using Document Loaders.
Split: The loaded data is divided into smaller chunks using Document Splitters.
Store: The chunks are stored and indexed for future searching, typically utilizing a VectorStore and an Embeddings model.
Retrieval and Generation: This is the core RAG chain that operates during runtime, handling user queries.
Retrieve: Relevant data chunks are fetched from the storage in response to a user query using a Retriever.
Generate: A ChatModel/LLM generates an answer by incorporating the user query and the retrieved data into a prompt.
Let's say you want to build a RAG pipeline using LangChain: You would need to write code to handle each of these sub-tasks, which can be quite complex and time-consuming.
How does dRAGon RaaG simplify the process?
So instead of having an app code like this:
You will simply use the following code with a dRAGon powered RAG pipeline:
dRAGon simplifies the process of building a RAG pipeline by providing a single interface to manage all these components by providing an OpenAI API compliant endpoint to interact with your RAG pipeline.
Components involved in dRAGon RaaG
Tutorials
Last updated
Was this helpful?