If you find this repository helpful, please consider giving it a star⭐️
Welcome to the comprehensive collection of advanced + agentic Retrieval-Augmented Generation (RAG) techniques.
RAG is a popular method that improves accuracy and relevance by finding the right information from reliable sources and transforming it into useful answers. This repository covers the most effective advanced + agentic RAG techniques with clear implementations and explanations.
The main goal of this repository is to provide a helpful resource for researchers and developers looking to use advanced RAG techniques in their projects. Building these techniques from scratch takes time, and finding proper evaluation methods can be challenging. This repository simplifies the process by offering ready-to-use implementations and guidance on how to evaluate them.
Note
This repository starts with naive RAG as a foundation and progresses to advanced and agentic techniques. It also includes research papers/references for each RAG technique, which you can explore for further reading.
Large Language Models are trained on a fixed dataset, which limits their ability to handle private or recent information. They can sometimes "hallucinate", providing incorrect yet believable answers. Fine-tuning can help but it is expensive and not ideal for retraining again and again on new data. The Retrieval-Augmented Generation (RAG) framework addresses this issue by using external documents to improve the LLM's responses through in-context learning. RAG ensures that the information provided by the LLM is not only contextually relevant but also accurate and up-to-date.
There are four main components in RAG:
Indexing: First, documents (in any format) are split into chunks, and embeddings for these chunks are created. These embeddings are then added to a vector store.
Retriever: Then, the retriever finds the most relevant documents based on the user's query, using techniques like vector similarity from the vector store.
Augment: After that, the Augment part combines the user's query with the retrieved context into a prompt, ensuring the LLM has the information needed to generate accurate responses.
Generate: Finally, the combined query and prompt are passed to the model, which then generates the final response to the user's query.
These components of RAG allow the model to access up-to-date, accurate information and generate responses based on external knowledge. However, to ensure RAG systems are functioning effectively, it’s essential to evaluate their performance.
Evaluating RAG applications is important for understanding how well these systems work. We can see how effectively they combine information retrieval with generative models by checking their accuracy and relevance. This evaluation helps improve RAG applications in tasks like text summarization, chatbots, and question-answering. It also identifies areas for improvement, ensuring that these systems provide trustworthy responses as information changes. Overall, effective evaluation helps optimize performance and builds confidence in RAG applications for real-world use. These notebooks contain an end-to-end RAG implementation + RAG evaluation part in Athina AI.
Here are the details of all the Advanced RAG techniques covered in this repository.
Here are the details of all the Agentic RAG techniques covered in this repository.
A quick demo of how each notebook works:
demo.mp4
First, clone this repository by using the following command:
git clone https://github.com/athina-ai/rag-cookbooks.git
Next, navigate to the project directory:
cd rag-cookbooks
Once you are in the 'rag-cookbooks' directory, follow the detailed implementation for each technique.
If you have a new technique or improvement to suggest, we welcome contributions from the community!
This project is licensed under MIT License