Advanced RAG With Semantic Caching, Semantic Routing, and Observability (Langchain, Ollama, Milvus, Redis, Langfuse, and Uptrain)
-
Semantic Caching:
- Semantic caching in Retrieval-Augmented Generation (RAG) systems enhances efficiency and relevance by storing semantically relevant information, reducing retrieval time and computational resources.
-
LLM Routing:
- LLM routing using LiteLLM involves directing user queries to the most appropriate language model based on factors like query complexity, domain specificity, or required response quality. This optimizes the use of computational resources and improves response accuracy.
-
Guardrails:
- Guardrails are mechanisms or guidelines implemented to ensure that a system, especially an AI or machine learning model, operates within safe, ethical, and intended boundaries, preventing unintended outcomes or misuse.
-
UpTrain:
- UpTrain is an open-source platform for evaluating and improving LLM applications.
-
Langfuse:
- Langfuse is an open-source platform for monitoring, evaluating, and improving LLM applications.
-
Clone the Repository:
git clone https://github.com/Major-wagh/Advanced-RAG.git cd Advanced-RAG
-
Install the requirements:
pip install -r requirements.txt
-
Run the Notebook:
Launch the Jupyter Notebook in your environment and open the project notebooks to start exploring and experimenting with the RAG techniques.
We welcome contributions to this project! Please read our contributing guidelines to get started.