Allow Retrieval-augmented generation (RAG) for better contextualization #363
alexcontor
started this conversation in
Feature Requests
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
In order to give actual context to the Chatbot rather than relying on the last messages combined within the user prompt (with the token limitations that this may infer if the conversation gets too long) it would be amazing if we could fit our RAG and embeddings into the actual LLM Inference.
Beta Was this translation helpful? Give feedback.
All reactions