-
Notifications
You must be signed in to change notification settings - Fork 1.7k
roadmap: Jan has threads based RAG #4817
New issue
Have a question about this project? No Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “No Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? No Sign in to your account
Comments
hey... after answer from the model, maybe an option also before (show the snippets as plain text) next step. you can set for every LLM model and embedd-Model and document with your own settings (systempromt, token, ...all) next step, implement more embedder models, bert-based are more or less same way, but it give some others qwen, jina and some ranking embedders... and one nice option from LM-Studio(not opensource) |
one options for embedding "chat" or "query" so more chat or force to keep all answer out of the doc! and one option (small button near the chat) to keep the snippets in VRAM after first answer or delete them. it should be set to delete by default when I ask a new question afterwards... |
maybe in next next step an option to send prefixes to embedder and |
Thank you for sharing this, @kalle07. It's truly helpful. |
@louis-menlo -- I started working on a simple document parsing implementation with a bit of RAG using @dan-menlo mentioned you have specific plans for implementing this via MCP. Can I confirm with you whether you would like me to add support for the things I mentioned or remove this as it will be done in a different way? |
cc @ramonpzg likely will drive this. |
Goal
Tasks
The text was updated successfully, but these errors were encountered: