Skip to content

roadmap: Jan has threads based RAG #4817

New issue

Have a question about this project? No Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “No Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? No Sign in to your account

Open
1 task
dan-menlo opened this issue Mar 20, 2025 · 6 comments
Open
1 task

roadmap: Jan has threads based RAG #4817

dan-menlo opened this issue Mar 20, 2025 · 6 comments
Assignees
Milestone

Comments

@dan-menlo
Copy link
Contributor

dan-menlo commented Mar 20, 2025

Goal

  • Have a simple "Chat with Docs"
    • Only for new conversations
    • LanceDB
  • But still support legacy RAG (let it be)

Tasks

  • Sync up w/ Louis
@dan-menlo dan-menlo added this to Menlo Mar 20, 2025
@github-project-automation github-project-automation bot moved this to Investigating in Menlo Mar 20, 2025
@kalle07
Copy link

kalle07 commented Mar 23, 2025

hey...
please start as simple as possible with nomic(as ambedder) and it will work well
if iam right some code you find here
https://github.com/nomic-ai/gpt4all
important that you can set embedding size and number of snippets
a fine start is "512token" and "4" snippets
create a collection and option to see my docs , how many words and how many verctors embedded (maybe live embedding)

after answer from the model, maybe an option also before (show the snippets as plain text)

next step. you can set for every LLM model and embedd-Model and document with your own settings (systempromt, token, ...all)
take a look Anything LM, they create workspaces for every settings

next step, implement more embedder models, bert-based are more or less same way, but it give some others qwen, jina and some ranking embedders...

and one nice option from LM-Studio(not opensource)
among other things they have an option for User-Poweruser-Developer (button below and instantly the layout and more option setting are changed)

@kalle07
Copy link

kalle07 commented Mar 28, 2025

one options for embedding "chat" or "query" so more chat or force to keep all answer out of the doc!

and one option (small button near the chat) to keep the snippets in VRAM after first answer or delete them.

it should be set to delete by default when I ask a new question afterwards...

@kalle07
Copy link

kalle07 commented Apr 11, 2025

maybe in next next step an option to send prefixes to embedder
https://www.youtube.com/watch?v=76EIC_RaDNw

and
re-ranker models (but i dont know what is the different)
one i have here
https://huggingface.co/kalle07/embedder_collection

@louis-menlo louis-menlo pinned this issue Apr 12, 2025
@louis-menlo louis-menlo moved this to Todo in Jan Apr 12, 2025
@louis-menlo louis-menlo changed the title roadmap: Jan's to transition legacy RAG roadmap: Jan has threads based RAG Apr 21, 2025
@louis-menlo
Copy link
Contributor

Thank you for sharing this, @kalle07. It's truly helpful.

@ramonpzg
Copy link
Contributor

@louis-menlo -- I started working on a simple document parsing implementation with a bit of RAG using llamaindex.ts and LanceDB on a separate branch based on the Tauri branch. This is mainly for enabling users to add documents, images, videos and audio while having it all in a db.lance file for storing and querying as users need it.

@dan-menlo mentioned you have specific plans for implementing this via MCP. Can I confirm with you whether you would like me to add support for the things I mentioned or remove this as it will be done in a different way?

@menloresearch menloresearch deleted a comment from david-menloai Apr 22, 2025
@louis-menlo
Copy link
Contributor

cc @ramonpzg likely will drive this.

No Sign up for free to join this conversation on GitHub. Already have an account? No Sign in to comment
Projects
Status: Todo
Status: Investigating
Development

No branches or pull requests

5 participants