-
-
Notifications
You must be signed in to change notification settings - Fork 840
WIP: Bookmark embeddings #834
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Deploying hoarder-docs with
|
Latest commit: |
1b52710
|
Status: | ✅ Deploy successful! |
Preview URL: | https://5bd08d4e.hoarder.pages.dev |
Branch Preview URL: | https://bookmark-embeddings.hoarder.pages.dev |
Deploying hoarder-landing with
|
Latest commit: |
1b52710
|
Status: | ✅ Deploy successful! |
Preview URL: | https://7b4c917a.hoarder-landing.pages.dev |
Branch Preview URL: | https://bookmark-embeddings.hoarder-landing.pages.dev |
Just linking #1315 here as I think it might be better to make it part of the first release that contains embeddings, otherwise we'll have to deal with the issue of resetting all the embeddings when switching to binary or not. |
I hope you won't mind but I was wondering what were the blockers for this much awaited feature. What is the state of the code, what are questions left to answer, what can the community do to help, etc. Because it's a major change, I'm afraid no one other than the maintainer would want to tackle it and so I think communication here would go a long way to make us help you without duplicating too much effort. |
@thiswillbeyourgithub So generating the embeddings currently work. The biggest blocker for this (beside my time) is me choosing what vector storage service to use. I initially was planning on Another caveat is that a lot of ollama users don't have an embeddings model set, so I'll need a way to communicate that as well. Nothing major left, but a lot of tiny things for a smooth rollout of such a feature. |
I don't think that the lack of embedding model on most user's setup is an issue because ollama use is opt in. As long as you haven't set an embedding model, you won't have embeddings search and that's fine. Fwiw: on ollama i've been happy with arctic embed 2 with quantization (not enabled by default, but ends up with a 450mb model or so) link. I would gladely contribute my Modelfile to help others set this up. Personaly, my money is on meillisearch. If you don't mind (i.e. don't tell me shortly not to do so) I'll take a look on their repo and maybe ask them directly for guidance. Also I'd say that for a smooth rollout it might be good to do a pre release/alpha/beta for a few weeks/month beforehand to iron out the misshaps. Edit: I took a look at meilisearch: what do you think of a docker one liner using alpine to execute an rm command on the meillisearch db then asking the user to force reindexing the db in karakeep? Edit: after reading issue 2570 of meillisearch (not linking to avoid making a fuss) I kinda get the feeling that they don't want to make the upgrade easy because they view "easy upgrade" as their moat. Not very reassuring but still a good product. I think the alpine way is a good solution. Or using a new volume entirely. |
I was suddenly reminded of that prominent quote in the readme:
So my opinion is now that, as long as there is no data loss (ever!), this disclaimer grants you the moral right to move faster on those roll out situations.
Just my 2 cents on priorities of course. I'm aware you're very busy and am sure you'll do what's best for us all. |
We are using the same OpenAI api-key and base url for both inference and embedding. I think have separate config for each of them would be better, as it allows for using different providers for each of these tasks. In addition, this also would remove the limitation of having to use a provider who supports both embeddings and chat. |
I sort of disagree as for more cutstomization you should instead run a litellm instance |
Litellm is one of the llm gateway providers that support embeddings , there are others that do not support embeddings but solely focus on llm inference (openrouter, cloudflare). Adding this would be good separation of concerns imo, if needed we can default to the inference parameters if the embedding params are not set. |
No description provided.