-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Add BM25 options to inference config #6893
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add BM25 options to inference config #6893
Conversation
e8c4b13
to
345c165
Compare
@@ -7,6 +7,36 @@ inference: | |||
address: "http://localhost:2114/api/v1/infer" | |||
timeout: 10 | |||
token: "98eb568c-dea5-4347-a3a5-583478983bb9" | |||
# Define custom models/vectorizer, which get handled directly in Qdrant. These do not require an `address` to be configured. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe they can, but it is up to the specific model params. In the case of BM25 it is not needed bc it is computed in qdrant directly.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What do you mean? custom_models
is supposed to only handle 'models' that are handled directly in Qdrant, like BM25.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I was thinking that maybe in the future users might be able to plug in their own embedding services via a custom config like this. In this case another default namespace might be more fitting like bm25/bm25-for-paragraphs
, or native/bm25
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am not convinced inference config is the right place for this settings.
Inference settings are global, while we would want to have it per-collection.
I propose to postpone this change and make it part of custom collection parameters instead.
Closing in favor of #6939 |
Depends on #6891
Adds configuration for BM25 vectorizer "models" in inference. One example model with the name
bm25
added in development config too.Highlights: