Skip to content

Conversation

JojiiOfficial
Copy link
Contributor

@JojiiOfficial JojiiOfficial commented Jul 17, 2025

Depends on #6891

Adds configuration for BM25 vectorizer "models" in inference. One example model with the name bm25 added in development config too.
Highlights:

  • Supports multiple models with different configs
  • Reuse of our tokenizers with full customization support

@JojiiOfficial JojiiOfficial marked this pull request as ready for review July 18, 2025 08:49
@JojiiOfficial JojiiOfficial changed the title [WIP] Add BM25 options to inference config Add BM25 options to inference config Jul 18, 2025
@@ -7,6 +7,36 @@ inference:
address: "http://localhost:2114/api/v1/infer"
timeout: 10
token: "98eb568c-dea5-4347-a3a5-583478983bb9"
# Define custom models/vectorizer, which get handled directly in Qdrant. These do not require an `address` to be configured.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe they can, but it is up to the specific model params. In the case of BM25 it is not needed bc it is computed in qdrant directly.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What do you mean? custom_models is supposed to only handle 'models' that are handled directly in Qdrant, like BM25.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I was thinking that maybe in the future users might be able to plug in their own embedding services via a custom config like this. In this case another default namespace might be more fitting like bm25/bm25-for-paragraphs, or native/bm25

@JojiiOfficial JojiiOfficial requested a review from coszio July 23, 2025 07:52
Copy link
Member

@generall generall left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I am not convinced inference config is the right place for this settings.
Inference settings are global, while we would want to have it per-collection.

I propose to postpone this change and make it part of custom collection parameters instead.

@JojiiOfficial
Copy link
Contributor Author

Closing in favor of #6939

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants