-
Notifications
You must be signed in to change notification settings - Fork 13k
Permalink
Choose a base ref
{{ refName }}
default
Choose a head ref
{{ refName }}
default
Comparing changes
Choose two branches to see what’s changed or to start a new pull request.
If you need to, you can also or
learn more about diff comparisons.
Open a pull request
Create a new pull request by comparing changes across two branches. If you need to, you can also .
Learn more about diff comparisons here.
base repository: ollama/ollama
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: v0.10.1
Could not load branches
Nothing to show
Loading
Could not load tags
Nothing to show
{{ refName }}
default
Loading
...
head repository: ollama/ollama
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: v0.11.0
Could not load branches
Nothing to show
Loading
Could not load tags
Nothing to show
{{ refName }}
default
Loading
- 19 commits
- 58 files changed
- 4 contributors
Commits on Jul 31, 2025
-
kvcache: Enable SWA to retain additional entries
Models that use sliding window attention can only resume a sequence from the cache if it falls within the saved windows. This works well if the next message picks up where the old one left off. However, it generally prevents a partial prefix match unless the entire conversation falls within the sliding window. This can be a problem with reasoning models where the traces are supposed to be removed from future messages, forcing the entire history to be re-evaluated. This change allows models to specify that a larger amount of the history be retained in memory, to allow more partial resumption. It still respects the window that the model was trained on for token generation.
Configuration menu - View commit details
-
Copy full SHA for 4183bb0 - Browse repository at this point
Copy the full SHA 4183bb0View commit details -
Configuration menu - View commit details
-
Copy full SHA for 4a8fc3f - Browse repository at this point
Copy the full SHA 4a8fc3fView commit details -
Configuration menu - View commit details
-
Copy full SHA for f1c7384 - Browse repository at this point
Copy the full SHA f1c7384View commit details
Commits on Aug 4, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 9950f6e - Browse repository at this point
Copy the full SHA 9950f6eView commit details -
Configuration menu - View commit details
-
Copy full SHA for 26ade3a - Browse repository at this point
Copy the full SHA 26ade3aView commit details -
Configuration menu - View commit details
-
Copy full SHA for 6ca094a - Browse repository at this point
Copy the full SHA 6ca094aView commit details -
Configuration menu - View commit details
-
Copy full SHA for c8ac4cc - Browse repository at this point
Copy the full SHA c8ac4ccView commit details -
Configuration menu - View commit details
-
Copy full SHA for 9679520 - Browse repository at this point
Copy the full SHA 9679520View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9d1de41 - Browse repository at this point
Copy the full SHA 9d1de41View commit details -
Configuration menu - View commit details
-
Copy full SHA for 9194874 - Browse repository at this point
Copy the full SHA 9194874View commit details -
This implements the Open Compute Microscaling (MX) FP4 format as a tensor type with backend implementations focusing on mulmat and mulmatid on CPU, CUDA, and Metal.
Configuration menu - View commit details
-
Copy full SHA for 4fb47ed - Browse repository at this point
Copy the full SHA 4fb47edView commit details -
This exercises various operations and shapes on both CPU and GPU (if detected on the system)
Configuration menu - View commit details
-
Copy full SHA for 0263ad9 - Browse repository at this point
Copy the full SHA 0263ad9View commit details -
Configuration menu - View commit details
-
Copy full SHA for e6f39bc - Browse repository at this point
Copy the full SHA e6f39bcView commit details -
Configuration menu - View commit details
-
Copy full SHA for 0ac1c0d - Browse repository at this point
Copy the full SHA 0ac1c0dView commit details -
Read 4 bytes at a time (8 elements) when performing mul_mat_vec_mxfp4
Configuration menu - View commit details
-
Copy full SHA for aa43da4 - Browse repository at this point
Copy the full SHA aa43da4View commit details -
mac: fix crash on old macos versions
cblas_sgemm is only supported on v13.3 and up, however bf16 is only supported on v14+ so we were falling back to ggml-blas and crashing on bf16 tensors. Checking for the function being null seems to be the simplest way to condittionally avoid registering the backend.
Configuration menu - View commit details
-
Copy full SHA for 6a68a17 - Browse repository at this point
Copy the full SHA 6a68a17View commit details -
server: Minimum context length for gptoss
This model requires a minimum context length of 8192 to function effectively. Users can set higher values through all normal mechanisms but lower values will be silently reset.
Configuration menu - View commit details
-
Copy full SHA for f5fd7cc - Browse repository at this point
Copy the full SHA f5fd7ccView commit details -
ggml: Multiply by numParallel for gptoss sliding window
When computing the graph size estimate, the context size is already multiplied by numParallel so estimates reflect that. However, since sliding window models use a smaller, fixed context size, they need to manually take numParallel into account.
Configuration menu - View commit details
-
Copy full SHA for 8306248 - Browse repository at this point
Copy the full SHA 8306248View commit details
Commits on Aug 5, 2025
-
Configuration menu - View commit details
-
Copy full SHA for d552068 - Browse repository at this point
Copy the full SHA d552068View commit details
Loading
This comparison is taking too long to generate.
Unfortunately it looks like we can’t render this comparison for you right now. It might be too big, or there might be something weird with your repository.
You can try running this command locally to see the comparison on your machine:
git diff v0.10.1...v0.11.0