CARVIEW |
Navigation Menu
-
Notifications
You must be signed in to change notification settings - Fork 12.6k
Releases: ollama/ollama
v0.10.0
6dcc5df
Compare
Ollama's new app
Ollama's new app is available for macOS and Windows: Download Ollama

What's Changed
ollama ps
will now show the context length of loaded models- Improved performance in
gemma3n
models by 2-3x - Parallel request processing now defaults to 1. For more details, see the FAQ
- Fixed issue where tool calling would not work correctly with
granite3.3
andmistral-nemo
models - Fixed issue where Ollama's tool calling would not work correctly if a tool's name was part of of another one, such as
add
andget_address
- Improved performance when using multiple GPUs by 10-30%
- Ollama's OpenAI-compatible API will now support WebP images
- Fixed issue where
ollama show
would report an error ollama run
will more gracefully display errors
New Contributors
- @sncix made their first contribution in #11189
- @mfornet made their first contribution in #11425
- @haiyuewa made their first contribution in #11427
- @warting made their first contribution in #11461
- @ycomiti made their first contribution in #11462
- @minxinyi made their first contribution in #11502
- @ruyut made their first contribution in #11528
Full Changelog: v0.9.6...v0.10.0
Assets 15
- sha256:dd90834fdc124b1292b805eae562feed5de15617ea58827375defb043814b04622.5 MB
2025-07-30T17:25:50Z - sha256:2b06828de4a6ff5dc401b3d10113784e8c07d74043106a6914f65ae78a30039440.8 MB
2025-07-30T17:25:47Z - sha256:fe45e523ed39f41e228d2507ac11679d77ffbcf8ebfb54d07c8748e18e68917a1.07 GB
2025-07-30T17:25:51Z - sha256:6bfe6b5eea83848a2af163ed36b4d0031c20d36463ef230bfa19313a56dfcfcc1.22 GB
2025-07-30T17:25:52Z - sha256:6437cb84a69ff1f382ca3ff86127e8566ae6eb44977c5fd8408bf253165a1d83434 MB
2025-07-30T17:25:53Z - sha256:d18b189a357f64ab88f022a7c6a538bb7901bd7b6b6aa74266fabdd880c0de47345 MB
2025-07-30T17:25:54Z - sha256:c43713794fb54ded923c095ed7e63a463f337caeb505d4794b180288476b22f31.04 GB
2025-07-30T17:25:55Z - sha256:9f8c185877df2df969229404a6ae38629624335cf8213af6e52492ffc15f3e1d246 MB
2025-07-30T17:25:47Z - sha256:3e6134aeec608509f39227449b6283509cbb5d5a831b9e7f57e8392fb12804f51.24 GB
2025-07-30T17:25:48Z - sha256:f142cd14f3ca6bc9b67515730212010cd3202663d3381112f15755b0dd711ff820.6 MB
2025-07-30T17:25:49Z -
2025-07-30T15:56:01Z -
2025-07-30T15:56:01Z - Loading
v0.9.6
43107b1
Compare
What's Changed
- Fixed styling issue in launch screen
tool_name
can now be provided in messages with"role": "tool"
using the/api/chat
endpoint
New Contributors
Full Changelog: v0.9.5...v0.9.6-rc0
Assets 15
v0.9.5
5d8c173
Compare
Updates to Ollama for macOS and Windows
A new version of Ollama's macOS and Windows applications are now available. New improvements to the apps will be introduced over the coming releases:

New features
Expose Ollama on the network
Ollama can now be exposed on the network, allowing others to access Ollama on other devices or even over the internet. This is useful for having Ollama running on a powerful Mac, PC or Linux computer while making it accessible to less powerful devices.
Model directory
The directory in which models are stored can now be modified! This allows models to be stored on external hard disks or alternative directories than the default.
Smaller footprint and faster starting on macOS
The macOS app is now a native application and starts much faster while requiring a much smaller installation.
Additional changes in 0.9.5
- Fixed issue where the
ollama
CLI would not be installed by Ollama on macOS on startup - Fixed issue where files in
ollama-darwin.tgz
were not notarized - Add NativeMind to Community Integrations by @xukecheng in #11242
- Ollama for macOS now requires version 12 (Monterey) or newer
New Contributors
- @xukecheng made their first contribution in #11242
Assets 15
v0.9.4
44b17d2
Compare
Updates to Ollama for macOS and Windows
A new version of Ollama's macOS and Windows applications are now available. New improvements to the apps will be introduced over the coming releases:

New features
Expose Ollama on the network
Ollama can now be exposed on the network, allowing others to access Ollama on other devices or even over the internet. This is useful for having Ollama running on a powerful Mac, PC or Linux computer while making it accessible to less powerful devices.
Model directory
The directory in which models are stored can now be modified! This allows models to be stored on external hard disks or alternative directories than the default.
Smaller footprint and faster starting on macOS
The macOS app is now a native application and starts much faster while requiring a much smaller installation.
What's Changed
- Reduced download size and startup time for Ollama on macOS
- Tool calling with empty parameters will now work correctly
- Fixed issue when quantizing models with the Gemma 3n architecture
- Ollama for macOS should not longer ask for root privileges when updating unless required
- Ollama for macOS now requires version 12 (Monterey) or newer
Full Changelog: v0.9.3...v0.9.4
Assets 15
v0.9.3
ba04902
Compare
Gemma 3n
Ollama now supports Gemma 3n.
Gemma 3n models are designed for efficient execution on everyday devices such as laptops, tablets or phones. These models were trained with data in over 140 spoken languages.
Effective 2B
ollama run gemma3n:e2b
Effective 4B
ollama run gemma3n:e4b
What's Changed
- Fixed issue where errors would not be properly reported on Apple Silicon Macs
- Ollama will now limit context length to what the model was trained against to avoid strange overflow behavior
New Contributors
Full Changelog: v0.9.2...v0.9.3
Assets 16
v0.9.2
ed567ef
Compare
What's Changed
- Fixed issue where tool calls without parameters would not be returned correctly
- Fixed
does not support generate
errors - Fixed issue where some special tokens would not be tokenized properly for some model architectures
New Contributors
Full Changelog: v0.9.1...v0.9.2
Assets 16
v0.9.1
5a8eb0e
Compare
Tool calling improvements
New tool calling support
The following models now support tool calling:
- DeepSeek-R1-2508 (671B model)
- Magistral
Tool calling reliability has also been improved for the following models:
To re-download the models, use ollama pull
.
New Ollama for macOS and Windows preview
A new version of Ollama's macOS and Windows applications are available to test for early feedback. New improvements to the apps will be introduced over the coming releases:
If you have feedback, please create an issue on GitHub with the app
label. These apps will automatically update themselves to future versions of Ollama, so you may have to redownload new preview versions in the future.

New features
Expose Ollama on the network
Ollama can now be exposed on the network, allowing others to access Ollama on other devices or even over the internet. This is useful for having Ollama running on a powerful Mac, PC or Linux computer while making it accessible to less powerful devices.
Allow local browser access
Enabling this allows websites to access your local installation of Ollama. This is handy for developing browser-based applications using Ollama's JavaScript library.
Model directory
The directory in which models are stored can now be modified! This allows models to be stored on external hard disks or alternative directories than the default.
Smaller footprint and faster starting on macOS
The macOS app is now a native application and starts much faster while requiring a much smaller installation.
What's Changed
- Magistral now supports disabling thinking mode. Note: it is also recommended to change the system prompt when doing so.
- Error messages that previously showed
POST predict
will now be more informative - Improved tool calling reliability for some models
- Fixed issue on Windows where
ollama run
would not start Ollama automatically
New Contributors
- @JasonHonKL made their first contribution in #10174
- @hwittenborn made their first contribution in #10998
- @krzysztofjeziorny made their first contribution in #10973
Full Changelog: v0.9.0...v0.9.1
Assets 16
v0.9.0
5f57b0e
Compare
New models
- DeepSeek-R1-2508: DeepSeek-R1 has received a minor version upgrade to DeepSeek-R1-0528 for the 8 billion parameter distilled model and the full 671 billion parameter model. In this update, DeepSeek R1 has significantly improved its reasoning and inference capabilities.
Thinking
Ollama now has the ability to enable or disable thinking. This gives users the flexibility to choose the modelβs thinking behavior for different applications and use cases.
When thinking is enabled, the output will separate the modelβs thinking from the modelβs output. When thinking is disabled, the model will not think and directly output the content.
Models that support thinking:
- DeepSeek R1
- Qwen 3
- more will be added under thinking models.
When running a model that supports thinking, Ollama will now display the model's thoughts:
% ollama run deepseek-r1
>>> How many Rs are in strawberry
Thinking...
First, I need to understand what the question is asking. It's asking how many letters 'R' are present in the word "strawberry."
Next, I'll examine each letter in the word individually.
I'll start from the beginning and count every occurrence of the letter 'R.'
After reviewing all the letters, I determine that there are three instances where the letter 'R' appears in the word "strawberry."
...done thinking.
There are three **Rs** in the word **"strawberry"**.
In Ollama's API, a model's thinking is now returned as a separate thinking
field for easy parsing:
{
"message": {
"role": "assistant",
"thinking": "First, I need to understand what the question is asking. It's asking how many letters 'R' are present in the word "strawberry...",
"content": "There are **3** instances of the letter **R** in the word **"strawberry."**"
}
}
Turning thinking on and off
In the API, thinking can be enabled by passing "think": true
and disabled by passing "think": false
curl https://localhost:11434/api/chat -d '{
"model": "deepseek-r1",
"messages": [
{
"role": "user",
"content": "Why is the sky blue?"
},
],
"think": true
}'
In Ollama's CLI, use /set think
and /set nothink
to enable and disable thinking.
What's Changed
- Add thinking support to Ollama
Full Changelog: v0.8.0...v0.9.0
Assets 14
v0.8.0
aa25aff
Compare
What's Changed
- Ollama will now stream responses with tool calls blog post
- Logs will now include better memory estimate debug information when running models in Ollama's engine.
New Contributors
- @hellotunamayo made their first contribution in #10790
Full Changelog: v0.7.1...v0.8.0
Assets 14
v0.7.1
884d260
Compare
What's Changed
- Improved model memory management to allocate sufficient memory to prevent crashes when running multimodal models in certain situations
- Enhanced memory estimation for models to prevent unintended memory offloading
ollama show
will now show...
when data is truncated- Fixed crash that would occur with
qwen2.5vl
- Fixed crash on Nvidia's CUDA for
llama3.2-vision
- Support for Alibaba's Qwen 3 and Qwen 2 architectures in Ollama's new multimodal engine
New Contributors
- @ronxldwilson made their first contribution in #10763
- @DarkCaster made their first contribution in #10779
Full Changelog: v0.7.0...v0.7.1