Ollama Now Runs Faster on Macs Thanks to Apple's MLX Framework - MacRumors
Skip to Content

Ollama Now Runs Faster on Macs Thanks to Apple's MLX Framework

Ollama, the popular app for running AI models locally on a computer, has released an update that takes advantage of Apple's own machine learning framework, MLX. The result is a hefty speed boost on Macs with Apple silicon.

ollama logo mac
According to Ollama, the new version processes prompts around 1.6 times faster (prefill speed) and nearly doubles the speed at which it generates responses (decode speed). Macs with M5-series chips are said to see the largest improvements, thanks to Apple's new GPU Neural Accelerators.

The update also includes smarter memory management, which should make AI-powered coding tools and chat assistants feel noticeably more responsive during extended use.

Ollama says the new performance boost should especially benefit macOS users who run personal assistants like OpenClaw or coding agents like Claude Code, OpenCode, or Codex.

The preview release is available to download as Ollama 0.19 – just make sure you have a Mac with more than 32GB of unified memory to run it. Support is currently limited to Alibaba's Qwen3.5, but Ollama says support for more AI models is planned.

Top Rated Comments

2 hours ago at 03:27 am
This is going to be some serious cash flow incoming for Apple in this year.
Score: 4 Votes (Like | Disagree)
1 hour ago at 04:02 am

This is going to be some serious cash flow incoming for Apple in this year.
I think this could be a major business for Apple - it’s way cheaper for a small business to buy a powerful Mac and run qwen 3.5 than pay for an enterprise license for a frontier model - and you don’t need to worry about privacy issues.
Score: 3 Votes (Like | Disagree)
Justin Cymbal Avatar
2 hours ago at 03:23 am
M-Series chips at work😎
Score: 3 Votes (Like | Disagree)
1 hour ago at 04:09 am
On device is definitely gonna be the future.

I can’t help but wonder if Apple looked ahead and foresaw this when developing the M series, or if they’ve lucked into it.
Score: 2 Votes (Like | Disagree)
Kirkster Avatar
50 minutes ago at 04:26 am
They are so far behind LM Studio. And only support for one model?
Score: 1 Votes (Like | Disagree)
1 hour ago at 04:10 am
I’m completely newbie here but would like to try local AI for documentation research and analysis, sort out files, translations.
Would Ollama be adequate ? And are more than 32 Go mandatory ?
Score: 1 Votes (Like | Disagree)