HomeTechnologyRunning local models on Macs gets faster with Ollama's MLX support

Running local models on Macs gets faster with Ollama's MLX support

TechnologyApril 1, 2026
1 min read
Running local models on Macs gets faster with Ollama's MLX support
Apple Silicon Macs get a performance boost thanks to better unified memory usage.

Ollama, a runtime system for operating large language models on a local computer, has introduced support for Apple's open source MLX framework for machine learning. Additionally, Ollama says it has improved caching performance and now supports Nvidia's NVFP4 format for model compression, making for much more efficient memory usage in certain models.

Combined, these developments promise significantly improved performance on Macs with Apple Silicon chips (M1 or later)—and the timing couldn't be better, as local models are starting to gain steam in ways they haven't before outside researcher and hobbyist communities.

The recent runaway success of OpenClaw—which raced its way to over 300,000 stars on GitHub, made headlines with experiments like Moltbook and became an obsession in China in particular—has many people experimenting with running models on their machines.

Read full article

Comments

Source: Ars Technica

Share this article

Related Articles

Apple names Johny Srouji as chief hardware officer
2026Apr 21

Apple names Johny Srouji as chief hardware officer

Apple has appointed Johny Srouji as its new chief hardware officer, "effective immediately," according to an announcement on Monday. He is stepping into the shoes of current hardware engineering head

Article1 min read
Read More
The Lenovo Legion Go S is RAMageddon’s latest victim
2026Apr 21

The Lenovo Legion Go S is RAMageddon’s latest victim

You can still find the Asus Xbox Ally X and the MSI Claw 8 AI Plus for $999 and $1,049 respectively, but Lenovo's Legion Go S has seemingly given up the fight. The best version of Lenovo's 8-inch hand

Article1 min read
Read More