AI & TechArtificial IntelligenceBigTech CompaniesNewswireTechnology

Ollama MLX Support Speeds Up Mac AI Models

▼ Summary

– Ollama now supports Apple’s MLX framework and has improved caching and model compression for more efficient memory usage.
– These updates significantly boost performance on Apple Silicon Macs, aligning with growing interest in local AI models.
– The popularity of projects like OpenClaw has increased experimentation with running models locally on personal machines.
– Developer frustration with cloud service costs and limits is driving more use of local coding models like those in Ollama.
– The new features are in preview, currently only supporting Alibaba’s Qwen3.5 model and requiring a Mac with at least 32GB of RAM.

The ability to run powerful large language models directly on a personal computer is becoming more accessible, thanks to key advancements in local AI runtimes. A major development is the new preview support for Apple’s open source MLX framework within the Ollama platform. This integration is specifically designed to accelerate AI model performance on Mac computers powered by Apple Silicon chips, such as the M1 and later generations. Alongside this framework support, Ollama has also enhanced its caching system and adopted Nvidia’s NVFP4 format for model compression, which collectively improve memory efficiency for certain models.

These technical upgrades arrive at a pivotal moment, as interest in local AI models expands beyond researchers and enthusiasts into the broader developer community. The recent phenomenon of projects like OpenClaw, which garnered massive attention on GitHub and sparked widespread experimentation, demonstrates a growing desire to operate models independently. Many developers are seeking alternatives to cloud-based services, often motivated by rate limits and the high cost of subscriptions for premium coding assistants. Running models locally provides greater control and can be a more economical long-term solution for intensive workflows.

Currently, the MLX support is available in a preview release within Ollama version 0.19. It is important to note that this initial implementation is limited to a single, very capable model: the 35-billion-parameter Qwen3.5 variant from Alibaba. The hardware requirements for utilizing this feature are substantial. Users must have a Mac with an Apple Silicon processor and, critically, a minimum of 32GB of RAM to handle the computational load. This specification highlights the demanding nature of cutting-edge local AI, even as the tools to manage it become more sophisticated.

The broader trend toward local model experimentation is being facilitated by improvements in developer tooling as well. Ollama’s recent expansion of its Visual Studio Code integration makes it easier for coders to incorporate these powerful models directly into their primary development environment. As performance on consumer hardware continues to improve through optimizations like MLX, the barrier to entry for running sophisticated AI locally will likely keep falling, opening new possibilities for personalized and private AI applications.

(Source: Ars Technica)

Topics

ollama updates 95% mlx framework support 90% apple silicon performance 88% memory efficiency 85% local coding models 83% local model adoption 82% openclaw success 80% model compression 79% developer frustration 78% qwen3.5 model 77%