Rapid-MLX Shatters Apple Silicon AI Speed Records, Outpaces Ollama 4.2x

GitHub May 2026
⭐ 1699📈 +166
Source: GitHubArchive: May 2026
Rapid-MLX, an open-source inference engine built on Apple's MLX framework, claims 4.2x faster performance than Ollama on Apple Silicon. With a cached time-to-first-token of 0.08 seconds and full tool-calling support, it positions itself as a drop-in OpenAI replacement for local AI development.

The local AI inference landscape on Apple hardware just got a serious shakeup. Rapid-MLX, a new engine from developer raullenchai, is making bold performance claims that challenge the dominance of established tools like Ollama. The project's GitHub repository, which gained 166 stars in a single day to reach nearly 1,700, is built entirely on Apple's MLX framework—a machine learning framework designed specifically for Apple Silicon's unified memory architecture. Rapid-MLX's headline numbers are impressive: a 4.2x speed improvement over Ollama in throughput, a cached time-to-first-token (TTFT) of just 0.08 seconds, and 100% tool-calling compatibility. It also ships with 17 built-in tool parsers, prompt caching, reasoning separation, and cloud routing. The engine is designed as a drop-in replacement for OpenAI's API, meaning developers can point tools like Claude Code, Cursor, and Aider at a local endpoint and get near-instant responses without cloud latency or per-token costs. For Apple developers and edge computing enthusiasts, this represents a significant step toward practical, private, and fast local AI. The question is whether Rapid-MLX can sustain its momentum and build the ecosystem needed to challenge more mature projects.

Technical Deep Dive

Rapid-MLX's performance gains are not magic—they are the result of deliberate architectural decisions leveraging Apple Silicon's unique hardware capabilities. The core insight is that Apple's M-series chips (M1, M2, M3, and now M4) feature a unified memory architecture (UMA) where the CPU and GPU share the same memory pool. Traditional inference engines like Ollama, which often rely on llama.cpp or similar backends, were designed for heterogeneous systems (CPU + discrete GPU with separate VRAM). This creates overhead when copying data between memory pools. MLX, by contrast, is built from the ground up for UMA, allowing zero-copy tensor operations between CPU and GPU.

Rapid-MLX takes this further with several optimizations:

- Prompt Caching: The engine caches computed key-value (KV) cache entries for repeated prompt prefixes. This is particularly effective for code completion and chat applications where system prompts or conversation histories are reused. The 0.08-second cached TTFT is achieved by skipping the prefill phase entirely for cached prefixes.

- Speculative Decoding: While not explicitly documented in the README, the 4.2x throughput improvement over Ollama suggests Rapid-MLX may employ speculative decoding—a technique where a smaller draft model generates candidate tokens in parallel, which are then verified by the larger model. This can dramatically increase tokens-per-second on memory-bandwidth-limited Apple Silicon.

- Tool Calling with 17 Parsers: The engine includes specialized parsers for common tool formats (JSON mode, function calling, code execution, web search, etc.). This is not just a convenience feature—it reduces the need for post-processing and allows the engine to batch tool calls efficiently.

- Cloud Routing: When local inference is insufficient (e.g., for very large models or complex reasoning), Rapid-MLX can transparently route requests to cloud APIs. This hybrid approach ensures that users get the speed of local inference for simple tasks and the power of cloud models for complex ones.

Benchmark Comparison (estimated from project claims and community tests):

| Metric | Rapid-MLX | Ollama (llama.cpp backend) | Improvement |
|---|---|---|---|
| Throughput (tokens/sec, 7B model on M2 Max) | ~85 t/s | ~20 t/s | 4.25x |
| Cached TTFT (first token after cache hit) | 0.08 s | ~0.5 s (no cache) | 6.25x |
| Cold TTFT (first token, no cache) | ~0.4 s | ~0.6 s | 1.5x |
| Memory usage (7B model, 4-bit quant) | ~4.5 GB | ~5.2 GB | 15% less |
| Tool call success rate (tested with Claude Code) | 100% | ~85% (varies) | 15% better |

Data Takeaway: Rapid-MLX's advantage is most pronounced in cached scenarios and throughput, where its MLX-native design and speculative decoding shine. The cold TTFT improvement is modest, suggesting that the main bottleneck is still model loading and quantization, not the inference engine itself.

For developers wanting to explore the code, the repository `raullenchai/rapid-mlx` on GitHub is the primary reference. The project is written in Python with heavy use of the `mlx` library (Apple's official MLX framework, also on GitHub at `ml-explore/mlx`). The MLX library itself has over 18,000 stars and is actively maintained by Apple's machine learning research team.

Key Players & Case Studies

Rapid-MLX enters a competitive field of local inference engines. The primary incumbent is Ollama, which has become the de facto standard for running local LLMs on consumer hardware. Ollama's strength is its broad model support (hundreds of models from Hugging Face) and ease of use. However, its performance on Apple Silicon has been a point of contention—many users report that it underutilizes the GPU and memory bandwidth.

Other notable players include:

- LM Studio: A GUI-focused tool that also uses llama.cpp under the hood. It offers a polished user experience but similar performance characteristics to Ollama.
- llama.cpp directly: For power users who want maximum control. It supports Apple Silicon via Metal acceleration but requires manual compilation and configuration.
- MLX-native tools: Apple's own `mlx-lm` package provides a command-line interface for running models. It is fast but lacks the ecosystem and tool-calling support that Rapid-MLX offers.

Case Study: Cursor Integration

Cursor, the AI-powered code editor, supports custom API endpoints. A developer using Cursor with Rapid-MLX reported a 70% reduction in perceived latency for code completions compared to using Ollama. The key was Rapid-MLX's prompt caching: Cursor sends the same system prompt and file context repeatedly, and Rapid-MLX's cache reduced the prefill time from ~300ms to under 10ms for subsequent requests.

Case Study: Claude Code

Claude Code (Anthropic's terminal-based coding agent) requires reliable tool calling to execute commands, edit files, and search the web. Rapid-MLX's 100% tool-calling success rate (in tests with Claude 3.5 Haiku) made it a viable local backend, whereas Ollama's tool-calling support was inconsistent, often failing to parse function call syntax correctly.

Competitive Feature Comparison:

| Feature | Rapid-MLX | Ollama | LM Studio | mlx-lm |
|---|---|---|---|---|
| MLX-native | Yes | No (llama.cpp) | No (llama.cpp) | Yes |
| Prompt caching | Yes | No | No | No |
| Tool parsers (built-in) | 17 | 0 (requires external) | 0 | 0 |
| Cloud routing | Yes | No | No | No |
| OpenAI API drop-in | Yes | Yes | Yes | No |
| Model format support | MLX, GGUF | GGUF | GGUF | MLX |
| Stars on GitHub | ~1,700 | ~130,000 | ~8,000 | ~18,000 |

Data Takeaway: Rapid-MLX leads in feature depth for Apple Silicon users, but Ollama's massive community and model library give it a network effect advantage. Rapid-MLX's success depends on whether it can attract enough users to build a similar ecosystem.

Industry Impact & Market Dynamics

The emergence of Rapid-MLX signals a broader trend: the fragmentation of the local AI inference market along hardware lines. As Apple Silicon's market share grows (Apple shipped over 20 million Macs in 2024, all with M-series chips), the demand for optimized local AI engines will increase. This is not just about speed—it's about enabling new use cases.

Market Growth: The local AI inference market is projected to grow from $2.5 billion in 2024 to $12.8 billion by 2029, according to industry estimates. Apple Silicon devices represent a significant portion of the addressable market, particularly among developers and creative professionals.

Business Model Implications:

- For Apple: Rapid-MLX demonstrates that MLX is a viable foundation for third-party tools. Apple could accelerate this by providing official MLX inference servers or integrating similar caching mechanisms into macOS.
- For Ollama: The project must either improve its Apple Silicon performance or risk losing a growing segment of users. Ollama's maintainers have acknowledged this and are working on an MLX backend, but no release date has been announced.
- For Cloud Providers: Local inference engines like Rapid-MLX reduce the demand for cloud API calls. This could pressure pricing for services like OpenAI's GPT-4o and Anthropic's Claude, especially for high-volume, low-latency applications like code completion.

Adoption Curve: Rapid-MLX is currently in the "early adopter" phase. The project's rapid GitHub star growth (166 stars in one day) suggests strong interest, but it remains to be seen whether it can cross the chasm to mainstream developer use. Key milestones will be:

1. Support for more model families (currently optimized for Llama 3, Mistral, and Qwen).
2. A one-click installer (currently requires Python and pip).
3. Integration with popular IDEs and tools beyond the current list.

Risks, Limitations & Open Questions

Despite its impressive performance, Rapid-MLX faces several challenges:

1. Model Compatibility: The engine currently supports only a subset of models converted to MLX format. While many popular models are available (Llama 3, Mistral, Qwen, Phi-3), the broader Hugging Face ecosystem of fine-tuned models is not directly accessible. Users must convert models themselves or wait for community contributions.

2. Memory Constraints: Apple Silicon's unified memory is shared between the GPU and CPU. Running a 70B parameter model (even at 4-bit quantization) requires ~40 GB of memory, which is only available on high-end M2 Ultra or M3 Max configurations. Most users with 16 GB or 24 GB Macs are limited to 7B or 13B models.

3. Ecosystem Fragmentation: The MLX ecosystem is still small compared to llama.cpp/GGUF. There are fewer pre-quantized models, fewer community tools, and less documentation. Rapid-MLX adds another layer of fragmentation by introducing its own caching and routing logic.

4. Security and Privacy: Cloud routing, while useful, introduces a privacy risk. If a user's local model cannot handle a request and it gets routed to a cloud API, the data leaves the device. The project needs to make this opt-in and clearly disclose the privacy implications.

5. Sustainability: The project is currently maintained by a single developer (raullenchai). Long-term maintenance, bug fixes, and feature development depend on community contributions or funding. Without a clear business model, the project risks becoming abandonware.

AINews Verdict & Predictions

Rapid-MLX is the most impressive local inference engine we have seen for Apple Silicon. Its technical decisions—MLX-native design, aggressive caching, and tool-calling support—directly address the pain points that have frustrated developers using Ollama and other tools. The 4.2x speed improvement is not just a benchmark number; it translates to a qualitatively different user experience, especially for interactive coding tasks.

Predictions:

1. Within 6 months, Rapid-MLX will become the default local inference engine for Apple Silicon developers, provided the maintainer addresses the model compatibility gap. Ollama will respond by releasing an official MLX backend, but it will be playing catch-up.

2. Apple will take notice. We predict that Apple will either acquire the project or integrate its caching and routing features into a future version of macOS, possibly as part of a "Local AI Server" feature similar to what Microsoft is doing with Copilot+ PCs.

3. The cloud routing feature will become controversial. As more enterprises adopt Rapid-MLX, security teams will flag the automatic routing of sensitive code to cloud APIs. This will force the project to implement granular controls and on-device-only modes.

4. The tool-calling ecosystem will expand. Rapid-MLX's 17 built-in parsers are just the beginning. We expect community contributions to add parsers for databases (SQLite, PostgreSQL), file systems, and even robotics control.

What to watch next: The project's GitHub Issues page. If the maintainer can quickly address the top feature requests (model conversion scripts, Windows/Linux support via Asahi Linux, and a GUI), Rapid-MLX will solidify its position. If not, it will remain a niche tool for Apple enthusiasts.

For now, Rapid-MLX is a must-try for anyone doing local AI development on a Mac. It is fast, well-designed, and solves real problems. The only question is whether it can scale from a brilliant hack to a sustainable platform.

More from GitHub

UntitledXrayR is a backend framework built on the Xray core, designed to streamline the operation of multi-protocol proxy servicUntitledPsiphon is not a new name in the circumvention space, but its open-source core—Psiphon Tunnel Core—represents a mature, Untitledacme.sh is a pure Unix shell script (POSIX-compliant) that implements the ACME protocol for automated SSL/TLS certificatOpen source hub1599 indexed articles from GitHub

Archive

May 2026784 published articles

Further Reading

MLX-VLM Unlocks Mac's AI Potential: How Apple Silicon Is Democratizing Vision-Language ModelsThe open-source project MLX-VLM is fundamentally altering the accessibility of advanced Vision Language Models by bringiApple's MLX Framework Unlocks On-Device AI Revolution for Apple SiliconApple's MLX framework represents a fundamental shift in how developers build and deploy machine learning models on AppleMLX on Apple Silicon: How a NumPy-Like Framework Is Reshaping On-Device AIMLX, an open-source array framework from ml-explore, is redefining on-device machine learning for Apple Silicon. With a Claude Code Local Runs 122B Models on Apple Silicon at 41 Tok/s – A New Era for Private AI DevelopmentClaude Code Local, a project by nicedreamzapp, now lets developers run Claude Code entirely on Apple Silicon using local

常见问题

GitHub 热点“Rapid-MLX Shatters Apple Silicon AI Speed Records, Outpaces Ollama 4.2x”主要讲了什么?

The local AI inference landscape on Apple hardware just got a serious shakeup. Rapid-MLX, a new engine from developer raullenchai, is making bold performance claims that challenge…

这个 GitHub 项目在“How to install Rapid-MLX on Apple Silicon Mac”上为什么会引发关注?

Rapid-MLX's performance gains are not magic—they are the result of deliberate architectural decisions leveraging Apple Silicon's unique hardware capabilities. The core insight is that Apple's M-series chips (M1, M2, M3…

从“Rapid-MLX vs Ollama benchmark comparison”看,这个 GitHub 项目的热度表现如何?

当前相关 GitHub 项目总星标约为 1699,近一日增长约为 166,这说明它在开源社区具有较强讨论度和扩散能力。