On-Device AI Roundup: Apple Silicon Is Quietly Becoming a Cloud Killer
The narrative around local AI inference has long been one of compromise — slower speeds, fragmented tooling, and the ever-present temptation to offload workloads to the cloud. That story is being rewritten. A new wave of hardware-native inference engines is demonstrating that Apple Silicon, when pus...

The narrative around local AI inference has long been one of compromise — slower speeds, fragmented tooling, and the ever-present temptation to offload workloads to the cloud. That story is being rewritten. A new wave of hardware-native inference engines is demonstrating that Apple Silicon, when pushed to its limits, can compete with — and in some cases outpace — cloud-backed pipelines. The latest and most technically ambitious entry comes from RunAnywhere (YC W26), whose MetalRT engine and open-source RCLI voice pipeline are drawing serious attention from the developer community.
---
The Performance Gap Is Closing — Fast
For years, the benchmark ceiling for Apple Silicon inference was set by tools like llama.cpp, Apple's MLX, and Ollama. These frameworks democratized local model execution, but they all share a fundamental limitation: abstraction layers that sit between the application and the GPU.
RunAnywhere's MetalRT takes a different approach. According to the team, the engine writes custom Metal compute shaders directly, pre-allocates all memory at initialization, and eliminates graph schedulers and runtime dispatchers entirely. The result, benchmarked on an M4 Max with 64 GB of unified memory, is striking:
- Qwen3-0.6B: 658 tokens/sec vs. 552 (MLX) and 295 (llama.cpp)
- Qwen3-4B: 186 tokens/sec vs. 170 (MLX) and 87 (llama.cpp)
- Time-to-first-token: 6.6 ms
The engine delivers a 1.67x improvement over llama.cpp and a 1.19x improvement over Apple MLX on LLM decode throughput — using the same model files. These aren't different models or quantization schemes; this is pure infrastructure efficiency at work.
---
Speech AI Gets a Major Speed Injection
LLM decode speed is only part of the story. MetalRT's most eye-catching numbers come from its speech processing benchmarks, which cover both speech-to-text (STT) and text-to-speech (TTS) — two workloads that are notoriously difficult to optimize in sequence.
On STT, the engine transcribed 70 seconds of audio in just 101 milliseconds, a rate of 714x real-time and 4.6x faster than mlx-whisper. On TTS, synthesis completes in 178 ms — 2.8x faster than mlx-audio and sherpa-onnx.
These numbers matter because voice AI is uniquely sensitive to latency compounding. As the RunAnywhere team explains:
> "In a voice pipeline, you're stacking three models in sequence. If each adds 200ms, you're at 600ms before the user hears a word, and that feels broken."
By optimizing all three modalities — STT, LLM, and TTS — within a single unified engine rather than stitching together separate runtimes, MetalRT achieves end-to-end voice response times that sit comfortably below the 200ms threshold where latency becomes perceptible to users. (Source)
---
Open-Source Tooling Lowers the Barrier to Entry
Performance benchmarks are compelling, but the team's decision to open-source RCLI under an MIT license may prove to be the more consequential contribution. RCLI is a fully on-device voice AI pipeline that chains all three modalities together — microphone input to spoken response — with no cloud dependency, no API keys, and no data leaving the device.
The technical architecture is worth noting:
- Three concurrent threads with lock-free ring buffers
- Double-buffered TTS to reduce output latency
- Local RAG retrieving over 5,000+ document chunks in approximately 4 ms
- 20 hot-swappable models and support for 38 macOS actions by voice
- A full-screen TUI with per-operation latency readouts for debugging
For developers who want to ship on-device voice features without building the underlying infrastructure from scratch, RCLI offers a functional, production-oriented starting point. It also falls back gracefully to llama.cpp when MetalRT is not installed, which broadens compatibility without sacrificing the performance ceiling for those who want it.
Installation is straightforward via Homebrew or a shell script, with approximately 1 GB of models downloaded during setup.
---
The Big Picture: Infrastructure, Not Just Hardware
The conversation around on-device AI has too often focused on model capability — whether a 4B parameter model can reason well enough to be useful. RunAnywhere's work reframes the conversation around inference infrastructure, which may be the more tractable problem.
As the team notes, most engineering teams default to cloud APIs not because local models are inadequate, but because local inference tooling has historically been fragmented, slow, or difficult to ship against. MetalRT's architecture — single engine, all modalities, no framework overhead — addresses the infrastructure gap directly.
This points to a broader trend: the commoditization of capable edge hardware (Apple Silicon being the most prominent example) is outpacing the development of software that fully exploits it. The teams that close this gap first are likely to define the on-device AI developer experience for the next several years.
The 235-point reception on Hacker News and 148 comments suggest that the developer community is hungry for exactly this kind of infrastructure-first thinking.
---
Outlook
The trajectory is clear. As Apple Silicon continues to advance in unified memory bandwidth and GPU core count, the performance delta between on-device and cloud inference will narrow further. What RunAnywhere is demonstrating today on an M4 Max will likely be achievable on mid-range consumer hardware within two to three product generations.
The more immediate question — one the team poses directly — is what developers will build once on-device AI is genuinely as fast as cloud. The answer will depend heavily on whether the open-source ecosystem can build on foundations like MetalRT and RCLI to create the same developer experience that cloud APIs currently offer. The infrastructure work is starting. The application layer is next.
---
Source: RunAnywhere / RCLI on GitHub | Originally discussed on Hacker News