Running the latest LLMs like Granite-4.0 and Qwen3 fully on ANE (Apple NPU)

Posted by Different-Effect-724@reddit | LocalLLaMA | View on Reddit | 11 comments

Last year, our two co-founders were invited by the Apple Data & Machine Learning Innovation (DMLI) team to share our work on on-device multimodal models for local AI agents. One of the questions that came up in that discussion was: Can the latest LLMs actually run end-to-end on the Apple Neural Engine?

After months of experimenting and building, NexaSDK now runs the latest LLMs like Granite-4.0, Qwen3, Gemma3, and Parakeet-v3, fully on ANE (Apple's NPU), powered by the NexaML engine.

For developers building local AI apps on Apple devices, this unlocks low-power, always-on, fast inference across Mac and iPhone (iOS SDK coming very soon).

Video shows performance running directly on ANE

https://reddit.com/link/1p0tko5/video/ur014yfw342g1/player

Links in comment.