Launching today
QuantaLLM

QuantaLLM

Intelligence. Powered by Your Handhelds. Anywhere

2 followers

Run large language models entirely on your Android phone. Powered by llama.cpp with Hexagon NPU acceleration and ONNX Runtime — 100% offline, fully private inference on ARM64.

QuantaLLM makers

Here are the founders, developers, designers and product people who worked on QuantaLLM