Launching today
QuantaLLM

QuantaLLM

Intelligence. Powered by Your Handhelds. Anywhere

2 followers

Run large language models entirely on your Android phone. Powered by llama.cpp with Hexagon NPU acceleration and ONNX Runtime — 100% offline, fully private inference on ARM64.
QuantaLLM gallery image
QuantaLLM gallery image
QuantaLLM gallery image
Free
Launch Team