Ollama

Ollama

The easiest way to run large language models locally

5.0
26 reviews

1.3K followers

Run Llama 2 and other models on macOS, with Windows and Linux coming soon. Customize and create your own.
This is the 4th launch from Ollama. View more

Ollama v0.19

Launching today
Massive local model speedup on Apple Silicon with MLX
Ollama v0.19 rebuilds Apple Silicon inference on top of MLX, bringing much faster local performance for coding and agent workflows. It also adds NVFP4 support and smarter cache reuse, snapshots, and eviction for more responsive sessions.
Ollama v0.19 gallery image
Ollama v0.19 gallery image
Ollama v0.19 gallery image
Ollama v0.19 gallery image
Free
Launch Team