Launching today
Chameleon

Chameleon

Run any LLM on demand — zero idle VRAM.

1 follower

Chameleon is a stateless AI runtime that becomes any LLM on demand. Instead of keeping models loaded, it routes each request to the best model, loads it just-in-time, executes, and fully unloads — resulting in zero idle VRAM usage. Run multiple models efficiently with one runtime, without wasting memory or restarting systems.
Chameleon gallery image
Free
Launch Team / Built With