Chameleon
p/chameleon-8
Run any LLM on demand β€” zero idle VRAM.
β€’0 reviewsβ€’2 followers
Start new thread
trending
Ashwanth Krishnaβ€’

8d ago

Chameleon - Run any LLM on demand β€” zero idle VRAM.

Chameleon is a stateless AI runtime that becomes any LLM on demand. Instead of keeping models loaded, it routes each request to the best model, loads it just-in-time, executes, and fully unloads β€” resulting in zero idle VRAM usage. Run multiple models efficiently with one runtime, without wasting memory or restarting systems.