OptiLLM

OptiLLM

Intelligent LLM Cost Optimization Platform

3 followers

OptiLLM automatically reduces LLM API costs by up to 50%+ without sacrificing quality. It routes each prompt to the cheapest capable model using ML classifiers, compresses tokens with LLMLingua-2, and caches semantically similar queries with FAISS vector search. Drop-in OpenAI-compatible proxy β€” no code changes needed. Includes evaluation tools, analytics dashboards, and custom router training to continuously optimize your cost-quality tradeoff.

OptiLLM makers

Here are the founders, developers, designers and product people who worked on OptiLLM