Timoa

Timoa

Lead DevSecOps, Accenture Interactive FR
liteLLM
Used as an LLM proxy, it allows the caching and load balancing between multiple AI services (Groq, OpenRouter, etc.) and even local with Ollama. It uses an OpenAI-compatible API that allows (when we can set the base URL) to use it in many apps or services. I use it configured with Langfuse which provides the performance analysis (monitoring) of each prompt/session.

What's great

caching and load balancing (1)OpenAI-compatible API (1)performance analysis integration (1)
18 views