Launching today
ZenMux

ZenMux

An Enterprise-Grade LLM Gateway with Automatic Compensation

250 followers

ZenMux is an enterprise-grade LLM gateway that makes AI simple and assured for developers through a unified API, smart routing, and an industry-first automatic compensation mechanism.
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
ZenMux gallery image
Free Options
Launch Team / Built With
Wispr Flow: Dictation That Works Everywhere
Wispr Flow: Dictation That Works Everywhere
Stop typing. Start speaking. 4x faster.
Promoted

What do you think? …

Haize Yu

Hey Product Hunt! 👋

I'm Haize Yu, CEO of ZenMux. We’ve been heads-down building an enterprise-grade LLM gateway that actually puts its money where its mouth is. I’m thrilled to finally get your feedback on it today.

Why we built this 

Scaling AI shouldn't feel like "fighting the infra." As builders, we grew tired of:

  • Juggling dozens of API keys and messy billing accounts.

  • Sudden "intelligence drops" or latency spikes in production.

  • Paying full price for hallucinations without any fallback. 😅

We thought: What if a gateway didn’t just route requests, but actually insured the outcome?

What ZenMux brings to your stack

  • Built-in Model Insurance: We’re the first to offer automatic credit compensation for poor outputs or high latency. We take the risk, so you don't have to.

  • Dual-Protocol Support: Full OpenAI & Anthropic compatibility. Works out-of-the-box with tools like Claude Code or Cline.

  • Transparent Quality (HLE): We conduct regular, open-source HLE (Human Last Exam) testing. We invest in these benchmarks to keep model routing honest.

  • High Availability: Multi-vendor redundancy means you’ll never hit a rate-limit ceiling.

  • Global Edge Network: Powered by Cloudflare for rock-solid stability worldwide.

Pricing that scales

  • Builder Plan: Predictable monthly subscriptions for steady development.

  • Pay-As-You-Go: No rate limits, no ceilings. Pure stability that scales freely with your traffic. Only pay for what you actually use.

Launch Special 

Bump up your credits! For a limited time: Top up $100, get a $10 bonus (10% extra).

One last thing... 

What’s the biggest "production nightmare" you've faced with LLMs? Drop a comment—I'm here all day to chat!

Stop worrying. Start building. 🚀

https://zenmux.ai

Luke Pioneero

Model insurance for AI infra? That’s new. Curious to try it.

Olivia Ma

@luke_pioneero Appreciate it! 🙏 You hit it — the model insurance is new, but honestly the best part is what comes with the payout: real edge cases from your own usage, ready to plug back in and make your product smarter.

Curious to hear what you think once you try it! 🚀

Jody
Maker

@luke_pioneero Thank you! We built it because we felt infra shouldn’t shift all risk to builders.

Haize Yu
@luke_pioneero This is a very special feature of ours. We noticed users asking, “Why should I pay when the model isn’t accurate?” After seeing this feedback many times, we started reflecting on what we could do—and this idea came out of it. Behind this feature is a push for us to improve stability and speed; otherwise, we’d lose a lot of money 😄
PRIYANKA MANDAL

Congrats on the launch, ZenMux.

While everyone is building on LLMs, you’re building the backbone. Unified, intelligent, and enterprise-ready, that’s how real AI infrastructure scales.

Wishing you powerful integrations and unstoppable momentum ahead.

Sophia L.
Used Zenmux for a while. Not Only API insurance works legit, (actual charged) pricing is 8% ~ 10% cheaper than openrouter. Best luck!
Victor Zhang

Excited to follow your journey. Great launch!

Olivia Ma

@victorzh Thanks! Appreciate it. Stoked to have you along for the ride — more coming soon!

Jody
Maker

@victorzh Thank you so much! Really appreciate the support 🙌

Ryan Thill

An auto-compensation LLM gateway will hit scale pain when “bad output” disputes and p99 latency spikes turn into noisy payout events without reproducible traces.

Best practice is OpenTelemetry GenAI semantic conventions plus per-request lineage (prompt hash, model, router decision, retries) and optional hedged requests or circuit breakers to tame tail latency.

How are you defining and verifying “poor quality” for payouts, and can customers export the full compensation case bundle for audit and fine-tuning?

Van de Vouchy
Hey Haize, that line about paying full price for hallucinations without any fallback is painfully relatable. Was there a specific moment in production where a model just tanked, gave you garbage output, and you realized you were still getting charged for it?
12
Next
Last