Basalt

Basalt

Reach 99% quality on your AI feature

5.0
1 review

1.8K followers

Ship AI that works, not AI that "feels right" 🛠 Prototype high-quality prompts and Agents from our UI or your codebase. ✅ Test & Evaluate LLM outputs through a dataset of scenarios. 🚀 Deploy AI features confidently with our SDK. 📊 Monitor & Optimize your AI with automated tracking & insights. 💡 Collaborate with your team for better AI-driven products. No more guesswork—just better AI, faster. 🔗
This is the 2nd launch from Basalt. View more

Basalt Agents

Evaluate AI workflows and reach 99% AI quality.
Basalt, #1 AI observability tool for teams, is launching its brand new Agent Builder : prototype, test, and deploy complex AI flows composed of multiple prompts, and run them through a dataset of scenarios.
Basalt Agents gallery image
Basalt Agents gallery image
Basalt Agents gallery image
Basalt Agents gallery image
Basalt Agents gallery image
Basalt Agents gallery image
Free Options
Launch Team / Built With
Migma AI
Migma AI
Lovable for Email
Promoted

What do you think? …

Victor Gross

Guys ! so happy to have you back on Product Hunt, the platform has powered up so much congrats !

Francois de Fitte

@vicgrss thanks Vic !!

Abdul Rehman

This looks amazing, François! 👀 How does the evaluation part work across different models?

Vladimir Lugovsky

Great launch! How does Basalt prevent overfitting on the evaluation dataset — e.g. if prompts start optimizing too much to test cases and lose generality?

Francois de Fitte

@vladimir_lugovsky great question !! To avoid overfitting, we recommend creating a dynamic dataset, meaning that you continuously enrich it with new test cases from your logs (something you can do from Basalt or programmatically) !

Steffan Bankier

Happy to discover the product! No more excel sheets for evals ;))

Francois de Fitte

@steffanb exactly !! Thanks :)

Ning Dong

Congrats on the launch! Curious if there's any example workflow with multimodal (video, speech) results?

Petter Magnusson

This looks solid! The evaluation-first approach is exactly what's needed. Curious how you handle workflows that need human approval gates between steps, not just eval metrics, but hard stops for review before continuing?