All activity
Dhia Kraiemleft a comment
This is a very clean idea especially like the “no app switching” angle. Out of curiosity, how do you currently test or validate the AI outputs (accuracy, consistency, multi-turn behavior)? We’re building an AI evaluation framework and would love to collaborate with teams like yours to test real agent workflows in the wild. If that sounds interesting, happy to chat, no pressure at all. Nice work
WebNote AI – Study Smarter on Any Webpage
webnoteaiJoin the discussion
Dhia Kraiemstarted a discussion
Honest question: how do you test your AI agents before production?
AI agents are being shipped faster than ever. But I keep noticing something worrying when talking to founders and engineers: Most teams don’t really test their agents they prompt them a few times and hope for the best. No structured evals. No multi-turn testing. No consistency checks. No guardrails. Just… vibes. This feels very similar to the early days of backend systems: no monitoring no...
