Social proof / engagement
We've just launched and would genuinely love feedback from this community. Especially if you're: Building LLM-powered products Working in AI/ML security A developer who's thought about prompt injection but hasn't found a clean solution yet First 5 people to sign up and email hello@bordair.io with code ProdHuntBordair get free Business tier access. What questions do you have? I'll answer...
The bigger picture
Prompt injection is the SQL injection of the AI era. In the early web, we shipped dynamic sites before understanding input validation. Entire security categories had to be invented after the fact - firewalls, WAFs, input sanitisation. We're at that same inflection point with LLMs. The difference is we can see it coming this time. Bordair is our attempt to build the security layer before it...
Technical credibility
For the technically curious - here's what's under the hood: Fine-tuned DistilBERT classifier, quantised to ONNX for <50ms inference Multimodal: text + image scanning via Claude vision API Dual-region deployment (EU + US) with latency-based routing Python and JS SDKs, one-line integration The ML model hits 99% accuracy on our test suite with zero false positives across 226 benign prompts. Happy...
The Castle
One of my favourite things we've shipped is Bordair's Castle 🏰 It's a 7-level prompt injection gauntlet - each level has an AI guard instructed to keep a password secret. Your job is to trick it into revealing it. It's not just a gimmick. The attack patterns we see in the Castle directly feed back into improving our detection models. Level 7 is Bordair himself. Nobody's beaten him yet. Give it...
The origin story
Hey Product Hunt 👋 Bordair started from a simple frustration: I kept seeing LLM apps ship with zero input validation - the same mistake we made with SQL injection in the early web. So I built the thing I wished existed. A real-time security layer that sits between user input and your LLM, blocking adversarial requests before they ever reach your model. Would love to hear - are you validating...
AI isn’t broken. It’s being tricked - and no one is fixing it properly
Most teams are still focused solely on prompt injection. But that’s already yesterday’s problem. The real threat to protect against is cross-modal injection. AI doesn’t just read text anymore. It processes: images PDFs audio files Every one of these = an attack surface. And here’s the issue: AI systems trust what they see. They don’t know if an instruction came from: a user hidden text in a PDF...
