1mo ago
Most AI safety tools use prompt-level filters:
- "Don't do anything harmful"
- LLM-as-judge (another model watching the first one)
- Retrieval-based guardrails
0
2
We just shipped v0.4.0 with @observe, a free, lightweight way to
see what your agents are doing.
from aip_protocol import passport, observe
We've all seen it - an agent goes rogue during a demo, a test agent
accesses production data, a chatbot sends emails it shouldn't.
What's the worst thing an AI agent has done in your system?
(Bonus points if it happened in production )