Wraith Academy
p/wraith-academy
Learn AI pentesting by attacking real LLM agents
0 reviews2 followers
Start new thread
trending

What's the weakest part of AI agent security right now?

Curious what the PH community thinks. There's a lot of noise about LLM safety, but "safety" and "security" get conflated constantly, and the actual attack surface on an agent in production is its own thing.

From where I sit, the most under-addressed failure modes seem to be:

  • Indirect prompt injection via retrieved content (RAG sources, tool outputs, even user-uploaded docs)

  • Tool/function abuse where the agent happily calls something it shouldn't

  • Trust boundaries between the agent and the systems it can touch

Wraith Academy - Learn AI pentesting by attacking real LLM agents

Free hands-on CTFs where you attack live LLM agents: prompt injection, tool abuse, RAG poisoning, and more. Built for security pros and AI engineers who want to understand how AI systems actually break, not just read about it.