SkeptAI
p/skeptai
The adversarial AI agent that challenges LLM outputs
0 reviews4 followers
Start new thread
trending
Daton Pope

2d ago

What LLM outputs have cost you the most when you acted on them without checking?

I can recall back to a time I was using Claude to assist me in creating an approach for conducting risk assessments. The example I used for my prompt was general, but industry specific enough for me to tell if something was off. The response I received was fine until I noticed that some information was referenced that made no sense. This caused me to think about all of the other times I received hallucinations as LLM outputs. Some I was lucky enough to catch, and some left me embarrassed as they slipped through the cracks. This is what led to the birth of SkeptAI. LLM outputs need a "Digital Devil's Advocate" at times. That's exactly what the CRIT (Challenge, Reveal, Interrogate, Transmit) framework is designed to do.

I'm curious to know some of you all's stories of hallucinations or false advice. What's the closest call you've had acting on an AI output that turned out to be wrong?

Daton Pope

20h ago

SkeptAI - The adversarial AI agent that challenges LLM outputs

SkeptAI is the adversarial reasoning layer that challenges AI outputs before you act on them. Paste any response from Claude, ChatGPT, or Gemini. CRIT runs four structured passes, then it takes action. CRIT generates a revised output with all critical findings addressed, runs web verification on factual claims inline, and exports a GitHub issue template if you need to escalate. We built this because LLMs optimize for confidence. CRIT optimizes for honesty.