ContextCheck

ContextCheck

Framework for testing and evaluating LLMs, RAG & chatbots.

16 followers

An open-source framework to test LLMs, RAGs & chatbots.It provides tools to automatically generate queries, request completions, detect regressions, perform penetration tests, and assess hallucinations, ensuring the robustness and reliability of these systems.
ContextCheck gallery image
ContextCheck gallery image
Free Options
Launch Team / Built With
Anima - Vibe Coding for Product Teams
Build websites and apps with AI that understands design.
Promoted

What do you think? …

edwinli
Hey Product Hunt! 👋 We built ContextCheck to solve a challenge many of us face: how do we reliably test LLMs and RAG systems? After battling with inconsistent outputs, hidden regressions, and sneaky hallucinations in production, we wanted a systematic way to validate AI systems before deployment. ContextCheck automates the heavy lifting - generating test queries, detecting regressions, and assessing hallucinations. Everything's configurable via YAML and fits right into your CI pipeline. We'd love to hear about your LLM testing challenges and workflows. What aspects of AI testing keep you up at night? How do you ensure reliability in production? If you find this useful, a ⭐️ on GitHub would mean a lot! It helps make the project more visible to other developers who might benefit from it.