Launching today
Peer
Peer answers health questions using real clinical evidence.
2 followers
Peer answers health questions using real clinical evidence.
2 followers
Peer answers health questions using real clinical evidence. It searches 9 medical databases, cites every claim to a specific study, and verifies each citation against its source. Evidence is graded by study design using fixed rules. Every response is independently scored for accuracy across six dimensions, and we test the same questions against leading AI tools to measure where we stand.





Hey everyone, Umar here. I built Peer.
This started from a simple frustration. Every time I looked into a health discussion, the answers were mostly anecdotes. Rarely actual studies, and almost never any context on how strong the evidence really is. So I built something I wanted to use myself.
Peer searches 9 medical databases as you ask the question. That includes PubMed, ClinicalTrials.gov with trial outcomes, openFDA, DailyMed, PubChem, FDA Orange Book, FDA UNII for substance identity, and major health organizations like WHO, CDC, NHS, and Mayo Clinic. It then writes a clear answer backed by actual studies.
The important part is what happens after:
Every citation is checked against the original source
Evidence is graded by study design using fixed rules
A separate system scores each response for accuracy
On evaluation, there is no standard benchmark for medical research retrieval, so we built our own using approaches adapted from existing healthcare AI evaluations. We use a set of 160 curated questions across 40+ categories, from supplements and drug safety to adversarial edge cases and prompt injection attempts.
Each answer is scored across six dimensions: factual accuracy, citation grounding, completeness, honest uncertainty, safety, and clarity. If an answer fails factual accuracy, it gets a zero, no matter how good the rest is. We also run a claim level verification step. Every answer is broken into individual claims, stripped of formatting, and each claim is independently checked against sources like PubMed.
We run the same questions against ChatGPT and Claude with web search enabled. Peer consistently performs best on our benchmark. We are not claiming perfection. We are saying we measure rigorously, and the results give us confidence.
If you try it, I would love to know where it feels unclear, where you do not trust it, or where it breaks.
Check it out for free: frompeer.com