Hallucination Detection
Verify that every claim in an LLM's answer is actually supported by provided evidence—robust to evidence ordering.
- Claim extraction Splits answers into atomic, verifiable claims
- QMV permutation probing Tests across evidence orderings for robustness
- Confidence metrics q_bar (mean) and q_lo (robust) support scores
- Order-sensitivity detection js_bound flags serialization-dependent judgments