
Case Studies
How organizations use confora labs to test, monitor, and govern their AI.
Underwriting
Insurance · Robustness
From AI Decisions to Measurable Risk
How automated testing revealed hidden robustness risks in LLM-assisted underwriting
The Context
Insurers deploy LLM-based systems to accelerate underwriting, extracting information from medical documents, summarizing histories, supporting decisions. But speed introduces risk: how do you know the summary is correct, the decisions fair, and the system EU AI Act-compliant?
The Approach
confora labs ran a systematic assessment using Confora Insight's robustness testing module. Using 10 adversarial alteration types, we generated perturbed test cases from real document conditions: rotation, blur, noise, contrast changes, and more.
The Finding
Rotated documents led to a sharp increase in extraction and summarization errors. Medical conditions were incorrectly omitted or added in underwriting summaries. This failure mode was invisible in normal operation: the system produced plausible output, but the content was wrong.
The Outcome
→ Document quality flagging before processing
→ Extraction pipeline retrained for rotation robustness
→ Robustness checks integrated into CI/CD pipeline
→ Findings documented in EU AI Act-aligned audit trail
→ Continuous monitoring set up to catch future regressions
Takeaway
AI systems that work in normal conditions can fail silently under realistic variations. Systematic testing turns invisible risks into documented, auditable findings.
Ready to test your AI?
Book a Demo
confora labs by spotixx GmbH


