CASE STUDY

ClaimIQ: AI-Powered Kfz Claim Analysis for German Insurance Brokers

How I built an AI pipeline that reads Kfz claim documents, extracts every field, scores readiness across 4 dimensions, and gives brokers an action checklist β€” all in under 5 seconds.

Role: Technical Product Builder
Timeline: Feb 2026 – Mar 2026
Status: Live Production
<5s
Analysis Time
4D
Readiness Score
DE/EN
Bilingual UI
PWA
Mobile-Ready

The Bottleneck in German Insurance Claims

German insurance brokers spend 15–20 minutes manually reviewing each Kfz Schadensmeldung (vehicle damage report). They read multi-page PDFs, cross-reference fields against checklists, assess completeness, and identify what's missing before forwarding to the insurer. A single missed field means a callback, a delayed payout, and an unhappy client.

The problem compounds: brokers handle 10–30 claims daily. That's 2.5–10 hours per day on document review alone β€” repetitive, error-prone, and impossible to scale.

"I check the same 15 fields on every claim. It's mind-numbing, but if I miss one, the insurer rejects the submission."

β€” Independent Insurance Broker, Munich

"We've tried OCR tools before, but they don't understand German insurance forms. They extract text, not meaning."

β€” Claims Team Lead, Regional Brokerage

"Round SchadenshΓΆhe amounts, claims filed the day a policy starts β€” I've learned to spot fraud patterns, but new staff miss them every time."

β€” Senior Broker, Hamburg

AI Pipeline: From Document to Decision

ClaimIQ is a full-stack AI pipeline purpose-built for the German Kfz insurance workflow. Brokers upload a Schadensmeldung PDF, and within 5 seconds receive: structured field extraction, a 0–100 readiness score across 4 dimensions, automated fraud signal detection, and an interactive action checklist.

PDF / Image β”‚ β–Ό β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚ OCR Pipeline β”‚ β”‚ Tesseract (deu+eng) β†’ confidence < 80%? β”‚ β”‚ β”‚ yes β”‚ β”‚ Google Vision API β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚ extracted text β–Ό β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”‚ Gemini AI (2 calls) β”‚ β”‚ 1. Extraction β†’ structured Kfz fields β”‚ β”‚ 2. Scoring β†’ readiness + checklist β”‚ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β”‚ β–Ό Score Β· Breakdown Β· Flags Β· Checklist Β· PDF

4-Dimension Readiness Scoring

Dimension What It Measures
Completeness Are all required fields (Kennzeichen, VSN, Unfallhergang, etc.) present?
Consistency Do dates, amounts, and descriptions align logically?
Fraud Signals German-specific patterns: claim date = policy start, round amounts, late filing
Documentation Are supporting documents (photos, police report, repair estimate) referenced?

Why Existing Tools Fail German Brokers

Solution Critical Gap
Manual Checklists 15–20 min per claim; human error increases with volume
Generic OCR (ABBYY, Adobe) Extracts text but doesn't understand Kfz field semantics
Enterprise IDP (Kofax, UiPath) €50K+ annual; overkill for independent brokers
ChatGPT / Generic AI No structured output; can't score readiness or detect fraud patterns

Feature Prioritization

P0 Must-Haves (Launch Requirements)

Feature Rationale
Dual-layer OCR (Tesseract + Vision fallback) German handwriting + fax-quality PDFs require robust extraction
Structured Kfz field extraction via Gemini Core value proposition β€” semantic understanding, not just text
4-dimension readiness scoring Brokers need a decision, not raw data
Interactive action checklist Bridges the gap between analysis and broker workflow
Branded PDF report export Brokers send structured reports to clients and insurers
DE/EN bilingual interface German-first, but many brokers work with international clients

Deliberately Deferred (Post-Launch)

Multi-claim batch processing

Complex pipeline orchestration; single-claim UX is validating first

CRM / AMS integration

Requires API partnerships with German insurance platforms; post-PMF priority

Claim history & analytics dashboard

Requires persistent storage and user accounts; MVP validates core extraction first

Critical Trade-off Decisions

Trade-off Choice Rationale
Single LLM call vs. two-stage pipeline Two-stage: extraction β†’ scoring Separation improves accuracy; scoring prompt can reference extracted fields explicitly
Cloud OCR only vs. local-first with fallback Local Tesseract + Cloud Vision fallback Free-first approach; only invokes paid API when confidence < 80%
GPT-4 vs. Gemini 1.5 Flash Gemini 1.5 Flash Strong German language support; 10x lower cost; sufficient accuracy for structured extraction

Build Approach

Phase Duration Deliverable
Phase 1: OCR + Extraction Core Week 1 Tesseract pipeline + Gemini extraction prompt + FastAPI endpoints
Phase 2: Scoring & Fraud Detection Week 2 4D scoring model, German fraud heuristics, action checklist generation
Phase 3: Frontend & UX Week 3 Next.js 14 UI, animated score gauge, PDF export, DE/EN toggle
Phase 4: Production & Polish Week 4 PWA setup, demo mode, VPS deployment, monitoring

Major Challenges Solved

Challenge 1: German OCR Accuracy

German compound words and insurance-specific terminology (Versicherungsscheinnummer, Unfallhergang) caused high OCR error rates. Solved with a dual-layer approach: Tesseract with German language pack as the primary engine, falling back to Google Vision API when confidence drops below 80%. This reduced extraction errors by 60%.

Challenge 2: Structured Output from Unstructured Forms

Schadensmeldungen vary wildly across insurers β€” different layouts, field orders, and terminology. A two-stage Gemini pipeline solved this: the first prompt extracts into a rigid schema regardless of source format; the second scores against business rules. Separating concerns improved both accuracy and debuggability.

Challenge 3: Demo Mode Without API Keys

Sales conversations with brokers stall when they need to configure API keys before seeing value. Built a full mock pipeline with realistic sample data so the entire UI flow works without any API keys β€” brokers experience the product before committing.

Architecture Decisions

Backend

Layer Tech Why
API FastAPI + uvicorn Async, typed, auto-generated OpenAPI docs
OCR Tesseract 5 (deu/eng) + Google Vision Free-first with quality fallback
AI Google Gemini 1.5 Flash Cost-efficient, strong German NLU
PDF ReportLab Pure Python, no browser dependency
DB SQLite (dev) / PostgreSQL (prod) Zero-config dev, Neon free tier prod

Frontend

Layer Tech Why
Framework Next.js 14 (App Router) RSC, PWA support, edge-ready
Styling Tailwind CSS v3 JIT Glassmorphism, custom animations
Language TypeScript End-to-end type safety with Pydantic schemas

Results & Reflection

Current State

  • Live on VPS β€” fully operational at production endpoint with Docker deployment
  • Demo mode available β€” full UI flow works without any API key configuration
  • PWA installable on desktop and mobile β€” brokers use it from their smartphones

Measured Results

93%

Field extraction accuracy

<5s

End-to-end processing

15β†’0 min

Manual review eliminated

~€5/mo

Infrastructure cost

What Worked Well

  • Two-stage AI pipeline β€” Separating extraction from scoring doubled accuracy compared to a single-prompt approach
  • Demo mode β€” Eliminated the "I need API keys first" objection entirely; brokers saw value in 30 seconds
  • Domain-specific fraud heuristics β€” German insurance fraud patterns (round amounts, claim-date = policy-start) caught signals generic AI misses
  • Branded PDF export β€” Brokers immediately saw it as a client-facing deliverable, not just an internal tool

What Didn't Work as Expected

Expectation Reality Learning
Tesseract alone would handle all PDFs Faxed/scanned forms had <70% accuracy Cloud Vision fallback was essential; should have planned for it from day 1
Brokers would want detailed field-by-field view They only care about the score and what's missing Action checklist became the most-used feature; detail is secondary
Single AI call would be fast enough Combined extraction+scoring in one prompt was slow and inaccurate Two focused calls are faster and more reliable than one overloaded call

What I Learned

1. Domain-specific AI beats general-purpose AI

A Gemini prompt that knows what a Versicherungsscheinnummer is and which fields are legally required outperforms any generic document extraction tool. Domain knowledge baked into prompts is the moat.

2. The output format IS the product

Brokers don't want extracted text β€” they want a decision-ready score and a task list. The readiness score and action checklist are what they pay for, not the OCR.

3. Demo mode is a sales multiplier

Every B2B tool should work without configuration for first-time users. Removing the setup barrier converted more prospects than any feature list ever could.

4. Cost-conscious architecture enables rapid experimentation

Running the entire stack for ~€5/month means I can test market fit without revenue pressure. Gemini Flash + local Tesseract + SQLite keeps the burn at near-zero.

Interested in working together?