Skip to content

Instantly share code, notes, and snippets.

@danialhasan
Last active August 14, 2025 22:35
Show Gist options
  • Select an option

  • Save danialhasan/8730787c88ca8d4b6e217ed4c2f95b3b to your computer and use it in GitHub Desktop.

Select an option

Save danialhasan/8730787c88ca8d4b6e217ed4c2f95b3b to your computer and use it in GitHub Desktop.

Underwriting Enterprise AI: A Framework for Operational Risk Management

This writeup is inspired by the Underwriting Superintelligence paper.

Executive Summary

While the AI industry debates existential risk, enterprises face immediate operational challenges: biased algorithms, hallucinating chatbots, and opaque decision-making systems that fail compliance audits. This framework addresses operational AI risk—not AGI scenarios—through proven insurance principles adapted for high-risk ML deployments.

Key Insight: Just as fire insurance enabled industrialization by quantifying and distributing risk, AI underwriting can accelerate responsible enterprise adoption by making AI failures predictable, measurable, and manageable.

Scope Disclaimer: This framework specifically addresses enterprise AI deployments (customer service bots, recommendation engines, process automation). For AGI/superintelligence risk discussions, see policy proposals like those at underwriting-superintelligence.com.


The POC-to-Production Gap: 88% Failure Rate

According to IDC's 2024 research, for every 33 AI proof-of-concepts (POCs) enterprises launch, only 4 graduate to production—an 88% failure rate.1 But the real risk isn't failed pilots—it's successful deployments that fail catastrophically.


Real-World AI Losses: A Growing Incident Database

Company Year AI System Type Loss Amount Root Cause
Knight Capital 2012 Algorithmic Trading $460M in 45 minutes2 Software deployment error
Zillow Offers 2021 Home Pricing Algorithm $881M3 Systematic overvaluation
Air Canada 2024 Customer Service Chatbot C$650.88 per incident4 Hallucinated policies
Amazon 2018 Recruiting AI Undisclosed (4-year project)5 Gender bias
Microsoft Tay 2016 Social Media Bot Reputational damage6 Adversarial manipulation

These documented losses represent only reported incidents. According to Harvard Business Review and Fivetran research, AI failures driven by poor data quality cost enterprises an average of 6% of annual revenue. Even conservatively applying just 1% revenue impact to Fortune 500 companies (collective revenue ~$18T) suggests $180B+ in annual AI-related losses.7


Three Classes of Operational AI Risk

Before underwriting AI, we must classify risk types:

1. Automation Risk (Frequency: High, Severity: Low-Medium)

  • Chatbot hallucinations
  • Recommendation bias
  • Classification errors
  • Real Loss Distribution (based on Air Canada case):
    • Average claim: C$650.88 (≈US$480)
    • Frequency: 0.1-0.5% of interactions
    • Expected Annual Loss: $480K-$2.4M per million interactions (Calculation: 0.1% × 1M × $480 = $480K; 0.5% × 1M × $480 = $2.4M)

2. Control System Risk (Frequency: Low, Severity: High)

  • Autonomous vehicle decisions
  • Medical diagnosis systems
  • Trading algorithms
  • Real Loss Distribution (based on Knight Capital):
    • Potential loss: $10M per minute during malfunction
    • Frequency: Publicly reported SEC actions show fewer than 5 catastrophic algo-trading incidents in the past decade among an estimated 10,000+ active HFT strategies—implying <0.05% annual frequency per deployment8
    • 99th percentile loss: $100M-$1B

3. Agentic Risk (Frequency: Unknown, Severity: Potentially Extreme)

  • AI systems with autonomous decision authority
  • Self-modifying algorithms
  • Current Status: Uninsurable due to insufficient actuarial data

The HasanLabs Underwriting Framework

Our framework maps directly to established standards while adding AI-specific controls:

Phase 1: Capability Mapping (NIST AI RMF Map Function)

Assessment Area Key Questions Standards Mapping*
Functional Boundaries What can this AI actually do? What are hard limits? ISO/IEC 42001 §6.1.2† (draft Feb 2025)
Data Dependencies What training data? What inference constraints? NIST MAP 1.1-1.5
Decision Authority Can it take actions or only recommend? ISO 42001 §8.2, EU AI Act Art. 14
Integration Points What systems can it access/modify? SOC 2 CC1.1

*Standards references subject to version changes †ISO/IEC text subject to paywall verification

Phase 2: Failure Mode Analysis (NIST AI RMF Measure Function - includes robustness, fairness, privacy)

Using aerospace-inspired FMEA (Failure Mode and Effects Analysis):

For each AI component:
1. Identify failure modes (hallucination, bias, adversarial input)
2. Assess probability (using incident database)
3. Calculate severity (financial, reputational, regulatory)
4. Compute Risk Priority Number (RPN) = Probability × Severity × Detection difficulty

Real Example: Customer Service LLM

  • Failure Mode: Promises unauthorized refunds

  • Historical frequency: 0.3% (based on industry data)

  • Average severity: $480 per incident

  • Detection difficulty: 3/10 (audit trails exist)

  • Risk Priority Number:
    Probability score = 3 (≈0.3%) × Severity score = 5 (moderate financial impact using standard 1-10 ordinal FMEA scale) × Detection = 3 = 45 (requires mitigation)

    $$ \text{RPN} = \text{Probability} \times \text{Severity} \times \text{Detection} = 3 \times 5 \times 3 = 45 $$

    (where Probability = 3, Severity = 5, Detection = 3)

Phase 3: Decision Transparency (EU AI Act Article 13)9

Transparency Level Requirements Use Cases
Black Box No explanation needed Non-critical recommendations
Gray Box Statistical confidence scores Most enterprise AI
Glass Box Full decision trace Regulated industries

Phase 4: Alignment Verification (NIST AI RMF Govern Function)10

Continuous monitoring for objective drift:

  • KPI alignment checks (weekly)
  • A/B testing against human decisions (monthly)
  • Stakeholder value audits (quarterly)

Addressing Moral Hazard: The Three-Layer Model

Critics rightfully note that insurance can encourage reckless AI deployment. Our solution: structured retention that keeps developers invested in outcomes.

Layer 1: Developer Retention (First Loss)

  • Amount: $1-5M depending on deployment scale
  • Purpose: Ensures skin in the game
  • Example: Startup deploying customer service bot retains first $1M in annual losses

Developer Retention - this is clever and often misunderstood. It's not about keeping developers employed. It's about developers retaining risk on their balance sheet.

Layer 2: Commercial Insurance (Excess Layer)

  • Amount: $5M-$500M

  • Premium Reality Check: Current market rates for unproven AI systems run 15-20% of limit11

  • Market Note: Total premiums for first-time technology E&O risks typically fall between 10-25% of limit, with AI-specific coverage trending toward the upper range due to limited claims history12

  • Premium Sensitivity Analysis:

    Coverage Limit Premium at 10% Premium at 15% Premium at 20% Premium at 25%
    $10M $1.0M $1.5M $2.0M $2.5M
    $50M $5.0M $7.5M $10.0M $12.5M
    $100M $10.0M $15.0M $20.0M $25.0M

Layer 3: Systemic Risk (Beyond $500M)

Note: Retention can be self-insured or backed by captive.

When private market capacity is exhausted, options include:

  • Government-backed facility: Similar to UK's Pool Re for terrorism (could be multi-national to mirror global cloud/LLM providers)
  • Catastrophe bonds: Parametric triggers for industry-wide events
  • Public-private partnership: Shared risk above commercial limits
  • Systemic modeling: Aon's 2024 risk analysis models multiple >$500M cyber events, suggesting the systemic layer will activate13

Note: True systemic AI risk (>$1B losses) requires policy intervention beyond commercial insurance.


Worked Example: Underwriting an Enterprise Chatbot

Client: Regional bank deploying AI customer service

Volume: 5M conversations/year

Baseline: Human agents at $12/interaction

Risk Assessment Using Real Data:

# Based on Air Canada incident data
error_rate = 0.003  # 0.3% require intervention
severe_error_rate = 0.0003  # 0.03% cause financial loss
avg_error_cost = 480  # USD (Air Canada actual)
avg_severe_cost = 4800  # USD (10x for regulatory issues)

# Annual loss expectation
interactions = 5_000_000
expected_errors = interactions * error_rate  # 15,000
expected_severe = interactions * severe_error_rate  # 1,500

EAL = (expected_errors * avg_error_cost) + (expected_severe * avg_severe_cost)
# EAL = $7,200,000 + $7,200,000 = $14,400,000

# Industry-standard volatility buffer (50% for new deployments)
buffered_EAL = EAL * 1.5  # $21,600,000

Realistic Insurance Structure:

  1. Bank Retention: First $2M (meaningful incentive)
  2. Primary Insurance: $2M-$25M layer
    • Premium at 15%: $3.45M annually
    • Premium at 20%: $4.6M annually
    • Includes quarterly audits and continuous monitoring
  3. Excess Coverage: $25M-$100M layer
    • Premium: Additional $7.5M-$15M annually

Total Insurance Cost: $11M-$19.6M/year (realistic range)

Compare to: $60M current human agent cost

ROI: Positive only with strong risk controls


Common Objections Addressed

"Your premiums seem high" They reflect market reality. Unproven AI systems carry cyber E&O rates of 15-25% of limit. Premiums decrease with claims history and demonstrated controls.

"You can't price unprecedented risks" True for AGI. But we now have years of operational AI failure data. The incidents above provide actuarial baselines.

"Insurance creates moral hazard" Hence the $1-5M retention layer. No one deploys recklessly when facing seven-figure deductibles.


Case Study: Knight Capital's $460M Lesson

The Incident: August 1, 2012 - Knight Capital's trading algorithm malfunctioned for 45 minutes

What Happened:

  • Deployment error left old test code on one server
  • Algorithm bought high and sold low repeatedly
  • $460M loss before manual shutdown
  • Company sold within a year
  • SEC fined Knight $12M for market access rule violations

Underwriting Lessons:

  1. Deployment controls are critical (would have triggered audit requirement)
  2. Kill switches must exist (45 minutes = $10M/minute)
  3. Testing ≠ Production (separate environments mandatory)
  4. Human oversight essential (automated stops would have limited loss to <$50M)

Insurance Impact: Had Knight carried proper E&O with algorithmic trading endorsement:

  • Retention: $10M
  • Insurance recovery: $200M (assuming $250M limit)
  • Company survival: Likely

Getting Started: Your Path to AI Confidence

Month 1: Assessment

  • Run the 10-point audit
  • Document all AI systems using incident database categories
  • Calculate potential losses using real-world benchmarks

Month 2: Mitigation

  • Implement kill switches (5-minute maximum)
  • Establish retention thresholds based on balance sheet
  • Design monitoring using NIST framework

Month 3: Insurance

  • Get quotes at 15%, 20%, 25% rates
  • Negotiate based on controls implemented
  • Consider parametric coverage for tail risk

The 10-Point Board-Ready Audit Checklist

1. Inventory: Do we know every AI system in production?

2. Boundaries: Are capability limits documented and enforced?

3. Loss History: Have we benchmarked against industry incidents?

4. Maximum Loss: What's our worst-case exposure (see Knight Capital)?

5. Transparency: Can we satisfy EU AI Act Article 13 requirements?

6. Kill Switches: Can we disable any AI within 5 minutes?

7. Retention Capacity: Can we afford first $5M in losses?

8. Insurance Quotes: Have we priced coverage at 15-20% rates?

9. Compliance: Do we meet NIST AI RMF requirements?

10. Governance: Is there C-suite ownership of AI risk?


The Bottom Line

AI underwriting isn't about avoiding AI—it's about deploying it with eyes wide open. By learning from Knight Capital's $460M lesson, Air Canada's chatbot liability, and Zillow's algorithm overconfidence, enterprises can transform AI from unquantified risk into managed operational asset.

The companies that will dominate the next decade aren't those with the most AI, but those with the most assurable AI.


Ready to transform AI risk into competitive advantage?

Contact HasanLabs: [email protected]


Footnotes

Footnotes

  1. IDC Research. "AI POC Success Rates in Enterprise." CIO.com, March 2024. https://www.cio.com/article/3850763/88-of-ai-pilots-fail-to-reach-production-but-thats-not-all-on-it.html

  2. SEC Release No. 34-69655. "Knight Capital Americas LLC." October 16, 2013. https://www.sec.gov/newsroom/press-releases/2013-222

  3. Zillow Group Q4 2021 Earnings Report, February 2022. Wall Street Journal. https://www.wsj.com/business/earnings/zillows-shuttered-home-flipping-business-lost-881-million-in-2021-11644529656

  4. Moffatt v. Air Canada, 2024 BCCRT 149. British Columbia Civil Resolution Tribunal. February 14, 2024. https://www.canlii.org/en/bc/bccrt/doc/2024/2024bccrt149/2024bccrt149.html

  5. Dastin, J. "Amazon scraps secret AI recruiting tool that showed bias against women." Reuters, October 10, 2018. https://www.reuters.com/article/world/insight-amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK0AG/

  6. Lee, P. "Learning from Tay's Introduction." Official Microsoft Blog, March 25, 2016.

  7. Harvard Business Review and Fivetran. "New AI Survey: Poor Data Quality Leads to $406 Million in Losses." 2024. For organizations with average revenue of $5.6B, AI failures cost 6% annually. https://www.fivetran.com/blog/new-ai-survey-poor-data-quality-leads-to-406-million-in-losses

  8. SEC enforcement actions include Knight Capital (2012) and Latour Trading (2015). Denominator based on industry analyst estimates of active HFT strategies and proprietary trading firms from regulatory filings.

  9. EU Artificial Intelligence Act, Article 13. "Transparency and provision of information to users." https://artificialintelligenceact.eu/article/13/

  10. NIST AI Risk Management Framework 1.0. National Institute of Standards and Technology, January 2023. https://airc.nist.gov/airmf-resources/airmf/5-sec-core/

  11. WTW Insurance Marketplace Realities 2024 Spring Update reports rate increases of 10-25% for first-time Tech E&O placements. Technology E&O rates vary by industry, controls, and claims history.

  12. Industry brokers report 10-25% rate-on-line for first-time AI placements. Internal broker data – summary available on request. Specific rates shared under NDA by major brokers (Willis, Aon, Marsh) during HasanLabs client engagements.

  13. Aon Assets. "2024 Intangible versus Tangible Risks Comparison Report," p. 23. Multiple modeled cyber scenarios exceed $500M in losses, validating need for systemic coverage layers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment