Back to Blog

Shadow AI is a Governance Failure: How to Build a “Paved Road” for Safe Experimentation

We have spent the last few weeks in this blog series diagnosing the “Shadow AI” crisis. We know why adjusters go rogue. We know the risks of inconsistency. We know the confusion around buying tools. But if you are a leader looking at this mess of unauthorized tools and risky prompts, you likely have one question: “How did we get here?”

It is tempting to blame the adjusters for breaking the rules. It may be tempting to blame Ops for buying wrappers. The truth is that Shadow AI is a governance failure. People go rogue when the right way is too hard. Innovation doesn’t stop if it takes three months to get a data science environment provisioned, or six months to get a contract signed. Instead, that innovation goes underground.

You don’t need more policies to fix this, you need a paved road.

The “Dirt Road” vs. The “Paved Road”

The concept of the “Paved Road” was popularized by the Netflix Engineering team (specifically in their How We Build Code at Netflix technical blog). Their philosophy was simple: Do not mandate tools. Instead, build a centralized platform that is so good, so easy, and so integrated that no rational engineer would want to build their own alternative. Create a centralized path that is so easy, safe, and effective that no one wants to go off-road.

In the context of Claims AI, the difference is stark.

The Dirt Road (Current state): An Ops leader has an idea to summarize medical files. To get it approved, they must navigate a gauntlet: an Architecture Review Board, a Privacy Impact Assessment, a Vendor Security Questionnaire, and a Legal procurement cycle. After four months of meetings, they are told the data science team “doesn’t have capacity until Q4.” The result? The adjusters under that Ops leader open ChatGPT on their phone.

The Paved Road (Target state): The adjuster logs into a secure internal portal. They select “Document Summarization” from a menu of pre-approved capabilities. They drag and drop their PDF. They get the answer immediately. The data never leaves the VPC. The prompt is pre-vetted.

Governance isn’t about building walls; it’s about building this road.

Step 1: The “Sandbox” (Safe Experimentation)

The first lane of your Paved Road is the Sandbox. This is a technical environment where innovation can happen without risking the enterprise.

Most carriers try to build “Production First,” meaning everything must be perfect before anyone sees it. This is a mistake. You need a dedicated environment where:

  1. Data is sanitized: Real claim data is scrubbed of PII (Personally Identifiable Information) or synthetic data is used.
  2. Access is open: Data Scientists and Super User adjusters can access raw models (like GPT-4 or Llama) via a secure API.
  3. Logs are on: Every prompt and output is recorded, not to punish, but to learn.

This Sandbox kills Shadow AI because it gives your “Scouts (the curious adjusters) a place to play that is actually better than the public tools.

Step 2: The “Pattern Library” (Standardized Governance)

Governance usually fails because every new idea triggers a start from scratch legal review:

  • “Can we use AI for subrogation?” Three-month legal review.
  • “Can we use AI for triage?” Three-month legal review.

On the Paved Road, you pre-approve patterns, not just individual projects.

Create a “Governance Menu” that defines pre-approved risk tiers:

  • Tier 1 (Green Light): Summarization, Translation, Internal Search. Requirement: No decision-making, human review required. Auto-Approve.
  • Tier 2 (Yellow Light): Drafting Correspondence, Sentiment Analysis. Requirement: Human-in-the-Loop edit required. Fast-Track Review.
  • Tier 3 (Red Light): Coverage Denials, Fraud Flagging, Reserve Setting. Requirement: Full Legal/Ethics Board Review.

When an Ops team comes to you with a “Tier 1” idea, you don’t say “let me check with Legal.” You say, “That’s on the Green Light list. Here is your API key. Go.”

Step 3: Transition to Production (The Tollbooth)

The final part of the road is the “Tollbooth.” This is the gate between the Sandbox and the live Claim File.

To cross this gate, the team must prove three things:

  1. The “Grounding” Check: Does the tool cite its sources (pointing to the specific page in the medical record)?
  2. The “Drift” Check: Is there a monitoring system in place to alert us if the model starts hallucinating?
  3. The “Human” Check: Is there a defined workflow for human review and intervention? 

If they have the toll, the gate opens. If not, they go back to the Sandbox.

Governance as a Product

Stop thinking of governance as a policeman. Start thinking of it as a product manager. Your customer is the Claims Department. Your product is the “Paved Road.” If your road is smooth, fast, and safe, the traffic will flow exactly where you want it. If your road is full of potholes and roadblocks, don’t be surprised when you see dust clouds rising from the fields next to the highway.

Chad Langford

Chad Langford is the Data Science Vice President at CLARA Analytics.

It's Easy To Get Started

Optimize claims outcomes with the power of AI

CLARA
Privacy Overview

This website uses cookies so that we can provide you with the best user experience possible. Cookie information is stored in your browser and performs functions such as recognising you when you return to our website and helping our team to understand which sections of the website you find most interesting and useful.