Zero Trust for AI: Building Baseline Trust Before You Lock Everything Down

Why traditional zero-trust fails for AI and how to establish baseline behavior first

← Back to Blog
AI Security

The zero-trust security model is foundational for traditional IT: assume breach, verify everything, grant least privilege. It works well for networks, applications, and infrastructure. But when your security team tries to apply zero-trust to AI systems, something breaks. And usually, it breaks spectacularly.

The problem is this: zero-trust assumes you know what "normal" looks like. For traditional systems, normal is deterministic—a user either should have access to a file or shouldn't. For AI systems, normal is probabilistic. A model's behaviour is inherently uncertain. You can't implement zero-trust for something you don't understand yet.

Why Zero-Trust Breaks for AI Systems

Zero-trust has three core pillars: verify identity, enforce least privilege, and assume breach. On the surface, these sound right for AI.

But the assumption embedded in zero-trust is that you can definitively answer: "Is this normal?" For traditional systems, you can. For AI systems, you cannot—not initially.

Consider a recommendation model in a retail bank. Zero-trust says: log all model inputs and outputs, verify they're legitimate, restrict data access to minimum needed. But what does "minimum needed" mean? The model needs customer transaction history to make recommendations. It also needs market data. Where's the line?

More fundamentally: what does "legitimate" output look like? If the model recommends an unusual investment product to a customer, is that anomalous (potential breach) or just normal model behaviour? Without baseline understanding, you can't know.

"Zero-trust assumes you know what normal is. For most AI systems, you don't. Not yet. That's the gap."

The Baseline Trust Model

Before you implement zero-trust for AI, you need to establish what normal looks like. This is the baseline trust model.

A baseline trust model answers these questions:

Building this requires observation. You run the model in a normal operating environment, log everything, and collect statistics. Only after you have months of baseline data can you detect anomalies with confidence.

The Three-Phase Approach

Phase 1: Establish Baseline (Months 1-3)

Deploy the model to production with comprehensive logging but minimal restrictions.

During this phase, security is intentionally loose. You're gathering signal. The model operates with generous data access. You don't restrict query patterns. The goal is to see how the model naturally behaves.

Phase 2: Define Normal (Months 3-6)

Analyze baseline data to establish statistical models of normal behaviour.

This is detective work. You're building a statistical model of "normal" that you can use to detect "abnormal".

Phase 3: Implement Zero-Trust Controls (Month 6+)

Now you lock things down based on what you learned.

Real-World Case Study: Australian Financial Services

A major Australian bank deployed an AI credit decisioning model. The security team, familiar with zero-trust, wanted to implement it immediately: log everything, restrict data access, flag anomalies.

Problem: after two weeks, the model stopped working. Why? The zero-trust implementation was blocking legitimate queries because the security team didn't understand the model's data requirements.

The model needed to query: customer transaction history (last 24 months), credit bureau data, market indicators, and internal risk models. But the security team didn't know which of these were critical. They restricted access to the most sensitive database (customer transactions) immediately.

Model accuracy dropped 15%. Business users complained. The project nearly failed.

What should have happened:

The bank eventually followed this approach and successfully deployed zero-trust controls by month 4. But they lost two months due to starting with the wrong assumption.

Mapping Data Flows

A critical part of establishing baseline trust is understanding data flows. Where does data come from? Where does it go?

For an AI model, this includes:

Map all of these before you lock things down. Then, zero-trust controls can be applied surgically: restrict access to what's necessary, monitor for deviations, and escalate on anomalies.

Practical Implementation

Here's a practical checklist for implementing baseline trust + zero-trust for AI:

  1. Deploy model with comprehensive logging (3 months)
  2. Analyse baseline data to define normal (3 months)
  3. Implement least-privilege data access based on analysis
  4. Deploy anomaly detection on inputs, outputs, and performance metrics
  5. Implement query rate limiting based on normal user patterns
  6. Set up alerts for deviations from baseline
  7. Monthly reviews: update baseline models as data distribution shifts

Key Takeaways