Skip to main content
Article

Securing Sensitive AI: Why Sensitive Models Must Stay Locked Down

By September 25, 2025No Comments
By Austin Amaya,
PhD, Director, Analytics & AI, ECS
and Matthew McDonald
Senior Director, D&I Technology and Innovation

The New Frontier of Risk: AI Models as Protected Assets

In 2025, the challenge isn’t building powerful AI models. It’s making sure they don’t become liabilities.

Across federal agencies, data scientists are training AI on some of the most sensitive information in existence: intelligence imagery, sensor feeds, and operational records. These models aren’t just smart systems; they’re restricted assets in their own right.

Why? Because unlike traditional software, AI doesn’t simply process data. It remembers it. And that memory can leak.

How AI Models Leak Sensitive Data

Even without access to raw training sets, adversaries can extract sensitive information directly from AI models themselves.

  • Black-box attacks: By submitting repeated queries and studying outputs, attackers can infer whether specific data points were in training (membership inference) or even approximate sensitive images (model inversion).
  • White-box attacks: With access to a model’s architecture and parameters, adversaries can analyze neurons and weights to reconstruct training data or reveal dataset properties such as hidden sensor characteristics.

The risk is clear: AI models trained on sensitive data inherit the same classification. Left unprotected, they can expose sources, methods, and capabilities, handing adversaries insights that years of intelligence collection sought to conceal.

ATTACK TYPE

Black-box attacks
(query-only)

WHAT HAPPENS

Adversary infers or reconstructs training examples

EXAMPLE RISK

“Was this satellite image in training?”
>> Leaks sensor characteristics

ATTACK TYPE

White-box attacks
(full access)

WHAT HAPPENS

Adversary mines neurons/weights for sensitive features

EXAMPLE RISK

“Which features drive this decision?”
>> Exposes vulnerabilities in critical infrastructure

From Cyber Defense to Model Defense

It follows that AI security requires the same rigor as cybersecurity. That means a layered defense posture, applied directly to models.

Differential Privacy:
Adds mathematical noise during training so no single example overly influences the model.

Data Scramblers (e.g., Disguised-Nets):
Transform inputs before training so models never “see” raw protected data.

Adversarial Testing:
Red-teaming with known attack methods to probe for leakage before deployment.

Output Sanitization:
Rounding or restricting outputs to make reverse-engineering harder.

Knowledge Distillation:
Training a “student” model from the outputs of a sensitive “teacher” model, reducing exposure of raw restricted details.

No one safeguard is enough. Agencies must combine techniques, just as they do with networks and infrastructure, to ensure resilience.

Can AI Models Trained on Sensitive Data Ever Be Downgraded?

Agencies often ask: if a model is trained on protected data, is it forever locked down?

The default answer must be yes. But in rare, mission-critical cases, downgrading may be considered, provided strict conditions are met.

  1. Define the sensitivity: Is classification tied to content (e.g., military platforms in imagery), source (e.g., sensor signatures), or metadata (e.g., resolution, geolocation)?
  2. Prove it won’t leak: Show that the model no longer contains or reveals the restricted element, backed by technical evidence.
  3. Validate with testing: Subject the model to adversarial red-teaming and formal privacy evaluation.
  4. Maintain oversight: Ensure ongoing monitoring and controls, just as with other sensitive systems.

Downgrading is possible, but it must be the exception, not the norm. Usability can never outweigh security.

The Future of Secure Federal AI

Getting this right requires expertise across multiple domains: machine learning, cybersecurity, and federal classification policy.

ECS helps agencies bridge these worlds by:

  • Determining when models must remain sensitive, and when downgrading might be safely pursued
  • Applying privacy-preserving training, adversarial testing, and secure deployment pipelines
  • Conducting rigorous risk assessments and red-teaming
  • Building trustable, usable AI that never compromises mission security

The next frontier of AI in national security is not just building smarter models, it’s building defendable models. Agencies that approach AI with the same discipline applied to cyber defense will unlock mission value while preventing mission risk.

Sitting on sensitive AI models? Let’s talk. ECS helps agencies evaluate and secure AI systems trained on sensitive data.

Share
WE'RE HIRING