The structural collapse of regulated enterprise security begins the moment governance loses line-of-sight to the data pipeline.
Across finance, healthcare, telecommunications and critical infrastructure, a new class of breach is forming — not from ransomware, not from credential compromise, but from unsanctioned artificial intelligence ingestion of sensitive data into third-party systems where neither the organisation nor its regulator has forensic visibility. The FCA's recent enforcement actions against UK banks for cloud-service misconfiguration; the SEC's 4-day breach notification rule tightening (effective 2023); the European Banking Authority's spotlight on AI governance gaps in NIS2 implementation — all signal the same structural failure: the control plane has become disconnected from the data plane. Legacy compliance infrastructure (SIEM, DLP, IAM policy engines, and encryption key management) assumes that data remains within bounded, auditable perimeters. Modern AI integration has shattered that assumption. The regulatory framework has not yet caught up. But the breach surface has already doubled.
The path to this moment is worth understanding — not as an operational problem to be solved with faster detection, but as an architectural inevitability that the industry's current security model could not have prevented.
The Industry Narrative: Shadow AI as Operational Convenience
In late 2023 and throughout 2024, multiple compliance officers and security leaders across FTSE 100 firms, US systemically important financial institutions (SIFIs), and large healthcare networks began reporting the same pattern. Development teams, data analytics groups, and risk departments had independently adopted generative AI tools — ChatGPT, Claude, internal large language models — for documentation, code review, threat modelling, and customer data summarisation. At first, these were desktop applications, occasional queries, seemingly low-risk. But as organisations moved to enterprise deployments (Copilot Pro subscriptions, OpenAI API integration, internal self-hosted Llama or Mistral instances), the data flow became systematic.
The UK Financial Conduct Authority issued specific guidance in Q1 2024 addressing "third-party AI model exposure and firm-level data governance", following informal market signals about undisclosed LLM training pipelines. The US Office of the Comptroller of the Currency (OCC) similarly highlighted AI governance as a priority examination theme in January 2024. Australia's APRA released draft guidance on AI risk for ADIs (Authorised Deposit-taking Institutions) and recommended stress testing for "model leakage scenarios".
But the catalysing event was the Snowflake tenant cascade breach of late 2024. Whilst the primary attack vector involved compromised credentials (a well-documented pattern in that incident class), the secondary exposure was structural: customer data had been exported to shared cloud infrastructure via LLM APIs for semi-automated anomaly detection and fraud model training, without explicit customer notification. The breach surface was thus amplified — not just the Snowflake warehouse itself, but the training datasets embedded in third-party AI services where the customer organisation no longer had access controls, audit logging, or data residency guarantees.
Similarly, in the Change Healthcare ransomware incident (February 2024), whilst the initial compromise was via an unpatched Citrix Netscaler, the post-compromise lateral movement and data exfiltration were accelerated by the presence of unaudited analytics workflows that had ingested patient records into cloud-native ETL pipelines connected to third-party ML inference services. The attacker's dwell time was compressed precisely because the control plane was fragmented.
More recently, the Scattered Spider intrusion at Marks & Spencer (2025) — a social engineering and lateral movement campaign — succeeded partly because identity compromise allowed attackers to assume the personas of data science teams and execute queries against data warehouses to export transaction records into "AI model training" workspaces on sanctioned but insufficiently audited cloud platforms.
The pattern is consistent: organisations adopted AI tools at the application layer without re-architecting the data governance and control planes to accommodate them. Compliance teams added policies ("AI tools must be approved before use"), but those policies created paper compliance, not structural resistance.
The Architectural Root Cause: Control Plane Fragmentation
The industry's response has been predictable — and inadequate. Vendors like Databricks, Palantir, and data-classification firms have released "AI data governance" layers atop existing SIEM/DLP infrastructure. Consultancies have published AI compliance frameworks (aligned to NIST AI RMF, ISO/IEC 42001, the EU's proposed AI Act). Auditors have extended their scope to include "LLM supply chain risk". But these are all control-plane augmentations. They assume the original architecture was sound.
It was not.
The fundamental issue is this: modern data governance (ISO 27001 Annex A.8, NIST CSF "Govern", DORA Article 28, FCA SM&CR operational resilience requirements) was designed for perimeter-bounded data systems. Access control, encryption key lifecycle, audit logging, and data classification all presume that data flows through known, inspectable channels — EDR on endpoints, SIEM aggregating logs, DLP rules matching sensitive data patterns against exfiltration attempts.
Generative AI rewired that assumption. An employee can now upload a dataset to a third-party API, receive a processed result, and the entire transaction occurs outside the organisation's data classification taxonomy and control plane. The SIEM logs the API call (if logging is configured), but does not see the data in flight. The DLP rule cannot inspect what happens inside the LLM's context window. The encryption key management system has no visibility into model weights trained on customer data. The audit trail is fragmented across the organisation's logs, the API provider's logs (often inaccessible without legal process), and the LLM provider's opaque infrastructure.
This is not a detection problem. You cannot SIEM your way out of architectural fragmentation. Deploying Sigma rules against "suspicious API calls to OpenAI endpoints" or configuring Snort signatures for "exfiltration to cloud storage" merely adds noise to the control plane. The data has already left the boundary.
The regulator has noticed. The UK PRA's recent "Operational Resilience" framework (which maps to DORA and NIS2 in spirit) requires firms to map data flows and identify points where third-party dependencies introduce uncontrolled risk. The SEC's breach notification rule, tightened to four days in 2023, has created an implicit demand for real-time forensic visibility — which is impossible if data is flowing into black-box AI services.
But the deepest issue is epistemic: organisations do not know what data is being processed by AI systems, because they never asked the systems to answer that question as an architectural invariant. Governance became a policy document, not a substrate property.
The PULSE Reading: Post-Breach Resistance Through Data-Plane Architecture
The PULSE doctrine rejects the premise that you can bolt governance onto a fragmented data plane. Instead, it proposes this: data governance must be embedded in the substrate itself, not overlaid via policy engines. Specifically:
Zero-Knowledge Substrate. The first principle is that sensitive data should never be available, even to authorised systems, in a form that permits onward export without cryptographic proof of purpose. This is not merely encryption at rest and in transit — encryption is a solved problem. It is functional separation: data accessible for legitimate analytics workflows remains cryptographically inaccessible to general-purpose AI APIs. The substrate enforces a privilege boundary that cannot be crossed by policy violation alone.
In practical terms, this means: an organisation should architect its analytics pipeline so that data flowing into ML models is transformed through a domain-specific data proxy that decrypts only for the intended operation, then re-encrypts or discards intermediate states. This is different from a data gateway (which merely logs access) or a data vault (which assumes the vault itself is trustworthy). The proxy is stateless, ephemeral, and cryptographically hostile to data exfiltration — exfiltration would require exfiltration of the decryption key itself, which would leave a forensic trace and activate an automated posture adjustment.
Control-Plane and Data-Plane Separation. Modern compliance frameworks blur this boundary. They assume that a policy engine (FCA rules, NIST controls, ISO 27001 clauses) determines data handling, and the system simply executes the policy. But policies are narratives, subject to interpretation, circumvention, and bureaucratic erosion. The PULSE model inverts this: the data plane is the truth. The control plane observes and adapts. This means: governance rules are not written once in a compliance document; they are instantiated as continuous properties of the data flow itself. A rule "patient records cannot flow to unapproved third parties" is not a checkbox on a compliance form; it is a cryptographic invariant enforced at the data layer, such that unauthorised egress is cryptographically impossible, not just detectable.
Domain-Specific Automation. The final layer is that AI governance must not be bolted onto legacy SIEM, SOAR, or DLP platforms. Those platforms are designed for forensic detection and incident response — they see an incident after it has occurred. Instead, governance of AI data flows requires domain-specific primitives embedded into the substrate: a data classification engine that understands which data categories are permissible within which model training contexts; a model lineage tracker that records not just which data entered a model, but what intermediate states were produced and where they flowed; and an adaptive posture system that continuously shifts the cryptographic boundaries around sensitive data as threat intelligence changes.
Practical Substrate Design: An Example
Consider a regulated healthcare organisation subject to HIPAA, UK GDPR, and increasingly NIS2. Today, it operates a cloud data warehouse (Snowflake or similar) where de-identified patient records are stored. It wants to train an ML model to predict patient readmission risk — a legitimate analytics use case. Here is how a standard compliance approach handles this:
- Data governance policy: "De-identified data only. Clinical teams must approve model training."
- Technical control: Role-based access control (RBAC) on the data warehouse. Only authorised users can query the table. DLP rule: "Block queries that return more than 10,000 rows."
- Audit: SIEM logs all queries. Compliance team reviews monthly.
Here is what happens in practice: an analyst queries the 10,000-row limit five times, exfiltrates 50,000 rows to a local file, uploads it to a cloud ML service with a justification memo ("exploratory training run"). The SIEM logs the queries (within policy). The DLP rule is not triggered (queries are under the limit). The exfiltration to the cloud service is logged as a "permitted API call" (the tool is on the approved list). The compliance team sees no violation. Six months later, the cloud service is breached, patient data is compromised, and the organisation faces regulatory action — not because the policy was violated, but because the policy could not express the actual constraint (no de-identified data can be exported to third-party ML services without re-encryption and explicit audit linkage).
A PULSE substrate approach would work differently. The de-identified data remains encrypted under a key that the data warehouse can decrypt for queries, but cannot export in plaintext. When an ML training job needs access, the substrate:
- Validates that the job is registered in a model lineage ledger (signed, tamper-evident).
- Spins up an ephemeral compute instance running in a sealed container.
- Decrypts the data only within that container, only for that job, and only for the declared training duration.
- The container is cryptographically hostile to exfiltration — any attempt to export plaintext data triggers a posture shift (the decryption key rotates, isolating that container; an anomaly signal is sent to the adaptive control plane).
- After training, the model weights are stored encrypted, with a cryptographic proof linking them to the specific dataset version used.
This substrate makes exfiltration require not just policy violation, but active cryptographic subversion. It shifts the attacker's burden from social engineering ("this is an approved use case") to cryptographic compromise. It moves the boundary from policy-driven to property-driven.
The Regulator's Dilemma and the Window
Regulators are aware that the compliance model is fractured, but they are behind the curve. The FCA's upcoming operational resilience assessments (PRA SS7/2024 and the follow-on prudential regimes) will require firms to map third-party AI dependencies. The SEC's AI disclosure rules (still in draft) will eventually require breach notification for "model training data exfiltration". But these are still compliance layer fixes — better policies, clearer reporting, faster detection.
They will not prevent the next Snowflake-scale incident. They will only ensure it is audited more thoroughly afterwards.
The window for substrate-level redesign is open now, before regulation crystallises around the assumption that AI governance is an overlay problem. Organisations that architect zero-knowledge data pipelines, implement cryptographic rather than policy-driven boundaries, and embed domain-specific AI governance into their data planes will move from "detection and response" to "post-breach resistance". Those that do not will face a regulatory environment in 2026-2027 where they are required to have that architecture, retrofitted at prohibitive cost, whilst their competitors — and the threat landscape — have already moved on.
The question is not whether your organisation will adopt AI. It is whether you will architect its data plane before, or after, the next breach.
Qualified operators seeking detailed substrate design briefings may request a confidential technical discussion under executed Mutual NDA.
Request a briefing under executed Mutual NDA.
PULSE engages only with verified counterparties. Strategic briefing material — reference architecture, regulatory mapping, deployment topology — is released after counter-execution of the NDA scoped to the recipient's evaluation purpose.
Request Briefing →