The API is not defended; it is merely observed.

The perimeter as enterprise security understood it for three decades—firewalls at the network edge, EDR on endpoints, DLP on egress points—is dead. What has replaced it is not a new perimeter, but a diffuse surface of application programming interfaces (APIs) that sits between your data and every other system on the internet, and most organisations have no architectural control over what those interfaces expose, filter, or log. APIs are not security tools. They are operational necessities. And when they become the de facto boundary of your organisation, the absence of a security control architecture designed specifically for that boundary becomes an existential vulnerability—one that no amount of detection, logging, or forensic hindsight will remediate.

This is not hyperbole born from consultancy incentive. The structure of modern data theft follows the architecture of modern data flow. Three recent incidents, each widely documented and each resulting in regulatory action, demonstrate this principle with uncomfortable clarity.

The Recorded Narrative: APIs as a Detection Problem

Between September 2023 and September 2024, Snowflake disclosed that multiple customers' data repositories had been compromised via stolen service account credentials, leading to exfiltration of billions of records from companies including LendingClub, Ticketmaster, and critically, the US Internal Revenue Service (IRS) — which lost taxpayer personal information for millions. The attack was not sophisticated in method: threat actors obtained credentials through infostealer malware deployed on developer machines, then used those credentials to call Snowflake's APIs with full legitimacy. No signatures were tripped. No IDS alert fired. The API calls were, by every observable metric, authentic.

The industry response was instantaneous and predictable: Snowflake released guidance on multi-factor authentication (MFA) enforcement, credential rotation, and network policy configuration. Customers were advised to enable audit logging and review access logs using Snowflake's QUERY_HISTORY table to detect lateral movement post-facto. The narrative framed the compromise as a detection and response problem—if only customers had been monitoring their API logs more diligently, the attack would have been caught faster, and the damage contained.

The Change Healthcare ransomware incident of February 2024, executed by the LockBit gang against one of America's largest healthcare claims processors, followed a similar arc. Access was gained through an unpatched Citrix NetScaler appliance (CVE-2023-4966), but once inside, the attackers moved laterally through the environment and exfiltrated data via APIs to extract claims, financial records, and personal health information from approximately 100 million individuals. The US Department of Health and Human Services Office for Civil Rights eventually assessed the breach as enabling the widest healthcare data compromise in American history. The remediation narrative, again: patch the appliance, segment the network, monitor API activity logs, and deploy Sigma rules for detecting suspicious bulk export patterns.

Most recently, the M&S (Marks & Spencer) incident disclosed in December 2024 involved the Scattered Spider threat group gaining access through compromised employee credentials and then systematically enumerating and calling APIs to extract customer and employee data across multiple retail and financial subsidiaries. Again, the API calls themselves were authorized and legitimate. No zero-day exploit was involved. The attacker moved through the environment with the privileges of a stolen account, and the APIs honoured those privileges without question.

Each of these incidents triggered regulatory attention: the IRS demanded forensic reports and timeline reconstruction; HHS issued guidance and pursued civil investigations under HIPAA; the UK Information Commissioner's Office (ICO) initiated investigations into M&S under GDPR and the Data Protection Act 2018. In each case, the regulator's focus rested not on the novelty of the attack technique, but on the absence of architectural controls that would have prevented the misuse of legitimate API credentials.

The industry's standard response—logging, alerting, post-incident forensics—treats the symptom, not the disease. It assumes that the API surface is fixed and known, that attackers must be detected in the act, and that the cost of detection is acceptable. None of these assumptions hold in practice.

The Structural Failure: APIs Without Owned Control

APIs are not firewalls. They are interfaces between systems. And when your organisation's data access is mediated entirely by APIs owned and operated by cloud vendors (Snowflake, Databricks, Salesforce, AWS, Azure, Google Cloud), you have outsourced your perimeter to a third party, and you cannot audit the control plane—the machinery that decides whether an API call is valid.

Consider what happened in the Snowflake case: the customer's data repository was not on their infrastructure. The API endpoint was not defined by the customer. The authentication mechanism was not defined by the customer. The audit log format was not defined by the customer. The customer could only observe what Snowflake permitted them to observe, and only after the fact. This is not a defensive posture. It is a transparency posture. And when the attacker has a valid credential and makes a legitimate API call, transparency offers no protection.

The Change Healthcare incident made this even clearer: once the attacker had a foothold inside the network perimeter, the APIs became frictionless data highways. Because healthcare systems have built their entire data exchange layer around APIs—HL7, FHIR, SFTP bridged to APIs—the attacker could request data via the very mechanisms designed for legitimate interoperability. The API calls were logged. They were, in retrospect, suspicious in pattern. But they were not blocked in real-time because the system architecture had no concept of "suspicious use of a legitimate privilege." It had only: is the credential valid? Yes. Is the request syntactically correct? Yes. Execute.

This is the fundamental architectural failure. Modern cloud-native and hybrid organisations have built their data perimeter as a set of APIs that assume:

  1. Credentials are secret and immutable (they are not; infostealer malware is commodity).
  2. The holder of a credential will use it only for the intended purpose (they will not; a compromised account is a compromised account).
  3. Monitoring and alerting will detect anomalies in real-time (they will not; bulk exports at 3 a.m. are noisy in a 24-hour operation).
  4. Post-incident forensics will be sufficient to satisfy regulators (increasingly, they are not; GDPR, HIPAA, DORA, NIS2, and emerging UK FCA SM&CR rules all demand preventive controls, not detective controls).

The NIST Cybersecurity Framework, ISO/IEC 27001, and APRA CPS 234 (for Australian financial services) all distinguish between detective controls and preventive controls. Modern regulatory expectation—reflected in the SEC's 4-day breach notification rule, NIS2's Article 21 on "security by design", and the FCA's Senior Managers & Certification Regime—has shifted decisively toward preventive architecture. The Snowflake and Change Healthcare incidents occurred precisely because organisations had built detective-only stacks, and regulators are now penalising that choice.

The PULSE Reading: Zero-Knowledge Substrate for APIs

An organisation that owns its API perimeter must architect APIs such that no single entity—not the API server, not the cloud platform, not an individual employee with stolen credentials—can unilaterally exfiltrate data. This is not achieved through logging and alerting. It is achieved through substrate-level design.

The principle is zero-knowledge: you cannot steal what is not there.

Applied to APIs, this means:

Data-plane isolation. The API that returns data to the client should not have visibility into the full dataset. The data-returning function should be architected as a query engine with built-in constraints that cannot be circumvented by parameter manipulation or privilege escalation. In practice, this means the API calls a cryptographically-bound query executor that enforces row-level and column-level access control at the cryptographic boundary, not the application logic boundary. If the API server itself is compromised, the query executor will not return unauthorised data. This is not implemented via SQL row-level security (RLS) or role-based access control (RBAC) in the database—those are enforced at the data layer and can be bypassed if the data layer is compromised. Instead, the data is encrypted at rest with encryption keys held outside the data plane, and the API request must be validated against a control plane that is architecturally separate, versioned, and auditable.

Continuous adversarial posture adjustment. Rather than assuming that an API credential is either valid or invalid, the API substrate should model credentials as time-limited, usage-limited, and geographically-constrained tokens. Not as a policy in a separate system, but as a cryptographic property of the credential itself. A stolen credential from a developer in London should become inert if used from an IP address in a country with no business operations. A credential issued for a specific query should fail if used to execute a different query. The overhead of this design is not performance—it is architectural thinking time upfront. Once implemented, the API is significantly harder to abuse because the boundaries are enforced before the data plane is touched.

Domain-specific automation. Legacy controls (EDR, SIEM, DLP, IDS/IPS) are bolted onto the infrastructure after the fact. They are network-layer or host-layer, not API-layer. They cannot see the semantic content of API requests. A SIEM rule that flags "bulk exports" is reactive and probabilistic. A domain-specific control embedded in the API substrate—a query cost governor that prevents single requests from iterating over more than a threshold of rows, or a rate limiter that is tied not to the number of requests per minute but to the quantity of data bytes returned per query—is preventive and deterministic. This control is not external observability; it is embedded logic.

Mutual authentication and cryptographic proof. Most API security relies on one-way authentication: the client authenticates to the server via a credential (API key, OAuth token, mTLS certificate). The server is trusted. But if the server is compromised, or if the server's logs are compromised, the attacker can read the logs and understand what data was exfiltrated. A zero-knowledge API architecture requires mutual authentication: the client proves its identity to the server, and the server proves its identity to the client, and all data exchange is encrypted with keys that neither party can decrypt unilaterally. This prevents not just unauthorised data theft, but unauthorised data observation by the server operator.

Architectural Principles in Practice

What does this look like concretely? Consider a financial institution that needs to expose customer account data to partner systems via API. Today's approach: deploy an API gateway (Kong, AWS API Gateway, Azure API Management), authenticate requests via OAuth2 or API keys, log all requests, and monitor for anomalies with a SIEM.

A PULSE architecture would instead:

This design has a direct consequence: a stolen credential (API key, OAuth token) can be misused only within the constraints encrypted into the credential itself. A compromised API server cannot exfiltrate data because it cannot decrypt it. A compromised database cannot leak audit logs because the logs do not contain plaintext data. An attacker with root access to the entire infrastructure still cannot bypass the HSM-protected encryption key.

The Regulatory Expectation Shift

Regulators are no longer satisfied with breach-response capability. The UK National Crime Agency's guidance on ransomware (published 2023), the FCA's guidance on third-party IT service providers (SYSC 13-1R), and the incoming NIS2 Directive (EU 2022/2555, enforcement deadline October 2024) all demand "technical and organisational measures" that prevent compromise, not merely detect it. APRA CPS 234 explicitly requires "loss prevention" as a control objective for API-mediated data access in Australian financial institutions.

The Snowflake incident is now cited in regulatory briefing materials as an example of inadequate architectural control. The M&S incident triggered ICO investigations specifically into whether the organisation had "appropriate technical measures" to prevent unauthorised API access. The Change Healthcare incident prompted HHS to issue guidance (May 2024) recommending that covered entities implement "encryption of sensitive data at rest and in transit" as a prerequisite for compliance—a shift from "recommended" to "expected baseline."

The Call to Qualified Operators

If your organisation's data flows through APIs you do not architecturally control, if your threat model assumes credentials will not be stolen, or if your defensive posture is primarily observational, you are operating under assumptions that regulators and incident evidence have already invalidated. Request a technical briefing under executed NDA to explore how sovereign API infrastructure, built on zero-knowledge substrate principles, can move your organisation from detection-dependent to prevention-assured.

Engagement

Request a briefing under executed Mutual NDA.

PULSE engages only with verified counterparties. Strategic briefing material — reference architecture, regulatory mapping, deployment topology — is released after counter-execution of the NDA scoped to the recipient's evaluation purpose.

Request Briefing →

Related Reading