The Patch Cycle Died When Firmware Stopped Being Deployed

The industry treats IoT device lifecycle management as a logistics problem—firmware versions, update schedules, rollback procedures—when it is actually an architectural catastrophe hiding in plain sight: the devices that move data, gate critical infrastructure, and authenticate transactions operate under an assumption of updateability that has never existed and will never exist at scale.

The consequence is that organisations now defend perimeters populated by devices that are permanently vulnerable by design. Not because patch cadences are slow—they are—but because the entire model of "ship, deploy, patch in perpetuity" is a fantasy. Every connected device in production is a fixed point in an adversary's threat model. And the industry has responded by building detection systems to watch them die slowly.

This is not a procurement problem. It is not a vendor roadmap problem. It is an infrastructure crisis dressed up as a monitoring problem, and the standard remediation—SOAR orchestration, device inventory platforms, firmware update dashboards—makes the situation worse, not better.

The Industry Narrative: Visibility Without Autonomy

The published account of IoT security failure runs like this: organisations cannot inventory their connected devices; vendors ship firmware with known CVEs; update deployment fails at scale; and consequently devices remain vulnerable for months or years. The remediation pathway is familiar—better asset discovery (CrowdStrike Falcon Endpoint, Tenable Nessus, Qualys, Rapid7 InsightVM), centralised update management (ManageEngine Mobile Device Manager, Cisco Meraki, VMware AirWatch), and enforcement of compliance via NIST SP 800-153 (Guidelines for Securing Wireless Local Area Networks) and ISO 27001 Annex A.12.6.1.

The regulatory backdrop amplifies this. DORA (Digital Operational Resilience Act) in the EU requires financial institutions to maintain an inventory of IT assets and demonstrate "third-party risk" controls by January 2025. NIS2 extends the notion to critical infrastructure operators—asset inventory becomes legally mandatory. NYDFS Part 500 (Cybersecurity Requirements for Financial Services Companies) mandates "multi-factor authentication" and "annual penetration testing," implicitly assuming that devices can be remediated within the audit cycle. APRA CPS 234 in Australia requires "risk management frameworks" that treat device lifecycle as a process control subject to AUD 1bn+ fines. The FCA's Senior Managers & Certification Regime in the UK makes compliance failures the personal liability of named executives.

Real incidents have made the failure visceral. The 2024 Synnovis healthcare ransomware attack (Change Healthcare's billing infrastructure subsidiary) did not begin with a zero-day in the EHR system itself—it pivoted through networked medical devices and administrative IoT equipment (thermostats, badge readers, printers) that had never received security updates. The devices were not targeted directly; they were lateral-movement waypoints with no monitoring, no inventory record, and no patch history. The incident cascaded across NHS trusts for weeks because the devices existed in a "shadow IT" zone, outside the scope of the centralised asset management system. Administrators could not identify them; vendors could not push patches to them; and forensic teams could not reimage them because the original firmware images had been discarded years prior.

Similarly, the 2023 Change Healthcare incident (affecting 100 million US health insurance records) involved older IoT-class devices (administrative workstations, legacy laboratory instruments) that remained online despite known vulnerabilities. The attacker used trivial reconnaissance—port scanning, default credential spray—to gain initial access. The devices had SNMP configured with community string "public"; they had SSH running on port 22 with vendor-default credentials; some were accessible via Telnet. None of this was "advanced." The reason the attack succeeded was simply that these devices existed outside any update regime. They were not patched because they could not be patched: their operating systems were no longer manufactured, their vendors had ceased business, and reimaging would have required downtime that the clinical operation could not absorb.

The standard response has been to build visibility. The Synnovis recovery involved deploying EDR (endpoint detection and response) agents to all accessible devices, ingesting syslog into a SIEM, and tuning YARA rules to detect lateral movement. The Change Healthcare investigation triggered thousands of device reimagings, firmware audits, and compliance attestations. Vendors released "IoT security modules"—bolt-on SOAR playbooks to detect anomalous firmware versions, and automated alert pipelines to identify unpatched devices. The industry published guidance: Gartner's "Asset Inventory Maturity Model," SANS ISC handler recommendations on "IoT baseline hardening," Krebs on Security's advocacy for vendor accountability.

All of this was noise. None of it addressed the underlying problem.

The Structural Failure: Updateability as Infrastructure Myth

The issue is not that organisations lack visibility into their IoT devices. The issue is that the devices should not exist in their current form.

Firmware updateability assumes several preconditions that are never satisfied in practice: (1) the vendor is still in business and actively maintaining the product; (2) the organisation has documented the original configuration and can test updates before deployment; (3) production downtime is acceptable and scheduled; (4) the device firmware is cryptographically signed and cannot be downgraded; (5) the update mechanism itself is not a threat surface (no code injection, no tampering during transit). In the real world, none of these conditions hold systematically. Vendors exit markets. Firmware signings are optional. Testing is deferred. Downtime is unacceptable. Update mechanisms are trivially compromised.

Consider what actually happens. A CVSS 9.2 vulnerability is published in a TCP/IP stack used by a 2015-era network-attached storage (NAS) device or an industrial control system (ICS) gateway. The vendor publishes a firmware update—maybe. The organisation's security team discovers the device exists (usually via a third-party scanner, never via their own inventory). They request the update from the asset owner. The asset owner says "the device is in production and cannot be rebooted." Weeks pass. The vulnerability is weaponised by a ransomware operator. The device becomes the entry point. By the time forensics teams examine the device, the firmware may have been overwritten, the device manufacturer may have gone out of business, and no one has the original configuration.

This is not a failure of process. It is a failure of architecture. The device was never designed to be continuously maintained. It was designed to be deployed-and-forgotten. The patch cycle was always aspirational. What the industry calls "legacy" devices are not antiques—they are the current state of the art for most IoT manufacturers. A 2019-vintage industrial gateway, deployed in 2021, with a 7-year useful life, will still be in production in 2028. The vulnerabilities discovered in that gateway in 2024, 2025, and 2026 will never be patched. There is no pathway to patch them. The organisation knows this. The vendor knows this. The regulatory framework assumes this is not happening while simultaneously imposing penalties when it does.

The PULSE doctrine recognises this as a design problem masquerading as a process problem. The standard response—device inventory, patch dashboards, SIEM correlation—deepens the failure by creating the illusion of control. Organisations invest millions in asset management platforms, only to discover that 30 per cent of their devices are unmanageable and a further 40 per cent are managed only through manual process (phone calls, spreadsheets, vendor portal logins). The cost of discovering a device's patch status often exceeds the cost of simply replacing it. And yet replacement is politically infeasible—the device is already paid for, embedded in operations, and "still working."

The correct response is not better visibility. It is architectural rejection of the premise.

Zero-Knowledge Substrate: The Only Viable IoT Architecture

The PULSE doctrine approaches this via a principle: design systems such that the compromise of any individual device does not propagate to the data plane or control plane.

This requires moving away from the notion that devices can or should be "secure individually." Instead, engineer systems in which every device is assumed to be potentially compromised at any moment, and design the network such that compromised devices are contained, not just detected.

Concretely, this means: (1) strict network segmentation at layer 3 and below—no device communicates with any other device outside its explicitly authorised peer group, and all inter-segment communication flows through a stateful appliance that validates every packet (not just inspects it). (2) Data-plane isolation—IoT devices never hold plaintext data at rest or in transit. Data is encrypted end-to-end with cryptographic keys managed outside the device. The device is a stateless conduit, not a repository. (3) Control-plane separation—firmware updates, configuration changes, and policy decisions are never pushed to devices. Instead, the device's functional parameters (which data flows to process, which transformations to apply, how long to retain state) are re-evaluated on every transaction, by systems external to the device. (4) Cryptographic authentication on every operation—not just on login. Every packet, every command, every data transfer requires proof that the issuer is authorised to issue it. This proof should be asymmetric and hardware-backed where feasible.

This is not novel in principle—it is standard in payments infrastructure (HSMs, hardware security modules) and military systems (air-gapped networks, signed commands). It is not standard in commercial IoT because it is operationally demanding. It requires rethinking device architecture from the ground up. But it is the only model that survives the reality that devices cannot be patched.

The Synnovis incident, examined through this lens, would have been contained to a single device. The badge readers and thermostats would have been isolated at layer 3, unable to communicate with the billing system, the EHR network, or clinical workstations. Even if they were compromised, they could not have been used as pivot points. Similarly, the Change Healthcare attack would have required the attacker to compromise not just a single legacy workstation, but the cryptographic identity material (which would not reside on the workstation itself, but in a managed, monitored, rotated secret store). This is not "impossible." It is simply different from how those organisations were actually structured.

Continuous Adversarial Drift: Post-Breach Resilience Without Detection

The second architectural principle is that device posture—firmware version, configuration state, network behaviour—should drift unpredictably and continuously, not remain static until a patch is applied.

This is counterintuitive: administrators typically seek consistency. All devices in a cohort should run the same firmware version. Configuration should be stable and documented. Network traffic should be "normal" and predictable. This is the perspective of the operator managing known systems. It is the perspective of the attacker identifying systems.

An alternative model treats drift as a feature. Every device's operational parameters—which peers it can communicate with, which data transformations it applies, which protocols it exposes—change on a schedule known only to the management plane. Not because the device is being "updated," but because the security posture is being continuously recalculated and re-imposed. An attacker who gains access to the device at time T cannot assume that the same access will be viable at time T+1 hour, because the device's functional envelope has shifted. This is not "intrusion detection." It is "intrusion rejection"—the device makes itself uninhabitable.

This requires architectural primitives that are rare in commercial IoT: a management plane that is cryptographically separate from the data plane, time-synchronised to within milliseconds, and capable of pushing cryptographic state updates to thousands of devices without centralised state. It requires that devices be stateless — they recompute their functional parameters from signed directives, not from stored configuration. It requires that the update mechanism itself not be a vector for attack—which means it must be impossible to push unsigned firmware to a device, and equally impossible to downgrade to an older firmware version.

The Medibank breach (2022) and Optus breach (2022) in Australia, both involving exfiltration of identity documents and health records, occurred because the organisations operated in a traditional "secure by inventory, detect by log analysis" model. Medibank's breach involved credential stuffing against a legacy application server; Optus's breach involved an unsecured API endpoint with overly permissive authentication. In both cases, the organisations had EDR, SIEM, and vulnerability scanning in place. What they lacked was the architecture to make the breach irreversible in the first few minutes. Once inside, the attacker had time to establish persistence, move laterally, and exfiltrate data. A system designed on post-breach resilience principles would have made the initial compromise—the credential stuff, the API access—essentially useless, because the attacker's access token, even if valid, would not have corresponded to any actual functional capability on the network.

Domain-Specific Automation and the Rejection of SOAR

The third principle is that remediation and adaptation must be hardwired into the infrastructure, not delegated to bolt-on orchestration platforms.

SOAR (Security Orchestration, Automation and Response) systems promise to codify incident response logic—if a device shows anomalous behaviour, automatically isolate it; if a vulnerability is discovered, automatically trigger a patch; if credentials are compromised, automatically revoke them. In theory, this is sound. In practice, SOAR deployments are notoriously fragile and operationally burdensome. They require constant tuning (false positives), constant updates (as threat patterns change), and constant manual intervention (because runbooks never cover the edge case). They are effective only in controlled environments with stable infrastructure.

IoT environments are neither stable nor controllable. SOAR systems in IoT deployments spend most of their time either generating false alarms (a device reboots after a crash, triggering "anomalous behaviour" alerts) or missing real compromises (an attacker that moves slowly, or adapts behaviour to match the "normal" pattern detected by the SOAR system). The fundamental issue is that SOAR treats security as a reaction problem—threats are detected, then automatically countered. But in IoT environments, detection is often too late.

The PULSE approach rejects SOAR as a primary control. Instead, security logic is pushed into the infrastructure: the network enforcer (firewall or zero-trust appliance) automatically blocks traffic that does not match the authorisation policy, not because the SOAR system detected a threat, but because the policy was never permissive enough to allow that traffic in the first place. The device firmware cannot be downgraded because the device firmware upgrade mechanism is cryptographically locked and the key is not stored on the device. The device cannot exfiltrate data because the network architecture does not allow the device to initiate outbound connections; all traffic is inbound from the management plane or intra-segment within its peer group.

This requires different infrastructure primitives—not SIEM, SOAR, or EDR, but rather: high-performance packet filtering at layer 3 and 4 with cryptographic validation; hardware security modules (HSMs) for key management; signed firmware distribution systems with hardware-backed verification; and network segmentation appliances capable of enforcing policies per-packet, not per-flow.

The Regulatory Reading: Compliance as Architecture, Not Process

The regulatory framework—DORA, NIS2, NYDFS Part 500, APRA CPS 234, FCA SM&CR—implicitly demands post-breach resilience without stating it directly. DORA's requirement for "third-party risk management" is not satisfied by vendor questionnaires and SIEM logs. It is satisfied by infrastructure that makes third-party compromise (e.g., vendor-supplied firmware, vendor-supplied update mechanisms) architecturally incapable of propagating to the organisation's data plane.

NIS2's requirement for "risk-based cybersecurity measures" is not a call for better threat detection. It is a call for systems that do not create risks in the first place—i.e., systems in which a compromised device is not a risk to the broader infrastructure, because the risk surface is designed to be zero.

NYDFS Part 500 and APRA CPS 234 impose fines and personal liability for breach. These impose a cost structure that makes it economically rational to invest in architecture that prevents breach—even expensive architecture—rather than pay regulatory fines after the fact.

The Synnovis, Change Healthcare, Medibank, and Optus incidents all triggered formal regulatory inquiries (by the ICO in the UK, by state attorneys general and HHS in the US, by the Australian Information Commissioner and the Privacy Commissioner). In all cases, the organisations acknowledged that the incidents could have been prevented by better device lifecycle management. In all cases, the recommended remediation was visibility and patch deployment. None of the regulators have yet demanded architectural change. But they will, because visibility and patch deployment do not work.

Engaged Operators: A Briefing Under NDA

Organisations managing critical data infrastructure or payment systems that operate IoT or legacy devices in production environments, and that recognise the architectural limits of detection-and-response approaches, should request a technical briefing under executed mutual NDA to evaluate post-breach resilience design principles applied to their specific asset classes and regulatory domain.

Engagement

Request a briefing under executed Mutual NDA.

PULSE engages only with verified counterparties. Strategic briefing material — reference architecture, regulatory mapping, deployment topology — is released after counter-execution of the NDA scoped to the recipient's evaluation purpose.

Request Briefing →

Related Reading