"

A nurse on a step-down unit glances at her tablet. The patient's SpO₂ alert arrived seconds late — delayed by a silent WebSocket fallback she had no way of seeing. In most cases, a few seconds is fine.

In critical cases, it could be fatal.

- This is the real-time clinical data gap.

This series is about clinical infrastructure. When a trading platform's streaming layer fails, users experience bad UX. When a remote patient monitoring platform's streaming layer fails, care teams experience delayed alerts.

We're going to spend three parts going deep on real-time data delivery for digital health platforms - the failure modes, the architecture, and the decision framework. The same rigour we applied to the Finance series, applied to a domain where the consequences of getting it wrong are different in kind.

Part 1 (you're reading it): The failure modes - why RPM platforms fail in hospital environments and what clinical alert latency actually means.
Part 2: The MQTT-to-dashboard bridge - what a production-grade real-time architecture for health tech actually looks like.
Part 3: Build vs. buy for health IT - the framework, with HIPAA compliance and on-prem deployment built in.

~40%Of ICU alarms go unanswered by staffPMC 2024
15–40%WebSocket failure rate in proxy-heavy networksLightstreamer
<2sEngineering target for real-time alert deliveryLightstreamer Broker

The Seconds Problem

A few seconds seems like a rounding error - in most software contexts. A web page that loads slowly is bad UX. Nobody gets hurt. A clinical alert that arrives seconds late is a different problem, and the clinical literature on time-to-treatment makes this concrete.

Clinical context
Sepsis progression, acute respiratory events, and cardiac arrhythmias are three categories where time-to-alert is a documented factor in outcomes.NEJM 2017Nature Med 2022 Early warning score systems - used in step-down units and general wards - are only as effective as the latency between sensor reading and care team notification.PubMed EWS Review A system that computes an early warning score in real time but delivers it 4–8 seconds late to the care team dashboard has degraded the clinical value of that system, even if it appears to be functioning normally.

Most RPM platforms don't measure their end-to-end alert latency in production. I couldn't find any citation to validate actual numbers, unless we get specific to the device level. I was looking for a real measure in a 400-bed hospital, on a WiFi network shared with PACS imaging systems, behind a proxy server, with 300 connected devices and 80 concurrent care team dashboard sessions.

So, the numbers in test simulations and actual production can be quite different (due to a variety of factors impacting real-time data). And the gap between these numbers is the real-time clinical data gap.

"A system that delivers accurate data on a polling interval is not a real-time monitoring system. It's a delayed monitoring system with a real-time UI."

Why Hospital Networks Break Everything

Hospital networks were designed for PACS imaging, EHR access, administrative systems, and staff communications - none of which have the continuous, low-latency streaming requirements of a patient monitoring system with hundreds of devices pushing vitals every few seconds.

The result is a network environment that systematically breaks the assumptions most streaming infrastructure is built on:

Hospital WiFi is not like office WiFi

Hospital WiFi environments are dense, contested, and often segmented. Medical devices, clinical workstations, staff mobile devices, and visitor networks all share infrastructure that was frequently installed in phases over years, with different vendors, different configurations, and often minimal QoS for real-time data applications.7SIGNAL Nurses moving between wards on tablets experience handoffs that drop connections. Devices in procedure rooms operate in RF environments that create intermittent connectivity. The WiFi that "works fine" for accessing an EHR creates silent failures for a WebSocket streaming connection.

Hospital proxies silently kill WebSocket upgrades

Enterprise hospital networks - particularly those operating under strict IT security policies - deploy proxy servers and deep packet inspection systems that intercept HTTP traffic. Standard WebSocket connections require an HTTP Upgrade handshake: a client sends an Upgrade: websocket header, and the server responds with 101 Switching Protocols. A significant proportion of hospital proxy configurations either strip the Upgrade header silently or block the protocol switch entirely.Lightstreamer Blog

The failure mode is insidious: the connection doesn't error. It falls back to HTTP polling without the client knowing. The RPM dashboard appears to be working. Vitals appear to be updating. They are - every 30 seconds, on a polling interval, instead of in real time. The care team has no indication that their "real-time" monitoring is running 25 seconds behind!

Technical note
In an analysis of enterprise network environments, WebSocket connections fail silently in proxy-heavy configurations at rates that vary from 15% to over 40% depending on the organisation's security posture.Lightstreamer Blog Hospital networks cluster toward the higher end of this range due to the combination of legacy infrastructure, strict security policies, and the presence of HIPAA-mandated traffic inspection requirements.HHS HIPAA Security Rule

Cellular handoffs break streaming sessions

Nurses, physicians, and care coordinators increasingly access monitoring dashboards on mobile devices - tablets and phones that move between WiFi zones and cellular coverage throughout a shift. A standard WebSocket connection is a stateful TCP session: when the underlying network changes, the session breaks. Reconnection requires a full handshake, and during the reconnection window - typically 2–15 seconds for a standard implementation - no alerts are delivered.

For a mobile care team member who steps into a stairwell, crosses a WiFi coverage gap, or moves from a ward to a procedure room, this is not a rare event. It is the normal operating condition of their workday.Celona

The Three Failure Modes

The hospital network environment produces three distinct failure modes in RPM streaming infrastructure. Every platform we've looked at hits at least one of these. Most hit all three.

FAILURE MODE 01The Silent Polling Fallback
Clinical scenario

A care team's monitoring dashboard shows vitals updating continuously. What nobody knows is that the WebSocket connection failed silently behind the hospital's proxy server 40 minutes ago. The platform fell back to HTTP polling. Data is now updating every 30 seconds. A deteriorating patient's early warning score has been trending upward for 28 seconds without triggering the alert threshold - because the latest reading hasn't arrived yet.

What's actually happening

HTTP Upgrade header stripped by proxy → client falls back to polling without error → polling interval (typically 15–60s) replaces continuous streaming → alert latency degrades from sub-second to polling-interval-length → no visible indication to the care team that monitoring quality has degraded.

FAILURE MODE 02The Reconnection Dead Zone
Clinical scenario

A nurse is reviewing a patient's trending vitals on her tablet when she walks from the general ward to the nurses' station in an adjacent corridor. Her tablet moves from one WiFi access point to another. The transition takes 3 seconds. During those 3 seconds, her monitoring dashboard shows the last known values - stale vitals - with no indication that the connection has dropped. The alert that fired during the transition is in a buffer that hasn't been delivered.

What's actually happening

TCP session breaks on network transition → WebSocket requires full reconnect → standard implementations retry with fixed interval (2–5s) → buffered alerts during reconnect window may be missed or delivered out of sequence → care team may act on stale data without knowing it's stale.

FAILURE MODE 03The Concurrent Connection Collapse
Clinical scenario

A hospital system deploys an RPM platform across a 200-bed ward. During a shift change, 60 care team members log in simultaneously to check on their patients before handoff. The platform's WebSocket server, sized for normal concurrent load, experiences a connection spike it wasn't designed for. Response times degrade across all sessions. Alert delivery latency spikes to 8–12 seconds. The period of degraded monitoring coincides exactly with the highest-risk moment of the clinical day: the shift handoff.

What's actually happening

WebSocket server connection pool saturation during load spike → queued connection requests increase per-session setup latency → active sessions experience throughput degradation → alert delivery latency increases proportional to queue depth → shift change, admission events, and code blue responses are all correlated load events.

The Hospital Network Reality

The three failure modes above share a common cause: standard streaming infrastructure was built for the public internet, not for hospital networks. The public internet is heterogeneous but generally permissive. Hospital networks are restrictive by design - and that restriction creates failure modes that only surface in production, in the actual environment.

Standard WebSocket
Office WiFi
Fiber desktop
Hospital WiFi
VPN / proxy
Cellular handoff
Air-gapped ward
Works in 2 of 6
Lightstreamer
Office WiFi
Fiber desktop
Hospital WiFi
VPN / proxy
Cellular handoff
Air-gapped (on-prem)
Works in 6 of 6

The diagram above represents the real-world connectivity profile that an RPM platform needs to support in a typical acute care hospital deployment. Standard WebSocket handles 2 of the 6 most common network environments. Lightstreamer - built from the ground up for exactly this problem - handles all 6, transparently, without the care team or the platform engineering team needing to configure anything.Lightstreamer StreamSense

The mechanism is Intelligent Streaming: automatic protocol negotiation that detects the best available transport for each client connection and falls back gracefully without breaking the session.Lightstreamer.com On a standard WiFi connection, it uses WebSocket. Behind a proxy that strips Upgrade headers, it uses HTTP streaming. In a highly restricted environment, it uses long polling. The care team's dashboard sees the same real-time data either way.

"Does the platform actually work in Bay 7 of a 200-bed ward at 2AM, on hospital WiFi, through a proxy server, on a nurse's tablet that just switched access points."

What Sub-Second Clinical Alert Delivery Actually Requires

The clinical target — sub-second end-to-end alert delivery from sensor reading to care team dashboard — is achievable in hospital environments. It requires getting four things right simultaneously.

First: last-mile transport resilience. The streaming layer must handle the full range of hospital network environments described above - proxies, WiFi handoffs, cellular transitions, restricted firewall rules — without degrading to polling and without requiring custom code for each environment. This is the problem Lightstreamer was specifically designed to solve.Lightstreamer Broker

Second: the MQTT bridge. Most RPM platforms use MQTT for device-to-cloud data collection - it's the right protocol for IoT devices with constrained power and connectivity. But MQTT was not designed for browser-to-dashboard delivery. The bridge between MQTT and the care team's web or mobile interface is where most platforms lose both latency and reliability. We'll go deep on this in Part 2 of this series.

Third: on-premise deployment capability. HIPAA compliance, hospital IT security policy, and data residency requirements frequently mandate that patient data not leave the hospital network perimeter.HHS Cloud & HIPAAInCountry Data Residency A cloud-based streaming solution that routes patient vitals through an external data center is a non-starter for many hospital system deployments. On-premise Lightstreamer deployment keeps all patient data within the hospital's own infrastructure - the streaming broker runs inside the perimeter, not outside it.

Fourth: graceful degradation without silent failure. The most dangerous failure mode is not the outright crash - it's the silent degradation where the dashboard appears functional but is actually running on stale data. A production-grade system must distinguish between "connected and streaming" and "connected but polling" and surface that distinction clearly to the care team.

Technical note
Lightstreamer's StreamSense technology addresses graceful degradation directly.Lightstreamer StreamSense Rather than silently falling back to a degraded transport without notification, it maintains explicit connection state that can be surfaced to the application layer. An RPM platform can show a care team member a clear indicator when their session is in a degraded state - allowing them to make clinical decisions with accurate information about the freshness of the data they're seeing.

The failure modes are documented. The fix must be architectural.

In Part 2, we go inside the architecture. Most RPM platforms use MQTT for device-to-cloud data collection - it's the right protocol for IoT devices. The problem is what happens between MQTT and the care team's dashboard.

Part 2 walks through the MQTT-to-Lightstreamer bridge in production detail: what it looks like, why MQTT alone doesn't solve the last-mile problem, and what the architecture needs to deliver to hit the clinical latency target. The bridge is simpler than it sounds - and it changes everything about what's possible in a hospital environment.

How does your current platform perform in a hospital network environment?

DP5 offers a free architecture consultation for digital health engineering teams.

Schedule Call for Free Consultation →
Healthcare Series - All Parts
01
The Real-Time Clinical Data Gap← you are here
Apr 17, 2026
02
MQTT, Hospital Networks, and the Last-Mile Problem
May 12, 2026
03
Build vs. Buy for Health IT: An Engineer's Framework
Jun 10, 2026 · soon
Next → Part 02 · May 12, 2026
MQTT, Hospital Networks, and the Last-Mile Problem
Most RPM platforms use MQTT for device-to-cloud. Here's why MQTT alone doesn't solve the last-mile delivery problem - and what a production-grade bridge architecture actually looks like.