Variations of this story have played out at trading platforms of every size - from scrappy fintechs to established brokerages - at exactly the moments that matter most: market open, FOMC announcements, earnings surprises, geopolitical shocks. The moments when your users need you most are the moments most likely to expose the hidden fault lines in your streaming infrastructure.
The painful irony? Most of the time, the platform works fine. It handles the Tuesday afternoon session beautifully. The team gains confidence. Architecture reviews get skipped. Technical debt gets deferred. And then - inevitably — something happens that nobody planned for.
This is the first in a three-part series on real-time data infrastructure for trading platforms. We're going deeper than most of these conversations go - past the benchmarks and the marketing copy, into the actual architectural decisions that separate the platforms that hold under pressure from the ones that fold.
Part 1 (you're reading it): Why streaming infrastructure fails — and the pattern nobody talks about.
Part 2: How Tier-1 firms actually architect real-time data layers at scale.
Part 3: The build vs. buy decision - a framework you can use this quarter.
The Architecture That Works - Until It Doesn't
Here's how most trading platforms arrive at their current streaming architecture: pragmatically. A small team needs to show prices updating in real time. Someone suggests WebSockets. It works. The demo looks great. Users grow. The team scales the same architecture with more servers, CPU, and memory.
For a long time, this is fine. WebSockets are genuinely good technology. They're bidirectional, low-latency, and supported in every browser. At 100 users, 1,000 users, even 10,000 users - the cracks may never show.
The cracks show at 40,000. Or when your mobile app users switch from WiFi to cellular mid-session. Or when your enterprise client behind a corporate proxy can't establish a WebSocket connection because their IT department's firewall strips upgrade headers. Or when a brief network hiccup causes 35,000 clients to try to reconnect within the same 10-second window.
"The build that works in staging never fully prepares you for what happens when 40,000 real humans, on 40,000 different network conditions, all need the same price at the same millisecond."
These are the normal operating environment of a trading platform at scale, not just edge cases. And they expose a fundamental truth about DIY streaming infrastructure: it's optimized for the average case, not the critical case.
The Three Fault Lines
After working with trading platforms across the fintech spectrum, three failure modes appear again and again - not as bugs, but as architectural inevitabilities baked into how most teams build streaming from scratch.
Fault Line 1: The Reconnection Storm. When connectivity drops — even briefly - clients don't politely wait to reconnect. They all try at once. A standard WebSocket server under a reconnection storm experiences exponential load at the exact moment it's least equipped to handle it. Without intelligent backoff, rate limiting, and connection queuing built into the streaming layer itself, this becomes a cascading failure.
Fault Line 2: The Bandwidth Trap. Streaming prices for 300 financial instruments to 40,000 users sounds straightforward until we do the math. If each instrument updates every 100ms, that's 3,000 updates per second - per user. Without intelligent data conflation - collapsing rapid-fire updates so that only the latest value is delivered - you're either drowning the network or building custom conflation logic on top of the WebSocket layer. Most teams do neither well.
Fault Line 3: The Last-Mile Reality. The infrastructure might be flawless - the AWS setup, the load balancers, the WebSocket clusters are all perfectly configured. And then a user tries to access your platform from inside a bank's corporate network. The proxy server strips the Upgrade: websocket header. The connection silently falls back to polling. Corporate proxies, hotel WiFi, and strict firewall environments. They're the standard operating environment of the highest-value enterprise clients.
The Invisible Tax on Your Engineering Team
There's a cost to DIY streaming infrastructure that never shows up on an AWS bill. It shows up in the sprint velocity. In the proportion of the senior engineers' time spent on infrastructure maintenance versus product development. In the number of on-call incidents that wake people up at 2AM for problems that have already been solved - by someone else, years ago.
The Pareto problem of DIY streaming: teams typically spend 80% of their streaming-related engineering effort on infrastructure - reconnection logic, bandwidth management, firewall handling, connection pooling - and 20% on the actual product features that users experience. The math should be the other way around.
"Every hour the best engineers spend maintaining WebSocket infrastructure is an hour they're not building the useful features that will actally generate revenue."
Consider what "building our own WebSocket infrastructure at scale" actually means in practice: custom reconnection and retry logic, server-side rate limiting, proxy detection and HTTP streaming fallback, delta compression and data conflation, mobile-specific connection handling, bandwidth throttling, multi-server session affinity. And then: maintaining all of it, evolving it as your user base grows, and debugging it at 2AM when it fails in a way you didn't anticipate.
What the Firms That Never Go Down Actually Do
Morgan Stanley. UBS. eToro. IG Group. ActivTrades. adesso - which manages 6 million financial instruments for the German market, pushing 500,000 to 600,000 price updates per second under normal conditions, spiking to 2 million per second at peak.
These firms share something: they've solved the streaming problem at the infrastructure layer rather than the application layer. Instead of building bespoke WebSocket infrastructure and patching it every time production reveals a new failure mode, they've deployed a streaming broker specifically designed for internet-scale last-mile delivery. The technology they share is Lightstreamer.
The Intelligent Streaming Difference
Lightstreamer isn't positioned as a WebSocket replacement. It's more accurate to think of it as a streaming broker that makes WebSockets work the way we always hoped they would - and falls back gracefully when they can't.
Intelligent Streaming and automatic protocol fallback. Lightstreamer automatically detects the best available transport for each client - WebSocket where available, HTTP streaming as a fallback, long polling as a last resort. This happens transparently. The corporate proxy silently killed the WebSocket? Lightstreamer routes around it.
Data conflation. When a financial instrument updates faster than a client can consume it, Lightstreamer automatically conflates updates, delivering only the latest value. This alone can reduce bandwidth consumption by 60–80% during high-volatility sessions.
Delta delivery. Rather than transmitting full instrument records on every update, Lightstreamer transmits only the fields that changed. For a stock tick where only the last price and volume changed, this means sending two values instead of thirty. At 500,000 updates per second, this adds up.
Bandwidth throttling. A mobile user on 4G has different bandwidth constraints than a desktop trader on fiber. Lightstreamer adapts delivery speed per client - slow connections receive real-time data at a rate their connection can actually sustain.
Not because they're lucky. Not because they have bigger teams. Because they made one architectural decision - early enough - that moved streaming reliability from a thing they maintain to a thing they depend on.
In Part 2, we go inside the architecture. How does adesso actually handle 2 million updates per second? What does the Lightstreamer–Kafka integration look like in a real trading platform? And what's the answer to the question every CTO eventually asks: could we just build this ourselves?
DP5 offers a free architecture consultation for trading platform engineering teams.
Schedule Call for Free Consultation →