Skip to main content
Technical Deep Dive
Engineering

Ultra-Low Latency Streaming: A Technical Deep Dive

Achieving sub-50ms streaming latency requires understanding the entire media pipeline—from encoder to viewer. This guide breaks down every component and provides actionable optimizations for real-time streaming.

18 min readJanuary 8, 2025By WAVE Engineering Team

Why Latency Matters

In traditional broadcasting, a 30-second delay between capture and display is acceptable—viewers at home watch the same feed, just slightly behind the live event. But modern interactive applications demand more. Sports bettors need odds that match what's happening on the field. Auction participants need to see bids in real-time. Gamers expect their audience to react to plays as they happen.

Ultra-low latency streaming—typically defined as sub-500ms end-to-end—enables these interactive experiences. But achieving it requires rethinking the entire streaming pipeline, from capture to delivery.

This guide provides a comprehensive technical overview of ultra-low latency streaming: where latency comes from, how different protocols handle it, and specific optimizations you can implement today.

Protocol Latency Comparison

Different streaming protocols offer vastly different latency characteristics. Choosing the right protocol is the first step to achieving your target latency.

ProtocolTypical LatencyReal-TimeInteractive
Traditional HLS15-30 seconds
Low Latency HLS (LL-HLS)2-5 seconds
RTMP to HLS3-10 seconds
SRT100-500ms
WebRTC (WHIP/WHEP)50-200ms
WAVE OMT<16ms

Where Latency Comes From

Understanding latency requires examining every step in the media pipeline. Each component contributes to total end-to-end delay.

Encoding

Contribution: 10-100ms

Use hardware encoding (NVENC/QuickSync) with ultrafast/zerolatency preset

Buffering

Contribution: 50-2000ms

Minimize jitter buffer size, use adaptive playback

Network Transit

Contribution: 5-100ms

Edge locations close to both broadcaster and viewers

Protocol Overhead

Contribution: 10-500ms

Use UDP-based protocols (WebRTC, SRT) instead of TCP

Transcoding

Contribution: 100-1000ms

Skip transcoding or use hardware-accelerated ABR

Decoding

Contribution: 10-50ms

Hardware decoding, optimized video player

Protocol Deep Dive

Each low-latency protocol makes different tradeoffs between latency, reliability, and ease of deployment.

WebRTC (WHIP/WHEP)

50-200ms

Web Real-Time Communication with standardized ingest (WHIP) and egress (WHEP) protocols.

Advantages

  • Browser-native, no plugins required
  • Excellent NAT traversal
  • Built-in adaptive bitrate
  • End-to-end encryption

Considerations

  • Complex server infrastructure
  • Higher bandwidth overhead
  • Limited codec support

Best For

Browser-based interactive streaming, video calls, auctions

SRT (Secure Reliable Transport)

100-500ms

Open-source protocol designed for reliable, low-latency video transport over unpredictable networks.

Advantages

  • Excellent error recovery
  • Works over public internet
  • AES encryption built-in
  • Wide encoder support

Considerations

  • Not browser-native
  • Requires player integration
  • Higher latency than WebRTC

Best For

Professional contribution, remote production, broadcast ingest

WAVE OMT Protocol

<16ms

Proprietary ultra-low latency protocol optimized for one-way media transport.

Advantages

  • Sub-16ms latency possible
  • Optimized for high-scale delivery
  • Seamless protocol fallback
  • Enterprise-grade reliability

Considerations

  • Requires WAVE platform
  • Limited to supported regions

Best For

Sports betting, financial trading, competitive gaming, live auctions

Encoder Configuration

Encoder settings have a significant impact on latency. Here are optimized configurations for different latency targets.

Ultra-Low Latency (<50ms)

Encoder:x264 ultrafast / NVENC LLHQ
Keyframe Interval:0.5-1 second
B-frames:0 (disabled)
Rate Control:CBR
Lookahead:0 (disabled)
Tune:zerolatency
Profile:baseline or main

Balanced (100-500ms)

Encoder:x264 veryfast / NVENC Quality
Keyframe Interval:2 seconds
B-frames:0-2
Rate Control:CBR or VBR
Lookahead:0-10 frames
Tune:zerolatency or default
Profile:main or high

Industry Requirements

Different industries have different latency requirements based on their interactive needs and regulatory considerations.

IndustryRequirementReasonRecommended Solution
Sports Betting<100msPrevent delayed viewers from exploiting oddsWebRTC with edge transcoding, OMT for premium
Live Auctions<500msReal-time bidding requires immediate feedbackWebRTC with fallback to LL-HLS
Esports<200msMatch viewer experience to gameplaySRT ingest with WebRTC delivery
Financial Trading<50msMarket-moving information deliveryOMT protocol with dedicated infrastructure
Interactive Education<1sNatural conversation flowWebRTC for interactive, HLS for scaled viewing
Remote Production<100msDirector feedback and coordinationSRT with bonded connections

Measuring & Monitoring Latency

You can't optimize what you can't measure. WAVE provides comprehensive tools for latency monitoring and troubleshooting.

WAVE Latency Dashboard

Real-time visualization of end-to-end latency across all protocols

  • Glass-to-glass latency
  • Network jitter
  • Buffer state
  • Frame drops

Protocol Analyzer

Deep packet inspection for troubleshooting delivery issues

  • Packet loss
  • Retransmissions
  • Congestion events
  • Bandwidth utilization

Viewer Experience Metrics

Aggregate latency data from actual viewer sessions

  • P50/P95/P99 latency
  • Buffering ratio
  • Start time
  • Quality switches

Key Takeaways

  • Choose the right protocol: WebRTC for browser-native, SRT for professional contribution, OMT for sub-16ms requirements.
  • Optimize encoding: Use hardware encoders with zerolatency tuning and minimal B-frames for fastest encode times.
  • Minimize buffering: Every buffer adds latency. Use just enough to handle network jitter, no more.
  • Deploy at the edge: Geographic proximity between servers and viewers reduces network transit time.
  • Measure continuously: Monitor glass-to-glass latency in production to catch regressions early.
  • Plan for fallback: Have higher-latency protocols ready for viewers who can't receive low-latency streams.

Ready to Achieve Ultra-Low Latency?

WAVE provides the infrastructure and protocols you need to deliver sub-50ms streaming at scale. Get started with a free trial today.

Ultra-Low Latency Streaming: A Technical Deep Dive | WAVE Blog