Part 2 - Multiview: Revolutionizing Sports Viewing

Part 2 - Multiview: Revolutionizing Sports Viewing

Part 2 - Multiview: Revolutionizing Sports Viewing

Part 2 - Multiview: Revolutionizing Sports Viewing

Jun 4, 2025

Jun 4, 2025

Jun 4, 2025

Jun 4, 2025

Tech Blog

This post builds on the concepts introduced in Multiview: Revolutionizing Sports Viewing – Part 1, where we explored how shifting viewer expectations and evolving media infrastructure are driving the adoption of multiview experiences in sports.


Under the Hood: The Technical Challenges of Multi-View

Creating a multi-camera experience that feels immediate, intuitive, and responsive is not simply a matter of adding more video feeds. Behind what appears to be a simple tap-to-switch gesture lies a dense layer of technical orchestration: a real-time choreography of timelines, data synchronization, content formats, and network adjustments. In this second part of our blog series, we reveal the invisible systems that power seamless camera switches, and how NativeWaves has architected a streaming infrastructure that brings the power of a live production studio into the hands of any mobile device user.

Let’s take a step beyond the viewer use cases explored in Part 1 and focus on the underlying technology:

  1. How do we keep multiple streams in perfect sync, even as fans jump between camera angles?

  2. What happens when streams come from different sources, with different formats?

  3. How do we map a simple user intent like “follow my favorite player” to the correct live feed, even as that player moves between zones, courts, or laps?

Solving these challenges meant building a layered architecture of precision time management, adaptive playback intelligence, and a semantic entity model that dynamically maps content to user expectation in real time.


Timing Is Everything: Orchestrating the Hive

In interactive sports streaming, where fans demand instant responsiveness and visual consistency across feeds, timing is not just a technical metric—it’s the entire experience. When a user taps to switch cameras or pushes the button to jump back 15 seconds, the expectation is not just speed but accuracy. Every stream must show the same moment in time, from a different perspective. Any drift breaks immersion instantly, regardless of a user’s device or network conditions.

Figure 1: Multi-Track and Multi-Object Timeline. This timeline shows multiple video and audio tracks, each containing multiple video and audio objects (e.g., HLS/DASH manifests with defined start/end) for redundancy. By keeping several objects available, the client can pick the best one based on timing and available content formats and continuously align playback to a shared reference timeline. The same system supports both live and VOD, seamlessly switching between objects based on content metadata and playback position.


Low-Latency Synchronization Engine

To address this, we built a low-latency synchronization engine that fits every stream to the individual experience on the client. Unlike conventional methods that let each player run on its own buffer logic, our system continuously aligns all playback sessions to a real-time reference. This gives us the flexibility needed for low-latency handling, because we don’t need to channel the entire experience through a remote backend service. Whether it’s the onboard view from a Formula 1 car, an aerial helicopter feed, or a pit-lane camera, all streams show the exact same slice of action on the device—even if they’re delivered through different protocols or from separate CDNs.


Continuous Metadata Validation

It’s not only about time sync; it’s about a living layer that constantly validates playback position—especially during user interactions like seeking, skipping, or switching. We match every content object locally against a metadata file—what we call the EXP manifest—which is continuously served by the backend. The EXP manifest consists of real-time references and durations of all available content objects, updated constantly based on stream duration changes. Since we have very low latency, we need these constant updates on the client to keep everything aligned with true real-time.

Matching that real-time on the client is another challenge. We’ve solved it with our own time-sync service, which enables the client to match its streams to backend time accordingly. The system supports both live and on-demand playback, as well as synchronized viewing across multiple devices. To deliver a seamless experience tailored to different scenarios, it uses context-aware timeline modes—Live, Sync, and VOD—each with specific playback behaviors.


Adaptive Strategies for Smooth Performance

Two adaptive strategies ensure smooth performance:

  • Catch-up Strategy: Keeps playback aligned with live or sync targets.

  • Buffering Strategy: Prioritizes buffering and stability when needed.

By clearly distinguishing between technical states and user-facing modes, the system avoids confusion and ensures reliable, responsive playback across a wide range of use cases. When a viewer rewinds to rewatch a football goal or jumps forward to catch up with live action, our algorithm:

  1. Applies the proper timeline state logic.

  2. Cross-checks the stream’s current timeline with the expected position on the reference timeline.

  3. Corrects any discrepancies on the fly.

This dynamic alignment allows fans to jump, switch, and replay without ever feeling detached from the live moment.

Looking Ahead: Having this powerful time synchronization engine also lets us couple multiple devices to view the same moment from different angles simultaneously in perfect sync. A user could watch a football match on their TV while their smartphone keeps a constant eye on the rival’s goal. (But that’s a topic for another blog post)


Taming the Chaos of Content Sources

Beyond synchronization, one of the most complex challenges in building a multi-view architecture is the diversity and unpredictability of content sources. In real-world deployments—especially when working with third-party production teams and heterogeneous hardware environments—you rarely receive clean, uniform streams. Instead, broadcasters may provide feeds in a wide variety of formats, encryption schemes, and delivery protocols.


Handling Multiple Origins

At NativeWaves, we handle multiple origins:

  • Broadcaster Feeds: Many main feeds already exist as HLS or MPEG-DASH streams.

  • In-House Encodings: Multi-view experiences rely on separate transcoding pipelines. These streams include dedicated audio commentary with embedded synchronization information, delivered via either externally hosted URLs or NativeWaves-hosted endpoints.


Stream Format and Protocol Fragmentation

Each stream can differ dramatically—including:

  • DRM Encryption: Some are encrypted using CENC (common on older Smart TVs) or CBCS (increasingly used on newer devices).

  • Delivery Protocols: We work across HLS, DASH, and WebRTC (primarily for live-only, ultra-low-latency use cases).

  • Playback Preferences: DASH tends to perform better on Android, while iOS favors HLS for its native support.

This fragmented landscape creates substantial technical overhead in terms of playback delay, buffer management, segment alignment, and error recovery. Yet, to the viewer, all content—regardless of origin or format—must appear perfectly synchronized and seamless. To bridge this gap, we’ve developed a robust ingestion and playback coordination layer that abstracts away these differences and delivers a unified experience across platforms.


From Feeds to Stories: The Entity System

While synchronization and stream management are foundational, the true power of our platform comes from a higher-level abstraction: the Entity System. This is where user intent meets editorial context. In traditional streaming, fans select video feeds manually—often labeled “Cam 1”, “Cam 2”, etc. But in today’s interactive ecosystems, users expect more meaningful control. They want to follow a driver, watch from a player’s perspective, or view the game through a favorite streamer’s eyes.

Making this work requires a semantic model that understands people, roles, zones, and moments—not just camera IDs.


Dynamic Mapping of Content to Entities

The Entity System acts as a dynamic mapping layer between:

  • Content Tracks & Objects (Streams): Individual camera feeds, each with its own technical attributes.

  • User-Facing Entities: Players, coaches, commentators, or any semantic concept a fan might want to follow.

Figure 2: Entity‐to‐Track Mapping. Top rows show Entity 1 and Entity 2. Below, three video tracks display colored segments indicating which entity is mapped to each track over time.

Control information can come from an independent data stream—for example, metadata about live game events, player positions, or editorial tags. These metadata feeds power intelligent automation at the client level, enabling low-latency switching decisions based on contextual game information.

Example: Football “Goalkeeper Cam”
  1. Suppose the user selects “Goalkeeper Cam” as their preferred view.

  2. Behind the scenes, a parallel data stream tracks the ball’s position on the field.

  3. If the ball approaches the left goal, the system switches to the left goal cam. If it shifts to the right, it transitions to the right goal cam—or defaults to a center cam when play is in midfield.

  4. All this can happen dynamically and client-side, reducing complexity on the server and delivering seamless responsiveness.

This kind of intelligent mapping, driven by event data, not only enhances the user experience but also opens up a broad range of personalized and automated storytelling modes. Because the video stream continues to arrive with minimal latency, the accompanying data empowers the system to orchestrate narrative logic—effortlessly and in real time.


Building for the Future

All of these systems—time synchronization engine, adaptive switching logic, and our entity system—form the invisible foundation of the NativeWaves experience. They are building blocks for more advanced capabilities:

  • Automated Highlight Reels: Because streams are precisely aligned, we can generate highlight reels that pull from multiple angles, creating rich, contextual moments without manual editing.

  • Cross-Device Synchronization: A TV can show the main feed while a smartphone displays a tactical view, both perfectly in sync.

  • Real-Time Co-Viewing: Friends or streamers can watch together, each controlling their own perspective but staying locked to the same timeline.

Most importantly, this architecture gives fans control—not just over what they see, but how they experience it. It also gives broadcasters the confidence that even the most complex interactions will be handled with precision, resilience, and quality.


Conclusion: Architecture as Experience

Delivering a seamless, real-time multi-camera experience isn’t just a technical win. It’s a strategic enabler. It:

  1. Unlocks Monetization – through dynamic ad placements and premium views.

  2. Deepens Fan Engagement – by allowing personalized storytelling.

  3. Differentiates Platforms – in an increasingly competitive streaming landscape.

At NativeWaves, we believe the future of sports viewing isn’t about broadcasting a single story—it’s about giving each fan the tools to shape their own. And the only way to do that at scale, across devices, networks, and sports, is with an architecture that treats timing, content, and context as first-class citizens.

This is how we move from passive watching to immersive participation—from feeds to experiences. And we’re just getting started.

Let's get in touch

Christof Haslauer

CEO & Founder, NativeWaves

"Our goal is simple: to be your first choice to deliver personalised, engaging live experiences."

Let's get in touch

Christof Haslauer

CEO & Founder, NativeWaves

"Our goal is simple: to be your first choice to deliver personalised, engaging live experiences."

Let's get in touch

Christof Haslauer

CEO & Founder, NativeWaves

"Our goal is simple: to be your first choice to deliver personalized, engaging live experiences."

Let's get in touch

Christof Haslauer

CEO & Founder, NativeWaves

"Our goal is simple: to be your first choice to deliver personalized, engaging live experiences."

NativeWaves AG

Ludwig-Bieringer-Platz 1
5071 Himmelreich, Austria
FN: 451545m, UID: ATU71154225
contact@nativewaves.com

© NativeWaves All rights reserved 2025

NativeWaves AG

Ludwig-Bieringer-Platz 1
5071 Himmelreich, Austria
FN: 451545m, UID: ATU71154225
contact@nativewaves.com

© NativeWaves All rights reserved 2025

NativeWaves AG

Ludwig-Bieringer-Platz 1
5071 Himmelreich, Austria

FN: 451545m,
UID: ATU71154225

contact@nativewaves.com

© NativeWaves All rights reserved 2025

NativeWaves AG

Ludwig-Bieringer-Platz 1
5071 Himmelreich, Austria

FN: 451545m,
UID: ATU71154225

contact@nativewaves.com

© NativeWaves All rights reserved 2025

NativeWaves AG

Ludwig-Bieringer-Platz 1
5071 Himmelreich, Austria

FN: 451545m, UID: ATU71154225

contact@nativewaves.com

© NativeWaves All rights reserved 2025

NativeWaves AG

Ludwig-Bieringer-Platz 1
5071 Himmelreich, Austria

FN: 451545m, UID: ATU71154225

contact@nativewaves.com

© NativeWaves All rights reserved 2025