From event to event it’s not a surprise that streaming traffic increases, but this look at the Wolrd Cup 2018 shows a very sharp rise beating many expecatations. Joachim Hengge tells us what hte World Cup looked like from Akamai’s perspective.
Joachim takes us through the stats for streaming the World Cup where they peaked at 23Tbps of throuhgput with nearly 10 million concurrent viewers. The bandwidth was significantly higher than the last World Cup but looking at the data, we can learn a few more things about the market.
After looking at a macth-by-match breakdown we look at a sytsem architecture for one customer who delivered the World Cup to highlight the importance of stable content ingest, latency and broadcast quality. Encoding and packaging into HLS with 4-second chunks were tasks done on site with the rest happening within Akamai and being fed to other CDNs. Joachim pulls this together into three key recommendations for anyone looking at streaming large events before delvingin to some Sweden-specific streaming stats where over 81% of feeds were played back at the highest quality.
This talk is from Streaming Tech Sweden, an annual conference run by Eyevinn Technology. Videos from the event are available to paid attendees but are released free of charge after several months. As with all videos on The Broadcast Knowledge, this is available free of charge after registering on the site.
Senior Product Manager, Media Services,
Discover the critical success factors the Broadcasters and platform owners, investing millions in building and upgrading OTT platforms, need to achieve to ensure they can compete successfully with a growing array of digital competitors and deliver compelling user experiences.
Many of these broadcasters are beginning to move from their initial OTT offerings to more mature services that can scale for the future, and answer the requirements of demanding viewers and regulators.
This webinar uncovers the essential parts of a flourishing OTT service, including:
– Delivering content at scale as more viewing and live events move to OTT
– Ensuring a class-leading user experience and quality
– Using analytics to maximise revenue and engagement
– Ensuring cost efficiency in the OTT workflow
– Securing platforms and content against piracy and malicious attacks
Low latency streaming is always a compromise, but what can be done to keep QOE high?
This on-demand webinar looks at CMAF and presents some real-world data on this low latency technique. The webinar starts by explaining that CMAF is a low-latency streaming technology similar to HLS and other streaming protocols where the idea is to deliver the video as small files. Olivier and Alain from Harmonic explain how this is done and look at some of the trade-offs and compromises that are needed and introduce techniques to keep QOE high. They also look at deployment in cloud vs. on premise.
Pieter-Jan Speelmans talks about play tradeoffs and optimisations within the player. CMAF allows the buffer to be reduced and whilst a bad network may mean you buffer is similar to ‘normal’, but in good networks, this buffer can be brought down significantly. He also talks about how ABR switching is impacted by GOP length even in CMAF.
Viaccess-Orca explains how DRM works with CMAF and looks at some of the challenges including licences acquisition time and overloading licence servers at the beginning of events. Akamai’s Will Law explains some benefits of CMAF and the near-real-time of chunk-based transfer (HTTP 1.1) and how downloading chunks at full speed leads to problems when the same broadband link is used by several clients.
There are lots of good talks on CMAF, but this is one of the few which talks about CMAF not as theory, but as is deployable today.
There are many ways to speed up live streaming and much work has gone in to reducing chunk lengths for HLS-style streaming, WebRTC has arrived on the scene and techniques to speed up chunk delivery are in production in CDNs around the world.
But we shouldn’t forget lower down in the detail, we have how the web sites are actually saved to customers – the venerable HTTP. Running on TCP/IP, HTTP packets are delivered using very thorough acknowledgement mechanisms within TCP/IP. Furthermore, it’s immune to spoofing attacks due to a three way handshake to set up the connection.
However, all this communication ads latency as even for low latency connections, these communications can add up to a significant latency and affect the speed of the throughout of the connection.
This talk introduces QUIC which is a replacement for HTTP developed by Google which uses UDP as its underlying delivery mechanism, thus avoiding much of this built-in two way comms.
At the Mile High Video event, Miroslav Ponec from Akamai introduces this protocol which is undergoing standardisation at the IETF explaining how it works and why it’s such a good idea.