Video: FOX – Uncompressed live sports in the cloud

Is using uncompressed video in the cloud with just 6 frames of latency to get there and back ready for production? WebRTC manages sub-second streaming in one direction and can even deliver AV1 in real-time. The key to getting down to a 100ms round trip is to move down to millisecond encoding and to use uncompressed video in the cloud. This video shows how it can be done.

Fox has a clear direction to move into the cloud and last year joined AWS to explains how they’ve put their delivery distribution into the cloud remuxing feeds for ATSC transmitters, satellite uplinks, cable headends and encoding for internet delivery, In this video, Fox’s Joel Williams, Evan Statton from AWS explain their work together making this a reality. Joel explains that latency is not a very hot topic for distribution as there are many distribution delays. The focus has been on getting the contribution feeds into playout and MCR monitoring quickly. After all, when people are counting down to an ad break, it needs to roll exactly on zero.

Evan explains the approach AWS has taken to solving this latency problem and starts with considering using SMPTE’s ST 2110 in the cloud. ST 2110 has video flows of at least 1 Gbps, typically and when implemented on-premise is typically built on a dedicated network with very strict timing. Cloud datacentres aren’t like that and Evan demonstrates this showing how across 8 video streams, there are video drops of several seconds which is clearly not acceptable. Amazon, however, has a product called ‘Scalable Reliable Datagram’ which is aimed at moving high bitrate data through their cloud. Using a very small retransmission buffer, it’s able to use multiple paths across the network to deliver uncompressed video in real-time. The retransmission buffer here being very small enables just enough healing to redeliver missing packets within the 16.7ms it takes to deliver a frame of 60fps video.

On top of SRD, AWS have introduced CDI, the Cloud Digital Interface, which is able to describe uncompressed video flows in a way already familiar to software developers. This ‘Audio Video Metadata’ layer handles flows in the same way as 2110, for instance keeping essences separate. Evan says this has helped vendors react favourably to this new technology. For them instead of using UDP, SRD can be used with CDI giving them not only normal video data structures but since SRD is implemented in the Nitro network card, packet processing is hidden from the application itself.

The final piece to the puzzle is keeping the journey into and out of the cloud low-latency. This is done using JPEG XS which has an encoding time of a few milliseconds. Rather than using RIST, for instance, to protect this on the way into the cloud, Fox is testing using ST 2022-7. 2022-7 takes in two identical streams on two network interfaces, typically. This way it should end up with two copies of each packet. Where one gets lost, there is still another available. This gives path redundancy which a single stream will never be able to offer. Overall, the test with Fox’s Arizona-based Technology Center is shown in the video to have only 6 frames of latency for the return trip. Assuming they used a California-based AWS data centre, the ping time may have been as low as two frames. This leaves four frames for 2022-7 buffers, XS encoding and uncompressed processing in the cloud.

Watch now!
Speakers

Joel Williams Joel Wiliams
VP of Architecutre & Engineering,
Fox Corporation
Evan Statton Evan Statton
Principal Architect, Media & Entertainment,
AWS

Where can SMPTE 2110 and NDI co-exist?

Our final look back at the most viewed articles of 2020 is a very prescient topic, that of live IP production. As we all know, this has very much come into focus during the pandemic. Those that already had an IP infrastructure found managing it remotely easier than those that needed to get in and move SDI cables when they needed to provision new video links. Moreover putting together live remote workflows is all the easier with video IP and traditionally the decision on whether to use SMPTE 2110 has been a case of whether you need to be in the cloud or not.

This article and video brought together Will Waters, an NDI expert from VizRT, Marc Risby from UK SI Boxer and Willem Vermost who was with the EBU. The conversation, pre-pandemic, focused on how to choose the video-over-IP technology which was best for you and really tried to find the ways in which broadcasters could benefit from both at the same time.

The Broadcast Knowledge Editor, Russell Trafford-Jones also moderated a SMPTE Webcast with VizRT going into the detail of NDI and how it can be deployed in the cloud.

Another important advance in 2020 was AWS’s release of CDI which is an implementation of SMPTE 2110 with enough proprietary adaptations to make it work within AWS. You can hear more about it in this video with David Griggs.

Click here to watch ‘Where can SMPTE 2110 and NDI co-exist?’

Video: Broadcast Playout Cloud Transformation

Playout has been gradually moving to the cloud for a number of years now. Famously Discovery moved all of their thematic playout to the cloud in 2018 and many have done the same since. As we saw the other day, Sky Italia are now seeing ‘code as infrastructure’ whereby automated API calls launch the in-cloud infrastructure they need as part of their linear playout.

In this video, we hear from Matt Westrup from A+E EMEA on how they’ve moved their playout to the cloud with their partner Amagi. Running 30 channels in Europe, Matt explains that due to some business uncertainty with a partner company, the need for a DR facility was identified. Talking to Srinivasan KA from Amagi, they were able to create this using Amagi’s product portfolio based in AWS. Matt explains that after the DR facility was set up, they moved quickly to full mirroring and ultimately they flipped the switch and they announced they were now broadcasting from the cloud.

Srinivasan KA explains that many companies take a similar route when working in the cloud. Sometimes a cost-effective DR facility is all they need, however it’s easy to replicate all your workflows in the cloud and have that on standby. This can be done by keeping the content in the cloud evergreen, running automation but keeping the playout functions switched off to save money which can be quickly brought online as needed. Srinivasan KA looks at the high-level diagram of the A+E operation showing how S3 holds the content, goes through a workflow to the CPU-powered playout and then is handed off using direct connect to affiliates and telcos using Amagi’s POPs.

Matt comments that this was relatively easy to do from a business perspective “No-one was investing massively in fixed infrastructure” and they’ve found they have been faster to market with a speed they’ve “never experienced before.” Needless to say, the move to the cloud also came into its own and provided a seamless move work home working during the pandemic. And, looking more longterm, A+E will continue to benefit from not having to manage the physical datacentre/serber room infrastructure.

The video finishe swith an overview of Broadcast in AWS from Andy Kane. He covers the main drivers for broadcasters moving to the cloud such as business agility, a preference with some companies for increasing Opex spending, increased ease in experiementing with new technologies/ways of engaging with customers, using a remote workforce among others. Andy covers an example broadcast flow using MediaConnect for contribution, MediaLive Statmux for distibution, redundancy strategies and other building blocks such as TAG multiviewers.

Watch now!
From the AWS Media Insights Webcast Series
Speakers

Andy Kane Andy Kane
Principal AI/ML Specialist Solutions Architect (Languages),
Amazon Web Services (AWS)
Matt Westrup Matt Westrup
VP Technology and Operations,
A+E EMEA
Srinivasan KA Srinivasan KA
Co-founder,
Amagi Corporation
Ian McPherson Ian McPherson
Partner Development Lead – Media & Entertainment,
Amazon Web Services (AWS)

Video: Live Production Forecast: Cloudy for the Foreseeable Future

Our ability to work remotely during the pandemic is thanks to the hard work of many people who have developed the technologies which have made it possible. Even before the pandemic struck, this work was still on-going and gaining momentum to overcome more challenges and more hurdles of working in IP both within the broadcast facility and in the cloud.

SMPTE’s Paul Briscoe moderates the discussion surrounding these on-going efforts to make the cloud a better place for broadcasters in this series of presentation from the SMPTE Toronto section. First in the order is Peter Wharton from TAG V.S. talking about ways to innovate workflows to better suit the cloud.

Peter first outlines the challenges of live cloud production, namely keeping latency low, signal quality high and managing the high bandwidths needed at the same time as keeping a handle on the costs. There is an increasing number of cloud-native solutions but how many are truly innovating? Don’t just move workflows into the cloud, advocates Peter, rather take this opportunity to embrace the cloud.

Working with the cloud will be built on new transport interfaces like RIST and SRT using a modular and open architecture. Scalability is the name of the game for ‘the cloud’ but the real trick is in building your workflows and technology so that you can scale during a live event.

Source: TAG V.S.

There are still obstacles to be overcome. Bandwidth for uncompressed video is one, with typical signals up to 3Gbps uncompressed which then drives very high data transfer costs. The lack of PTP in the cloud makes ST 2110 workflows difficult, similarly the lack of multicast.

Tackling bandwidth, Peter looks at the low-latency ways to compress video such as NDI, NDI|HX, JPEG XS and Amazon’s lossless CDI. Peter talks us through some of the considerations in choosing the right codec for the task in hand.

Finishing his talk, Peter asks if this isn’t time for a radical change. Why not rethink the entire process and embrace latency? Peter gives an example of a colour grading workflow which has been able to switch from on-prem colour grading on very high-spec computers to running this same, incredibly intensive process in the cloud. The company’s able to spin up thousands of CPUs in the cloud and use spot pricing to create temporary, low cost, extremely powerful computers. This has brought waiting times down for jobs to be processed significantly and has reduced the cost of processing an order of magnitude.

Lastly Peter looks further to the future examining how saturating the stadium with cameras could change the way we operate cameras. With 360-degree coverage of the stadium, the position of the camera can be changed virtually by AI allowing camera operators to be remote from the stadium. There is already work to develop this from Canon and Intel. Whilst this may not be able to replace all camera operators, sports is the home of bleeding-edge technology. How long can it resist the technology to create any camera angle?

Source: intoPIX

Jean-Baptiste Lorent is next from intoPIX to explain what JPEG XS is. A new, ultra-low-latency, codec it meets the challenges of the industry’s move to IP, its increasing desire to move data rather than people and the continuing trend of COTS servers and cloud infrastructure to be part of the real-time production chain.

As Peter covered, uncompressed data rates are very high. The Tokyo Olympics will be filmed in 8K which racks up close to 80Gbps for 120fps footage. So with JPEG XS standing for Xtra Small and Xtra Speed, it’s no surprise that this new ISO standard is being leant on to help.

Tested as visually lossless to 7 or more encode generations and with latency only a few lines of video, JPEG XS works well in multi-stage live workflows. Jean-Baptiste explains that it’s low complexity and can work well on FPGAs and on CPUs.

JPEG XS can support up to 16-bit values, any chroma and any colour space. It’s been standardised to be carried in MPEG TSes, in SMPTE ST 2110 as 2110-22, over RTP (pending) within HEIF file containers and more. Worst case bitrates are 200Mbps for 1080i, 390Mbps for 1080p60 and 1.4Gbps for 2160p60.

Evolution of Standards-Based IP Workflows Ground-To-Cloud

Last in the presentations is John Mailhot from Imagine Communications and also co-chair of an activity group at the VSF working on standardising interfaces for passing media place to place. Within the data plane, it would be better to avoid vendors repeatedly writing similar drivers. Between ground and cloud, how do we standardise video arriving and the data you need around that. Similarly standardising new technologies like Amazon’s CDI is important.

John outlines the aim of having an interoperability point within the cloud above the low-level data transfer, closer to 7 than to 1 in the OSI model. This work is being done within AIMS, VSF, SMPTE and other organisations based on existing technologies.

Q&A
The video finishes with a Q&A and includes comments from AWS’s Evan Statton whose talk on CDI that evening is not part of this video. The questions cover comparing NDI with JPEG XS, how CDI uses networking to achieve high bandwidths and high reliability, the balance between minimising network and minimising CPU depending on workflow, the increasingly agile nature of broadcast infrastructure, the need for PTP in the cloud plus the pros and cons of standards versus specifications.

Watch now!
Speakers

Peter Wharton Peter Wharton
Director Corporate Strategy, TAG V.S.
President, Happy Robotz
Vice President of Membership, SMPTE
Jean-Baptiste Lorent Jean-Baptiste Lorent
Director Marketing & Sales,
intoPIX
John Mailhot John Mailhot
Co-Chair Cloud-Gounrd-Cloud-Ground Activity Group, VSF
Directory & NMOS Steering Member, AMWA
Systems Architect for IP Convergence, Imagine Communcations
Paul Briscoe Moderator: Paul Briscoe
Canadian Regional Governor, SMPTE
Consultant, Televisionary Consulting
Evan Statton Evan Statton
Principal Architect, Media & Entertainment
Amazon Web Services