Video: Delivering Better Manifests with Effective VMAF

Measuring video quality is done daily around the world between two video assets. But what happens when you want to take the aggregate quality of a whole manifest? With VMAF being a well regarded metric, how can we use that in an automatic way to get the overview we need?

In this talk, Nick Chadwick from Mux shares the examples and scripts he’s been using to analyse videos. Starting with an example where everything is equal other than quality, he explains the difficulties in choosing the ‘better’ option when the variables are much less correlated. For instance, Nick also examines the situations where a video is clearly better, but where the benefit is outweighed by the minimal quality benefit and the disproportionately high bitrate requirement.

So with all of this complexity, it feels like comparing manifests may be a complexity too far, particularly where one manifest has 5 renditions, the other only 4. The question being, how do you create an aggregate video quality metric and determine whether that missing rendition is a detriment or a benefit?

Before unveiling the final solution, Nick makes the point of looking at how people are going to be using the service. Depending on the demographic and the devices people tend to use for that service, you will find different consumption ratios for the various parts of the ABR ladder. For instance, some services may see very high usage on 2nd screens which, in this case, may take low-resolution video and also lot of ‘TV’ size renditions at 1080p50 or above with little in between. Similarly other services may seldom ever see the highest resolutions being used, percentage-wise. This shows us that it’s important not only to look at the quality of each rendition but how likely it is to be seen.

To bring these thoughts together into a coherent conclusion, Nick unveils an open-source analyser which takes into account not only the VMAF score and the resolution but also the likely viewership such that we can now start to compare, for a given service, the relative merits of different ABR ladders.

The talk ends with Nick answering questions on the tendency to see jumps between different resolutions – for instance if we over-optimise and only have two renditions, it would be easy to see the switch – how to compare videos of different resolutions and also on his example user data.

Watch now!
Speakers

Nick Chadwick Nick Chadwick
Software Engineer,
Mux

Video: Towards Measuring Perceptual Video Quality & Why

In the ongoing battle to find the minimum bitrate for good looking video, automation is key to achieving this quickly and cheaply. However, metrics like PSNR don’t always give the best answers meaning that eyes are still better the job than silicon.

In this talk from the Demuxed conference, Intel’s Vasavee Vijayaraghavan shows us examples of computer analysis failing to identify lowest bitrate leaving the encoder spending many megabits encoding video so that it looks imperceptibly better. Further more it’s clear that MOS – the Mean Opinion Score – which has a well defined protocol behind it continues to produce the best results, though setting up and co-ordinating takes orders of magnitude more time and money.

Vasavee shows how she’s managed to develop a hybrid workflow which combines metrics and MOS scores to get much of the benefit of computer-generated metrics fed into the manual MOS process. This allows a much more targeted subjective perceptual quality MOS process thereby speeding up the whole process but still getting that human touch where it’s most valuable.

Watch now!
Speaker

Vasavee Vijayaraghavan Vasavee Vijayaraghavan
Cloud Media Solutions Architect,
Intel

Video: An introduction to Biological Compression

The search for better codecs is ever lasting so it’s no surprise that with AI’s recent advances, we see a codec based on AI/machine learning. The AI approach not only frees up the maths from, say, upscaling using a fixed algorithm to doing it however it sees fit, but also gives it a holistic view of the image.

Considering the image as a whole whilst encoding it allows the encoder to better apportion bitrate and detail to the needed areas whereas other codecs have trouble breaking out of the procedural ‘one block at a time’ mode which tends to treat each macro block separately.

Co-founder of Deep Render Aim Christian Besenbruch gives us examples of his company’s ‘biological compression’ codec against the latest BGP codec which is an HEVC-based still image codec which delivers smaller images than JPEG.

Watch now!
Speaker

Christian Besenbruch Christian Besenbruch
Co-Founder,
Deep Render AI

Video: Quantitative Evaluation and Attribute of Overall Brightness in a HDR World

HDR has long being heralded as a highly compelling and effective technology as high dynamic range can improve video of any resolution and much better mimics the natural world. HDR continues its relatively slow growth into real-world use, but continues to show progress.

HDR is so compelling because it can feed our senses more light and it’s no secret that TV shops know we like nice, bright pictures on our TV sets. But the reality of production in HDR is that you have to contend with human eyes which have a great ability to see dark and bright images – but not at the same time. The total ability of the eye to simultaneously distinguish brightness is about 12 stops, which is only two thirds of its non-simultaneous total range.
 

 
The fact that our eyes constantly adapt and, let’s face it, interpret what they see, makes understanding brightness in videos tricky. There are dependencies on overall brightness of a picture at any one moment, the previous recent brightness, the brightness of local adjacent parts of the image, the ambient background and much more to consider.

Selios Ploumis steps into this world of varying brightness to creat a ways of quantitatively evaluating brightness for HDR. The starting place is the Average Picture Level (APL) which is what the SDR world uses to indicate brightness. With the greater dynamic range in HDR and the way this is implemented, it’s not clear that APL is up to the job.

Stelios explains his work in analysing APL in SDR and HDR and shows the times that simply taking the average of a picture can trick you into seeing two images as practically the same, whereas the brain clearly sees one as more ‘bright’ than the other. On the same track, he also explains ways in which we can work to differentiate signals better, for instance taking in to account the spread of the brightness values as opposed to APL’s normalised average of all pixels’ values.

The talk wraps up with a description of how the testing was carried out and a summary of the proposals to improve the quantitive analysis of HDR video.

Watch now!
Speakers

Stelios Ploumis Stelios Ploumis
PhD Research Candidate
MTT Innovation Inc.