Versatile Video Coding: Compression Tools for UHD and 360° VideoMathias Wien
The document discusses the development of the Versatile Video Coding (VVC) standard. It describes how a call for proposals was issued to develop coding tools beyond HEVC. 46 proposals were submitted across standard dynamic range, high dynamic range, and 360-degree video categories. The proposals were evaluated through subjective testing and shown to provide over 40% bitrate reduction compared to HEVC and over 10% reduction compared to the Joint Exploration Model, with the best proposals demonstrating visual quality equal or better than HEVC at higher bitrates. Seven proposals were identified as significantly better than the Joint Exploration Model. This marked the starting point for developing the VVC standard based on the selected coding tools from the top-performing proposals
Trends and Recent Developments in Video Coding StandardizationMathias Wien
This document summarizes a tutorial on trends and recent developments in video coding standardization. It discusses the history of video coding standards organizations and the standards they have developed. These include MPEG-1, H.261, H.262, H.264, H.265 and the upcoming H.266 Versatile Video Coding standard. The document outlines the tutorial, which will cover topics like video resolutions, current compression techniques, VVC, and future trends in areas like multi-camera coding.
ICME 2016 - High Efficiency Video Coding - Coding Tools and Specification: HE...Mathias Wien
The tutorial covers the complete HEVC standard, including all currently defined extensions (range extensions, scalability, multi-view, 3D video coding, and screen content coding). It further covers the state of the current activities on Free-Viewpoint Television and on High Dynamic Range + Wide Color Gamut Coding. The standard is assessed from various perspectives, including an algorithmic view on the video coding layer as well as a high-level / system-layer view on the network abstraction layer and the overall structure. The discussion includes a detailed treatment of the HEVC layer concept which allows for seamless incorporation of spatial and quality scalability as well as multi-view, 3D, or FTV extensions. The essential concepts and the coding tools comprised in each of the extensions are detailed and explained in the context of their respective application space. The tutorial further discusses the basic structure of specification text from a more abstract point of view as well as by concrete example in HEVC. For all mentioned perspectives, the tutorial develops the topic in a step-by-step fashion and gradually introduces concepts, algorithms, and terminology. Examples are provided at all levels of the presentation illustrating the concepts and deepening the understanding of the presented technology. Various demos are presented to visualize the algorithmic advancement. The tutorial is based on the book “High Efficiency Video Coding: Coding Tools and Specification” by the tutorial speaker which currently covers HEVC version 1. The tutorial shall enable the participants to understand the design principles and concepts behind the specification of HEVC. They shall recognize and understand the innovation of HEVC compared to the previous standards (esp. H.264/AVC) and regard the extensible nature of the specification design.
In October 2017, ISO/IEC JCT1 SC29/WG11 MPEG and ITU-T SG16/Q6 VCEG have jointly published a Call for Proposals on Video Compression with Capability beyond HEVC and its current extensions. It is targeting at a new generation of video compression technology that has substantially higher compression capability than the existing HEVC standard. The responses to the call are evaluated in April 2018, forming the kick-off for a new standardization activity in the Joint Video Experts Team (JVET) of VCEG and MPEG, with a target of finalization by the end of the year 2020. Three categories of video are addressed: Standard dynamic range video (SDR), high dynamic range video (HDR), and 360° video. While SDR and HDR cover variants of conventional video to be displayed e.g. on a suitable TV screen at very high resolution (UHD), the 360° category targets at videos capturing a full-degree surround view of the scene. This enables an immersive video experience with the possibility to look around in the rendered scene, e.g. when viewed using a head-mounted display. This application triggers various technical challenges which need to be addressed in terms of compression, encoding, transport, and rendering. The talk summarizes the current state of the complete standardization project. Focussing on the SDR and 360° video categories, it highlights the development of selected coding tools compared to the state of the art. Representative examples of the new technological challenges as well as corresponding proposed solutions are presented.
Perceptual evaluation of Immersive Media – From video quality towards a holi...Alpen-Adria-Universität
The document discusses perceptual evaluation and quality of experience (QoE) for immersive media such as 360-degree video and interactive virtual environments. It presents several models and frameworks for evaluating various aspects of immersive media quality, including video quality prediction, suitability of content, 360-degree video viewing behavior, and a holistic QoE evaluation framework. The document also describes several open datasets and reference implementations for models related to immersive media quality evaluation.
Versatile Video Coding: Compression Tools for UHD and 360° VideoMathias Wien
The document discusses the development of the Versatile Video Coding (VVC) standard. It describes how a call for proposals was issued to develop coding tools beyond HEVC. 46 proposals were submitted across standard dynamic range, high dynamic range, and 360-degree video categories. The proposals were evaluated through subjective testing and shown to provide over 40% bitrate reduction compared to HEVC and over 10% reduction compared to the Joint Exploration Model, with the best proposals demonstrating visual quality equal or better than HEVC at higher bitrates. Seven proposals were identified as significantly better than the Joint Exploration Model. This marked the starting point for developing the VVC standard based on the selected coding tools from the top-performing proposals
Trends and Recent Developments in Video Coding StandardizationMathias Wien
This document summarizes a tutorial on trends and recent developments in video coding standardization. It discusses the history of video coding standards organizations and the standards they have developed. These include MPEG-1, H.261, H.262, H.264, H.265 and the upcoming H.266 Versatile Video Coding standard. The document outlines the tutorial, which will cover topics like video resolutions, current compression techniques, VVC, and future trends in areas like multi-camera coding.
ICME 2016 - High Efficiency Video Coding - Coding Tools and Specification: HE...Mathias Wien
The tutorial covers the complete HEVC standard, including all currently defined extensions (range extensions, scalability, multi-view, 3D video coding, and screen content coding). It further covers the state of the current activities on Free-Viewpoint Television and on High Dynamic Range + Wide Color Gamut Coding. The standard is assessed from various perspectives, including an algorithmic view on the video coding layer as well as a high-level / system-layer view on the network abstraction layer and the overall structure. The discussion includes a detailed treatment of the HEVC layer concept which allows for seamless incorporation of spatial and quality scalability as well as multi-view, 3D, or FTV extensions. The essential concepts and the coding tools comprised in each of the extensions are detailed and explained in the context of their respective application space. The tutorial further discusses the basic structure of specification text from a more abstract point of view as well as by concrete example in HEVC. For all mentioned perspectives, the tutorial develops the topic in a step-by-step fashion and gradually introduces concepts, algorithms, and terminology. Examples are provided at all levels of the presentation illustrating the concepts and deepening the understanding of the presented technology. Various demos are presented to visualize the algorithmic advancement. The tutorial is based on the book “High Efficiency Video Coding: Coding Tools and Specification” by the tutorial speaker which currently covers HEVC version 1. The tutorial shall enable the participants to understand the design principles and concepts behind the specification of HEVC. They shall recognize and understand the innovation of HEVC compared to the previous standards (esp. H.264/AVC) and regard the extensible nature of the specification design.
In October 2017, ISO/IEC JCT1 SC29/WG11 MPEG and ITU-T SG16/Q6 VCEG have jointly published a Call for Proposals on Video Compression with Capability beyond HEVC and its current extensions. It is targeting at a new generation of video compression technology that has substantially higher compression capability than the existing HEVC standard. The responses to the call are evaluated in April 2018, forming the kick-off for a new standardization activity in the Joint Video Experts Team (JVET) of VCEG and MPEG, with a target of finalization by the end of the year 2020. Three categories of video are addressed: Standard dynamic range video (SDR), high dynamic range video (HDR), and 360° video. While SDR and HDR cover variants of conventional video to be displayed e.g. on a suitable TV screen at very high resolution (UHD), the 360° category targets at videos capturing a full-degree surround view of the scene. This enables an immersive video experience with the possibility to look around in the rendered scene, e.g. when viewed using a head-mounted display. This application triggers various technical challenges which need to be addressed in terms of compression, encoding, transport, and rendering. The talk summarizes the current state of the complete standardization project. Focussing on the SDR and 360° video categories, it highlights the development of selected coding tools compared to the state of the art. Representative examples of the new technological challenges as well as corresponding proposed solutions are presented.
Perceptual evaluation of Immersive Media – From video quality towards a holi...Alpen-Adria-Universität
The document discusses perceptual evaluation and quality of experience (QoE) for immersive media such as 360-degree video and interactive virtual environments. It presents several models and frameworks for evaluating various aspects of immersive media quality, including video quality prediction, suitability of content, 360-degree video viewing behavior, and a holistic QoE evaluation framework. The document also describes several open datasets and reference implementations for models related to immersive media quality evaluation.
Comparison of compression efficiency between HEVC and VP9 based on subjective...Touradj Ebrahimi
These are the slides of my presentation at SPIE Optics + Photonics 2014 Applications of Digital Image Processing XXXVII. The paper itself can be downloaded from SPIE Digital Library. For people in hurry, a pre-print version is available at: http://infoscience.epfl.ch/record/200925?ln=en
In this contribution, we present selected novel approaches and results of our research work in the \ATHENA Christian Doppler Laboratory (Adaptive Streaming over HTTP and Emerging Networked Multimedia Services), a major research project at our department jointly funded by public sources and industry. By putting this work also into the context of related ongoing research activities, we aim at working out where HTTP Adaptive Streaming is currently heading.
Chris Varekamp (Philips Group Innovation, Research): Depth estimation, Proces...AugmentedWorldExpo
A talk from the Develop Track at AWE USA 2018 - the World's #1 XR Conference & Expo in Santa Clara, California May 30- June 1, 2018.
Chris Varekamp (Philips Group Innovation, Research): Depth estimation, Processing & Rendering for Dynamic 6DoF VR
In this talk I will discuss how a real-time depth-based processing chain can be built using our experience in stereo-to-depth conversion for autostereoscopic displays.
http://AugmentedWorldExpo.com
The document provides an overview of the High Efficiency Video Coding (HEVC) standard. Some key points:
- HEVC was created as a new video compression standard to address the growing needs of higher resolution video content and more efficient compression compared to prior standards like H.264.
- It achieves 50% bitrate reduction over H.264 for the same visual quality or improved quality at the same bitrate.
- The standard uses a block-based coding structure with coding tree units and supports intra-frame and inter-frame coding with motion estimation/compensation.
- It introduces more intra-prediction modes and block sizes along with improved transforms, quantization, and entropy coding.
Bitmovin AV1/VVC Presentation_Streaming Media East by Christian FeldmannBitmovin Inc
This document provides an overview and update on AV1 and VVC video coding standards. It summarizes the novel technical features of each, including AV1 improvements like overlapped block motion compensation and affine motion, and VVC features like triangle partitioning and decoder-side motion refinement. Performance results show VVC provides around 30% better compression than HM 16.20, while AV1 is around 20% better. Both standards are still in development and neither has wide adoption yet. VVC licensing is unknown while AV1 remains free and open.
Tutorial on Point Cloud Compression and standardisationRufael Mekuria
Tutorial on Point Cloud Compression and standardisation given at IEEE VCIP 2017 in december. I provide the techniques for point cloud compression and the designed quality metrics and codecs in my PhD at CWI. I detail the standardisation activity on point cloud compression that I started in 2014 and that started in 2017 involving all mobile device makers like Apple, Huawei, Sony, Samsung and Nokia.
Excerpts from the HEVC / H265 Hands-on course.
This parts of the course explains how to download the reference code (HM) compile it configure it and analyze the video output
MPEG Immersive Media
By Thomas, Director, Technical Standards at Qualcomm
at 2nd ITU-T Mini-Workshop on Immersive Live Experience (ILE) in 19 January 2017
This document discusses various methods of image resampling, including downsampling, upsampling, and fractional resampling. It describes traditional blind resampling methods like nearest neighbor, bilinear, bicubic, and Lanczos interpolation. It also covers content-aware resampling techniques like seam carving and new edge-directed interpolation (NEDI) that aim to minimize information loss. For each method, it explains the approach, artifacts, filter kernels, and provides visual examples comparing the output images. The document concludes by discussing tradeoffs between methods and opportunities for future improvement in content-aware resizing.
The document discusses the concept of the Tactile Internet with Human-in-the-Loop. It aims to democratize access to skills and expertise for people of all backgrounds and abilities. This goes beyond the current Internet's goal of providing access to information regardless of location or time. The document outlines a vision for two-way skills transfer between humans and machines using multimodal feedback over 5G networks. It discusses challenges like differing neural time delays for multisensory perception and individual differences in processing that affect perception and action. The Center for Tactile Internet's research agenda involves understanding multisensory goal-directed processing neurocognitively, modeling perception and action, and expertise in related fields to advance human-technology interactions
The document provides an overview of the High Efficiency Video Coding (HEVC) standard. It was developed jointly by ISO/IEC and ITU-T to provide roughly half the bit-rate of H.264/AVC for the same subjective quality. Key aspects of HEVC include use of larger block sizes, intra-picture prediction with 33 directional modes, motion vectors with quarter-sample precision, transform sizes from 4x4 to 32x32, adaptive coefficient scanning, in-loop filtering including deblocking and sample adaptive offset, and support for lossless and transform skipping modes. Many companies are starting to support HEVC in their video products and services.
The document discusses the new HEVC/H.265 video compression standard and its benefits for ultra high definition video. Key points:
- HEVC is 50% more efficient than H.264/MPEG-4 AVC, allowing a 50% reduction in bandwidth. It can support resolutions up to 8K.
- Tests show HEVC achieves 50-75% lower bitrates than H.264 for ultra high definition video, while maintaining comparable quality.
- HEVC's increased efficiency comes from processing video in 64x64 pixel blocks rather than 16x16, and parallel processing of video frames. This requires powerful multi-core processors.
- The improved compression enables
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...Ijripublishers Ijri
This document discusses techniques for improving video compression efficiency for surveillance videos. It proposes modifying the architecture of scalable video coding to make it surveillance-centric by allowing adaptive rate-distortion optimization at the GOP level based on whether events of interest are present. Experimental results show foreground detection and updating of background adaptively over time to improve compression. Future work includes further enhancing selective motion estimation techniques to improve processing efficiency without degrading video quality.
This document provides an overview and comparison of the H.264 and HEVC video coding standards. It describes the key features and innovations that allow each standard to compress video more efficiently than previous standards. H.264 introduced features like adaptive block sizes, multi-frame prediction, quarter-pixel motion compensation and loop filtering that improved compression performance over prior standards. HEVC aims to further increase compression efficiency through innovations such as larger coding tree blocks, additional intra-prediction modes, and improved entropy coding. The document analyzes these standards to understand how their new coding tools enable significantly higher compression ratios and support for new applications like higher resolution video.
Quality Optimization of Live Streaming Services over HTTP with Reinforcement ...Alpen-Adria-Universität
Recent years have seen tremendous growth in HTTP adaptive live video traffic over the Internet. In the presence of highly dynamic network conditions and diverse request patterns, existing yet simple hand-crafted heuristic approaches for serving client requests at the network edge might incur a large overhead and significant increase in time complexity. Therefore, these approaches might fail in delivering acceptable Quality of Experience (QoE) to end users. To bridge this gap, we propose ROPL, a learning-based client request management solution at the edge that leverages the power of the recent breakthroughs in deep reinforcement learning, to serve requests of concurrent users joining various HTTP-based live video channels. ROPL is able to react quickly to any changes in the environment, performing accurate decisions to serve clients requests, which results in achieving satisfactory user QoE. We validate the efficiency of ROPL through trace-driven simulations and a real-world setup. Experimental results from real-world scenarios confirm that ROPL outperforms existing heuristic-based approaches in terms of QoE, with a factor up to 3.7×.
This document summarizes the outcomes of the ITU-T Study Group 16 meeting held from 16-27 January 2017 in Geneva. It discusses the number of contributions reviewed and recommendations approved. It provides details on the collaborative work done with other groups on topics like video coding, IPTV, accessibility, e-health, and more. Major accomplishments included completing new recommendations on scalable vector graphics, speech translation services, and requirements for areas like vehicle gateways and content delivery networks. Work also progressed on developing standards for immersive media, visual surveillance, and blockchain technologies.
This document provides an overview of HEVC (High Efficiency Video Coding) including:
- HEVC aims to provide roughly half the bitrate of H.264/AVC at the same quality.
- It uses block-based hybrid video coding with improved intra-prediction, transform, quantization and entropy coding techniques.
- HEVC supports a wide range of resolutions, color spaces and bit depths for 4K and beyond.
Universal media access as proposed almost two decades ago is now reality. We can generate, distribute, share, and consume any media content, anywhere, anytime, and with/on any device. A technical breakthrough was the adaptive streaming over HTTP resulting in the standardization of MPEG-DASH, which is now successfully deployed in a plethora of environments. The next big thing in adaptive media streaming is virtual reality applications and, specifically, omnidirectional (360°) media streaming, which is currently built on top of the existing adaptive streaming ecosystems. This tutorial provides a detailed overview of adaptive streaming of both traditional and omnidirectional media. The tutorial focuses on the basic principles and paradigms for adaptive streaming as well as on already deployed content generation, distribution, and consumption workflows. Additionally, the tutorial provides insights into standards and emerging technologies in the adaptive streaming space. Finally, the tutorial includes the latest approaches for immersive media streaming enabling 6DoF DASH through Point Cloud Compression (PCC) and concludes with open research issues and industry efforts in this domain. More information available at: https://multimediacommunication.blogspot.com/2019/07/acmmm19-tutorial-journey-towards-fully.html
The document summarizes the H.264/MPEG4 Advanced Video Coding standard. It was developed by the Joint Video Team (JVT) which included the ITU-T Video Coding Experts Group and ISO/IEC Moving Picture Experts Group. The goals were to improve coding efficiency, enhance error robustness, simplify syntax, and increase flexibility. Key features of H.264 include enhanced motion prediction, small block size transforms, an adaptive in-loop deblocking filter, and improved entropy coding. These allow for approximately 50% bit rate savings over prior standards. Evaluation showed it can achieve near transparent quality at lower bit rates than previous standards.
Machine learning approaches are being explored for video compression. Conservative approaches replace individual MPEG blocks with deep learning blocks, while disruptive end-to-end approaches aim to replace the entire MPEG chain. Optical flow networks can exploit temporal redundancy by estimating motion between frames. Fully neural network-based video compression models jointly learn motion estimation, motion compression, and residual compression in an end-to-end optimized framework. However, performance gains must be balanced against increased complexity, and neural network approaches are not yet mature enough to be included in video compression standards.
Comparison of compression efficiency between HEVC and VP9 based on subjective...Touradj Ebrahimi
These are the slides of my presentation at SPIE Optics + Photonics 2014 Applications of Digital Image Processing XXXVII. The paper itself can be downloaded from SPIE Digital Library. For people in hurry, a pre-print version is available at: http://infoscience.epfl.ch/record/200925?ln=en
In this contribution, we present selected novel approaches and results of our research work in the \ATHENA Christian Doppler Laboratory (Adaptive Streaming over HTTP and Emerging Networked Multimedia Services), a major research project at our department jointly funded by public sources and industry. By putting this work also into the context of related ongoing research activities, we aim at working out where HTTP Adaptive Streaming is currently heading.
Chris Varekamp (Philips Group Innovation, Research): Depth estimation, Proces...AugmentedWorldExpo
A talk from the Develop Track at AWE USA 2018 - the World's #1 XR Conference & Expo in Santa Clara, California May 30- June 1, 2018.
Chris Varekamp (Philips Group Innovation, Research): Depth estimation, Processing & Rendering for Dynamic 6DoF VR
In this talk I will discuss how a real-time depth-based processing chain can be built using our experience in stereo-to-depth conversion for autostereoscopic displays.
http://AugmentedWorldExpo.com
The document provides an overview of the High Efficiency Video Coding (HEVC) standard. Some key points:
- HEVC was created as a new video compression standard to address the growing needs of higher resolution video content and more efficient compression compared to prior standards like H.264.
- It achieves 50% bitrate reduction over H.264 for the same visual quality or improved quality at the same bitrate.
- The standard uses a block-based coding structure with coding tree units and supports intra-frame and inter-frame coding with motion estimation/compensation.
- It introduces more intra-prediction modes and block sizes along with improved transforms, quantization, and entropy coding.
Bitmovin AV1/VVC Presentation_Streaming Media East by Christian FeldmannBitmovin Inc
This document provides an overview and update on AV1 and VVC video coding standards. It summarizes the novel technical features of each, including AV1 improvements like overlapped block motion compensation and affine motion, and VVC features like triangle partitioning and decoder-side motion refinement. Performance results show VVC provides around 30% better compression than HM 16.20, while AV1 is around 20% better. Both standards are still in development and neither has wide adoption yet. VVC licensing is unknown while AV1 remains free and open.
Tutorial on Point Cloud Compression and standardisationRufael Mekuria
Tutorial on Point Cloud Compression and standardisation given at IEEE VCIP 2017 in december. I provide the techniques for point cloud compression and the designed quality metrics and codecs in my PhD at CWI. I detail the standardisation activity on point cloud compression that I started in 2014 and that started in 2017 involving all mobile device makers like Apple, Huawei, Sony, Samsung and Nokia.
Excerpts from the HEVC / H265 Hands-on course.
This parts of the course explains how to download the reference code (HM) compile it configure it and analyze the video output
MPEG Immersive Media
By Thomas, Director, Technical Standards at Qualcomm
at 2nd ITU-T Mini-Workshop on Immersive Live Experience (ILE) in 19 January 2017
This document discusses various methods of image resampling, including downsampling, upsampling, and fractional resampling. It describes traditional blind resampling methods like nearest neighbor, bilinear, bicubic, and Lanczos interpolation. It also covers content-aware resampling techniques like seam carving and new edge-directed interpolation (NEDI) that aim to minimize information loss. For each method, it explains the approach, artifacts, filter kernels, and provides visual examples comparing the output images. The document concludes by discussing tradeoffs between methods and opportunities for future improvement in content-aware resizing.
The document discusses the concept of the Tactile Internet with Human-in-the-Loop. It aims to democratize access to skills and expertise for people of all backgrounds and abilities. This goes beyond the current Internet's goal of providing access to information regardless of location or time. The document outlines a vision for two-way skills transfer between humans and machines using multimodal feedback over 5G networks. It discusses challenges like differing neural time delays for multisensory perception and individual differences in processing that affect perception and action. The Center for Tactile Internet's research agenda involves understanding multisensory goal-directed processing neurocognitively, modeling perception and action, and expertise in related fields to advance human-technology interactions
The document provides an overview of the High Efficiency Video Coding (HEVC) standard. It was developed jointly by ISO/IEC and ITU-T to provide roughly half the bit-rate of H.264/AVC for the same subjective quality. Key aspects of HEVC include use of larger block sizes, intra-picture prediction with 33 directional modes, motion vectors with quarter-sample precision, transform sizes from 4x4 to 32x32, adaptive coefficient scanning, in-loop filtering including deblocking and sample adaptive offset, and support for lossless and transform skipping modes. Many companies are starting to support HEVC in their video products and services.
The document discusses the new HEVC/H.265 video compression standard and its benefits for ultra high definition video. Key points:
- HEVC is 50% more efficient than H.264/MPEG-4 AVC, allowing a 50% reduction in bandwidth. It can support resolutions up to 8K.
- Tests show HEVC achieves 50-75% lower bitrates than H.264 for ultra high definition video, while maintaining comparable quality.
- HEVC's increased efficiency comes from processing video in 64x64 pixel blocks rather than 16x16, and parallel processing of video frames. This requires powerful multi-core processors.
- The improved compression enables
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...Ijripublishers Ijri
This document discusses techniques for improving video compression efficiency for surveillance videos. It proposes modifying the architecture of scalable video coding to make it surveillance-centric by allowing adaptive rate-distortion optimization at the GOP level based on whether events of interest are present. Experimental results show foreground detection and updating of background adaptively over time to improve compression. Future work includes further enhancing selective motion estimation techniques to improve processing efficiency without degrading video quality.
This document provides an overview and comparison of the H.264 and HEVC video coding standards. It describes the key features and innovations that allow each standard to compress video more efficiently than previous standards. H.264 introduced features like adaptive block sizes, multi-frame prediction, quarter-pixel motion compensation and loop filtering that improved compression performance over prior standards. HEVC aims to further increase compression efficiency through innovations such as larger coding tree blocks, additional intra-prediction modes, and improved entropy coding. The document analyzes these standards to understand how their new coding tools enable significantly higher compression ratios and support for new applications like higher resolution video.
Quality Optimization of Live Streaming Services over HTTP with Reinforcement ...Alpen-Adria-Universität
Recent years have seen tremendous growth in HTTP adaptive live video traffic over the Internet. In the presence of highly dynamic network conditions and diverse request patterns, existing yet simple hand-crafted heuristic approaches for serving client requests at the network edge might incur a large overhead and significant increase in time complexity. Therefore, these approaches might fail in delivering acceptable Quality of Experience (QoE) to end users. To bridge this gap, we propose ROPL, a learning-based client request management solution at the edge that leverages the power of the recent breakthroughs in deep reinforcement learning, to serve requests of concurrent users joining various HTTP-based live video channels. ROPL is able to react quickly to any changes in the environment, performing accurate decisions to serve clients requests, which results in achieving satisfactory user QoE. We validate the efficiency of ROPL through trace-driven simulations and a real-world setup. Experimental results from real-world scenarios confirm that ROPL outperforms existing heuristic-based approaches in terms of QoE, with a factor up to 3.7×.
This document summarizes the outcomes of the ITU-T Study Group 16 meeting held from 16-27 January 2017 in Geneva. It discusses the number of contributions reviewed and recommendations approved. It provides details on the collaborative work done with other groups on topics like video coding, IPTV, accessibility, e-health, and more. Major accomplishments included completing new recommendations on scalable vector graphics, speech translation services, and requirements for areas like vehicle gateways and content delivery networks. Work also progressed on developing standards for immersive media, visual surveillance, and blockchain technologies.
This document provides an overview of HEVC (High Efficiency Video Coding) including:
- HEVC aims to provide roughly half the bitrate of H.264/AVC at the same quality.
- It uses block-based hybrid video coding with improved intra-prediction, transform, quantization and entropy coding techniques.
- HEVC supports a wide range of resolutions, color spaces and bit depths for 4K and beyond.
Universal media access as proposed almost two decades ago is now reality. We can generate, distribute, share, and consume any media content, anywhere, anytime, and with/on any device. A technical breakthrough was the adaptive streaming over HTTP resulting in the standardization of MPEG-DASH, which is now successfully deployed in a plethora of environments. The next big thing in adaptive media streaming is virtual reality applications and, specifically, omnidirectional (360°) media streaming, which is currently built on top of the existing adaptive streaming ecosystems. This tutorial provides a detailed overview of adaptive streaming of both traditional and omnidirectional media. The tutorial focuses on the basic principles and paradigms for adaptive streaming as well as on already deployed content generation, distribution, and consumption workflows. Additionally, the tutorial provides insights into standards and emerging technologies in the adaptive streaming space. Finally, the tutorial includes the latest approaches for immersive media streaming enabling 6DoF DASH through Point Cloud Compression (PCC) and concludes with open research issues and industry efforts in this domain. More information available at: https://multimediacommunication.blogspot.com/2019/07/acmmm19-tutorial-journey-towards-fully.html
The document summarizes the H.264/MPEG4 Advanced Video Coding standard. It was developed by the Joint Video Team (JVT) which included the ITU-T Video Coding Experts Group and ISO/IEC Moving Picture Experts Group. The goals were to improve coding efficiency, enhance error robustness, simplify syntax, and increase flexibility. Key features of H.264 include enhanced motion prediction, small block size transforms, an adaptive in-loop deblocking filter, and improved entropy coding. These allow for approximately 50% bit rate savings over prior standards. Evaluation showed it can achieve near transparent quality at lower bit rates than previous standards.
Machine learning approaches are being explored for video compression. Conservative approaches replace individual MPEG blocks with deep learning blocks, while disruptive end-to-end approaches aim to replace the entire MPEG chain. Optical flow networks can exploit temporal redundancy by estimating motion between frames. Fully neural network-based video compression models jointly learn motion estimation, motion compression, and residual compression in an end-to-end optimized framework. However, performance gains must be balanced against increased complexity, and neural network approaches are not yet mature enough to be included in video compression standards.
Tutorial High Efficiency Video Coding Coding - Tools and Specification.pdfssuserc5a4dd
This document provides an outline for a tutorial on High Efficiency Video Coding (HEVC). It discusses the motivation for developing a new video coding standard to support higher resolutions and bandwidth efficiency. It describes the formation of the Joint Collaborative Team on Video Coding (JCT-VC) by MPEG and VCEG to develop the HEVC specification. It also gives an overview of the hybrid coding scheme used in HEVC and other video coding standards, including prediction, transform coding of residuals, and entropy coding.
Design and Analysis of Quantization Based Low Bit Rate Encoding Systemijtsrd
This document summarizes research on developing a low bit rate encoding system for video compression using vector quantization. It first discusses how vector quantization can achieve high compression ratios and has been used widely in image and speech coding. It then describes the methodology used, which involves taking video frames as input, downsampling the frames to extract pixels, applying vector quantization, and detecting edges on the compressed frames to check compression quality. Finally, it discusses the results of testing the approach on MATLAB and presents conclusions on the advantages of the proposed algorithm for very low bit rate video coding applications.
Video services are evolving from traditional two-dimensional video to virtual reality and holograms, which offer six degrees of freedom to users, enabling them to freely move around in a scene and change focus as desired. However, this increase in freedom translates into stringent requirements in terms of ultra-high bandwidth (in the order of Gigabits per second) and minimal latency (in the order of milliseconds). To realize such immersive services, the network transport, as well as the video representation and encoding, have to be fundamentally enhanced. The purpose of this tutorial article is to provide an elaborate introduction to the creation, streaming, and evaluation of immersive video. Moreover, it aims to provide lessons learned and to point at promising research paths to enable truly interactive immersive video applications toward holography.
Jiri ece-01-03 adaptive temporal averaging and frame prediction based surveil...Ijripublishers Ijri
Global interconnect planning becomes a challenge as semiconductor technology continuously scales. Because of the increasing wire resistance and higher capacitive coupling in smaller features, the delay of global interconnects becomes large compared with the delay of a logic gate, introducing a huge performance gap that needs to be resolved A novel equalized global link architecture and driver– receiver co design flow are proposed for high-speed and low-energy on-chip communication by utilizing a continuous-time linear equalizer (CTLE). The proposed global link is analyzed using a linear system method, and the formula of CTLE eye opening is derived to provide high-level design guidelines and insights.
Compared with the separate driver–receiver design flow, over 50% energy reduction is observed.
Insight toolkit을 이용한 삼차원 흉부 CT 영상분석 및 폐결절 검출 시스템Wookjin Choi
The document discusses Insight Toolkit (ITK), an open-source software library for image analysis. ITK can be used for tasks like image segmentation, registration, and processing. It has algorithms for segmentation like watersheds, level sets, and thresholding. The document then describes using ITK to develop a system for detecting lung nodules in 3D chest CT scans. The system performs lung segmentation, finds nodule candidates, extracts features, and aims to classify nodules versus false positives using methods like linear discriminant analysis. Experimental data comes from the LIDC lung cancer CT scan database.
This document discusses video quality analysis for H.264 based on the human visual system. It proposes an improved video quality assessment method that adds color comparison to structural similarity measurement. The method separates similarity measurement into four comparisons: luminance, contrast, structure, and color. Experimental results on video sets with two distortion types show the proposed method's quality scores are more consistent with visual quality than classical methods. It also discusses the H.264 video coding standard and provides examples of encoding and decoding experimental results.
The document provides an overview of the Dirac video codec and compares its performance to H.264/MPEG-4 AVC. Dirac is an open source video compression format developed by the BBC that uses wavelet transforms and arithmetic coding. It achieves compression performance close to H.264/AVC at lower bitrates, with less complexity, though H.264 provides slightly better compression at higher resolutions. Testing showed Dirac performs better than H.264 at low bitrates for QCIF sequences.
This document provides an activity and research report for Marco Cagnazzo from September 2013. It summarizes his teaching activities from 2004-present, which include courses on information theory, multimedia signal processing, compression techniques, and digital video/multimedia at various universities. It also lists his PhD student supervision and involvement in research projects related to video coding optimization, adaptive image compression, and robust video streaming. His main research themes are described as motion representation, 3D video coding, and distributed video coding. Bibliometric data on his publications and other scholarly activities are also presented.
This document summarizes the state-of-the-art in scalable video compression approaches based on wavelet transforms. It describes the basic principles of wavelet-based scalable video coding (WSVC), including enabling spatial, temporal, and quality scalability. Spatial scalability is enabled through multiresolution transforms like the discrete wavelet transform (DWT) and lifting schemes. Temporal scalability is achieved through motion-compensated temporal filtering. Quality scalability can be realized by progressive bitplane coding of wavelet subbands. The document reviews different WSVC architectures and compares their performance to the scalable video coding (SVC) standard.
Are you Digitized Files Really OK? Levels of QC and Film Digitization (SCHALL...FIAT/IFTA
This document discusses different types of quality control (QC) that can be performed during the digitization of video, film, and audiovisual content. It describes standards compliance QC to ensure files follow encoding standards. Audiovisual migration QC detects defects introduced during migration to ensure high quality. Content and metadata QC checks that files are complete and metadata is consistent. Case studies on large scale digitization projects at ORF and Indiana University illustrate the QC workflows and statistics on detected defects. Automated analysis detects more issues than VTR-based methods, and the results are complementary.
Introduction to Video Compression Techniques - Anurag JainVideoguy
The document provides an overview of video compression techniques and standards. It discusses the motivation for video compression to reduce data sizes for storage and transmission. It then reviews several key video compression standards including H.261, H.263, MPEG-1, MPEG-2, MPEG-4, H.264 and others. For each standard, it summarizes the goals, features, applications and technical details like motion compensation methods, block sizes, and bitrate ranges.
Distributed Adaptation Decision-Taking Framework and Scalable Video Coding Tu...mgrafl
The document discusses research into distributed adaptation frameworks and Scalable Video Coding (SVC) tunneling for edge and in-network media adaptation. It presents the ALICANTE adaptation framework, which introduces new home-box and content-aware network layers. The framework uses an Adaptation Decision-Taking Framework (ADTF) to coordinate adaptation decisions across modules at the content source, network edges, and within the network. It also utilizes SVC tunneling to enable media adaptation at network elements. The research aims to improve network resource utilization while maintaining quality of experience. Initial results show SVC tunneling can reduce bandwidth requirements compared to simulcasting.
Video Compression Using Block By Block Basis Salience DetectionIRJET Journal
This document presents a method for video compression using block-by-block salience detection. It aims to reduce noticeable coding artifacts in non-region of interest (ROI) parts of video frames by optimizing the saliency-related Lagrange parameter possibly on a block-by-block basis. The proposed method detects ROI using a visual saliency model and encodes ROI blocks with higher quality than non-ROI blocks. It then separates each frame into blocks and uses a conjugate gradient algorithm to iteratively update weight coefficients and minimize a cost function, compressing each block losslessly based on its saliency. An experiment found the proposed method improved visual quality over other perceptual video coding methods according to metrics like eye-tracking weighted PSNR and
This document discusses post-processing and rate distortion algorithms for the VP8 video codec. It first provides background on the need for post-processing algorithms to reduce blocking artifacts in compressed video, and for rate control algorithms to regulate bitrates and achieve high video quality within bandwidth constraints. It then summarizes existing in-loop deblocking filters and post-processing algorithms. A novel optimal post-processing/in-loop filtering algorithm is described that can achieve better performance than H.264/AVC or VP8 by computing optimal filter coefficients. Finally, a proposed rate distortion optimization algorithm for VP8 is discussed to improve its rate control and coding efficiency.
This document describes the implementation and analysis of a real-time scalable high efficiency video coding (SHVC) decoder. The SHVC decoder is based on an open source HEVC decoder and supports parallel processing across tiles, slices and wavefronts. Experimental results show that decoding two SHVC layers introduces 43-77% additional complexity compared to HEVC. Using parallel processing across four threads provides a speedup of 2-4x and enables real-time decoding of 1080p SHVC video at over 50 frames per second. Future work includes supporting more than two layers and the base AVC layer.
The document discusses methods for objective and subjective video quality assessment and speech enhancement. It covers four parts: (1) a classification and review of no-reference visual quality assessment methods, (2) no-reference and reduced-reference methods for video quality assessment including neural network and support vector machine approaches, (3) subjective methods for video quality assessment including studies on low resolution videos and crowdsourcing, and (4) speech enhancement techniques including spectral center-of-gravity based demodulation and convex optimization based demodulation. The document evaluates various computational models and machine learning techniques for video and speech quality assessment.
The document discusses techniques for constant bit rate video streaming over packet switching networks. It proposes adapting variable bit rate video to a constant bit rate by controlling the video encoder's output rate based on buffer level feedback. This allows transporting video over networks using constant bit rate channels while avoiding network congestion issues. The key techniques involve bit allocation to each coding unit based on buffer status, and adjusting encoder quantization parameters to encode units with the allocated bits. Simulation results show the approach maintains constant compression ratio and peak signal-to-noise ratio while varying group of picture size and quality settings.
Similaire à PCS 2019 Panel on Emerging Video Coding Standards: Overview on the Emerging Versatile Video Coding Standard (20)
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Trusted Execution Environment for Decentralized Process MiningLucaBarbaro3
Presentation of the paper "Trusted Execution Environment for Decentralized Process Mining" given during the CAiSE 2024 Conference in Cyprus on June 7, 2024.
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Nunit vs XUnit vs MSTest Differences Between These Unit Testing Frameworks.pdfflufftailshop
When it comes to unit testing in the .NET ecosystem, developers have a wide range of options available. Among the most popular choices are NUnit, XUnit, and MSTest. These unit testing frameworks provide essential tools and features to help ensure the quality and reliability of code. However, understanding the differences between these frameworks is crucial for selecting the most suitable one for your projects.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
leewayhertz.com-AI in predictive maintenance Use cases technologies benefits ...alexjohnson7307
Predictive maintenance is a proactive approach that anticipates equipment failures before they happen. At the forefront of this innovative strategy is Artificial Intelligence (AI), which brings unprecedented precision and efficiency. AI in predictive maintenance is transforming industries by reducing downtime, minimizing costs, and enhancing productivity.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
PCS 2019 Panel on Emerging Video Coding Standards: Overview on the Emerging Versatile Video Coding Standard
1. Brief Overview on the Emerging
Versatile Video Coding Standard
… as seen by an individual
Picture Coding Symposium 2019, Ningbo, China
Mathias Wien
Lehrstuhl für Bildverarbeitung
RWTH Aachen University
wien@lfb.rwth-aachen.de
2. Versatile Video Coding | Panel on Emerging Video Codecs | Picture Coding Symposium 2019 | Ningbo, China
15.11.2019 | Mathias Wien | Institute of Imaging and Computer Vision | RWTH Aachen University
2
• Joint Video Experts Team of ISO/IEC MPEG and ITU-T VCEG, est. in
Oct. 2015 as Joint Video Exploration Team, renamed Apr. 2018 at CfP eval.
• Joint Exploration Model: 7 iterations
Investigating coding tools enhanced relative to HEVC test model HM
• Call for Evidence (CfE), issued Apr. 2017, evaluated in Jul. 2017
Significant improvement for JEM and independent submissions
• Call for Proposls (CfP), issued Oct. 2017, evaluated in Apr. 2018
at10th JVET meeting
Kick-off for VVC, definition of name, VVC draft 1 and test model 1 (VTM1)
Evolution towards VVC
Evolution of JVET docs and participants
3. Versatile Video Coding | Panel on Emerging Video Codecs | Picture Coding Symposium 2019 | Ningbo, China
15.11.2019 | Mathias Wien | Institute of Imaging and Computer Vision | RWTH Aachen University
3
• Entropy Coding
CABAC
Multi-hypothesis probability estimation
• Coding structures
4 × 4 to 128 × 128 Coding Unit (CU) size
Quad+ternary+binary tree partitioning
Dual-tree (separate trees for luma/chroma)
Partial coverage of prediction regions by transforms (SBT,ISP)
• Inter Coding
16th-sample prec. for derived MVs | Locally adaptive MV
resolution: signaling at ¼, 1, 4 sample precision
8-tap DCTIF + alt. 6-tap switched IF for ½ sample, 6-tap for affine
subblock MC
Block-level CU weights, Triangular prediction with merge
Combined inter/intra prediction (CIIP)
Affine prediction refinement with optical flow (PROF),
Bi-directional optical flow prediction (BDOF)
Subblock-based temporal merge, affine MV control point MVP,
decoder-side MV refinement, history-based MVP, symmetrical
MVD, Merge mode with MVD (MMVD)
• Intra Coding
DC, planar, 67 intra prediction directions signaled, 93 after wide
angle mapping | signaling: 6 most probable modes + remainder
Matrix-based intra prediction
Multi reference lines intra prediction
Intra subpartitions (ISP)
Position-dependent intra prediction sample filtering
Cross-component linear prediction
• Residual Coding
DCT2 4×4-64×64, multi-transform selection (DST7/DCT8, ≤32),
inter: subblock transform (SBT), low-frequency non-separable
transform (LFNST), transform skip, 4:4:4 adaptive colour transform
Coefficient coding: HEVC + parity-based binarization for dependent
quantization (DQ)
• Loop Filters
Deblocking filter, sample adaptive offset (SAO)
Adaptive loop filter (ALF), luma mapping with chroma scaling
VVC Coding Tools in a Nutshell
MV/MVP/MVD: motion vector / predictor / difference | DCTIF: Discrete
cosine transform derived interpolation filter | DST: Discrete sine transform
4. Versatile Video Coding | Panel on Emerging Video Codecs | Picture Coding Symposium 2019 | Ningbo, China
15.11.2019 | Mathias Wien | Institute of Imaging and Computer Vision | RWTH Aachen University
4
• Parameter sets
Video parameter set, sequence parameter set,
picture parameter set (VPS, SPS, PPS)
Adaptation parameter set (APS), picture header
• Picture types
IDR, CRA, RADL, RASL, Trail, STSA
Gradual decoding refresh (GDR)
• Picture partitioning
Slices, tiles
Subpictures, raster-scan slices of multiple tiles,
rectangular slices within tiles
• Screen content coding
Palette mode for 4:4:4
Intra block copy (IBC)
Residual coding for transform skip
Block-based Delta PCM (BDPCM)
• More versatility
Reference picture resampling (RPR)
Spatial, quality and temporal scalability
Bitstream extraction and merging without
VCL NAL unit rewriting
Film grain synthesis SEI message
VVC Versatility Aspects
IDR: instantaneous decoding refresh | CRA: clean random access | RADL/RASL: random access decodable/skipped leading pic. | STSA: step-wise temporal sublayer access
5. Versatile Video Coding | Panel on Emerging Video Codecs | Picture Coding Symposium 2019 | Ningbo, China
15.11.2019 | Mathias Wien | Institute of Imaging and Computer Vision | RWTH Aachen University
5
• Assessment according to JVET Common Testing Conditions [JVET-M1010]
Rate-distortion performance, encoder / decoder run-time of VVC reference software implementation
• Systematic evaluation throughout standardization project
All Intra, Random Access, Low Delay B configurations
Comparison to HEVC reference software HM: Comparable rate-distortion based encoder implementations
Versatile Video Coding – Compression Performance
6. Thanks for your attention!
Mathias Wien
Lehrstuhl für Bildverarbeitung
RWTH Aachen University
wien@lfb.rwth-aachen.de