skip to main content
10.1145/3510450.3517275acmconferencesArticle/Chapter ViewAbstractPublication PagesmhvConference Proceedingsconference-collections
short-paper

CAMBI: introduction and latest advances

Published: 17 March 2022 Publication History

Abstract

Banding is manifested as false contours in otherwise smooth regions in an image or a video. There are many different reasons for banding artifacts to occur, however, one of the most prominent causes is the quantization inside a video encoder. Compared to other types of artifacts common in video processing, e.g. blur, ringing, or blockiness, only a relatively small change of the original pixel values can produce an easily noticeable and visually very annoying case of banding. This property makes it very difficult for banding to be captured by generic objective quality metrics such as PSNR or VMAF [10] which brings a need for a distortion specific detector targeted directly at banding artifacts.
Most of the previous attempts to solve this problem tried to tackle it as false segments or false edges detection. Both block-based [7, 14] and pixel-based [2, 3, 15] segmentation methods have been tried in the first category, while the edge-based methods exploited different local statistics such as gradients, contrast, or entropy [4, 6, 9, 13]. The main difficulty for all of these approaches is distinguishing between the real and false edges or segments. Recently, banding detection has also been addressed by deep neural networks [8].
The above mentioned approaches have been developed for 8-bit content and mostly tuned towards the banding artifacts occurring in user-generated images and videos. Moreover, they do not address the potential presence of dithering - an intentionally inserted noise used to randomize the error caused by quantization. Dithering is commonly used during bit-depth conversion and is often enabled by default in popular image and video processing tools, such as ffmpeg [12]. Despite being highly effective in reducing the perceived banding, dithering does not suppress the false contours completely, and thus needs to be factored in a reliable banding detector.
Our goal was, therefore, to develop an algorithm capable of evaluating perceived banding in professionally generated videos processed in ways relevant to the adaptive streaming scenario (i.e. video compression and scaling). The requirements also included ability to capture the effect of dithering and to work on both 8-bit and 10-bit content.
We hereby present CAMBI, a Contrast Aware Multiscale Banding Index. CAMBI is a white-box solution to the above described problem derived from basic principles of human vision with just a few, perceptually-motivated, parameters. The first version was introduced at PCS'2021 [11]. Here, we also present several improvements made since then.
There are three main steps in CAMBI - input preprocessing, multiscale banding confidence calculation, and spatio-temporal pooling. Although it has been shown that chromatic banding exists [5], like most past works, we assume that most of the banding can be captured in the luma channel. The preprocessing step, therefore, consists of luma channel extraction followed by filtering to account for dithering and a spatial mask computation to exclude regions with textures. Banding confidence is calculated for 4 brightness level differences on 5 scales, taking into account contrast perception of human visual system. This creates 20 banding confidence maps per frame that are pooled spatially considering only a certain percentage of highest banding confidence. Such mechanism ensures that even the banding appearing in relatively small area of the frame is captured proportionally to its perceptual importance. Finally, the scores from different video frames are pooled into a single banding index.
To test the accuracy of CAMBI, we conducted a subjective test on 86 video clips created from 9 different sources from Netflix catalog using different levels of compression and scaling with and without dithering. The ground-truth mean opinion scores (MOS) were obtained from 26 observers who were asked to rate the annoyance of the banding in the scene on the continuous impairment scale annotated with 5 equidistant labels (imperceptible, perceptible but not annoying, slightly annoying, annoying, very annoying) [1]. CAMBI achieved a correlation exceeding 0.94 in terms of both Pearson Linear Correlation Coefficient (PLCC) and Spearman Rank Order Correlation Coefficient (SROCC), significantly outperforming state-of-the-art banding detectors for our use-case.
CAMBI is currently used alongside VMAF in our production to improve the quality of encodes prone to banding. In the future, we are planning to integrate it into VMAF as one of the features to make it capable of accurately evaluating video quality in the presence of banding as well as other artifacts.

References

[1]
ITU-T Recommendation P. 913. 2016. ITU (2016).
[2]
Gary Baugh, Anil Kokaram, and François Pitié. 2014. Advanced video debanding. In Proceedings of the 11th European Conference on Visual Media Production. 1--10.
[3]
Sitaram Bhagavathy, Joan Llach, and Jiefu Zhai. 2009. Multiscale probabilistic dithering for suppressing contour artifacts in digital images. IEEE Transactions on Image Processing 18, 9 (2009), 1936--1945.
[4]
Scott J Daly and Xiaofan Feng. 2004. Decontouring: Prevention and removal of false contour artifacts. In Human Vision and Electronic Imaging IX, Vol. 5292. International Society for Optics and Photonics.
[5]
Gyorgy Denes, George Ash, Huameng Fang, and Rafał K Mantiuk. 2019. A visual model for predicting chromatic banding artifacts. Electronic Imaging 2019, 12 (2019), 212--1.
[6]
Qin Huang, Hui Yong Kim, Wen-Jiin Tsai, Se Yoon Jeong, Jin Soo Choi, and C-C Jay Kuo. 2016. Understanding and removal of false contour in hevc compressed images. IEEE Transactions on Circuits and Systems for Video Technology 28, 2 (2016), 378--391.
[7]
Xin Jin, Satoshi Goto, and King Ngi Ngan. 2011. Composite model-based dc dithering for suppressing contour artifacts in decompressed video. IEEE Transactions on Image Processing 20, 8 (2011), 2110--2121.
[8]
Akshay Kapoor, Jatin Sapra, and Zhou Wang. 2021. Capturing banding in images: Database construction and objective assessment. In International Conference on Acoustics, Signal, and Speech Processing (ICASSP). IEEE.
[9]
Ji Won Lee, Bo Ra Lim, Rae-Hong Park, Jae-Seung Kim, and Wonseok Ahn. 2006. Two-stage false contour detection using directional contrast and its application to adaptive false contour reduction. IEEE Transactions on Consumer Electronics 52, 1 (2006), 179--188.
[10]
Zhi Li, Anne Aaron, Ioannis Katsavounidis, Anush Moorthy, and Megha Manohara. 2016. Toward a practical perceptual video quality metric. The Netflix Tech Blog 6 (2016), 2.
[11]
Pulkit Tandon, Mariana Afonso, Joel Sole, and Lukáš Krasula. 2021. CAMBI: Contrast-aware Multiscale Banding Index. In 2021 Picture Coding Symposium (PCS). 1--5.
[12]
Suramya Tomar. 2006. Converting video formats with FFmpeg. Linux Journal 2006, 146 (2006), 10.
[13]
Zhengzhong Tu, Jessie Lin, Yilin Wang, Balu Adsumilli, and Alan C Bovik. 2020. Bband Index: a No-Reference Banding Artifact Predictor. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2712--2716.
[14]
Yanxiang Wang, Charith Abhayaratne, Rajitha Weerakkody, and Marta Mrak. 2014. Multi-scale dithering for contouring artefacts removal in compressed UHD video sequences. In 2014 IEEE Global Conference on Signal and Information Processing (GlobalSIP). IEEE, 1014--1018.
[15]
Yilin Wang, Sang-Uok Kum, Chao Chen, and Anil Kokaram. 2016. A perceptual visibility metric for banding artifacts. In 2016 IEEE International Conference on Image Processing (ICIP). IEEE, 2067--2071.

Cited By

View all
  • (2024)A Novel Energy-Efficient Coding Based on Coordinated Group Signal Transformation for Image Compression in Energy-Starved SystemsApplied Sciences10.3390/app1410417614:10(4176)Online publication date: 15-May-2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MHV '22: Proceedings of the 1st Mile-High Video Conference
March 2022
150 pages
ISBN:9781450392228
DOI:10.1145/3510450
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 17 March 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. CAMBI
  2. VMAF
  3. banding
  4. quality assessment
  5. video compression

Qualifiers

  • Short-paper

Conference

MHV '22
Sponsor:
MHV '22: Mile-High Video Conference
March 1 - 3, 2022
Colorado, Denver

Upcoming Conference

MHV '25
Mile-High Video Conference
February 18 - 20, 2025
Denver , CO , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)8
  • Downloads (Last 6 weeks)2
Reflects downloads up to 14 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)A Novel Energy-Efficient Coding Based on Coordinated Group Signal Transformation for Image Compression in Energy-Starved SystemsApplied Sciences10.3390/app1410417614:10(4176)Online publication date: 15-May-2024

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media