Next Article in Journal
Dynamical Sampling with Langevin Normalization Flows
Previous Article in Journal
Quantification of Information Exchange in Idealized and Climate System Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Perceptual Video Coding Scheme Using Just Noticeable Distortion Model Based on Entropy Filter

1
Faculty of Information Science and Engineering, Ningbo University, No. 818, Ningbo 315211, China
2
School of Electrical and Electronic Engineering, Chongqing University of Technology, Chongqing 400054, China
*
Author to whom correspondence should be addressed.
Submission received: 16 October 2019 / Revised: 4 November 2019 / Accepted: 5 November 2019 / Published: 8 November 2019
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
Because perceptual video coding (PVC) can reduce bitrates with negligible visual quality loss in video compression, a PVC scheme based on just noticeable distortion (JND) model is proposed for ultra-high definition video. Firstly, the proposed JND model is designed, considering the spatial JND characteristics such as contrast sensitivity, luminance adaptation and saliency weight factor. Secondly, in order to perform precise JND suppression, the Gauss differential entropy (GDE) filter is designed to divide the image into smooth and complex texture region. Thirdly, through incorporating the proposed JND model into the encoding process, the transform coefficients are suppressed in harmonization with the transform/quantization process of high efficiency video coding (HEVC). In order to achieve the JND suppression effectively, a distortion compensation factor and distortion compensation control factor are incorporated to control the extent of distortion in the rate distortion optimization process. The experimental results show that the proposed PVC scheme can achieve a remarkable bitrate reduction of 32.98% for low delay (LD) configuration and 28.61% for random access (RA) configuration with a negligible subjective quality loss. Meanwhile, the proposed method only causes about average 12.94% and 22.45% encoding time increase under LD and RA configuration compared with an HEVC reference software, respectively.

1. Introduction

Ultra-high definition (UHD) video provides viewers with enhanced visual experience via a wider field of view (FOV) and more exquisite frames than high definition (HD) video [1]. UHD video is widely applied in various fields, education, amusement, sports, etc. [2,3]. Unfortunately, the large bandwidth and storage space are required for realization of these UHD video applications [4,5]. These problems limit the development and application of UHD video. Besides spatial and temporal redundancies, visual redundancy widely exists in UHD videos. Hence, the perceptual video coding (PVC) scheme can be utilized to further exploit the visual redundancy, and improve compression performance.
The key of PVC is to determine the distortion that users can just notice. Therefore, the study of just noticeable distortion (JND) and PVC become a hot topic and has attracted much interest. So far, various PVC schemes based on JND have been proposed and utilized for HD image/video compression [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]. The JND models can be categorized into two types: pixel and subband (e.g., discrete cosine transform (DCT), wavelet) domain approaches.
Specifically, the pixel-based JND profiles are established in the pixel domain, considering luminance masking (LM) and contrast masking (CM) effects [6,7,8,9]. Yang et al. proposed a pixel-based JND profile by considering the overlapping effect between LM and CM effects [6]. Chen et al. incorporated a foveated masking model into the JND profile and the JND values were weighted with visual eccentricity [7]. Liu et al. improved the work in [6], where the JND model was more precisely estimated by decomposing edge pixels from texture regions [8]. Wu et al. proposed a JND profile with free energy principle, where an autoregressive model was deduced to imitate the active prediction of the human visual system (HVS) with Bayesian inference [9]. Meanwhile, the disorderly concealment effect was estimated for the JND model. Several stereoscopic JND profiles have been established in the pixel domain [10,11,12]. Zhao et al. established a binocular JND profile by considering the asymmetric binocular vision property [10]. Zhang et al. proposed a JND profile for autostereoscopic display by considering the view position [11]. Wang et al. considered binocular spatial sensitivity in stereoscopic image quality assessment and utilized the [10] to measure the joint visibility of a pair distortions in the two views of stereoscopic image [12].
The pixel-based JND profiles cannot ensure conformance with perceptual characteristics of human eyes, because they do not individually take the sensitivity of HVS for each frequency component into consideration. In contrast, the subband-based JND profiles are modeled in subband domains and show better performance. In particular, the JND profiles in the DCT domain are more appropriate to image/video compression schemes [13,14,15,16,17,18,19,20]. Ahumada and Peterson proposed a JND based on DCT-domain quantization matrices which is known as the DCTune algorithm [13]. Watson applied the 8 × 8 JND-based DCT quantization matrix to a Joint Photographic Experts Group (JPEG) coder [14]. Wei and Ngan et al. proposed a CM-JND model based on 8 × 8 block. There were different CM-JND values for three image patch types of plain, edge, and texture [15]. Ma et al. extended [15] to work even at a 16 × 16 block-sized DCT kernel [16]. Luo et al. proposed a JND model to suppress the magnitudes of transform coefficients [17]. Bae et al. proposed a JND model in a DCT domain [18]. They proposed an elaborate JND model for variable block-sized transforms in a monochrome image [19]. They also incorporated foveated masking effects into a DCT-based total JND profile in [20].
Recently, many related works focus on the DCT-based JND profiles for high efficiency video coding (HEVC)-complaint PVC scheme [21,22,23,24,25,26]. Kim et al. proposed an HEVC-compliant PVC scheme based on JND models for variable block-sized transform kernels [21]. In addition, they proposed a compensation factor to tackle the distortions caused by transform coefficients suppression with JND profile. Bae et al. proposed an HEVC-based perceptually adaptive video coding method using a DCT-based local distortion detection probability (LDDP) model [22]. The proportion of suppression was modulated by the LDDP model. The models in [21,22] did not take the quantization parameter (QP) into account. Kim et al. proposed a JND model with the convolutional neural network. Moreover, the QP was considered in encoding process [23]. However, the distortion increased when the transform coefficients were suppressed in rate distortion optimization (RDO) process. Zhang et al. proposed a novel perception-based quantization to remove nonvisible information in high dynamic range (HDR) color pixels by exploiting luminance masking [24]. Through this method, the coding performance of HEVC was improved for HDR content. Jiang et al. proposed a saliency-based quantization control algorithm to reduce the bitrate of HEVC encoding processing [25]. Zeeshan et al. proposed a HEVC-compliant PVC scheme using entropy based visual saliency model to adjust the quantization parameter at the coding tree unit (CTU) level [26].
The conventional JND models are designed and utilized for PVC of videos with eight-bit depth. The existing JND models for videos with 10-bit depth are directly derived from the conventional JND models [6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26]. They are only a simple extension of the luminance range. In this work, we perform psychophysical experiments, and reveal that JND models with the simple extension do not conform well with perceptual distortion of images/videos with 10-bit depth. Hence, we proposed a JND model for the UHD images/videos with 10-bit depth based on psychophysical experiments. Following this, a perceptually HEVC-complaint PVC scheme based on the proposed JND model is proposed for improving the encoding efficiency of UHD videos with 10-bit depth. The main contributions of the work are as follows:
Because the conventional JND models are inappropriate for perceptual distortions of UHD videos with 10-bit depth, an LM effect is proposed. To consider the JND threshold values in the texture region, the Gauss differential entropy (GDE) filter is designed to divide the image into smooth and complex texture region. Because the visual focus is not fixed in a specific region when viewing the UHD videos, a saliency factor is incorporated into the proposed JND model. Moreover, the edge of the image is protected in the saliency region.
A perceptual friendly HEVC-complaint PVC scheme is proposed. Meanwhile, a distortion compensation factor (DCF) and a distortion compensation control factor (DCCF) are incorporated into the RDO process for improving the encoding efficiency of UHD videos with 10-bit depth.
The remainder of the paper is organized as follows. In Section 2, we introduce a new JND model that takes luminance masking effect of UHD videos with 10-bit depth, saliency factor, Gauss differential entropy filter into account. In Section 3, a HEVC-compliant PVC scheme is designed via the proposed JND model. Section 4 presents the subjective and objective experimental results for the proposed HEVC-complaint PVC schemes. Section 5 concludes our work.

2. The Proposed Transform-Domain Suppressed JND Model

Usually, the SKIP mode is an efficient inter prediction technology, and it does not incorporate transforms in HEVC. Therefore, the proposed JND model based on a DCT domain is applied for the non-SKIP mode for various block sizes (4 × 4–32 × 32) of the coding unit (CU) and the transform unit (TU). Figure 1 shows the overall framework of the proposed HEVC-compliant PVC scheme. The symbols are described in Table 1. The PVC scheme includes two parts, the proposed JND model and original encoder. The block diagrams of the two parts are respectively surrounded by red and blue dotted boxes. The JND model is used to suppress the transform coefficients.
The “red dotted box” of Figure 1 shows the proposed JND model. To estimate the JND threshold values for transform coefficients suppression, the original TU blocks of input image are applied to calculate the μp for the LM effect. Here, the LM effects are calculated as modulation factors. Then, LM effects are multiplied to a basic contrast sensitivity function (CSF). In order to control the proposed JND threshold values range in the smooth and complex region of image, the GDE filter is incorporated into JLM model. Finally, the JLM model after the filter multiply the saliency factor for the FOV of UHD.
To effectively apply the transform coefficient based on JND suppression, the DCF and DCCF are proposed to compensate the distortion introduced by JGDE-S model at different QPs. Following this, they are incorporated into the RDO process of HEVC in the “blue box” of Figure 1.

2.1. The Proposed JND Model by GDE Filter and the Saliency Factor

An appropriate transform-based JND model for UHD image/video with 10-bit depth is proposed in this section. Firstly, the basic framework of transform domain JND model is modeled based on the CSF and proposed 10-bit LM effect. Following this, the proposed JND threshold values are suppressed by the GDE filter. Besides, the saliency factor is incorporated into the proposed JND model.

2.1.1. The Basic Framework of the JND Model

The proposed basic framework of JND model is constructed based on the HCSF and MFLM, which is expressed as:
J L M ( w i , j , φ i , j , μ p ) = H C S F ( w i , j , φ i , j ) M F L M ( μ p )
where wi,j is cycles per degree (cpd) in spatial frequency for the (i, j)-th DCT coefficient, which is given by:
w i , j = ( 1 / 2 M ) ( i / θ x ) 2 + ( j / θ y ) 2
where θx and θy indicate the horizontal and vertical visual angles of a pixel. The M is the transform sizes (M = four to 32 for HEVC). The pixel aspect ratio is one for most of the display screens, and the θ is identical in the horizontal and vertical directions [15], which is given by:
θ x = θ y = θ = 2 arctan ( 1 R V H H 2 )
where RVH is the ratio of the view distance to the screen height, and H is the number of pixels in the screen height. From (2) and (3), it is found that the view distance or the pixel resolution of display screen increases, the spatial frequency also increases for each DCT component.
In (1), the directional angle φi,j between φi,0 and φ0,j at the (i, j)-th DCT coefficient is taken into account for HCSF modeling. Usually, the HVS is more sensitive to the distortions along the horizontal and vertical directions (i or j = 0) than the diagonal (i = j) direction in spatial frequency, the character is called the oblique effect [27]. The φi,j is given by:
φ i , j = arcsin ( 2 w i , 0 w 0 , j w i , j 2 )
the μp is the normalized pixel intensity of image block, which is defined as:
μ p = ( 1 / K M 2 ) y M x M I ( x , y )
where I(x, y) is the pixel intensity at (x, y).
K = 2 b i t 1
where K is the maximum pixel intensity (255 in 8-bit image format and 1023 in 10-bit image format).
The CSF quantifies the characteristics of human visual perception, which has different sensitivities at different spatial frequencies. In this paper, we adopt HCSF threshold in the transform domain [16], which is given by:
H C S F ( w i , j , φ i , j ) = 1 ϕ i ϕ j exp ( c w i , j ) / ( a + b w i , j ) r + ( 1 r ) cos 2 φ i , j
where φi and φj are the normalization factors for transform coefficients, which is given by:
ϕ k = { 1 / M ,       k = 0 2 / M ,     k > 0
where the model parameter values of a = 1.33, b = 0.11, and c = 0.18 are defined as [15] in (7).

2.1.2. The Proposed 10-Bit Luminance Masking Effect

The LM effect is modeled as a human visual function by which JND threshold values change along with the average pixel intensity. The existing JND models only extends the luminance range to adapt the images/videos with 10-bit depth, but do not fully consider the visual difference with increasing the luminance range.
The proposed 10-bit LM effect is obtained with psychophysical experiments by increasing the distortion amplitudes of the DCT coefficients until subjects start to perceive the distortions. Bae et al. designed a subjective experiment to obtain a LM effect in DCT domain [18]. In this paper, the subjective experiment is followed as in [18], however there are two differences with Bae’s work. The first one is that Bae et al. added distortion into selected 15 DCT coefficients in a lower triangle frequency zone of 8 × 8 DCT and perceived the distortion in the DCT domain. However, we directly added the distortions into all DCT coefficients and perceived the distortions in the luminance domain. Our method can better reflect the distortions of the DCT domain coefficients. The other is that we test the subjective experiments in a wider luminance region (0 to 1023) than Bae’s (0 to 255). Therefore, in order to get a more perceptual friendly 10-bit LM effect, the following subjective experiments are carried out as
Initialization: Monitor shows test images resided in parafovea regions (P2 in Figure 2). The experimental environments and conditions are listed in Table 2.
Step 1: A subject is noticed where the distortion will be injected by making the region dark (Figure 2a).
Step 2: The subject gradually increases the amplitude of distortion for the DCT coefficient until the viewer perceives the distortion in luminance domain, such as P1 in Figure 2b. Each measured JND value is a value that 50% of subjects start to perceive as the corresponding distortion.
Step 3: Step 1 and Step 2 are iteratively conducted by going back to Step 1 with the different presentation, until all test presentations are finished.
Figure 3 shows the experimental results of the subjective experiment. The circle represents the JND threshold values obtained from the subjective experiment. Normalized pixel intensity of the N × N TU blocks given by (5) are displayed in the x-axis and JND threshold values are in the y-axis. Zhang et al. exploit several parabolas to approximate the subjective data curve to build an LM effect model [28]. Therefore, we also use the method to fit subjective data, the proposed 10-bit MFLM model is expressed as:
M F L M ( μ p ) = { A 1 ( 1 ( μ p / 0.2 ) α ) + B , μ p 0.2 A 2 ( μ p 0.2 ) β + B , μ p > 0.2
where the parameters can be obtained by subjective experiments and data fitting, then A1 = 4, A2 = 5, B = 1.5, α = 0.5, and β = 0.8.
Figure 4 shows that the proposed MFLM model compared with the proposed LM effect model by Zhang et al. and Kim et al. The luminance range of the Zhang’s and Kim’s is between 0 and 255, and the JND threshold values are between 1 and 3.2 in Figure 4. When the Zhang’s and Kim’s LM effect models are utilized to the 10-bit luminance range, the range of JND threshold values are consistent with 8-bit JND threshold values. Through subjective experiments, the proposed MFLM model does not only extend the luminance range, but also widens the JND threshold range than the existing LM effect models. As shown in Figure 4, the MFLM threshold values are between 1.5 and 5.7.

2.2. The Proposed 10-Bit JND Model Is Suppressed by the GDE Filter

With the dramatic improvement of UHD resolution, the image texture details are more fine and closer to the natural scene. To analyze the differences between pixels of UHD image/videos, the GDE is utilized in this paper. Meanwhile, the designed GDE filter is not only applied to divide the image into the smooth and complex texture region, but also it is utilized to control the extent of imperceptible noise in different regions of the image. Moreover, the texture of the image has not been considered in the proposed JLM model and the fixed quantization matrices are used, which limits the ability to perform precise JND suppression. The GDE is widely utilized in the information theory [29], and is generally defined as:
H ( x ) = + p ( x ) log 2 p ( x ) d x
where x = I(i, j) expresses the (i, j)-th position pixel, P(x) is a probability density function for arbitrary pixels, which accord with the Gaussian distribution as follows:
p ( x ) = 1 2 π σ 2 e ( x μ ) 2 2 σ 2 ,         < x < +
where μ is mean and σ is variance of the image block.
Therefore, (10) and (11) are utilized to deduced as:
H ( x ) = + p ( x ) log 2 p ( x ) d x = 1 2 log 2 ( 2 π e σ 2 )
where the GDE is only related to the σ. Then, it is expressed as:
H ( σ ) = log 2 σ + 1 2 log 2 ( 2 π e )
The σ-H(σ) curves of ShakeNDry video sequences with eight-bit and 10-bit depth are shown in Figure 5a,b, moreover they display an approximately logarithmic distribution. The small variance indicates that the texture of current image block is smooth and allows less imperceptible noise. On the contrary, the texture of the current image block is more complex and allows for more imperceptible noise. Thus, the H(σ) reflects the capacity of imperceptible noise in different regions.
The maximum curvature of the whole curve is taken as the “first-turning” point, which is used as the threshold to distinguish smooth regions from texture complex regions. Usually, let ρ be the curvature of the curve, and is expressed as:
ρ = | H ( σ ) | ( 1 + H ( σ ) 2 ) 2 3
where H′ and H″ are the first derivative and second derivative of σ respectively.
Finally, the “first-turning” point at (σth, ρ(σth)) which is the maximum curvature of the curve is obtained as follows:
( σ t h , ρ ( σ t h ) ) = M a x i m u m ( ρ ( σ ) )
where the σth is a threshold to divide the image into complex texture and smooth region. In order to preserve more smooth regions, the “second-turning” point is defined as the threshold. The “second-turning” point is the maximum curvature of the curve from the “first-turning” point to the end, then (14) and (15) can be used to obtain as σth ≈ 2.0 and ρ (σth) ≈ 3.0, which are in the intersection of two red lines as shown in Figure 5a,b.
Table 3 lists statistics of image segmentation region. Specifically, the σ of image blocks is smaller than σth, the current image blocks are in the smooth region, otherwise are in the complex texture or edge region. As shown in Table 3, the proportion of the smooth region of the video with 8-bit depth is larger than the texture region. However, for the video with 8-bit depth, the proportion of texture region of the UHD video is smaller than the texture region of HD video, because of UHD video containing more pixels in the same content image area. As shown in Table 3, the proportion of texture region is larger than smooth region for UHD image with 10-bit depth. Essentially, the UHD images can contain more texture details for the 10-bit grey luminance range.
Given the above analysis, the GDE filter is modeled as:
H G D E ( x ) = { H ( x ) , σ > σ t h σ , σ σ t h     & &     H ( x ) 0 H ( x ) , σ σ t h     & &     H ( x ) > 0
As shown in Figure 6, one frame of the TrafficFlow video sequence is segmented by HGDE at 8-bit depth and 10-bit depth respectively. As shown in Figure 6b,d, the texture or edge is white region, and the smooth region is black region. Noticeably, there are more white regions in 10-bit segmentation image than in 8-bit segmentation image, since the image with 10-bit depth contains more texture details than image with 8-bit depth. Because the human eyes are not sensitive to the complex texture details, the complex texture details are essentially considered as noises for HVS. Thus, the image with 10-bit depth has a capacity for more perception distortions than the image with 8-bit depth. Meanwhile, the JND threshold values are mainly exploited in complex texture regions of the image.
The GDE is expressed as the capacity of noise in different regions of image. Therefore, the JND threshold values can be suppressed to a certain range by HGDE without causing perceptual distortion, so the J′LM(μp) is obtained by HGDE filter and it is modeled as:
J L M ( μ p ) = { H ( x ) , σ > σ t h       & &       J L M ( μ p ) > H ( x ) J L M ( μ p ) , σ > σ t h       & &       J L M ( μ p ) H ( x ) σ , σ σ t h     & &         H ( x ) 0       & &     J L M ( μ p ) > H ( x ) H ( x ) , σ σ t h       & &       H ( x ) > 0       & &     J L M ( μ p ) > H ( x ) J L M ( μ p ) , σ σ t h     & &   H ( x ) > 0         & &     J L M ( μ p ) H ( x )
As shown in Figure 7, the proposed JLM threshold values are “blue stars” and the HGDE threshold values are “red stars”. The proposed JND threshold values increase along with variances, but they all suppressed by HGDE with imperceptible distortions.

2.3. Saliency Factor

The human eyes do not view all UHD video contents. Furthermore, the visual focus will not be fixed in a specific region. Thus, the saliency region is utilized to simulate the change of visual focus when human eyes view the UHD image/video. Moreover, it better reflects the performance of the proposed JND model in a global image.

2.3.1. Saliency Region Extraction Algorithm Based on DCT Domain

Hou et al. proposed a simple salient region extraction algorithm using the Fourier transform (FT) [30]. In order to maintain consistency with the HEVC coding structure and simplify the complexity of the algorithm, the DCT transform is used to replace the FT transform according to Hou’s method. Different from Hou’s method, the saliency region is modeled as:
S ( x , y ) = g ( x , y ) D C T 1 [ sin ( R ( u , v ) + θ ( u , v ) ) ] 2
where the g(x, y) is a gaussian filter, which is utilized to smooth the saliency map for better visual effects. The DCT−1(∙) is inverse discrete cosine transform. The residual spectrum of image R(u, v) is extract in spectral domain and is given by:
R ( u , v ) = L ( u , v ) h n ( u , v ) L ( u , v )
where hn(∙) represents the mean filter. The logarithmic spectrum of image on a log-log scale is defined as:
L ( u , v ) = log ( A ( u , v ) )
I(x, y) represents one image, the amplitude spectrum of image A(u, v) is obtained by:
A ( u , v ) = ( D C T ( I ( x , y ) ) )
and phase spectrum of image θ(u, v) is obtained by:
θ ( u , v ) = ( D C T ( I ( x , y ) ) )
where DCT(∙) is denoted as a discrete cosine transform.
Figure 8b is the extracted saliency region based on Hou’s method and Figure 8c is the saliency region based on the proposed saliency extraction algorithm. As shown in Figure 8b,c, the saliency region of the proposed algorithm is consistent with Hou’s.

2.3.2. The Proposed JND Model Based on the Saliency Weight Factor

Through using the proposed saliency region extraction algorithm, we continue to obtain the JND model based on saliency weight factor. Firstly, the object map is expressed as:
μ s = N N S ( x , y ) / ( N N )
where μ is N × N the normalized intensity of saliency.
K t h = β μ s
The selection of Kth is a trade-off problem between false alarm and the neglect of objects. According to [30] and subject experiments, we set β = 1.5 empirically. Thus, the object map O(x, y) in a saliency is defined as:
O ( x , y ) = { 1 , i f     S ( x , y ) K t h 0 , o t h e r w i s e
According to the characteristics of human subjective perception, small JND thresholds are utilized to saliency regions, while the large JND thresholds are utilized to non-saliency regions. Therefore, the saliency regions are inversely proportional to saliency weight factor γ and the γ is given by:
γ = A α μ 0
where the parameter α is measure factor. μ0 ∈ [0, 1] is normalized intensity of binary saliency image block with a size of N × N and expressed as:
μ o = N N O ( x , y ) / ( N × N )
According to (17), the JND threshold values of the current image block are not larger than the HGDE. Thus, the conditional relation is obtained as:
0 < γ J L M J L M H G D E ( x )
where J′LM is suppressed by HGDE, then the range of values γ can be deduced as:
0 < γ 1
According to (28) and (29), the range of parameter α can be obtained, that is α ∈ [0, 1], then the parameter A can be deduced as:
A = 1
in order to obtain a large control range, the parameter α = 0.9.
Therefore, the JND model based on saliency weight factor is given by:
J G D E S = γ J L M
As Figure 9b shows, one frame of CatRobot1 UHD video sequence is contaminated by the proposed JND values. Compared with Figure 9a, there are no obvious subjective visual distortion loss in Figure 9b. There are three colored (red, yellow and blue) boxes in Figure 9 to show the visual differences clearly. Figure 9c shows the distribution of J′LM threshold values, where the white regions represent that the JND values are more lager than the JND values in black or grey regions. Because the HGDE divides the image into complex texture regions and smooth regions, JND threshold values are mainly embedded in texture regions or edge regions of the image. Figure 9d shows the distribution of proposed JGDE-S threshold values. There are some saliency weight factors are incorporated into the J′LM model. Figure 9f shows the saliency image of Figure 9a, and as shown in Figure 9d, the edge region of saliency object is not embedded or embedded with small JND threshold values. Therefore, the edge regions of saliency object are protected to avoid the distortion of the edge region which could often cause significantly visual perception. Figure 9e shows that the difference region is the same as Figure 9b.
The temporal masking is not considered and incorporated into our proposed JND model through weighing encoding time and efficiency. Because the reuse of CU or TU block motion vectors are obtained in the HEVC recursively coding process, they are uncertain true motion vectors. Meanwhile, the precise temporal JND values are hard to calculate with these motion vectors, which may cause to bring additional visible distortions. Moreover, the calculation of the temporal masking model is computationally heavy because of the HEVC recursively coding process.

3. A JND-Based HEVC-Complaint PVC Scheme Using Distortion Compensation

3.1. Overall Architecture of the Proposed PVC Scheme

The proposed JGDE-S model is incorporated into an HEVC Test model (HM 16.9) reference software [31], and the JND-based perceptually HEVC-complaint video coding is proposed. Figure 10 shows the whole flowchart of proposed PVC encoder. Firstly, a CU block is split into some TU blocks according to the rate distortion (RD) cost. If the parent TU block is split into four sub-TU blocks further, then each sub-TU block is decided to continue splitting by RD cost of the parent TU and its children TUs respectively, else the transform coding process is carried out directly. The transform coding process is mainly divided into DCT transform and quantization. As shown in the “red dot box” of Figure 10, the average pixel value and the variance of the current N × N TU block is calculated, then the saliency factor is obtained in DCT domain, finally the proposed JGDE-S threshold is modeled.
Figure 11 shows the JND suppression example for 1-D transform coefficients. It is the distribution before JND suppression in Figure 11a, where the red dot line is represented the JND threshold. As (32) shows, if the amplitudes of transform coefficients |C(n, i, j)| < JGDE-S, then the after suppression transform coefficients |C′(n, i, j)| is set to zero. Otherwise, the |C′(n, i, j)| is equal to |C(n, i, j)| − JGDE-S. Through the above decision, the transform coefficients of complex texture regions or unperceived regions by the HVS are set to zero, and the transform coefficients of smooth regions or perceived regions are decreased by JND suppression. Since the zero or smaller transform coefficients are discarded in the quantization processing, the encoding bitrates are saved, and the coding compression efficiency is further to improved.
| C ( n , i , j ) | = { 0 , | C ( n , i , j ) | J G D E S 0 | C ( n , i , j ) | J G D E S , o t h e r w i s e
As Figure 12 shows, the “red bars” are the absolute values of 1-D transform coefficients before suppression and the “blue bars” are the absolute values of 1-D transform coefficients after suppression with proposed JND threshold. As Figure 12 shows, we can draw the following two conclusions. First, the amplitudes of “blue bars” are all lower than the “red bars” because of the suppression with proposed JND threshold values. Second, the distributions of the “blue bar” are sparser than “red bars”, because the lower absolute values of coefficients than the JND threshold values are suppressed to zero. This explains that parts of the transform coefficients are suppressed by proposed JND threshold values.

3.2. The Proposed JND-Based HEVC-Complaint PVC Scheme using Distortion Compensation

Because the residuals of CU blocks are utilized in HEVC recursively coding process, the subsequent coding performance is influenced on the reconstructed image quality. For the transform coefficient based on the JND suppression method, a certain distortion of the reconstructed image will be produced, and then the coding errors are accumulated recursively in the encoding process. Therefore, how to control the distortion caused by JND suppression is the key to determining the final coding performance. Obviously, the distortions directly affect the RD cost of encoding process, and RD cost is the basis for judging the selection of encoding mode and CU partition. In (33), the RD cost function JRDO is composed of distortion D, bitrate R and Lagrange factor λ, which has a relationship with quantization step.
J R D O = D + λ R
Kim et al. control the distortion caused by JND suppression through incorporating the distortion compensation factor into RDO of the encoding process [21]. However, Kim’s method needs to calculate the dequantization DCT coefficients without JND suppression, which increases computational complexity. The proposed distortion compensation factor only needs to calculate the absolute value of the difference between the DCT coefficients before and after suppression. Moreover, the distortion compensation control factor (DCCF) under different QPs are introduced to control the extent of distortion compensation more efficiently.
J R D O = 1 / ψ q ε D + λ R
Equation (34) is the proposed rate-distortion (RD) cost function J′RDO, where ε is the distortion compensation factor (DCF) and ψq is DCCF.
The DCF is calculated by (35), where the JGDE-S(n, i, j) represents the (i, j)-th JND threshold values of the n-th block.
ε = N N J G S E S ( n , i , j ) / N N Δ C ( n , i , j )
The ∆C″(n, i, j) is the absolute value of the difference between transform coefficient before suppression and transform coefficient after suppression and is defined as (36). The DCF is large when ∆C″(n, i, j) is small, which means that the distortion caused by JND suppression is small at this time. Therefore, a large distortion D and small DCCF are allowed in the RD cost function. Otherwise the distortion caused by JND suppression is large, then the distortion D in the RD cost function needs to be reduced.
Δ C ( n , i , j ) = | C ( n , i , j ) C ( n , i , j ) |
Usually, the coding distortion of the video is smaller when using a small QP than the use of a large QP. As shown in Figure 13a,b, the subjective and objective (PSNR) quality of encoded frame when using QP = 19 is higher than that of encoded frame when using QP = 40. The encoding process causes high coding distortion because of using a large QP, and there is no chance to further suppress perceptual redundancy. Therefore, the large DCF is adopted to the RDO process. Otherwise, the small DCF is adopted to the RDO process and more bitrates are saved. Therefore, a DCCF is applied to control the extent of DCF under different QPs. When the QP value is greater than 40, because the objective and subjective quality of encoded video decreases significantly, the small DCCF is utilized. In order to get the relationship between DCCF and QPs, the designed decision is shown in (37), where q is a vector of QPs and q = [22 27 32 37]. Therefore, ψq represents the difference of PSNR between q ≠ 37 and q = 37 or the difference of PSNR between q = 37 and q = 40.
ψ q = { P S N R q P S N R 37 , q 37 P S N R q P S N R 40 , q = 37
In Figure 14, the statistics are come from different scenarios of HD and UHD sequences at different QPs under the random access (RA) and low delay (LD) configurations. As shown in Figure 14, we can obtain the values of ψq (in ordinate) are large at small QPs and the curves are smooth at high QPs. Essentially, the objective and subjective quality of encoded video are high at the small QP and more perceptual distortions are acceptable. Thus, the DCF is reduced and DCCF should be increased appropriately. On the contrary, in order to ensure the subjective and objective quality of the encoded video, and allow for small perceptual distortion, the DCF should be increased while the DCCF should be reduced.
From the above statistical results, average values of ψq are set to DCCF under different QPs. Figure 15 shows the polynomial fitting curves based on the statistics in Figure 14. The ψq decrease with increasing QPs, and ψq is larger in LD configuration than in RA configuration, because the encoding usually operates with smaller prediction errors under the RA configuration than that under the LD configuration for different QPs. The ψq is given by:
ψ q = { 0.013 Q P 2 1.109 Q P + 24.641 , f o r       R A   0.015 Q P 2 1.294 Q P + 28.310 , f o r       L D  
where (38) shows the fact that the high ψq is utilized at small QPs. On the contrary, the small ψq is utilized at high QPs. It is consistent with the results shown in Figure 14.

4. Experimental Results

4.1. Verification of the Proposed JGDE-S Model

The HD and UHD videos are mainly provided by [32] and Joint Video Exploration Team (JVET). For a wide FOV of 4K UHD video/image, it is not appropriate to apply the adjectival categorical judgment (ACJ) method, which shows the reference images (left) and test images (right) at the same time. To verify the effectiveness of the proposed JGDE-S model for 4K UHD video/image, subjective viewing tests are conducted based on double-stimulus continuous quality-scale (DSCQS) where A (a reference sequence) and B (a sequence to be compared) are pseudo-randomly ordered for each presentation [33,34]. For the subjective of still picture, the sequence A consists of one frame from 4K UHD video sequence. Then sequence A is contaminated by (40) to form sequence B as shown in Figure 16. However, a 3–4 s sequence and five repetitions (voting during the last two) may be appropriate for still picture [33,34]. The display condition is the same as the one in Table 2. The viewing distance is set to 2.1 m which is appropriate for a 55-inch 4K UHD display. A total of 15 subjects have participated in the subjective quality assessment experiments where all of them have normal vision power.
According to [33,34], the both A and B are evaluated with subjective voting scores ranged from 0 to 100 for the worst and the best visual qualities, respectively. The differential mean opinion score (DMOS) value is defined as:
D M O S = M O S P V C M O S O R I
where MOSPVC and MOSORI are the measured mean opinion score (MOS) values from the image contaminated by the JND and the original image of sequence.
According to (40), the C(n, i, j) represent the (i, j)-th DCT coefficient of the n-th block, the C′(n, i, j) is the (i, j)-th DCT coefficient of the n block that is contaminated by random noise, Srand(n, i, j) is random noise with the value of +1 or −1 randomly.
C ( n , i , j ) = C ( n , i , j ) + S r a n d ( n , i , j ) J N D
Table 4 lists the comparison of the Kim’s CM-JND and the proposed JGDE-S models in terms of PSNR and DMOS values. As shown in Table 4, the average PSNR values of the Kim’s CM-JND and the proposed JND model are 52.77 dB and 49.82 dB respectively, and their corresponding DMOS values are both −0.1. That is, the proposed JND model produces 2.95 dB lower average PSNR at the same visual quality level compared with the Kim’s CM-JND model. The DMOS value of −0.1 indicates that most of subjects rarely distinguish the distorted images compared to the original ones. The results show that the proposed JGDE-S model is more appropriate to image with 10-bit depth than Kim’ JND model.

4.2. Objective and Subjective Performance Evaluation for the Proposed PVC Scheme

To verify the effectiveness of the proposed HEVC-compliant PVC scheme, the proposed JGDE-S model is implemented into HM 16.9, and it is compared with the original HM 16.9 and Kim’s PVC scheme [21]. Because Kim’s PVC is proposed in harmonization with HM 11.0 and also used in the RDO-based encoding process, the Kim’s PVC is re-implemented into HM 16.9 for fair comparison in this paper. The test sequence used for the experiments include eight different scenes of 4k UHD video sequences in the 4:2:0 color format. Because 10-bit depth is used as the mainstream for 4K UHD videos, the LD and RA configuration based on 10-bit depth are used for all experiments with a set of fixed quantization parameters, such as QP = 22, 27, 32, and 37.
We compare the original HM 16.9, Kim’s PVC and proposed PVC in terms of bitrate reduction and encoding time to verify the objective RD performance and encoder complexity. Then we also use DMOS values to verify subjective quality assessment. The bitrate reduction is represented as (41), between the original HM16.9 and the proposed (Kim’s and our) PVC scheme.
Δ R = R P V C R o r i R o r i × 100 %
and encoding time is represented as:
Δ T = T P V C T o r i T o r i × 100 %
where Rori and Tori are the bitrates and encoded time produced by the original HM 16.9. where RPVC and TPVC are the bitrates and encoded time produced by proposed or Kim’s PVC scheme. The proposed PVC scheme are also compared via subjective quality assessment experiments. For this, the DSCQS method is employed for subjective evaluation and the experimental set as Section 4.1. For movie video sequence, A and B are respectively presented to last 10 s and 3 s dummy videos are inserted between them as Figure 16 shows, where one test sequence is a reconstructed sequence encoded by the original HM 16.9, and the other is a reconstructed sequence encoded by proposed PVC or Kim’s PVC. The presentation order of the two test sequences are randomly selected for each presentation. The subjective voting scores are set the same as Section 4.1.
Table 5, Table 6 and Table 7 show the objective and subjective test results for the proposed PVC scheme, the Kim’s PVC scheme and the original HM 16.9 under LD and RA Main10 profiles, respectively. For performance in terms of bitrate reduction, the proposed PVC scheme outperforms Kim’s PVC scheme for all the test sequences and all the four QP values, yielding the average ∆R = 32.98%, the maximum ∆R = 80.89% under the LD profile and the average ∆R = 28.61%, the maximum ∆R = 66.04% under the RA profile for the ‘DaylightRoad2’ sequence at QP = 22. The maxim bitrate reduction obtained from the ‘DaylightRoad2’ sequence is due to the fact that it contains richer scenes and a lot of complex texture regions throughout all the frames. Therefore, it contains more noise which is insensitive to visual perception so that the proposed PVC scheme effectively suppresses the transform coefficients in those regions.
Table 5 and Table 6 show that the ∆R values are usually decreased as QP values increase for both PVC schemes. This is because the distortions introduced by quantization errors are high enough with QP values increasing, in order to maintain high objective quality, thus resulting in almost no room for coefficient suppression and making the coefficient suppression relatively less contribute to the total bitrate reduction. Compared to the Kim’s PVC scheme, the proposed PVC scheme achieves remarkably higher bitrate reductions with higher objective and subjective quality than Kim’s for all test sequences at different QPs. In order to reach a large extent of suppression for DCT coefficients, the JND threshold values are scaled to a large threshold values in Kim’s PVC. Therefore, the reconstruction distortions of TU residual block are increased to cause the lower objective and subjective quality than proposed PVC scheme. Given that the DCCF is utilized in the proposed PVC scheme to control the extern of distortions, the proposed PVC scheme more effectively suppresses the transform coefficients than the Kim’s JND suppression. In particular, this is because the objective and subjective video quality encoded by proposed PVC scheme is higher than Kim’s PVC scheme.
It is noticed that the average bitrate reduction ratios for the test PVC schemes under the RA profile are slightly smaller than those obtained under the LD profile. This is because the encoding under the RA profile usually operates with smaller prediction errors than that under the LD profile. That is, there exist smaller chances for both the proposed JGDE-S suppression and Kim’s suppression on the transform coefficients of residues under the RA profile.
As shown in Table 5 and Table 6, the proposed PVC scheme increases the total encoding time only with average 12.94% and 22.45% under LD Main10 and RA Main10 profiles compared to the original HM 16.9, respectively. Because the temporal masking is not considered and the proposed simple distortion compensation factor is utilized in our PVC scheme, the proposed PVC scheme is lower in complexity than Kim’s.
Table 7 shows that the PSNR is a little degradation than the original PSNR for the proposed PVC scheme, and Kim’s PVC scheme. However, the PSNR of the proposed PVC scheme is higher than Kim’s. Usually the objective quality is not very well correlated to perceived visual quality, thus subjective quality is also an important measurement to assessment PVC scheme. For the subjective test results shown in Table 7, Kim’s and the proposed PVC scheme have statistically almost zero average DMOS values, yielding rarely visible distortion. The subjective quality of the proposed PVC scheme is higher than Kim’s PVC scheme. For example, the character details and edges are clearer compare with Kim’s PVC scheme in the yellow boxes of Figure 17 and Figure 18, due to the effect of different QPs on distortion considered in the proposed PVC scheme.
From the objective and subjective test results, the proposed PVC scheme achieves significant bitrate reduction at the same perceptual quality with little increment in encoder complexity compared to the original HM 16.9, and outperforms the Kim’s PVC scheme about two times the average bitrate reduction ratio. The superiority of the proposed PVC scheme comes from the fact that the proposed JND model is more appropriate to UHD videos with 10-bit depth and the designed DCCF which minimize the bitrate for different QPs under high PSNR. Thus, the proposed PVC scheme can reduce more bitrates than Kim’s PVC scheme under the higher PSNR.

5. Conclusions

A perceptual friendly JND-based HEVC-compliant PVC scheme is proposed in this paper. To better reflect the UHD image/video perception characteristics of HVS, a simple JGDE-S model is proposed by appropriately subjective experiments. For example, a LM effect for UHD image/video with 10-bit depth is modeled, and it is suppressed by GDE filter which control the extent of imperceptible noise for image. In addition, a DCT-based saliency factor is added into proposed JND model according to the FOV of UHD image/video. A simple transform coefficient suppression method with JND model is proposed in harmonization with the transform and quantization process of HM16.9 in a HEVC-complaint manner. Perceptual distortion is also appropriately compensated in the RDO process by incorporating the proposed DCF. In order to efficiently control the extent of the distortion compensation based on different QPs, the DCCF is also incorporated to RDO process. In objective and subjective experiments, the proposed HEVC-compliant PVC scheme yielded a remarkable bitrate reduction of 32.98% average for LD configuration and 28.61% average for RA configuration with a negligible subjective quality loss. It only causes an average of 12.94% and 22.45% encoding time increase under LD and RA configuration compared to the original HM 16.9, respectively. However, the proposed PVC scheme can reduce more bitrates with higher objective and subjective quality assessment and is computationally faster than the Kim’s PVC. There are still some shortages and improvements in this paper, and we conclude as follow: (1) The proposed JND model is not considered in the chroma domain. (2) We only considered that the encoding OPs lead to the different extent of perceptual distortion in the RDO process. The other encoding characters are explored in our future work.

Author Contributions

X.C. designed and realized the algorithm and drafted the manuscript. Z.P. co-designed the algorithm and polished the manuscript. G.J. co-designed the algorithm. F.C. analyzed the experimental results. M.Y. polished the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant Nos. 61771269, 61620106012, 61871247, and 61671258, the Natural Science Foundation of Zhejiang Province, China under No. LY17F010005, and the K. C. Wong Magna Fund in Ningbo University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cheon, M.; Lee, J.S. Subjective and objective quality assessment of compressed 4K UHD videos for Immersive Experience. IEEE Trans. Circuits Syst. Video Technol. 2018, 28, 1467–1480. [Google Scholar] [CrossRef]
  2. Bae, S.-H.; Kim, J.; Kim, M.; Cho, S.; Choi, J.S. Assessments of Subjective Video Quality on HEVC-Encoded 4K-UHD Video for Beyond-HDTV Broadcasting Services. IEEE Trans. Broadcast. 2013, 59, 209–222. [Google Scholar] [CrossRef]
  3. Wang, M.; Xie, W.; Meng, X.; Zeng, H.; Ngan, K.N.; Chen, Z. UHD Video Coding: A Light-Weight Learning-Based Fast Super-Block Approach. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 3083–3094. [Google Scholar] [CrossRef]
  4. Koike, Y.; Inoue, A. High-speed graded-index plastic optical fibers and their simple interconnects for 4K/8K video transmission. J. Lightwave Technol. 2018, 34, 1551–1555. [Google Scholar] [CrossRef]
  5. Kim, Y.; Choi, J.-S.; Kim, M. 2X Super-Resolution Hardware Using Edge-Orientation-Based Linear Mapping for Real-Time 4K UHD 60 fps Video Applications. IEEE Trans. Circuits Syst. II Express Briefs 2018, 65, 1274–1278. [Google Scholar] [CrossRef]
  6. Yang, X.; Ling, W.; Lu, Z.; Ong, E.; Yao, S. Just noticeable distortion model and its applications in video coding. Signal Process. Image Commun. 2005, 20, 662–680. [Google Scholar] [CrossRef]
  7. Chen, Z.; Guillemot, C. Perceptually-Friendly H.264/AVC Video Coding Based on Foveated Just-Noticeable-Distortion Model. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 806–819. [Google Scholar] [CrossRef]
  8. Liu, A.; Lin, W.; Paul, M.; Deng, C.; Zhang, F. Just Noticeable Difference for Images With Decomposition Model for Separating Edge and Textured Regions. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 1648–1652. [Google Scholar] [CrossRef]
  9. Wu, J.; Shi, G.; Lin, W.; Liu, A.; Qi, F. Just Noticeable Difference Estimation for Images With Free-Energy Principle. IEEE Trans. Multimedia 2013, 15, 1705–1710. [Google Scholar] [CrossRef]
  10. Zhao, Y.; Chen, Z.; Zhu, C.; Tan, Y.P.; Yu, L. Binocular just noticeable-difference model for stereoscopic images. IEEE Signal. Process. Lett. 2011, 18, 19–22. [Google Scholar] [CrossRef]
  11. Zhang, L.; Peng, Q.; Wang, Q.H.; Wu, X. Stereoscopic perceptual video coding based on just-noticeable-distortion profile. IEEE Trans. Broadcast. 2011, 57, 572–581. [Google Scholar] [CrossRef]
  12. Wang, X.; Kwong, S.; Zhang, Y. Considering binocular spatial sensitivity in stereoscopic image quality assessment. In Proceedings of the 2011 Visual Communications and Image Processing (VCIP), Tainan, Taiwan, 6–9 November 2011; p. 36. [Google Scholar]
  13. Ahumada, A.J.; Peterson, H.A. Luminance-model-based DCT quantization for color image compression. In Proceedings of the SPIE—The International Society for Optical Engineering, San Diego, CA, USA, 30 July–1 August 1997; Volume 1666, pp. 365–374. [Google Scholar]
  14. Watson, A.B. DCTune: A technique for visual optimization of DCT quantization matrices for individual images. In Society for Information Display Digest of Technical Papers; XXIV; Society for Information Display: San Jose, CA, USA, 1993; Volume 24, pp. 946–949. [Google Scholar]
  15. Wei, Z.; Ngan, K.N. Spatio-temporal just noticeable distortion profile for grey scale image/video in DCT domain. IEEE Trans. Circuits Syst. Video Technol. 2009, 19, 337–346. [Google Scholar]
  16. Ma, L.; Ngan, K.N.; Zhang, F.; Li, S. Adaptive block-size transform based just-noticeable difference model for images/videos. Signal. Process. Image Commun. 2011, 26, 162–174. [Google Scholar] [CrossRef]
  17. Luo, Z.; Song, L.; Zheng, S.; Ling, N.H. 264/advanced video control perceptual optimization coding based on JND-directed coefficient suppression. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 935–948. [Google Scholar] [CrossRef]
  18. Bae, S.H.; Kim, M. A Novel DCT-Based JND Model for Luminance Adaptation Effect in DCT Frequency. IEEE Signal Process. Lett. 2013, 20, 893–896. [Google Scholar]
  19. Bae, S.H.; Kim, M. A novel generalized DCT-based JND profile based on an elaborate CM-JND model for variable block-sized transforms in monochrome images. IEEE Trans. Image Process. 2014, 23, 3227–3240. [Google Scholar] [PubMed]
  20. Bae, S.H.; Kim, M. A DCT-based total JND profile for spatiotemporal and foveated masking effects. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1196–1207. [Google Scholar] [CrossRef]
  21. Kim, J.; Bae, S.-H.; Kim, M. An HEVC-Compliant Perceptual Video Coding Scheme based on JND Models for Variable Block-sized Transform Kernels. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 1786–1800. [Google Scholar] [CrossRef]
  22. Bae, S.-H.; Kim, J.; Kim, M. HEVC-Based Perceptually Adaptive Video Coding Using a DCT-Based Local Distortion Detection Probability Model. IEEE Trans. Image Process. 2016, 25, 3343–3357. [Google Scholar] [CrossRef]
  23. Ki, S.; Bae, S.-H.; Kim, M.; Ko, H. Learning-Based Just-Noticeable-Quantization—Distortion Modeling for Perceptual Video Coding. IEEE Trans. Image Process. 2018, 27, 3178–3193. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Naccari, M.; Agrafiotis, D.; Mrak, M. High dynamic range video compression exploiting luminance masking. IEEE Trans. Circuits Syst. Video Technol. 2016, 26, 950–964. [Google Scholar] [CrossRef]
  25. Jiang, X.; Song, T.; Zhu, D.; Katayama, T.; Wang, L. Quality-Oriented Perceptual HEVC Based on the Spatiotemporal Saliency Detection Model. Entropy 2019, 21, 165. [Google Scholar] [CrossRef]
  26. Zeeshan, M.; Majid, M. High Efficiency Video Coding Compliant Perceptual Video Coding Using Entropy Based Visual Saliency Model. Entropy 2019, 21, 964. [Google Scholar] [CrossRef]
  27. Rust, B.W.; Rushmeier, H.E. A new representation of the contrast sensitivity function for human vision. In Proceedings of the International Conference on Imaging Science, Systems, and Technology, Las Vegas, NV, USA, 30 June–3 July 1997; pp. 1–15. [Google Scholar]
  28. Zhang, X.; Lin, W.; Xue, P. Improved estimation for just-noticeable visual distortion. Signal Process. 2005, 85, 795–808. [Google Scholar] [CrossRef]
  29. Thomas, M.C.; Joy, A.T. Elements of Information Theory, 2nd ed.; Wiley-Blackwell: New York, NY, USA, 1991; pp. 224–231. [Google Scholar]
  30. Hou, X.; Zhang, L. Saliency detection: A spectral residual approach. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  31. HM. HM Reference Software 16.9. Available online: https://hevc.hhi.fraunhofer.de/svn/svn_HEVCSoftware/tags/HM-16.9/ (accessed on 15 June 2018).
  32. Petrinšak, L.; Grbić, R.; Kaprocki, Z.; Subotić, M. Challenges in 4K UHD multimedia device testing. In Proceedings of the 2017 Zooming innovation in Consumer Electronics International Conference (ZINC), Novi Sad, Serbia, 31 May–1 June 2017; pp. 5–8. [Google Scholar]
  33. Series, B.T. Methodology for the Subjective Assessment of the Quality of Television Pictures. Recommendation ITU-R BT. Available online: http://www.itu.int/dms_pubrec/itu-r/rec/bt/R-REC-BT.500-12-200909-S!!PDF-E.pdf (accessed on 28 December 2018).
  34. Parameter Values for Ultra-High Definition Television Systems for Production and International Programme Exchange. Recommendation ITU-R BT. Available online: https://www.itu.int/dms_pubrec/itu-r/rec/bt/R-REC-BT.2020-2-201510-I!!PDF-E.pdf (accessed on 2 March 2019).
Figure 1. Overall framework of perceptual video coding with the JND suppression.
Figure 1. Overall framework of perceptual video coding with the JND suppression.
Entropy 21 01095 g001
Figure 2. A test image exemplar, (a) a test image patch in a fovea region (P1), (b) the distortions injected into P1.
Figure 2. A test image exemplar, (a) a test image patch in a fovea region (P1), (b) the distortions injected into P1.
Entropy 21 01095 g002
Figure 3. The subject experiment.
Figure 3. The subject experiment.
Entropy 21 01095 g003
Figure 4. Comparison with the other luminance masking (LM) effect model in the grey luminance range.
Figure 4. Comparison with the other luminance masking (LM) effect model in the grey luminance range.
Entropy 21 01095 g004
Figure 5. The GDE for one frame of ShakeNDry sequence; (a) 8-bit depth, (b) 10-bit depth
Figure 5. The GDE for one frame of ShakeNDry sequence; (a) 8-bit depth, (b) 10-bit depth
Entropy 21 01095 g005
Figure 6. One frame of the TrafficFlow sequence is segmented by HGDE (a) original grey image with 8-bit depth, (b) segmentation image with eight-bit depth, (c) original grey image with 10-bit depth, (d) segmentation image with 10-bit depth.
Figure 6. One frame of the TrafficFlow sequence is segmented by HGDE (a) original grey image with 8-bit depth, (b) segmentation image with eight-bit depth, (c) original grey image with 10-bit depth, (d) segmentation image with 10-bit depth.
Entropy 21 01095 g006
Figure 7. The JND threshold values distribution of TrafficFlow 3840 × 2160 video sequence with 10-bit depth.
Figure 7. The JND threshold values distribution of TrafficFlow 3840 × 2160 video sequence with 10-bit depth.
Entropy 21 01095 g007
Figure 8. Comparison of the extracted saliency region with Hou’s and the proposed method; (a) original image, (b) based on FT, (c) based on DCT.
Figure 8. Comparison of the extracted saliency region with Hou’s and the proposed method; (a) original image, (b) based on FT, (c) based on DCT.
Entropy 21 01095 g008
Figure 9. The fourth frame of the CatRobot1 (3840 × 2160) video sequence is contaminated by the proposed JGDE-S model. (a) Original image, (b) contaminated image, (c) J′LM image (without saliency weight factor), (d) JGDE-S image (with saliency weight factor), (e) difference image of (a) and (b), (f) saliency image.
Figure 9. The fourth frame of the CatRobot1 (3840 × 2160) video sequence is contaminated by the proposed JGDE-S model. (a) Original image, (b) contaminated image, (c) J′LM image (without saliency weight factor), (d) JGDE-S image (with saliency weight factor), (e) difference image of (a) and (b), (f) saliency image.
Entropy 21 01095 g009
Figure 10. Flowchart of the proposed perceptual visual coding (PVC) encoder.
Figure 10. Flowchart of the proposed perceptual visual coding (PVC) encoder.
Entropy 21 01095 g010
Figure 11. JND-directed suppression example for 1-D transform coefficients. (a) Before suppression, (b) after suppression.
Figure 11. JND-directed suppression example for 1-D transform coefficients. (a) Before suppression, (b) after suppression.
Entropy 21 01095 g011
Figure 12. A 32 × 32 residual transform unit (TU) coding block of the BasketballPass video sequence under the random access (RA) main10 configuration at quantization parameter (QP) = 22.
Figure 12. A 32 × 32 residual transform unit (TU) coding block of the BasketballPass video sequence under the random access (RA) main10 configuration at quantization parameter (QP) = 22.
Entropy 21 01095 g012
Figure 13. The 34th encoded frame of the Cactus video sequence with QP = 19 and QP = 40 under the RA mian10 configuration of HM 16.9. (a) The subjective and objective quality (PSNR) of the encoded frame is 40.10 dB with QP = 19; (b) the PSNR of encoded frame is 31.30dB with QP = 40.
Figure 13. The 34th encoded frame of the Cactus video sequence with QP = 19 and QP = 40 under the RA mian10 configuration of HM 16.9. (a) The subjective and objective quality (PSNR) of the encoded frame is 40.10 dB with QP = 19; (b) the PSNR of encoded frame is 31.30dB with QP = 40.
Entropy 21 01095 g013
Figure 14. ψq with different QPs; (a) based on RA configuration, (b) based LD configuration.
Figure 14. ψq with different QPs; (a) based on RA configuration, (b) based LD configuration.
Entropy 21 01095 g014
Figure 15. The polynomial fitting curves based on RA and LD configuration.
Figure 15. The polynomial fitting curves based on RA and LD configuration.
Entropy 21 01095 g015
Figure 16. The double-stimulus continuous quality-scale (DSCQS) method for the movie video sequence.
Figure 16. The double-stimulus continuous quality-scale (DSCQS) method for the movie video sequence.
Entropy 21 01095 g016
Figure 17. The 12th encoded frame of ToddlerFountain2 video sequence with QP = 32 under the LD mian10 configuration. (a) The encoded image with original HM 16.9, (b) the encoded image with Kim’ PVC, (c) the encoded image with proposed PVC.
Figure 17. The 12th encoded frame of ToddlerFountain2 video sequence with QP = 32 under the LD mian10 configuration. (a) The encoded image with original HM 16.9, (b) the encoded image with Kim’ PVC, (c) the encoded image with proposed PVC.
Entropy 21 01095 g017
Figure 18. The 47th encoded frame of Tango2 video sequence with QP = 32 under the RA mian10 configuration. (a) The encoded image with original HM 16.9, (b) the encoded image with Kim’s PVC, (c) the encoded image with the proposed PVC.
Figure 18. The 47th encoded frame of Tango2 video sequence with QP = 32 under the RA mian10 configuration. (a) The encoded image with original HM 16.9, (b) the encoded image with Kim’s PVC, (c) the encoded image with the proposed PVC.
Entropy 21 01095 g018
Table 1. Description of the symbols in Figure 1.
Table 1. Description of the symbols in Figure 1.
SymbolDescription
HCSF(wi,j, φi,j)Basic contrast sensitivity function (CSF) model
MFLM(wi,j, μp)Luminance masking effects
JLMThe proposed luminance masking effects
HGDEThe proposed Gauss differential entropy (GDE) filter
S(wi,j)Modulation factor for saliency
JGDE-SThe proposed JND model
wi,j,
φi,j,
μp
Cycles per degree,
Directional angle,
Average pixel intensity of N × N TU block
Table 2. Experimental setup.
Table 2. Experimental setup.
Display InformationDisplayCHANGHONG (55Q3R)
Type and SizeLED 55’’
Resolution3840 × 2160
Dynamic contrast ratio7000:1
Number of subjects15 subjects (male: 10, female: 5, average age: 25, min: 22, max age: 33)
Viewing distance1.2 m
Ambient illuminance200 lux
DCT transform optionFloating point, 8 × 8 block size
Test image size (P2)256 × 256 pixels (covering a parafovea region)
Test patch size (P1)32 × 32 pixels (resided in a fovea region)
Table 3. Statistics of image segmentation region.
Table 3. Statistics of image segmentation region.
One Frame of Ultra-High Definition (UHD) Video SequenceTexture
> σth
Smooth
σth
Bosphorus (1920 × 1080-8bit)10.9%89.1%
Bosphorus (3840 × 2160-8bit)3.9%96.1%
Bosphorus (3840 × 2160-10bit)96.0%4.0%
ShakeNDry (1920 × 1080-8bit)36.8%63.2%
ShakeNDry (3840 × 2160-8bit)11.9%88.1%
ShakeNDry (3840 × 2160-10bit)95.6%4.4%
HoneyBee (1920 × 1080-8bit)12.6%87.4%
HoneyBee (3840 × 2160-8bit)3.4%96.6%
HoneyBee (3840 × 2160-10bit)99.9%0.1%
Table 4. Subjective experiments for 10-bit 4K UHD image.
Table 4. Subjective experiments for 10-bit 4K UHD image.
Test Image
(3840 × 2160)
10-Bit Depth
KimPro.
PSNR (dB)Differential Mean Opinion Score (DMOS)PSNR (dB)DMOS
Beauty50.42−0.148.25−0.1
Bosphorus53.56−0.150.82−0.0
HoneyBee53.28−0.049.05−0.0
Jockey53.03−0.149.40−0.1
ReadySetGo52.75−0.049.95−0.1
ShakeNDry52.78−0.048.71−0.0
YachtRide53.62−0.150.66−0.1
TrafficFlow52.74−0.149.48−0.1
Campfire52.74−0.052.08−0.0
Avg.52.77−0.149.82−0.1
Table 5. Bit rate and encoding complexity comparison of LD configuration.
Table 5. Bit rate and encoding complexity comparison of LD configuration.
SequenceQPBitrates (kpbs)R (%)T (%)
HMKimPro.KimPro.KimPro.
Tango22237,17325,33913,945−31.83−62.499.924.68
2711,28293277441−17.33−34.0517.618.70
32571554855150−4.02−9.8924.9322.54
37320531503139−1.72−2.0624.9423.95
Drums10022152,61355,72637,854−63.49−75.205.44−12.7
2737,86126,51617,501−29.96−53.7815.338.01
3215,04113,74712,419−8.60−17.4324.3920.10
37744172256850−2.90−7.9425.6222.19
Campfire22100,79851,27934,100−49.13−66.1710.68−6.70
2729,61224,23615,683−18.15−47.0415.655.47
3213,14112,45311,208−5.24−14.7129.0918.88
37735971526585−2.81−10.5230.2422.13
Toddler-Fountain222272,351184,980175,151−32.08−35.6913.128.91
27133,18879,83178,904−40.06−40.7616.4910.88
3260,89938,52651,470−36.74−15.4819.6524.38
3726,59618,69024,087−29.73−9.4322.7527.09
CatRobot2244,76325,64914,713−42.70−67.137.62−1.80
2712,31699807414−18.97−39.8010.7610.24
32608256925097−6.41−16.2025.8521.83
37330632222930−2.54−11.3726.0123.85
Traffic-Flow22984860943448−38.12−64.999.631.49
27285523631616−17.23−43.4015.5415.54
32125411321085−9.73−13.4825.3521.52
37630628614−0.32−2.5426.0523.73
Daylight-Road22283,29927,49715,922−66.99−80.897.84−9.72
2714,54999707173−31.42−50.7015.267.48
32660858315127−11.76−22.4126.1921.88
37336430542881−9.22−14.3627.4924.45
Roller-coaster22231,68922,89312,844−27.76−59.474.89−2.20
2711,67210,4506333−10.47−45.7413.105.32
32520050404457−3.77−14.2923.3119.22
37268826462528−1.56−5.9524.8922.84
Avg.-−21.02−32.9818.6112.94
Table 6. Bit rate and encoding complexity comparison of RA configuration.
Table 6. Bit rate and encoding complexity comparison of RA configuration.
SequenceQPBitrates (kpbs)R (%)T (%)
HMKimPro.KimPro.KimPro.
Tango22235,84822,11516,659−38.31−53.5317.748.34
2710,51192247399−12.24−29.6121.0223.32
32538650674776−5.92−11.3338.4132.22
37307230402917−1.04−5.0539.0834.53
Drums10022125,14570,48945,109−43.67−63.958.65−3.01
2736,42926,44921,078−27.40−42.1420.2618.51
3215,60814,09013,229−9.73−15.2441.0229.86
37800378707414−1.61−7.3641.4431.83
Campfire2279,10050,37034,004−36.32−57.0111.221.93
2726,39521,73314,980−17.66−43.2519.6313.85
3211,41410,3659655−9.19−15.4143.8428.96
37648562235765−4.04−11.1046.7033.55
Toddler-Fountain222236,874171,444151,512−27.62−36.0325.5610.90
27109,78578,27267,705−28.70−38.3324.6515.26
3249,08536,67541,910−25.28−14.6225.5430.67
3721,77618,75619,609−13.87−9.9532.5734.04
CatRobot2246,84235,78320,886−23.61−55.4111.378.47
2713,24511,5269017−12.98−31.9211.6424.02
32675662425749−7.61−14.9141.9033.67
37380437643409−1.05−10.3842.0635.20
Traffic-Flow2214,15111,2775928−20.31−58.1118.4613.10
27396533392676−15.79−32.5131.5821.22
32185017541627−5.19−12.0544.4734.81
37971942923−2.97−4.9444.1736.76
Daylight-Road22280,46043,42427,325−46.03−66.0411.87−2.45
2715,40112,1099201−21.38−40.2628.3818.03
32724261425873−15.19−18.9037.6828.90
37382135293365−7.64−11.9338.5732.01
Roller-coaster22228,27420,33614,506−28.08−48.6916.339.80
2711,18590316887−19.26−38.4317.7319.30
32531950114544−5.79−14.5738.3130.30
37289128382816−1.83−2.5940.7530.66
Avg.----−16.79−28.6129.1422.45
Table 7. Objective and subjective quality comparison of LD and RA configurations.
Table 7. Objective and subjective quality comparison of LD and RA configurations.
ConfigurationLDRA
SequenceQPPSNR (dB)DMOSPSNR (dB)DMOS
HMKimPro.KimPro.HMKimPro.KimPro.
Tango22240.4638.9338.91−0.1−0.140.3138.8039.26−0.1−0.1
2739.5938.0137.85−0.6−0.239.5737.9438.40−0.6−0.3
3238.4636.6337.56−0.8−0.138.5636.6437.85−0.8−0.1
3736.9035.0536.14−0.8−0.337.1535.1436.55−0.8−0.5
Drums1002242.3137.0838.050.00.041.4737.2138.630.00.0
2739.3035.5736.31−0.2−0.139.1035.8637.13−0.2−0.1
3236.9734.0135.91−0.2−0.237.0834.2436.23−0.2−0.1
3734.6232.3333.88−0.4−0.234.9432.6134.29−0.4−0.2
Campfire2240.2836.0036.46−0.2−0.239.3235.7636.50−0.4−0.3
2737.3834.6735.20−0.3−0.337.0734.3935.11−0.3−0.5
3235.6733.2434.86−0.4−0.435.4332.9434.59−0.5−0.5
3733.8831.9033.29−0.4−0.333.6531.6733.02−0.5−0.3
Toddler-Fountain22240.7832.1336.290.00.039.7831.6835.370.00.0
2737.0629.8932.90−0.2−0.135.8829.6232.12−0.2−0.1
3233.5228.2932.41−0.3−0.232.5329.6331.52−0.3−0.2
3730.5926.9930.01−0.5−0.329.9026.7329.34−0.5−0.3
CatRobot2240.2838.2138.280.00.040.2338.1439.030.00.0
2738.9637.1136.95−0.10.039.1737.1837.93−0.30.0
3237.3435.6536.42−0.30.237.7535.8237.02−0.40.2
3735.3133.8434.56−0.5−0.135.9034.0635.26−0.7−0.1
Traffic-Flow2239.8438.7038.48−0.1−0.439.8938.4938.90−0.1−0.5
2738.6237.8037.46−0.2−0.438.8237.6637.97−0.3−0.5
3237.2736.5836.66−0.1−0.337.5536.3437.05−0.6−0.3
3735.6335.0135.120.0−0.236.0234.8035.61−0.7−0.2
Daylight-Road22238.0436.0336.130.00.037.9136.0236.630.00.0
2736.8735.1235.11−0.10.036.9535.3035.84−0.20.0
3235.6533.9834.75−0.4−0.135.8934.2635.23−0.5−0.1
3734.0332.5933.28−0.5−0.234.4132.8533.86−0.7−0.2
Roller-coaster22245.1741.1441.18−0.2−0.144.7641.2242.22−0.2−0.1
2742.8139.3439.09−0.4−0.342.7039.5740.13−0.5−0.3
3240.1937.4038.54−0.5−0.340.3337.5939.06−0.5−0.3
3737.6135.3536.06−0.6−0.437.9235.4536.98−0.7−0.4
Avg.-37.8635.1436.07−0.3−0.237.7535.1836.39−0.4−0.2

Share and Cite

MDPI and ACS Style

Cui, X.; Peng, Z.; Jiang, G.; Chen, F.; Yu, M. Perceptual Video Coding Scheme Using Just Noticeable Distortion Model Based on Entropy Filter. Entropy 2019, 21, 1095. https://0-doi-org.brum.beds.ac.uk/10.3390/e21111095

AMA Style

Cui X, Peng Z, Jiang G, Chen F, Yu M. Perceptual Video Coding Scheme Using Just Noticeable Distortion Model Based on Entropy Filter. Entropy. 2019; 21(11):1095. https://0-doi-org.brum.beds.ac.uk/10.3390/e21111095

Chicago/Turabian Style

Cui, Xin, Zongju Peng, Gangyi Jiang, Fen Chen, and Mei Yu. 2019. "Perceptual Video Coding Scheme Using Just Noticeable Distortion Model Based on Entropy Filter" Entropy 21, no. 11: 1095. https://0-doi-org.brum.beds.ac.uk/10.3390/e21111095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop