Next Article in Journal
Symmetric Face Normalization
Previous Article in Journal
Improving Accuracy of the Kalman Filter Algorithm in Dynamic Conditions Using ANN-Based Learning Module
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

No-Reference Image Quality Assessment with Local Gradient Orientations

Department of Computer and Control Engineering, Rzeszow University of Technology, W. Pola 2, 35-959 Rzeszow, Poland
Submission received: 22 December 2018 / Revised: 12 January 2019 / Accepted: 13 January 2019 / Published: 16 January 2019

Abstract

:
Image processing methods often introduce distortions, which affect the way an image is subjectively perceived by a human observer. To avoid inconvenient subjective tests in cases in which reference images are not available, it is desirable to develop an automatic no-reference image quality assessment (NR-IQA) technique. In this paper, a novel NR-IQA technique is proposed in which the distributions of local gradient orientations in image regions of different sizes are used to characterize an image. To evaluate the objective quality of an image, its luminance and chrominance channels are processed, as well as their high-order derivatives. Finally, statistics of used perceptual features are mapped to subjective scores by the support vector regression (SVR) technique. The extensive experimental evaluation on six popular IQA benchmark datasets reveals that the proposed technique is highly correlated with subjective scores and outperforms related state-of-the-art hand-crafted and deep learning approaches.

The recent advancement of digital imaging has stimulated a tremendous growth in the use of visual information for communication [1,2,3]. Therefore, it is essential to develop reliable automatic image quality assessment (IQA) measures for the evaluation of results of image processing methods for the acquisition, storage, transmission, restoration, or enhancement. The main role of IQA measures is to provide the objective assessment and replace cumbersome tests with human subjects [4]. The IQA measures are divided into three categories, based on the availability of reference images [1,5,6]: full-reference (FR), reduced-reference (RR), and no-reference (NR) approaches. In an FR-IQA measure, a distorted image is compared with its reference image, while only some statistics of the distortion-free image are available in the RR-IQA case. The peak signal-to-noise ratio (PSNR) is often used as an FR-IQA model due to its simplicity. However, it weakly correlates with human perception [5]. Therefore, more suitable measures have been developed that employ structural information [7], image statistical properties [8], visual saliency maps [9,10], structure and contrast changes [11], phase congruency [12], distortion distribution [13], or other measures [14,15]. The RR-IQA measures use only a part of reference data [16].
In this work, the discussion is confined to NR approaches, which are considered challenging and highly desired due to their applicability in absence of reference images. Many measures are devoted to evaluating the perceptual quality of images distorted by Gaussian white noise, JPEG compression, contrast change, or Gaussian blur [17,18]. Since their practical application is limited and based on the prior knowledge of distortion types, general-purpose NR methods have been developed. Many of them assume that statistical regularities of natural images can be reflected by natural scene statistics (NSS), as the Human Visual System (HVS) is very sensitive to local regularities [19]. Consequently, NSS characteristics of distorted images are used for the IQA in many domains, e.g., Discrete Cosine Transform (DCT) [20], wavelet [21,22], or spatial [23]. A variety of gradient-based features are often employed to model NSS [24]. Furthermore, the use of perceptual features [25,26,27] or image patches [28] can be found in the literature.Since the supervised learning bridges image statistics with the perceptual quality, it is often applied to obtain a model used for the quality prediction. For the learning, the Support Vector Regression (SVR) [20,21,23,27], neural networks [24], or random forests [29] are applied. In methods that do not use supervised learning, distortion types are modeled with a set of centroids of quality levels or NSS from multiple cues [30,31]. Another direction is to employ a pseudo-reference image which is created and compared with a distorted image with blockiness, sharpness, and noisiness metrics [32]. Recently, many NR-IQA approaches which use deep neural network (DNN) architectures have been introduced. They merge the feature extraction and quality prediction steps. However, they suffer from a small number of training examples available in IQA benchmark datasets or use complex architectures that are devoted to image recognition tasks. To overcome these limitations, most of them use image patches [33,34], train models using FR-IQA measures instead of subjective scores [34,35], or perform fine-tuning to adopt an architecture to the IQA [36]. Interestingly, some DNN-based approaches use features introduced in earlier methods [35].
The HVS is sensitive to local structures, which are often described using local binary patterns (LBP) and gradient-based statistics. However, a spatial distribution of LBP may not be able to capture more complex structures [37]. Thus, statistics extracted from gradient maps often occur in conjunction with other approaches to improve the IQA performance [27,38]. Such techniques use global distributions of gradient magnitude maps [25], relative gradient orientations or magnitude [24,39].
To describe an image and efficiently take into account local gradient orientations, Histogram of Oriented Gradients (HOG) descriptor can be used [40]. However, the HOG produces high-dimensional feature vectors, which are devoted to object recognition tasks due to their discriminative capabilities. Consequently, considering its application to the IQA, it is worth noticing that an original image content of an assessed image, which is described using the HOG, may influence the quality prediction performance. The descriptor also strongly depends on the size of processed image blocks and the described neighborhood. Since the image gradient orientation captured by the HOG and its relevance for the NR-IQA is interesting and still seems largely uninvestigated, in this paper, a novel no-reference technique for the image quality assessment with a SEt of Histogram of OriEnted GRadients (HOG) descriptors [40] (SEER) is introduced. In the SEER, in contrary to a widely accepted application of the HOG, an image is described by a set of descriptors which are obtained taking into account different local neighborhoods. In other words, each feature vector is composed of histograms of gradient orientations calculated for image regions (cells), which are arranged together with their neighboring cells in blocks. The descriptors in the set consider different sizes of image regions and blocks. Then, each descriptor is characterized by a histogram, seen as perceptual features. In a typical image recognition system, high-dimensional descriptors are often compared with each other. However, to apply the HOG to the NR-IQA, such comparison cannot take place since a distortion-free image and its descriptors are unavailable. Furthermore, the feature vectors in the HOG are designed to discriminate objects in images. Therefore, in the SEER, to train the SVR model, statistics of descriptors are employed instead of high-dimensional vectors. To improve the IQA performance, the method processes luminance and chrominance channels of an image, as well as their high-order derivatives.
The extensive experimental evaluation of the introduced NR measure against the related state-of-the-art techniques on six popular large-scale IQA benchmark datasets, which contain various distortion types, demonstrates that the SEER provides the higher quality prediction accuracy than compared NR models and is consistent with subjective scores. The method was evaluated against hand-crafted and deep learning NR measures.
The rest of this paper is arranged as follows. Section 1 reviews previous work on NR-IQA and Section 2 describes the proposed measure. Section 3 presents and discusses the experimental results obtained for the SEER and the related measures on TID2013 [41], TID2008 [42], CSIQ [43], LIVE [7], and LIVE In the Wild Image Quality Challenge (LIVE WIQC) [44] datasets. Finally, Section 4 concludes the paper.

1. Related Work

In this section, a brief review of previous studies closely related to the introduced work is presented.
The introduced SEER is based on gradient processing. However, there are many works which employ other features. For example, Moorthy and Bovik proposed a two-stage framework, in which distortion type is predicted and then an image is evaluated [21]. Saad et al. trained a probabilistic model with DTC-based NSS as a single-stage framework [20]. The generalized Gaussian distribution used for capturing NSS with locally normalized luminance coefficients is employed in Blind/Referenceless Image Spatial QUality Evaluator (BRISQUE) [23]. A scheme which combines artifact-specific metrics and employs a generalized Laplace distribution of the difference of two adjacent pixel values in an image was introduced by Fang et al. [45]. The measure uses three transductive 2 k-nearest neighbor algorithms to map the metrics into subjective scores. Normalized gradients magnitude and Laplacian of Gaussian responses were jointly used by Xue et al. [25]. Le et al. [46], in turn, used a histogram of local binary patterns (LBP) obtained for a gradient map. They also used LBP extracted from texture and structural maps [38]. A more advanced gradient-based image descriptor, Speeded-Up Robust Features (SURF), is employed in the measure proposed by Oszust [27]. In that work, the sample mean, standard deviation, entropy, skewness, kurtosis, and histogram variance for the assessed image, the image filtered with Prewitt operators, and their SURF features are used. In Optimized filteRing with binAry desCriptor for bLind imagE quality assessment technique (ORACLE) [47], in turn, a data-driven filtering based on the appearance of Features From Accelerated Segment Test (FAST) in grayscale images is proposed. The histograms of Fast Retina Keypoint (FREAK) descriptors for keypoints detected in filtered images are used to characterize the assessed content. The sample mean, standard deviation, and histogram variance of raw local patches describing FAST keypoints in images filtered with the bilaplacian operator in the YCbCr color space are used in stATistics of pixel blocks of local fEatuRes (RATER) measure [48]. Unlabeled data for learning Gabor features and modeling an image using the soft-assignment coding with the max pooling are employed in [49]. In another method that uses a codebook, High Order Statistics Aggregation (HOSA) [28], K-means clustering of normalized image patches and their description with the low and high order statistics are considered. The HOSA uses the soft assignment for an image representation and trains the SVR model for the prediction. Screen content images were assessed by Lu and Li [50] using orientation selectivity mechanism for extraction of orientation features.
Gradient-based techniques are often employed to provide effective IQA measures [11,24,25,46]. These measures use global distributions of gradient magnitude maps [25], relative gradient orientations or magnitude [24,39]. For example, in the Oriented Gradients Image Quality Assessment (OG-IQA) index, in which a correlation structure of image gradient orientations [24] is employed to train a quality model with AdaBoosting-backpropagation neural network. In that work, histograms of gradient magnitude, relative gradient orientation, and relative gradient magnitude maps are used and characterized using the histogram variance. These gradient-based measures do not take into account local gradient distributions. Such distributions are used in the HOG to provide a feature vector composed of 1-D histograms of gradient directions of pixels within image regions (cells) [40]. In the literature, there are FR-IQA measures which compare a reference image with the distorted image, calculating a distance between corresponding HOG vectors [51] or produce a weight map for the FR SSIM index [52]. In the technique proposed by Ahn et al. [18], in turn, blurred images are assessed by comparing HOG vectors approximated by a random sample consensus set (RANSAC).
Taking into account the referred works, it can be stated that the effectiveness of the HOG in NR image quality prediction remains largely uninvestigated, and a promising application of this descriptor to the NR-IQA is introduced for the first time in this paper.

2. Proposed NR Measure

In this section, the feasibility of using the HOG for the NR-IQA is investigated. The image processing steps in the SEER are explained in details in Figure 1, while in Figure 2 the block diagram of the proposed method is shown. As illustrated, the measure uses the HOG and provides feature vectors for luminance and chrominance channels of a distorted image, as well as for channels filtered using two bilaplacian operators. The measure uses differently sized image regions and the way the regions are combined together in the HOG descriptors. In the SEER, the K HOG features are obtained for each channel and their histograms are used to train a quality prediction model.

2.1. Local Gradient Orientations

The HVS is sensitive to variations in local structures [19,39]. Such sensitivity and the resultant subjective perception of an image are related to local semantic structural information which forms primitives in V1 [39]. Local distributions of gradients are used in the HOG to characterize an image [40]. Specifically, a 2D image I is convolved with 1-D Prewitt filters in the horizontal ( h P h = [ 1 , 0 , 1 ] ) and vertical ( h P v = [ 1 , 0 , 1 ] T ) directions to obtain gradients. Then, the edge magnitude is calculated as g = g x 2 + g y 2 , and its orientation is obtained as θ = a r c t a n ( g y / g x ) . The orientation is then transformed to [ 0 , 180 ] degrees range, ensuring that the opposite directions are assigned the same angle. The I is divided into adjacent, non-overlapping cells of size C C 1 × C 2 and, for each cell, the gradient orientations are binned into o bins with votes based on their magnitudes. To reduce aliasing, each pixel contributes to adjacent bins a fraction of its gradient magnitude. To reduce contrast changes, in turn, histograms for cells are normalized. Since gradient magnitudes carry the information about the described object in a region, to preserve the information, cells are grouped into overlapping blocks. There are B B 1 × B 2 cells in each block. Then, B 1 × B 2 cell histograms are concatenated and normalized using L 2 -norm. Finally, the obtained features for blocks are concatenated and normalized. This normalization makes the descriptor invariant against overall image contrast [40]. The division of an image into blocks and cells is shown in Figure 3. The number of overlapping cells between adjacent blocks is equal to A A 1 × A 2 , where A 1 × A 2 = [ B 1 × B 2 ] / 2 . The number of dimensions in the resulting feature vector D is calculated as:
l e n ( D ) = o B 1 B 2 · M C 1 B 1 B 1 A 1 + 1 · N C 2 B 2 B 2 A 2 + 1 ,
where M and N denote the height and width of an image, respectively. If A 1 is equal to B 1 , or A 2 to B 2 , the length of the vector is calculated without the subtraction in the denominator. For an exemplary image I 512 × 512 , the lengths of the feature vectors for the descriptors D ( C 2 × 2 , B 2 × 2 ) and D ( C 4 × 4 , B 2 × 2 ) , calculated with o = 9 , are 2,340,900 and 580,644, respectively. Such high-dimensional feature vectors are designed for robust object recognition and they cannot be used to train a quality model with the SVR. Therefore, it is shown in this paper that their histograms are sensitive to image degradation, consequently leading to the efficient application of the HOG to the NR-IQA.

2.2. Feature Extraction

In the introduced NR measure, a distorted RGB image is converted to the YCbCr color space. The YCbCr color space is specified in ITU-R BT.601 and selected as the preferred format for video broadcasting with the efficient use of the channel bandwidth [53]. Following the finding that the image filtering can enhance the image quality prediction of NR-IQA measures [47] and the bilapacian operator can be used to capture more information about described image regions for the IQA than it can be achieved with other filters [48], the SEER uses the YCbCr color space and filters color channels of the assessed image with two bilaplacian kernels Δ a 2 and Δ b 2 . The bilaplacian is obtained by convolving two Laplacian kernels. The following kernels are used:
Δ 1 = 0 1 0 1 4 1 0 1 0 , Δ 2 = 1 2 1 2 4 2 1 2 1 , Δ 3 = 1 0 1 0 4 0 1 0 1 , and Δ 4 = 2 1 2 1 4 1 2 1 2 .
Finally, Δ a 2 = Δ 1 Δ 3 and Δ b 2 = Δ 2 Δ 4 , where “*” denotes the convolution. Consequently, a filtered channel (e.g., Δ a 2 ( Y ) ) is obtained using the convolution Δ a 2 Y .
To show that the HOG technique provides different feature vectors for YCbCr components, Figure 4 presents an exemplary distorted image and the visualization of the HOG obtained for a small square image patch. For the visualization, a grid of rose plots is used. In the grid, each rose expresses the distribution of gradient orientations within a cell. The length of a petal indicates the contribution of a given orientation within the cell histogram and displays two times o petals. As presented, the feature vectors for the descriptor D ( C 2 × 2 , B 2 × 2 ) have different shapes for different YCbCr channels. The figure also contains image patches filtered with two used bilaplacian operators Δ a 2 and Δ b 2 .
As shown in Figure 4, the differences between descriptors for channels seem to justify the need for their joint application in order to extract more information about the described image regions. Furthermore, since the size of a local structure in an image cannot be determined in advance, the method uses several HOG descriptors with cells and blocks of different sizes. The HOG descriptor produces high-dimensional features, depending on its focus on a local appearance of objects within an image. Thus, in this paper, each obtained feature vector is characterized using its histogram h ( D ) . The histogram is used since it captures distributions of natural images [24]. The values in the feature vectors for images are in the range [ 0 , 1 ] . Therefore, the number of bins in the histogram and its variance is calculated by dividing the range into d intervals and using them for the determination of bin centers.
To illustrate that the used statistics can characterize the distortion severity of images, they are computed for two sequences of images distorted with Gaussian blur and Gaussian noise [41]. Here, the descriptors D ( C 2 × 2 , B 2 × 2 ) and D ( C 4 × 4 , B 2 × 2 ) are used. For the computation of the histogram, 10 bins are applied. As shown in Figure 5, the histogram responds consistently to the distortion type and its severity.
In the SEER, for YCbCr channels and channels filtered with two bilaplacian operators, feature vectors for the HOG descriptors are obtained and concatenated. This can be written as D = [ D Y , D Cb , D C r , D Δ a 2 Y , D Δ a 2 C b , D Δ a 2 C r , D Δ b 2 Y , D Δ b 2 C b , D Δ b 2 C r ] or D i , where i = 1 , , 9 . There are nine HOG descriptors per ith image, with different C and B. Hence, the HOG descriptors for the I can be written as D I = [ D 1 I , D 2 I , , D 9 I ] . Finally, the perceptual feature vector is obtained as F = [ h ( D 1 1 ) , h ( D 2 1 ) , , h ( D 1 2 ) , h ( D 2 2 ) , , h ( D 9 9 ) ] . The parameters of HOG descriptors used in the SEER are discussed in details in Section 3.6. Since the distortions affect images across scales [23], the method also processes an input image which is downsampled by a factor of two to improve the quality prediction. Finally, the concatenated histograms for an input image and its downscaled version are used to train the SVR model. The SVR model which maps the vector F into subjective ratings S is obtained using LIBSVM library [54]. Here, as in many NR techniques [24,46], the radial basis function (RBF) kernel is employed. Finally, given an input image, the model predicts its quality and provides the objective score Q.

3. Experimental Results and Discussion

3.1. Datasets and Protocol

In this work, to evaluate an NR metric, six publicly available large IQA datasets were used: TID2013 [41], TID2008 [42], CSIQ [43], LIVE [7], LIVE WIQC [44], and the Waterloo Exploration Database (WE) [55]. The first four datasets are typically selected to evaluate recently introduced measures [34,56,57], while the WE dataset, similar to the Group MAD Competition [58,59], is tied with a novel evaluation methodology. The LIVE WIQC dataset contains subjective scores collected in an uncontrolled manner using the Amazon Mechanical Turk. However, it contains images captured with a mobile camera and can be used for the evaluation of NR measures using real images. The TID2013 dataset is the largest and the most demanding public IQA benchmark. It consists of 3000 distorted images and covers 24 image distortion types. The other datasets contain only half (TID2008) or less than one-third of the number of the images in TID2013 (CSIQ and LIVE). The LIVE dataset contains popular distortions, such as JPEG compression, JPEG2000 compression, Gaussian blur, white noise, or simulated fast fading Rayleigh channel. It is worth noticing that some distortion types in TID2013 can be regarded as multiple, e.g., lossy compression of noisy images. Interestingly, most existing general-purpose NR measures are designed to provide an acceptable performance on the LIVE [32]. As shown in Section 3.3, some of them experience a drop in the performance on datasets that contain more diverse distortions. The datasets contain high-quality images, their distorted images, and related subjective scores. The subjective scores obtained in tests with human subjects are denoted as mean opinion scores (MOS) or differential MOS (DMOS). The WE dataset contains images distorted with JPEG compression, JPEG2000 compression, white Gaussian noise, and Gaussian blur. However, it does not contain subjective scores. The datasets are characterized in Table 1.
To measure the consistency of the prediction results provided by an IQA measure with subjective ratings, the following four indices were considered [60]: the Spearman’s Rank Correlation Coefficient (SRCC), Kendall Rank order Correlation Coefficient (KRCC), Pearson linear Correlation Coefficient (PCC), and Root Mean Square Error (RMSE). PCC and RMSE were calculated after a nonlinear mapping between the vectors of objective and subjective scores, Q p and S (MOS or DMOS), respectively. For the mapping, the following function was used [60]: Q p = β 1 1 2 1 1 + exp ( β 2 ( Q β 3 ) ) + β 4 Q + β 5 , where β = [ β 1 , β 2 , , β 5 ] are parameters of the fitted regression model, Q is the objective score, and Q p is the fitted score.
The evaluation protocol associated with the WE dataset requires: the calculation of pristine/distorted image discriminability test (D-test), listwise ranking consistency test (L-test), and pairwise preference consistency test (P-test) [55]. The D-test uses predictions of a model to classify distorted and pristine images. Larger values of D denote better separability. The L-test, in turn, adopts the average SRCC to quantify the ranking consistency among distorted images. Finally, the P-test compares preference predictions of IQA models on pairs of images whose quality is clearly discriminable [55], dividing the number of image pairs with correctly predicted concordance by the number of image pairs. The values obtained in tests lie in [ 0 , 1 ] .

3.2. Model Training

The proposed method should be trained to obtain the SVR model used for the quality prediction. Therefore, a typical protocol used for the validation of NR techniques was adopted, in which each image dataset was divided into disjoint training and testing subsets, i.e., distorted images of 80% reference images were used in training and the remaining 20% of images were used for testing [28,32,46]. Then, to avoid bias and fairly compare a measure with other measures, the performance of each method was reported in terms of the median values of SRCC, KRCC, PCC, and RMSE over 100 training-testing iterations [61].

3.3. Performance on Individual Datasets

The performance of the presented NR measure was compared with those of the related state-of-the-art techniques. The following NR measures were considered: (i) HOSA [28]; (ii) BPRI [32]; (iii) BRISQUE [24]; (iv) IL-NIQE [31]; (v) SISBLIM [62]; (vi) OG-IQA [24]; (vii) GWH-GLBP [46]; and (viii) RATER [48]. Among NR measures, which, to some extent, are similar to the SEER, the RATER, HOSA, and IL-NIQE assess color images, while the OG-IQA incorporates image gradients. The GWH-GLBP and SISBLIM are designed for the evaluation of images with multiple distortions. The RATER and BPRI are recently introduced general-purpose measures. However, due to distortion-specific steps in BPRI, its range of applicability is larger than those of distortion-specific measures, but it seems to be confined to several distortion types. As reported by Zhang et al., the IL-NIQE is superior to the BLIINDS2, DIIVINE, CORNIA, NIQE, BRISQUE, and QAC [31]. The HOSA, in turn, is reported to outperform the GM-LOG, BRISQUE, or IL-NIQE [28].
The experimental evaluation was conducted on five IQA datasets using the protocol shown in Section 3.1. For the fair comparison, the parameters of all learning-based techniques were obtained aiming at their best performance [25,31]. The methods were run using publicly available implementations. For the SVR, the popular LIBSVM library was used [54]. The parameters of AdaBoosting BP neural network in the OG-IQA were also determined. The methods that do not require training (BPRI, IL-NIQE, and SISBLIM) were evaluated using the defined testing subsets of images in datasets. The SEER was run with o = 36 and d = 30 (see Section 3.6).
Table 2 summarizes the results on IQA datasets, where the best result for each performance index is written in bold. The table also contains average values for SRCC, KRCC, and PCC. The RMSE was not averaged due to the different range of values in the benchmarks.
As demonstrated, the introduced NR measure outperformed the state-of-the-art measures on four IQA benchmarks, i.e., the TID2013, TID2008, CSIQ, and LIVE. In the case of the LIVE WIQC dataset, the RATER and BRISQUE were slightly better than SEER, which was the third-best measure. Remarkably, the proposed measure had the highest average performance across the datasets, which confirms its usability.
To determine whether the relative performance differences between the measures are statistically significant, the Wilcoxon rank-sum test was used. The test measures the equivalence of the median values of independent samples with a 5% significance level. Here, the null hypothesis assumes that the SRCC values of compared metrics are drawn from a population with equal medians. The results are shown in Table 3, where the symbols “−1”, “0” and “1” denote that the IQA measure in the column was statistically better with a confidence greater than 95%, indistinguishable, or worse than the SEER on a given IQA dataset, respectively. The findings are consistent with conclusions drawn from the previous experiments, i.e., the SEER was statistically better than other measures on the TID2013, TID2008, and LIVE. Taking into account the results for the CSIQ, the SEER was on pair with the RATER, as well as on pair with the RATER and BRISQUE for the LIVE WIQC.
The measures were evaluated using the methodology associated with the WE database [55]. Here, the BRISQUE and HOSA were compared with the SEER. Since the IL-NIQE is reported to provide superior performance using this methodology [55], it is also added to the comparison. The results reported for BRISQUE, HOSA, and IL-NIQE in [55] are presented in Table 4. The SEER, as other techniques, was trained on LIVE dataset. Interestingly, the IL-NIQE, which is much less correlated with subjective scores on LIVE than other measures, yielded encouraging L and P values. This may indicate that the relationship between distortion types, their levels, and used tests in the used methodology may require further attention, leading to its possible improvement. For example, the P values of the measures whose performances on the databases with subjective scores were significantly different seem to be close to its upper limit. Furthermore, the assumption that image quality degrades monotonically with the distortion levels may not be true for all distortion types and content of images, which may influence the results of the L-test [55]. However, it can be seen that all measures provided acceptable performance in the tests on the WE database.
The comparative evaluation of the SEER with state-of-the-art NR methods that use DNN or other neural networks (NN) architectures was based on published results. Due to the large complexity of the models, and the unavailability of learning source codes for some of them, such comparison is very popular. Consequently, many papers report the performance of measures on the basis of results obtained in 10 random training–testing splits on only one or two IQA benchmarks. Furthermore, the coherent comparison of DNN-based methods is often impeded by the exclusion of some distortion types, which can make the use of the largest IQA datasets, such as the TID2013 or TID2008, superfluous. Table 5 contains the comparison of published median values of SRCC and PCC for NN-based NR methods with those obtained for the SEER in 10 training–testing splits. Other performance indices, i.e., KRCC and RMSE, are seldom reported in the referenced works. The results for the TID2008 are not presented since the referenced works were not evaluated on this dataset. As reported, the SEER clearly outperformed other measures on the most demanding datasets, such as the TID2013 or CSIQ. The NN-based measures were better than the SEER on the LIVE dataset. The worse results of the compared models on the datasets that contain considerably more distortions can be attributed to the lack of a sufficient number of training samples or imperfections of used architectures [36,57]. It can be concluded that the introduced hand-crafted NR measure is highly competitive to the recently introduced techniques based on DNN or NN architectures.

3.4. Performance across Datasets

To verify whether the proposed measure is independent of a dataset, a cross-dataset validation was performed. The NR measures were trained on one IQA dataset and tested on the remaining datasets. In the experiment, the learning-based measures were compared on four datasets. The results, in terms of SRCC, reported in Table 6, reveal that the SEER maintained acceptable generalization capability across the datasets. The best results were with bold type. In general, it achieved the best average SRCC.

3.5. Computational Complexity

The computational complexity of a given method was analyzed in terms of the average time taken to assess an image ( 512 × 384 ) from the TID2013 dataset. The experiments were performed on a 3.3 GHz Intel Core CPU with 16 GB RAM system running on Microsoft Windows 7 64 bit. For all compared methods, their Matlab implementations were used. As demonstrated in Table 7, the SEER is of moderate complexity. It was faster than IL-NIQE and slower than RATER and HOSA, which also process color images. The execution time of the SEER strongly depends on its parameters. However, unlike other measures, the SEER can be easily run in parallel, since, for the assessed color image, the nine used HOG descriptors were executed for 18 resulting images, including YCbCr color channels, their filtered images, and downsampled images. A simultaneous run of the computation of the HOG for these images using an efficient native implementation (e.g., C++) with a GPU code of the HOG could shorten the execution time of the method up to 162 times.

3.6. Metric Configuration and Contribution of Features

In the SEER, as explained in Section 2.2, an evaluated image is characterized using the histogram of K HOG descriptors. Since the HOG can be used with different parameters, their choice and the resulting IQA performance of the introduced measure requires investigation. SRCC was used as the quality index, taking into account that the remaining criteria similarly indicate the performance of the method.
Figure 6a contains SRCC values obtained for the SEER with only one HOG descriptor, using the protocol introduced in the previous section, on the TID2013 dataset. TID2013 was used due to its size and the number of image distortions. In the experiment, the performance of the SEER with many different HOG parameters was investigated. The presented HOG descriptors were run with o = 36 and d = 30 . In the IQA, a detailed description of a pixel neighborhood is needed, confirmed by various NR approaches using LBP in which a pixel is characterized by its 8 neighbors. Therefore, HOG configurations with smaller pixel areas were taken into account to determine the IQA performance of the SEER with them. The nine HOG descriptors presented in Figure 6a were selected to be used jointly. In the case of single application of the descriptor D ( C 1 × 1 , B 1 × 1 ) , the method resembles approaches in which statistics for a gradient map, among other features, are used for the quality prediction (e.g., [27]). To show that such joint application of the HOG descriptors is beneficial, Figure 6b reports the performance of the method with 1 , 2 , , 9 HOG descriptors. Since the presentation of all possible combinations of these descriptors is unfeasible, the performance is reported for the descriptors added in the order presented in Figure 6a. Interestingly, the SEER with only two HOG descriptors delivered promising performance, outperforming the state-of-the-art measures evaluated in the previous section. It can be assumed that an application of dimensionality reduction or feature selection techniques, which would indicate the most influential descriptors or their statistics, may provide better results. However, these techniques can also deepen the dataset-dependency. Therefore, in this study, the selected K = 9 HOG descriptors were used in the SEER.
In the proposed method, the descriptors D ( C 1 × 3 , B 1 × 3 ) , D ( C 3 × 1 , B 3 × 1 ) , D ( C 1 × 1 , B 1 × 1 ) and D ( C 2 × 2 , B 1 × 1 ) contributed the most to its performance (see Figure 6a). Apart from the size of the cell and the arrangement of cells in the block, the computation of the feature vector F in the SEER requires two additional parameters, o and d. Therefore, the performance of the measure was evaluated taking into account their variability. The results in terms of SRCC on the TID2013 are shown in Figure 6c. They reveal that these two parameters almost did not affect the performance of the SEER. In other experiments with the method, o = 36 and d = 30 are used.
Since the measure uses YcbCr color space and filtered images, Table 8 presents their contribution to its performance in terms of SRCC on the TID2013. As reported, color channels contributed similarly. However, the channels filtered using the bilaplacian operators carry the most information that can be used for the quality prediction. As reported, the channels alone could not provide satisfactory SRCC performance, which confirms their complementary relationship and justifies their joint application in the SEER.

4. Conclusions

In this work, a novel NR-IQA technique has been presented. The introduced SEER incorporates the distribution of local intensity gradients. The histogram of a set of these descriptors, obtained for YCbCr channels of a distorted image, as well as its channels filtered with bilaplacian operators, are used as perceptual features to train an SVR model for the image quality prediction. It has been shown that the use of a descriptor, which takes into account different sizes of described image regions and their mutual relationship, is beneficial to the IQA, as obtained feature vectors respond consistently to the distortion type and its severity. Furthermore, it has been demonstrated that, to deliver superior performance, the information extracted from YCbCr channels in the bilaplacian domain should be employed. The introduced technique was evaluated and compared with the state-of-the-art NR hand-crafted and NN-based measures on six IQA datasets. The experimental results demonstrate that overall the SEER outperforms the compared NR measures in terms of prediction accuracy and generalization capability.
The future work on the SEER will focus on an application of a keypoint detector that indicates image regions for the description [27,67]. Another promising direction of research is to train a deep learning model using features provided by the SEER, similar to the work in [35].
The Matlab code of the proposed approach is publicly available at http://marosz.kia.prz.edu.pl/SEER.html.

Funding

This research received no external funding.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Chandler, D.M. Seven Challenges in Image Quality Assessment: Past, Present, and Future Research. ISRN Signal Process. 2013, 2013, 905685. [Google Scholar] [CrossRef]
  2. Korus, P.; Białas, J.; Dziech, A. Towards Practical Self-Embedding for JPEG-Compressed Digital Images. IEEE Trans. Multimed. 2015, 17, 157–170. [Google Scholar] [CrossRef]
  3. Kim, Y.G.; Lee, W.O.; Kim, K.W.; Hong, H.G.; Park, K.R. Performance Enhancement of Face Recognition in Smart TV Using Symmetrical Fuzzy-Based Quality Assessment. Symmetry 2015, 7, 1475–1518. [Google Scholar] [CrossRef] [Green Version]
  4. Zhu, W.; Zhai, G.; Hu, M.; Liu, J.; Yang, X. Arrow’s Impossibility Theorem inspired subjective image quality assessment approach. Signal Process. 2018, 145, 193–201. [Google Scholar] [CrossRef]
  5. Lin, W.; Kuo, C.C.J. Perceptual visual quality metrics: A survey. J. Vis. Commun. Image Represent. 2011, 22, 297–312. [Google Scholar] [CrossRef] [Green Version]
  6. Gabarda, S.; Cristobal, G.; Goel, N. Anisotropic blind image quality assessment: Survey and analysis with current methods. J. Vis. Commun. Image Represent. 2018, 52, 101–105. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  8. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef] [Green Version]
  9. Zhang, L.; Shen, Y.; Li, H. VSI: A Visual Saliency-Induced Index for Perceptual Image Quality Assessment. IEEE Trans. Image Process. 2014, 23, 4270–4281. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, S.; Li, P.; Xu, X.; Li, L.; Chang, C.C. No-Reference Image Blur Assessment Based on Response Function of Singular Values. Symmetry 2018, 10, 304. [Google Scholar] [CrossRef]
  11. Liu, A.; Lin, W.; Narwaria, M. Image Quality Assessment Based on Gradient Similarity. IEEE Trans. Image Process. 2012, 21, 1500–1512. [Google Scholar] [CrossRef]
  12. Saha, A.; Wu, Q.J. Perceptual image quality assessment using phase deviation sensitive energy features. Signal Process. 2013, 93, 3182–3191. [Google Scholar] [CrossRef]
  13. Gu, K.; Wang, S.; Zhai, G.; Lin, W.; Yang, X.; Zhang, W. Analysis of Distortion Distribution for Pooling in Image Quality Prediction. IEEE Trans. Broadcast. 2016, 62, 446–456. [Google Scholar] [CrossRef]
  14. Okarma, K. Quality assessment of images with multiple distortions using combined metrics. Elektron. Elektrotech. 2014, 20, 128–131. [Google Scholar] [CrossRef]
  15. Oszust, M. Decision Fusion for Image Quality Assessment using an Optimization Approach. IEEE Signal Proc. Lett. 2016, 23, 65–69. [Google Scholar] [CrossRef]
  16. Wu, J.; Liu, Y.; Li, L.; Shi, G. Attended Visual Content Degradation Based Reduced Reference Image Quality Assessment. IEEE Access 2018, 6, 12493–12504. [Google Scholar] [CrossRef]
  17. Ospina-Borras, J.; Restrepo, H.D.B. Non-reference assessment of sharpness in blur/noise degraded images. J. Vis. Commun. Image Represent. 2016, 39, 142–151. [Google Scholar] [CrossRef]
  18. Ahn, S.; Park, J.; Chong, J. Blurring Image Quality Assessment Method Based on Histogram of Gradient. In Proceedings of the 19th Brazilian Symposium on Multimedia and the Web, Salvador, Brazil, 5–8 November 2013; ACM: New York, NY, USA, 2013; pp. 181–184. [Google Scholar] [CrossRef]
  19. Gerhard, H.E.; Wichmann, F.A.; Bethge, M. How Sensitive Is the Human Visual System to the Local Statistics of Natural Images? PLoS Comput. Biol. 2013, 9, 1–15. [Google Scholar] [CrossRef]
  20. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind Image Quality Assessment: A Natural Scene Statistics Approach in the DCT Domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef] [PubMed]
  21. Moorthy, A.K.; Bovik, A.C. Blind Image Quality Assessment: From Natural Scene Statistics to Perceptual Quality. IEEE Trans. Image Process. 2011, 20, 3350–3364. [Google Scholar] [CrossRef]
  22. Tang, L.; Li, L.; Sun, K.; Xia, Z.; Gu, K.; Qian, J. An efficient and effective blind camera image quality metric via modeling quaternion wavelet coefficients. J. Vis. Commun. Image Represent. 2017, 49, 204–212. [Google Scholar] [CrossRef]
  23. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Liu, L.; Hua, Y.; Zhao, Q.; Huang, H.; Bovik, A.C. Blind image quality assessment by relative gradient statistics and adaboosting neural network. Signal Process. Image 2016, 40, 1–15. [Google Scholar] [CrossRef]
  25. Xue, W.; Mou, X.; Zhang, L.; Bovik, A.C.; Feng, X. Blind Image Quality Assessment Using Joint Statistics of Gradient Magnitude and Laplacian Features. IEEE Trans. Image Process. 2014, 23, 4850–4862. [Google Scholar] [CrossRef]
  26. Li, Q.; Lin, W.; Xu, J.; Fang, Y. Blind Image Quality Assessment Using Statistical Structural and Luminance Features. IEEE Trans. Multimed. 2016, 18, 2457–2469. [Google Scholar] [CrossRef]
  27. Oszust, M. No-Reference Image Quality Assessment Using Image Statistics and Robust Feature Descriptors. IEEE Signal Proc. Lett. 2017, 24, 1656–1660. [Google Scholar] [CrossRef]
  28. Xu, J.; Ye, P.; Li, Q.; Du, H.; Liu, Y.; Doermann, D. Blind Image Quality Assessment Based on High Order Statistics Aggregation. IEEE Trans. Image Process. 2016, 25, 4444–4457. [Google Scholar] [CrossRef]
  29. Zhang, L.; Gu, Z.; Liu, X.; Li, H.; Lu, J. Training Quality-Aware Filters for No-Reference Image Quality Assessment. IEEE MultiMedia 2014, 21, 67–75. [Google Scholar] [CrossRef]
  30. Xue, W.; Zhang, L.; Mou, X. Learning without Human Scores for Blind Image Quality Assessment. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 995–1002. [Google Scholar] [CrossRef]
  31. Zhang, L.; Zhang, L.; Bovik, A.C. A Feature-Enriched Completely Blind Image Quality Evaluator. IEEE Trans. Image Process. 2015, 24, 2579–2591. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Min, X.; Gu, K.; Zhai, G.; Liu, J.; Yang, X.; Chen, C.W. Blind Quality Assessment Based on Pseudo Reference Image. IEEE Trans. Multimed. 2017, PP, 1. [Google Scholar] [CrossRef]
  33. Bosse, S.; Maniry, D.; Wiegand, T.; Samek, W. A deep neural network for image quality assessment. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3773–3777. [Google Scholar] [CrossRef]
  34. Kim, J.; Lee, S. Fully Deep Blind Image Quality Predictor. IEEE J. Sel. Top. Signal 2017, 11, 206–220. [Google Scholar] [CrossRef]
  35. Ma, K.; Liu, W.; Liu, T.; Wang, Z.; Tao, D. dipIQ: Blind Image Quality Assessment by Learning-to-Rank Discriminable Image Pairs. IEEE Trans. Image Process. 2017, 26, 3951–3964. [Google Scholar] [CrossRef]
  36. Zeng, H.; Zhang, L.; Bovik, A.C. A Probabilistic Quality Representation Approach to Deep Blind Image Quality Prediction. arXiv, 2017; arXiv:1708.08190. [Google Scholar]
  37. Xie, X.; Zhang, Y.; Wu, J.; Shi, G.; Dong, W. Bag-of-words feature representation for blind image quality assessment with local quantized pattern. Neurocomputing 2017, 266, 176–187. [Google Scholar] [CrossRef]
  38. Li, Q.; Lin, W.; Fang, Y. BSD: Blind image quality assessment based on structural degradation. Neurocomputing 2017, 236, 93–103. [Google Scholar] [CrossRef]
  39. Zhou, W.; Qiu, W.; Wu, M.W. Utilizing Dictionary Learning and Machine Learning for Blind Quality Assessment of 3-D Images. IEEE Trans. Broadcast. 2017, 63, 404–415. [Google Scholar] [CrossRef]
  40. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar] [CrossRef]
  41. Ponomarenko, N.; Jin, L.; Ieremeiev, O.; Lukin, V.; Egiazarian, K.; Astola, J.; Vozel, B.; Chehdi, K.; Carli, M.; Battisti, F.; Kuo, C.C.J. Image database TID2013: peculiarities results and perspectives. Signal Process. Image 2015, 30, 57–77. [Google Scholar] [CrossRef]
  42. Ponomarenko, N.; Lukin, V.; Zelensky, A.; Egiazarian, K.; Carli, M.; Battisti, F. TID2008—A Database for Evaluation of Full-Reference Visual Quality Assessment Metrics. Adv. Mod. Radioelectron. 2009, 10, 30–45. [Google Scholar]
  43. Larson, E.C.; Chandler, D.M. Most apparent distortion: Full-reference image quality assessment and the role of strategy. J. Electron. Imaging 2010, 19, 011006. [Google Scholar] [CrossRef]
  44. Ghadiyaram, D.; Bovik, A.C. Massive Online Crowdsourced Study of Subjective and Objective Picture Quality. IEEE Trans. Image Process. 2016, 25, 372–387. [Google Scholar] [CrossRef]
  45. Fang, R.; Al-Bayaty, R.; Wu, D. BNB Method for No-Reference Image Quality Assessment. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1381–1391. [Google Scholar] [CrossRef]
  46. Li, Q.; Lin, W.; Fang, Y. No-Reference Quality Assessment for Multiply-Distorted Images in Gradient Domain. IEEE Signal Proc. Lett. 2016, 23, 541–545. [Google Scholar] [CrossRef]
  47. Oszust, M. Optimized Filtering With Binary Descriptor for Blind Image Quality Assessment. IEEE Access 2018, 6, 42917–42929. [Google Scholar] [CrossRef]
  48. Oszust, M. No-reference image quality assessment with local features and high-order derivatives. J. Vis. Commun. Image Represent. 2018, 56, 15–26. [Google Scholar] [CrossRef]
  49. Ye, P.; Kumar, J.; Kang, L.; Doermann, D. Unsupervised feature learning framework for no-reference image quality assessment. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1098–1105. [Google Scholar] [CrossRef]
  50. Lu, N.; Li, G. Blind quality assessment for screen content images by orientation selectivity mechanism. Signal Process. 2018, 145, 225–232. [Google Scholar] [CrossRef]
  51. Wang, Y.; Jiang, T.; Ma, S.; Gao, W. Image quality assessment based on local orientation distributions. In Proceedings of the 28th Picture Coding Symposium, Nagoya, Japan, 8–10 December 2010; pp. 274–277. [Google Scholar] [CrossRef]
  52. Yang, Y.; Tu, D.; Cheng, G. Image quality assessment using Histograms of Oriented Gradients. In Proceedings of the 2013 Fourth International Conference on Intelligent Control and Information Processing (ICICIP), Beijing, China, 9–11 June 2013; pp. 555–559. [Google Scholar] [CrossRef]
  53. International Telecommunications Union. ITU-R Recommendation BT. 601-5: Studio Encoding Parameters of Digital Television for Standard 4:3 and Wide-screen 16:9 Aspect Ratios. 2011. Available online: http://www.itu.int/dms_pubrec/itu-r/rec/bt/r-rec-bt.601-7-201103-i!!pdf-e.pdf (accessed on 1 December 2018).
  54. Chang, C.C.; Lin, C.J. LIBSVM: A Library for Support Vector Machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  55. Ma, K.; Duanmu, Z.; Wu, Q.; Wang, Z.; Yong, H.; Li, H.; Zhang, L. Waterloo Exploration Database: New Challenges for Image Quality Assessment Models. IEEE Trans. Image Process. 2017, 26, 1004–1016. [Google Scholar] [CrossRef]
  56. Shi, Z.; Zhang, J.; Cao, Q.; Pang, K.; Luo, T. Full-reference image quality assessment based on image segmentation with edge feature. Signal Process. 2018, 145, 99–105. [Google Scholar] [CrossRef]
  57. Kim, J.; Zeng, H.; Ghadiyaram, D.; Lee, S.; Zhang, L.; Bovik, A.C. Deep Convolutional Neural Models for Picture-Quality Prediction: Challenges and Solutions to Data-Driven Image Quality Assessment. IEEE Signal Proc. Mag. 2017, 34, 130–141. [Google Scholar] [CrossRef]
  58. Ma, K.; Wu, Q.; Wang, Z.; Duanmu, Z.; Yong, H.; Li, H.; Zhang, L. Group MAD Competition—A New Methodology to Compare Objective Image Quality Models. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  59. Ma, K.; Duanmu, Z.; Wang, Z.; Wu, Q.; Liu, W.; Yong, H.; Li, H.; Zhang, L. Group Maximum Differentiation Competition: Model Comparison with Few Samples. IEEE Trans. Pattern Anal. Mach. Intell. 2018. [Google Scholar] [CrossRef]
  60. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A Statistical Evaluation of Recent Full Reference Image Quality Assessment Algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef] [PubMed]
  61. Lu, Q.; Zhou, W.; Li, H. A no-reference Image sharpness metric based on structural information using sparse representation. Inf. Sci. 2016, 369, 334–346. [Google Scholar] [CrossRef]
  62. Gu, K.; Zhai, G.; Yang, X.; Zhang, W. Hybrid No-Reference Quality Metric for Singly and Multiply Distorted Images. IEEE Trans. Broadcast. 2014, 60, 555–567. [Google Scholar] [CrossRef]
  63. Fan, C.; Zhang, Y.; Feng, L.; Jiang, Q. No Reference Image Quality Assessment based on Multi-Expert Convolutional Neural Networks. IEEE Access 2018, 6, 8934–8943. [Google Scholar] [CrossRef]
  64. Liu, X.; van de Weijer, J.; Bagdanov, A.D. RankIQA: Learning from Rankings for No-reference Image Quality Assessment. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  65. Ma, K.; Liu, W.; Zhang, K.; Duanmu, Z.; Wang, Z.; Zuo, W. End-to-End Blind Image Quality Assessment Using Deep Neural Networks. IEEE Trans. Image Process. 2018, 27, 1202–1213. [Google Scholar] [CrossRef]
  66. Kang, L.; Ye, P.; Li, Y.; Doermann, D. Convolutional Neural Networks for No-Reference Image Quality Assessment. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1733–1740. [Google Scholar] [CrossRef]
  67. Matusiak, K.; Skulimowski, P.; Strumillo, P. Unbiased evaluation of keypoint detectors with respect to rotation invariance. IET Comput. Vis. 2017, 11, 507–516. [Google Scholar] [CrossRef]
Figure 1. Image processing steps in the SEER.
Figure 1. Image processing steps in the SEER.
Symmetry 11 00095 g001
Figure 2. Block diagram of the method.
Figure 2. Block diagram of the method.
Symmetry 11 00095 g002
Figure 3. Division of an image into blocks and cells in the HOG.
Figure 3. Division of an image into blocks and cells in the HOG.
Symmetry 11 00095 g003
Figure 4. Influence of the Gaussian blur on HOG descriptors obtained for YCbCr channels of an image. The D ( C 2 × 2 , B 2 × 2 ) descriptor is visualized for the selected image patch.
Figure 4. Influence of the Gaussian blur on HOG descriptors obtained for YCbCr channels of an image. The D ( C 2 × 2 , B 2 × 2 ) descriptor is visualized for the selected image patch.
Symmetry 11 00095 g004
Figure 5. Influence of distortions on features used in the SEER for exemplary images distorted with: Gaussian blur (ae); and Gaussian noise (lp). The images are ordered by the distortion severity (from left to right). Below distorted images, the histograms for two descriptors are shown: D ( C 2 × 2 , B 2 × 2 ) denoted by D a (fh,rt); and D ( C 4 × 4 , B 2 × 2 ) denoted by D b (ik,uw). Colors are assigned to distorted images.
Figure 5. Influence of distortions on features used in the SEER for exemplary images distorted with: Gaussian blur (ae); and Gaussian noise (lp). The images are ordered by the distortion severity (from left to right). Below distorted images, the histograms for two descriptors are shown: D ( C 2 × 2 , B 2 × 2 ) denoted by D a (fh,rt); and D ( C 4 × 4 , B 2 × 2 ) denoted by D b (ik,uw). Colors are assigned to distorted images.
Symmetry 11 00095 g005
Figure 6. Influence of parameters of the SEER on its performance: (a) sizes of cells and their arrangement in blocks in the HOG descriptor; (b) the number of used descriptors added in the precedence shown in (a); and (c) the number of orientation bins o and intervals d, respectively.
Figure 6. Influence of parameters of the SEER on its performance: (a) sizes of cells and their arrangement in blocks in the HOG descriptor; (b) the number of used descriptors added in the precedence shown in (a); and (c) the number of orientation bins o and intervals d, respectively.
Symmetry 11 00095 g006
Table 1. Image quality assessment datasets.
Table 1. Image quality assessment datasets.
DatasetRef. ImagesDist. ImagesDist. TypesScore TypeDist. Types in an Image
TID201325300024MOS1, 2
TID200825170017MOS1
CSIQ308666DMOS1
LIVE297795DMOS1
LIVE WIQCNA1162NAMOSNA, multiple
WE474494,8804None1
Table 2. Performance evaluation on individual datasets.
Table 2. Performance evaluation on individual datasets.
HOSABPRIBRISQUEIL-NIQESISBLIMOG-IQAGWH-GLBPRATERSEER
TID2013, 3000 images
SRCC0.71320.22220.55510.51260.32120.48550.48350.82690.8362
KRCC0.53920.15270.39880.36310.22530.34730.34830.64110.6571
PCC0.78230.46600.64860.63070.49200.62280.64480.84090.8588
RMSE0.77341.09460.94220.96791.07340.97120.94700.67030.6327
TID2008, 1700 images
SRCC0.77320.18250.60660.15100.24270.58020.52080.82570.8444
KRCC0.59350.12910.44230.10050.17050.41910.37160.64960.6767
PCC0.81360.47470.67590.19840.45670.66660.65400.83620.8704
RMSE0.77321.18010.98311.31571.19301.00241.01780.73610.6599
CSIQ, 866 images
SRCC0.82900.56790.86080.86830.69460.76890.76930.89830.9037
KRCC0.64000.42380.68010.68520.51250.57590.58580.72400.7374
PCC0.84730.72500.88510.88600.70440.80640.81580.92110.9218
RMSE0.14330.17810.12500.12540.19010.15890.15770.10240.0997
LIVE, 779 images
SRCC0.94080.88260.93910.89930.09560.91590.87310.94220.9512
KRCC0.79220.72110.79230.72000.05960.76380.69190.79870.8140
PCC0.94150.88080.94270.90610.19240.91950.89180.94280.9534
RMSE9.157913.0028.952211.56726.77710.80112.3268.94128.2862
LIVE WIQC, 1162 images
SRCC0.54810.17000.60490.19170.42800.47020.54590.60330.6016
KRCC0.37340.11400.42760.12890.29420.32230.38630.42770.4241
PCC0.58530.29690.64220.19300.50380.51340.59180.62850.6293
RMSE16.37619.28915.49419.73017.34917.24516.30615.74815.856
Overall direct
SRCC0.81410.46380.74040.60780.33850.68760.66170.87330.8839
KRCC0.64120.35670.57840.46720.24200.52650.49940.70340.7213
PCC0.84620.63660.78810.65530.46140.75380.75160.88530.9011
Overall weighted
SRCC0.73820.31350.64960.46220.33960.58200.57500.81220.8215
KRCC0.56370.23160.48640.34160.23950.42930.42250.63590.6510
PCC0.78290.51470.71160.52310.47920.66780.68400.82680.8430
Table 3. SRCC-based statistical significance tests.
Table 3. SRCC-based statistical significance tests.
DatasetHOSABPRIBRISQUEIL-NIQESISBLIMOG-IQAGWH-GLBPRATER
TID201311111111
TID200811111111
CSIQ11111110
LIVE11111111
LIVE WIQC11011110
Table 4. Evaluation of measures on WE Database [55].
Table 4. Evaluation of measures on WE Database [55].
D-testL-testP-test
BRISQUE [24]0.92040.97720.9930
IL-NIQE [31]0.90840.99260.9927
HOSA [28]0.91750.96470.9983
SEER0.85470.94660.9895
Table 5. Performance comparison between the SEER and NR measures which use NN architectures.
Table 5. Performance comparison between the SEER and NR measures which use NN architectures.
BIECON [34]PQR [36] (S CNN)PQR [36] (ResNet50)PQR [36] (AlexNet)I-wise CNN [57]IQAMSCN [63]RIQA +FT [64]MEON [65]DeepIQA [35]P-wise [33]CNN [66]SEER
TID2013, 3000 images
SRCC0.7170.6920.7400.5740.800-0.7800.8080.761--0.868
PCC0.7620.7500.7980.6690.802------0.872
CSIQ, 866 images
SRCC0.8150.9080.8730.8710.812------0.901
PCC0.8230.9270.9010.8960.791------0.920
LIVE, 779 images
SRCC0.9580.9640.9650.9550.9630.9530.981--0.9600.9560.937
PCC0.9600.9660.9710.9640.9640.957---0.9720.9530.941
Note: The comparison is based on published median values of SRCC and PCC over 10 training-testing iterations.
Table 6. Cross-dataset performance of learning-based NR measures in terms of SRCC.
Table 6. Cross-dataset performance of learning-based NR measures in terms of SRCC.
TestingHOSABRISQUEOG-IQAGWH-GLBPRATERSEER
Training on TID2013
TID20080.8390.7520.5800.9360.9270.953
CSIQ0.6100.6220.5810.3070.6840.687
LIVE0.8370.8110.8480.5140.7880.836
Training on TID2008
TID20130.7720.6560.5070.8070.7530.816
CSIQ0.6070.5950.4750.3160.6480.672
LIVE0.8240.8350.7970.4890.7500.818
Training on CSIQ
TID20130.5340.4140.3350.1390.4300.446
TID20080.4850.5010.3220.1650.4880.449
LIVE0.9040.6890.8300.5330.8580.928
Training on LIVE
TID20130.4680.3600.3150.3100.3600.400
TID20080.4100.3170.2550.3040.3190.395
CSIQ0.5840.5970.5830.4780.6780.755
Average
-0.6560.5960.5360.4410.6400.680
Table 7. Average run-time.
Table 7. Average run-time.
NR MeasureRun-Time (in Seconds)
HOSA0.440
BPRI0.997
BRISQUE0.049
IL-NIQE8.200
SISBLIM2.200
OG-IQA3.770
GWH-GLBP0.064
RATER0.168
SEER, o = 36 2.322
SEER, o = 9 1.121
Table 8. Contribution of YCbCr channels with filtering to the performance of the SEER on the TID2013 in terms of SRCC.
Table 8. Contribution of YCbCr channels with filtering to the performance of the SEER on the TID2013 in terms of SRCC.
Described ImagesSRCC
Y0.717
C b 0.728
C r 0.729
Y , C b , C r 0.727
Δ a 2 ( Y , C b , C r ) 0.788
Δ b 2 ( Y , C b , C r ) 0.808

Share and Cite

MDPI and ACS Style

Oszust, M. No-Reference Image Quality Assessment with Local Gradient Orientations. Symmetry 2019, 11, 95. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11010095

AMA Style

Oszust M. No-Reference Image Quality Assessment with Local Gradient Orientations. Symmetry. 2019; 11(1):95. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11010095

Chicago/Turabian Style

Oszust, Mariusz. 2019. "No-Reference Image Quality Assessment with Local Gradient Orientations" Symmetry 11, no. 1: 95. https://0-doi-org.brum.beds.ac.uk/10.3390/sym11010095

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop