Next Article in Journal
MHD Flow and Heat Transfer over Vertical Stretching Sheet with Heat Sink or Source Effect
Previous Article in Journal
A Numerical Simulation of Silver–Water Nanofluid Flow with Impacts of Newtonian Heating and Homogeneous–Heterogeneous Reactions Past a Nonlinear Stretched Cylinder
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Center-Emphasized Visual Saliency and a Contrast-Based Full Reference Image Quality Index

Department of Computer Science and Engineering, Kyung Hee University Global Campus, Yongin 17104, Korea
*
Author to whom correspondence should be addressed.
Symmetry 2019, 11(3), 296; https://doi.org/10.3390/sym11030296
Submission received: 28 December 2018 / Revised: 14 February 2019 / Accepted: 22 February 2019 / Published: 26 February 2019

Abstract

:
Objective image quality assessment (IQA) is imperative in the current multimedia-intensive world, in order to assess the visual quality of an image at close to a human level of ability. Many parameters such as color intensity, structure, sharpness, contrast, presence of an object, etc., draw human attention to an image. Psychological vision research suggests that human vision is biased to the center area of an image and display screen. As a result, if the center part contains any visually salient information, it draws human attention even more and any distortion in that part will be better perceived than other parts. To the best of our knowledge, previous IQA methods have not considered this fact. In this paper, we propose a full reference image quality assessment (FR-IQA) approach using visual saliency and contrast; however, we give extra attention to the center by increasing the sensitivity of the similarity maps in that region. We evaluated our method on three large-scale popular benchmark databases used by most of the current IQA researchers (TID2008, CSIQ and LIVE), having a total of 3345 distorted images with 28 different kinds of distortions. Our method is compared with 13 state-of-the-art approaches. This comparison reveals the stronger correlation of our method with human-evaluated values. The prediction-of-quality score is consistent for distortion specific as well as distortion independent cases. Moreover, faster processing makes it applicable to any real-time application.

1. Introduction

Computer-based automatic image quality assessment (IQA) has been sought after for decades because numerous image and video applications need this assessment to automate their quality maintenance. To date, IQA research has been significantly advanced, however, this is still an active area of research to bring the methods closer to human-level ability. In the literature, there are three principle IQA approaches. No-reference image quality assessment (NR-IQA) uses a single distorted image without any reference image, whereas in reduced-reference IQA (RR-IQA), partial information of a reference image is given. The third category is full-reference IQA (FR-IQA) where the complete reference image is given along with the distorted one. In this paper, we deal with FR-IQA.
Table 1 presents a brief survey on several state-of-the-art IQA approaches, which we compared in this paper. We made a separate column for the pooling strategy because recent IQA research tends to combine multiple features where pooling plays an important role. Early pixel-based, faster IQA methods such as mean squared error (MSE) and peak signal-to-noise ratio (PSNR) consider neither the human visual system (HVS) nor any aspects of human perception. Thus, those approaches fail to achieve good correlation with human assessment [1,2]. Two images with the same PSNR or MSE may be perceived in totally different ways by a human observer. However, humans are the ultimate receiver of images; as a result, the search for methods that can achieve a closer correlation with humans is ongoing. Wang et al., in their revolutionary work on the structural similarity index or SSIM [3], argued that human visual perception is highly sensitive to structural information. The SSIM index incorporates luminance, contrast, and structural comparison information and achieves a very good correlation with the mean opinion scores (MOS) of human observers. Inspired by the success of SSIM, several extended versions, such as the multi-scale structural similarity for image quality assessment (MS-SSIM) [4] and Information content weighting for perceptual image quality assessment (IW-SSIM) [5], were proposed by the same research group. IW-SSIM utilizes an image pyramid to decompose the original and distorted images into versions of varying scales, and then computes the information content from the images. Finally, it finds the quality score using the information content as a weighting function.
Based on shared information between the reference and distorted images, Sheikh et al. proposed the information fidelity criteria (IFC) [6] and the visual information fidelity (VIF) [7]. The most apparent distortion (MAD) approach [8] separates images based on the distortion and applies either a detection-based strategy or an appearance-based strategy. Some of the methods, such as the noise quality measure (NQM) [9] and the visual signal-to-noise ratio (VSNR) [10], take into account the HVS by incorporating interactions among different visual signals. In contrast, other approaches, including the popular feature similarity index or FSIM [11], emphasize phase congruency [12,13,14]. FSIM uses the image gradient as a secondary feature and local quality maps are weighted by phase congruency to obtain the final score. The image gradient has been used effectively in a number of other works [15,16]. Xue et al., in their gradient magnitude standard deviation (GMSD) [17], used the gradient magnitude with a different pooling strategy, by applying the standard deviation, and Alaei et al. adopted a similar approach for assessing document images [18]. Both examples prove the effectiveness of standard deviation pooling, however, the authors of the GMSD approach showed that standard deviation (SD) pooling is not effective for all types of methods. Wang et al. proposed the multi-scale contrast similarity deviation (MCSD) metod [19], which can be termed as a continuation of SSIM and MS-SSIM, since it also uses the root mean square (RMS) contrast similarity; however, they employed standard deviation pooling for the final score.
Meanwhile, inspired by vision-related psychological research, visual saliency (VS)-based IQA methods [20,21], which utilize different kinds of visual saliencies [22,23,24], have attracted researchers’ attention. In the visual saliency index (VSI) method [23], VS is used as both a quality map and the weighting function at the pooling stage. The spectral residual similarity index (SR-SIM) [24] uses the spectral residual saliency, which makes the approach very fast while maintaining a competitive correlation with the mean opinion score. Combining VS with other features has also become popular [25,26]. Li et al. proposed an approach that combines VS and FSIM while, recently, Jia et al. used contrast and spectral residual saliency as well as summation-based SD pooling.
In the context of the HVS, center bias in early eye movements is an established fact in psychological vision research [27,28,29,30]. Bindemann found that eye movement is biased not only to the scene center but also to the screen center [31]. As a result, if a scene appears at the center of the screen, it will receive the most attention. For example, in Figure 1, the human eye will first move to the Block05 region and if that part has visually important information then it will attract even more attention. As a result, people will be more sensitive to the distortions in this region. To the best of our knowledge, there is no research in IQA considering this center bias for quality assessment.
In this paper, we propose a new method for IQA which accounts for the center emphasis in HVS. In the proposed method, we first obtain both the contrast and VS similarity maps for the entire image. To give center emphasis, we find the VS similarity map of the mid-region and apply element-wise multiplication in the mid-part to raise the similarity deviation there. However, for the contrast similarity, we apply element-wise squaring in the center part. Contrast is a local quality map, so we do not calculate the contrast of the mid-area separately. On the other hand, VS is a global quality map, and thus it is calculated differently in the mid-region. The final score is obtained by performing weighted summation of the standard deviations on both of the similarity maps; further details with mathematical equations are given in Section 3.
We evaluated our proposed method on three popular benchmark databases for IQA research and compared it with 13 other state-of-the-art approaches. The results in terms of the correlations with human evaluated scores show that the method proposed by us outperforms the other approaches, with a reasonable amount of processing time.
This paper is organized as follows. Section 2 describes some underlying theories and related techniques. Section 3 explains the proposed center-emphasized assessment approach, and the results with relevant discussions are presented in Section 4. Finally, the paper is concluded in Section 5.

2. Background

In this section, we briefly review the underlying theories on which the content of this paper relies, including the spectral residual visual saliency similarity, contrast similarity, standard deviation pooling, and relevant evaluation metrics.

2.1. Spectral Residual Visual Saliency Similarity

In the human visual system (HVS), some interesting or salient regions of an image receive more attention than other parts. Detection of image saliency itself is an active field in vision research. As such, the human is more sensitive to these salient parts and any distortion in these parts attracts more intense attention, which makes it an important feature in IQA methods. There are a lot of saliency detection techniques available [32], among which spectral residual saliency detection [22] is a very fast approach. We adopt the saliency map generator described in the SR-SIM [24] and the visual saliency plus contrast (VSP) [26] approaches.
For an image f ( x , y ) , according to Reference [22], the spectral residual saliency (SRS) is computed as follows:
M f ( u , v ) = a b s [ F { f ( x , y ) } ( u , v ) ]
A f ( u , v ) = a n g l e [ F { f ( x , y ) } ( u , v ) ]
L M ( u , v ) = log { M f ( u , v ) }
R f ( u , v ) = L M ( u , v ) h n ( u , v ) L M ( u , v )
S R S ( x , y ) = g ( x , y ) [ F 1 { exp ( R f + j A f ) } ( x , y ) ] 2 ,
where F and F 1 denote the Fourier and inverse Fourier transforms, respectively; a b s ( . ) and a n g l e ( . ) return the magnitude and argument of a complex number, respectively; h n ( u , v ) is an n × n mean filter; g ( x , y ) is a Gaussian filter; and ∗ denotes the convolution operation.
In this way, we calculate the SRS for both the reference and distorted images denoted by S R S r ( x , y ) and S R S d ( x , y ) , respectively. Then, the spectral residual visual saliency similarity ( S R V S S ( r , d ) ) is calculated as:
S R V S S ( r , d ) = 2 S R S r ( x , y ) S R S d ( x , y ) + c 1 S R S r ( x , y ) 2 + S R S d ( x , y ) 2 + c 1 ,
where ⊙ is the element-wise multiplication, 2 is the element-wise squaring and c 1 is a positive constant used to increase the calculation stability.

2.2. Contrast Similarity

Contrast is a basic perceptual attribute of an image [33] which varies greatly over the image, and the contrast map (CM) contains the spatial distribution of those varying values. There are many ways devised for calculating CMs and in this paper, we adopted the RMS contrast from SSIM, since it achieves better performance for natural images. The RMS contrast map C X of an image signal X is given by:
C X = 1 N 1 i = 1 N ( x i μ X ) 2 1 2 ,
where N is the total number of pixels in the image, x i is the intensity of pixel i, and μ X is the mean intensity, defined as:
μ X = 1 N i = 1 N x i .
Again, using Equation (7), we obtain the contrast maps C r and C d for both reference and distorted images, respectively, and find the contrast similarity C S ( r , d ) as follows:
C S ( r , d ) = 2 C r C d + c 2 C r 2 + C d 2 + c 2 ,
where similar to Equation (6), ⊙ is the element-wise multiplication, 2 is the element-wise squaring and c 2 is a positive constant used to increase the calculation stability.

2.3. Standard Deviation Pooling

As discussed in the introduction, standard deviation (SD) pooling achieves very good performance in specific cases and is adopted by several successful methods. Jia et al. conducted an experiment with several other combinations of pooling and found that SD pooling provides the best correlation. The final quality score ( Q S ) is calculated using the following equation:
Q S = 1 w 1 + w 2 w 1 × s t d { S R V S S ( r , d ) } + w 2 × s t d { C S ( r , d ) } ,
where w 1 and w 2 are weighting factors that specify the importance of VSS and CS, respectively. The standard deviations in the above equation are defined as:
s t d ( S R V S S ( r , d ) ) = 1 M i = 1 M ( S R V S S i μ S R V S S ) 2 1 2
s t d ( C S ( r , d ) ) = 1 M i = 1 M ( C S i μ C S ) 2 1 2 ,
where M is the number of total elements in the similarity matrices; S R V S S i and C S i are the ith items; μ S R V S S and μ C S are the mean values of the S R V S S ( r , d ) and C S ( r , d ) , respectively, and are given by:
μ S R V S S = 1 M i = 1 M S R V S S i
μ C S = 1 M i = 1 M C S i

2.4. Evaluation Metrics

The performance of any IQA method is usually measured by the mean squared error and several other correlations with the subjective scores that are human evaluated values usually in the form of MOS, or their differential DMOS. However, to apply linear correlation, the two compared values should be on the same scale and perfectly linearly correlated [34]. To ensure better fairness, before applying the linear correlation measurements, a logistic mapping function is used to convert the objective scores. We use the following nonlinear regression model as suggested by Sheikh [35].
q = β 1 1 2 1 1 + exp ( β 2 ( q β 3 ) ) + β 4 q + β 5 ,
where q is the objective score calculated by a IQA method, q is the mapped value, and β i , i = 1 , 2 , 3 , 4 , 5 are the parameters that are tuned based on the relationship between objective and subjective scores. We utilized the MATLAB function n l i n f i t to find the optimal parameters. After the mapping is done, the subjective scores are then used with these mapped scores to find the correlation coefficients.
One of the widely adopted basic correlations is Pearson’s linear correlation coefficient (PLCC) which is defined as follows:
P L C C ( o , s ) = i = 1 m ( o i μ o ) ( s i μ s ) i = 1 m ( o i μ o ) 2 1 2 i = 1 m ( s i μ s ) 2 1 2 ,
where m is the number of distorted images; o and s are vectors of the objective and subjective scores, respectively; and μ o and μ s are the mean scores, defined by:
μ o = 1 m i = 1 m o i
μ s = 1 m i = 1 m s i .
In our case, the objective scores o are actually the mapped scores using Equation (15). If the nonlinear mapping in Equation (15) is to be avoided, then rank order coefficients can be used. The most popular Spearman’s rank-order correlation coefficient (SROCC) is defined as:
S R O C C ( o , s ) = P L C C ( r a n k ( o ) , r a n k ( s ) ) .
The function r a n k ( ) of a vector returns a rank-vector, where the i-th entry contains the relative rank of the i-th item in the original vector.
Another popularly adopted rank order coefficient is Kendall’s rank-order correlation coefficient (KROCC), which is given as below:
K R O C C ( o , s ) = C D m ( m 1 ) / 2 ,
where C is the number of concordant pairs that are consistently correlated between objective and subjective scores; and D is the number of discordant pairs.
The root mean square error (RMSE) is also commonly adopted and is defined as:
R M S E ( o , s ) = 1 m i = 1 m ( o i s i ) 2 1 2
A larger value for PLCC, SROCC, and KROCC indicates that the corresponding method is better. On the other hand, a smaller value of the RMSE is a sign of a superior IQA.

3. Proposed Center-Emphasized Quality Assessment

The general flow diagram of our proposed method is presented in Figure 2. At first, the center parts of both the reference and distorted images are extracted. To do this, we split the image in 3 × 3 image blocks as shown in Figure 1, and the fifth block, which resides in the middle both horizontally and vertically, is taken as the center area. If the original image dimension is ( H × W ) , then the corresponding dimension for the center block becomes ( H m i d × W m i d ) , where:
H m i d = [ H 3 ] and W m i d = [ W 3 ] .
The center block is defined as a rectangular area identified by two corner points ( x m i n , y m i n ) and ( x m a x , y m a x ) , where:
x m i n = [ H 3 ] , y m i n = [ W 3 ] , x m a x = [ H 3 ] + H m i d , and y m a x = [ W 3 ] + W m i d
First, the saliency similarity maps for the full images and middle images are found using Equations (1)–(6) and are denoted as V S S and V S S m i d , respectively. Simultaneously, the contrast similarity map for the full-size, C S , is also obtained. As discussed before, we do not derive the CS map for middle images.
Then, we increase the sensitivity of the center area within both of the maps. Let V S S ( m i d ) and C S ( m i d ) be the center areas of V S S and C S , respectively. The updated middle parts will be determined as follows:
V S S ( m i d ) = V S S ( m i d ) V S S m i d
C S ( m i d ) = C S ( m i d ) C S ( m i d ) ,
where ⊙ is the element-wise multiplication.
With the updated middle portion, we obtain the finalized maps V S S f i n a l and C S f i n a l , and using Equation (10), we calculate the final quality score of the proposed method C E Q I as:
C E Q I = 1 w 1 + w 2 w 1 × s t d ( V S S f i n a l ) + w 2 × s t d ( C S f i n a l )
The MATLAB code is publicly available to test the algorithm and can be found online at http://layek.khu.ac.kr/CEQI.

4. Results and Analysis

Experiments were carried out on three popular benchmark databases for IQA research—TID2008 [36], CSIQ [37] and LIVE [38]. Our approach was compared with 13 other state-of-the-art approaches as listed in Table 1. Basic information about the databases is given in Table 2 and the distortion information is recorded in Table 3.
For performance comparison, we use four commonly adopted metrics—Spearman’s rank-order correlation coefficient (SROCC), Kendall’s rank-order correlation coefficient (KROCC), Pearson’s linear correlation coefficient (PLCC), and the root mean square error (RMSE)—which we defined in Section 2.4.
Table 4 compares the four metrics among the different IQA models, for all of the three databases. The top three values for each metric are typed in boldfaced and light-gray shaded; the top value is colored blue; the second highest value is colored red, and the third highest value is colored black. However, in the case of RMSE, coloring is done in a reverse way, i.e., the lowest value is colored in blue and so on, since a lower RMSE implies a better method. We see that, for the biggest database, TID2008, our proposed method outperforms all other methods in all metrics. For the other two databases, it achieves competitive performance. We calculated the weighted averages of the SROCC, KROCC, PLCC, and RMSE using the number of distorted images to find the overall performance, as proposed in Reference [5]. It can be noticed that, compared to VSI and VSP, our approach shown better prediction accuracy with (1.09%, 0.3%)-point, (2.44%, 0.39%)-point and (2.19%, 0.22%)-point higher overall SROC, KROC and PLCC values, respectively. The overall ranking based on performance is shown in Table 5.
Table 6 compares the SROCC performance for all distortion types; please refer to Table 3 for a description of the abbreviations. We see that different methods perform better for different distortions and performance even varies between databases. This is the case because images are not affected equally by a specific type of distortion—it depends on the color, salient regions, and perhaps a combination of many other factors. Still, distortion-wise comparison gives us a good understanding of whether an IQA method is biased to some noise type or not. It can be seen that the proposed CEQI performs consistently well for all types of distortion; it is not too biased to any specific type of distortion, while retaining an average performance within the top 3 methods.
Figure 3 shows scatter plots of the predicted scores for different IQA approaches with the MOS/DMOS values on the TID2008 database. These results show that CEQI’s prediction is consistent compared to other methods, while providing a better correlation. We do not include PSNR because its predictions are very inconsistent. NQM is also discarded for the same reason, although its performance is not as inconsistent as PSNR.
Although the prime consideration of an IQA model is the performance of its prediction, having a low computational cost is also a desirable feature, especially for a real-time system. We evaluated the various IQA models with MATLAB R2017b using a computer equipped with an Intel(R) Core(TM) i5-4670 CPU with a 3.40 GHz processor and 16 GB of RAM. The MATLAB codes provided by the authors were used and elapsed time was recorded using the traditional tic-toc function. The results of these tests are shown in Table 7. As expected, PSNR has the lowest computation time. Surprisingly, the gradient magnitude similarity division model can process 263.05 images per second with satisfactory performance (rank 4 as shown in Table 5). VIF shows very good performance for the LIVE database where it is the best-performing IQA, but it can only process 1.79 images per second on average, which makes it inappropriate for real-time systems or systems with low processing capability. On the other hand, CEQI takes 15.25 ms to process an image, with the capability of processing 65.51 images per second. This frame rate meets the need for almost all kinds of real-time systems.

5. Conclusions

In this paper, we considered the center bias of HVS and proposed a full-reference image quality assessment method, CEQI, combining visual saliency and contrast. We placed extra emphasis on the center part of the image so that any degradation within the center region results in more effects than other regions. The proposed approach was compared with other state-of-the-art IQA models and it outperforms other competing methods in most cases. Comparing individual distortion types, the proposed method gives consistent scores. Additionally, the running time is suitable for real-time applications. The center emphasis makes the method more balanced and robust. We believe that this center emphasis will enhance the performance of other existing IQA models, including no-reference and reduced-reference approaches. In our future work, we will investigate these possibilities.

Author Contributions

This paper represents the result of collaborative teamwork. Conceptualization, M.A.L.; Funding acquisition, E.-N.H.; Software, M.A.L. and A.F.M.S.U.; Visualization, M.A.L. and L.P.T.; Writing—original draft, M.A.L.; Writing—review & editing, M.A.L., E.-N.H. and T.C.

Funding

This work was supported by Institute for Information & communications Technology Promotion (IITP) grant funded by the Korea government (MSIT) (No. 2017-0-00294, Service mobility support distributed cloud technology).

Acknowledgments

This work was supported by Institute for Information & communications Technology Promotion (IITP) grant funded by the Korea government (MSIT) (No. 2017-0-00294, Service mobility support distributed cloud technology).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Girod, B. Psychovisual aspects of image processing: What’s wrong with mean squared error? In Proceedings of the Seventh Workshop on IEEE Multidimensional Signal Processing, Lake Placid, NY, USA, 23–25 September 1991; p. 2. [Google Scholar]
  2. Wang, Z.; Bovik, A.C. Mean squared error: Love it or leave it? A new look at signal fidelity measures. IEEE Signal Process. Mag. 2009, 26, 98–117. [Google Scholar] [CrossRef]
  3. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; IEEE: New York, NY, USA, 2003; Volume 2, pp. 1398–1402. [Google Scholar]
  5. Wang, Z.; Li, Q. Information content weighting for perceptual image quality assessment. IEEE Trans. Image Process. 2011, 20, 1185–1198. [Google Scholar] [CrossRef] [PubMed]
  6. Sheikh, H.R.; Bovik, A.C.; De Veciana, G. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans. Image Process. 2005, 14, 2117–2128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Sheikh, H.R.; Bovik, A.C. A visual information fidelity approach to video quality assessment. In Proceedings of the First International Workshop on Video Processing and Quality Metrics for Consumer Electronics, Doubletree Paradise Valley Resort, Scottsdale, AZ, USA, 23–25 January 2005; pp. 23–25. [Google Scholar]
  8. Larson, E.C.; Chandler, D.M. Most apparent distortion: Full-reference image quality assessment and the role of strategy. J. Electron. Imaging 2010, 19, 011006. [Google Scholar]
  9. Damera-Venkata, N.; Kite, T.D.; Geisler, W.S.; Evans, B.L.; Bovik, A.C. Image quality assessment based on a degradation model. IEEE Trans. Image Process. 2000, 9, 636–650. [Google Scholar] [CrossRef] [PubMed]
  10. Chandler, D.M.; Hemami, S.S. VSNR: A wavelet-based visual signal-to-noise ratio for natural images. IEEE Trans. Image Process. 2007, 16, 2284–2298. [Google Scholar] [CrossRef] [PubMed]
  11. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A feature similarity index for image quality assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed]
  12. Kovesi, P. Image features from phase congruency. Videre 1999, 1, 1–26. [Google Scholar]
  13. Liu, Z.; Laganière, R. Phase congruence measurement for image similarity assessment. Pattern Recognit. Lett. 2007, 28, 166–172. [Google Scholar] [CrossRef]
  14. Saha, A.; Wu, Q.J. Perceptual image quality assessment using phase deviation sensitive energy features. Signal Process. 2013, 93, 3182–3191. [Google Scholar] [CrossRef]
  15. Chen, G.H.; Yang, C.L.; Xie, S.L. Gradient-based structural similarity for image quality assessment. In Proceedings of the 2006 IEEE International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; IEEE: New York, NY, USA, 2006; pp. 2929–2932. [Google Scholar]
  16. Zhu, J.; Wang, N. Image quality assessment by visual gradient similarity. IEEE Trans. Image Process. 2012, 21, 919–933. [Google Scholar] [PubMed]
  17. Xue, W.; Zhang, L.; Mou, X.; Bovik, A.C. Gradient magnitude similarity deviation: A highly efficient perceptual image quality index. IEEE Trans. Image Process. 2014, 23, 684–695. [Google Scholar] [CrossRef] [PubMed]
  18. Alaei, A.; Conte, D.; Raveaux, R. Document image quality assessment based on improved gradient magnitude similarity deviation. In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Nancy, France, 23–26 August 2015; pp. 176–180. [Google Scholar]
  19. Wang, T.; Zhang, L.; Jia, H.; Li, B.; Shu, H. Multiscale contrast similarity deviation: An effective and efficient index for perceptual image quality assessment. Signal Process. Image Commun. 2016, 45, 1–9. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, L.; Shen, Y.; Li, H. VSI: A visual saliency-induced index for perceptual image quality assessment. IEEE Trans. Image Process. 2014, 23, 4270–4281. [Google Scholar] [CrossRef] [PubMed]
  21. Ma, Q.; Zhang, L. Saliency-based image quality assessment criterion. In International Conference on Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2008; pp. 1124–1133. [Google Scholar]
  22. Hou, X.; Zhang, L. Saliency detection: A spectral residual approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2007 (CVPR’07), Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  23. Duan, L.; Wu, C.; Miao, J.; Qing, L.; Fu, Y. Visual saliency detection by spatially weighted dissimilarity. In Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 473–480. [Google Scholar]
  24. Zhang, L.; Li, H. SR-SIM: A fast and high performance IQA index based on spectral residual. In Proceedings of the 2012 19th IEEE International Conference on Image Processing (ICIP), Orland, FL, USA, 30 September–3 October 2012; pp. 1473–1476. [Google Scholar]
  25. Li, A.; She, X.; Sun, Q. Color image quality assessment combining saliency and fsim. In Proceedings of the Fifth International Conference on Digital Image Processing (ICDIP 2013), Beijing, China, 21–22 April 2013; International Society for Optics and Photonics: Bellingham, WA, USA, 2013; Volume 8878, p. 88780. [Google Scholar]
  26. Jia, H.; Zhang, L.; Wang, T. Contrast and visual saliency similarity-induced index for assessing image quality. IEEE Access 2018, 6, 65885–65893. [Google Scholar] [CrossRef]
  27. Langford, R.C. How people look at pictures, a study of the psychology of perception in art. J. Educ. Psychol. 1936, 27, 397–398. [Google Scholar] [CrossRef]
  28. Mannan, S.; Ruddock, K.; Wooding, D. Fixation sequences made during visual examination of briefly presented 2D images. Spat. Vis. 1997, 11, 157–178. [Google Scholar] [CrossRef] [PubMed]
  29. Parkhurst, D.; Law, K.; Niebur, E. Modeling the role of salience in the allocation of overt visual attention. Vis. Res. 2002, 42, 107–123. [Google Scholar] [CrossRef] [Green Version]
  30. Tatler, B.W. The central fixation bias in scene viewing: Selecting an optimal viewing position independently of motor biases and image feature distributions. J. Vis. 2007, 7, 4. [Google Scholar] [CrossRef] [PubMed]
  31. Bindemann, M. Scene and screen center bias early eye movements in scene viewing. Vis. Res. 2010, 50, 2577–2587. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Cong, R.; Lei, J.; Fu, H.; Cheng, M.M.; Lin, W.; Huang, Q. Review of visual saliency detection with comprehensive information. arXiv, 2018; arXiv:1803.03391. [Google Scholar] [CrossRef]
  33. Peli, E. Contrast in complex images. JOSA A 1990, 7, 2032–2040. [Google Scholar] [CrossRef]
  34. Ding, Y. General Framework of Image Quality Assessment. In Visual Quality Assessment for Natural and Medical Image; Springer: Berlin/Heidelberg, Germany, 2018; pp. 45–62. [Google Scholar]
  35. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Proc. 2006, 15, 3440–3451. [Google Scholar] [CrossRef]
  36. Ponomarenko, N.; Lukin, V.; Zelensky, A.; Egiazarian, K.; Carli, M.; Battisti, F. TID2008—A database for evaluation of full-reference visual quality assessment metrics. Adv. Modern Radioelectron. 2009, 10, 30–45. [Google Scholar]
  37. Larson, E.C.; Chandler, D. Categorical Image Quality (CSIQ) Database. 2010. Available online: http://vision.eng.shizuoka.ac.jp/ (accessed on 23 December 2018).
  38. Sheikh, H. LIVE Image Quality Assessment Database Release 2. 2005. Available online: http://live. ece. utexas. edu/research/quality (accessed on 20 December 2018).
Figure 1. The image ’Sailing.bmp’ is split into nine blocks, where Block05 is the center area.
Figure 1. The image ’Sailing.bmp’ is split into nine blocks, where Block05 is the center area.
Symmetry 11 00296 g001
Figure 2. Flow diagram of the proposed center-emphasized approach.
Figure 2. Flow diagram of the proposed center-emphasized approach.
Symmetry 11 00296 g002
Figure 3. Scatter plots of the mean opinion scores (MOS/DMOS) versus scores predicted by different methods on the TID2008 database. The black curves are obtained by the nonlinear fitting based on Equation (15).
Figure 3. Scatter plots of the mean opinion scores (MOS/DMOS) versus scores predicted by different methods on the TID2008 database. The black curves are obtained by the nonlinear fitting based on Equation (15).
Symmetry 11 00296 g003
Table 1. Overview of the compared image quality assessment (IQA) methods.
Table 1. Overview of the compared image quality assessment (IQA) methods.
IQA MethodPrinciple ConsiderationPooling UsedComments
PSNRPixel-by-pixel errorAveragePrimitive method, does not consider HVS, poor correlation with humans, low computation, widely used.
[3] SSIMLuminance, contrast, structureAverageMilestone method to consider structural information that better represents HVS.
[4] MS-SSIMMulti-scale structureWeighted sumAn extension to SSIM that incorporates variations of viewing conditions and capable of multi-scale assessment.
[5] IW-SSIMInformation content extractionWeighted by information contentMain emphasis is on information content extraction which is applicable to other methods as well, uses image pyramid.
[7] MADMost visible distortionWeighted productA novel strategy consisting of two phases to detect the most apparent distortion.
[11] FSIMPhase congruency, Gradient magnitudeWeighted averageA state-of-the art IQA approach using phase congruency and the gradient magnitude weighted by phase congruency to calculate the final score.
[17] GMSDImage gradientStandard deviationVery fast assessment approach after PSNR showing competitive performance with other state-of-the-art approaches.
[23] VSIVisual saliency (VS)Weighted averageIntroduced visual saliency for IQA where VS is used for both the quality map and weighting function at the pooling stage.
[19] MCSDMulti-scale contrastStandard deviationAnother fast method next to GMSD, but providing better results.
[7] VIFVisual information extractionAverageQuantifies the extracted reference information from a distorted image.
[9] NQMDistortion and noiseSquared sumConsiders HVS using distortion and noise, better than PSNR.
[24] SR-SIMSaliency and gradientWeighted averageUses spectral residual saliency and image gradient, which shows competitive performance.
[26] VSPSaliency and contrastWeighted sum of deviationsUses deviation pooling in a combined method.
CEQI (Proposed)Saliency, contrast, center emphasisWeighted sum of deviationsGives special attention to the center area.
PSNR: peak signal-to-noise ratio; SSIM: structural similarity index; MS-SSIM: multi-scale structural similarity; IW-SSIM: information content weighted structural similarity; MAD: most apparent distortion; FSIM: feature similarity index; GMSD: gradient magnitude standard deviation; VSI: visual saliency index; MCSD: multi-scale contrast similarity deviation; VIF: visual information fidelity; NQM: noise quality measure; SR-SIM: spectral residual similarity index; VSP: visual saliency plus; CEQI: center-emphasized quality index; HVS: human visual system.
Table 2. Basic information about the databases used for the experiments.
Table 2. Basic information about the databases used for the experiments.
DatasetReference ImagesDistorted ImagesDistortion TypesNo. of Subjects
TID200825170017838
CSIQ30866635
LIVE297795161
Table 3. Types of distortion used in each database.
Table 3. Types of distortion used in each database.
TID2008CSIQLIVEType of DistortionAbbreviation
YYYAdditive Gaussian noiseAGN
Y--Additive noise in color componentsANC
Y--Spatially correlated noiseSCN
Y--Masked noiseMN
Y--High frequency noiseHFN
Y--Impulse noiseIN
Y--Quantization noiseQN
YYYGaussian blurGB
Y--Image denoisingIDN
YYYJPEG compressionJPEG
YYYJPEG2000 compressionJP2K
Y--JPEG transmission errorsJGTE
Y--JPEG2000 transmission errorsJ2TE
Y--Non-eccentricity pattern noiseNEPN
Y--Local block-wise distortions of different intensityLBD
Y--Mean shift (intensity shift)MS
YY-Contrast changeCTC
--YFast fading RayleighFF
-Y-Additive pink Gaussian noiseAWPN
Table 4. Performance comparison of IQA methods on three databases.
Table 4. Performance comparison of IQA methods on three databases.
DatasetMetricPSNRSSIMMS-SSIMIW-SSIMMADFSIMGMSDVSIMCSDVIFNQMSRSIMVSPCEQI (Prop.)
TID 2008SROCC0.52450.77490.85420.85590.83400.88050.89070.89790.89110.74910.62430.89130.90010.9069
KROCC0.36960.57680.65680.66360.64450.69460.70920.71230.71330.58600.46080.71490.72150.7307
PLCC0.53090.77320.84510.85790.83060.87380.87880.87620.88440.80840.60850.88670.89620.9014
RMSE1.13720.85110.71730.68950.74730.65250.64040.64660.62630.78991.06490.62050.59530.5810
CSIQSROCC0.83880.87550.91320.92120.94660.92420.95700.94220.95920.91940.74360.93180.95790.9563
KROCC0.63510.69000.73860.75220.79630.75610.81220.78500.81710.75320.56480.77180.81710.8138
PLCC0.82760.86120.89910.91440.95020.91200.95410.92790.95600.92570.74330.92500.95890.9565
RMSE0.14740.13340.11490.10630.08180.10770.07860.09790.07700.09930.17560.09980.07450.0766
LIVESROCC0.87650.94600.95120.96040.95670.96100.95460.94640.96030.97190.85450.95580.95730.9577
KROCC0.70120.80570.81810.83790.82900.83800.82360.80000.83500.85710.69380.81900.82970.8307
PLCC0.91320.93850.94680.95150.94930.94920.95110.94310.95400.97230.87730.94530.95230.9534
RMSE9.41977.98387.43807.11167.26907.27627.13747.68566.93295.403011.09417.54347.05066.9730
OVERALLSROCC0.69860.84680.89540.90080.89540.91340.92460.92210.92690.85230.71710.91900.93000.9330
KROCC0.52620.66780.72140.73350.73260.74930.76600.75430.77230.70180.55070.75760.77480.7787
PLCC0.70910.84040.88640.89760.89260.90400.91720.90730.92110.88240.71580.91230.92700.9292
RMSE3.18802.65012.43042.32462.38992.35282.30152.46092.23771.89813.62372.40952.25492.2270
– For each row, the first, second and third-ranked performances are highlighted respectively in blue, red and black colors; – For SROCC, KROCC and PLCC metrics, the higher the value, the better the method whereas for RMSE a lower score is better.
Table 5. Overall performance ranking of the compared IQA methods.
Table 5. Overall performance ranking of the compared IQA methods.
IQASROCCKROCCPLCCRMSE
PSNR14141413
SSIM12121212
MS-SSIM9101010
IW-SSIM8886
MAD10998
FSIM7777
GMSD4445
VSI56611
MCSD3333
VIF1111111
NQM13131314
SRSIM6559
VSP2224
CEQI (Proposed)1112
– Number 1 indicates the best performing method and 14 is the worst.
Table 6. Distortion-wise SROCC performance comparison of the IQA methods on three databases.
Table 6. Distortion-wise SROCC performance comparison of the IQA methods on three databases.
DatasetDist.PSNRSSIMMS-SSIMIW-SSIMMADFSIMGMSDVSIMCSDVIFNQMSRSIMVSPCEQI (Prop.)
TID 2008AGN0.91150.81090.80860.78690.83840.85620.91800.92400.91870.87970.76790.89990.92020.9203
ANC0.90680.80290.80540.79200.83070.85270.89770.91180.88980.87570.75960.89520.89690.9014
SCN0.92290.81450.82090.77140.86800.84870.91280.93510.92100.86980.77200.90840.90580.9060
MN0.84870.77950.81070.80880.73360.80210.73470.80110.73210.86830.70710.79060.77620.7835
HFN0.93230.87740.87340.87030.88750.91530.91730.92580.91800.90750.90300.91970.91900.9204
IN0.91770.67320.69070.64650.05790.74520.66110.82980.68930.83270.77710.76670.69810.7176
QN0.87000.85310.85890.81770.81600.85640.88750.87310.89520.88130.83170.83540.88970.8809
GB0.86730.95440.95630.96360.91960.94720.89680.95290.88800.95400.88460.95490.92960.9320
IDN0.93810.95300.95820.94730.94330.96030.97520.96930.97660.91830.94500.96670.96930.9706
JPEG0.90110.92520.93220.92070.93270.93700.95250.96160.94860.91680.90750.94110.94450.9467
JP2K0.83010.96300.97000.97380.97070.97730.97950.98480.97870.97090.95310.98040.97740.9762
JGTE0.76640.86780.86810.85880.86610.87080.86210.91600.76810.85850.73590.88770.88930.8985
J2TE0.77650.85770.86060.82030.83940.85440.88250.89420.89460.85010.74120.89070.87040.8752
NEPN0.59310.71070.73770.77240.82980.74920.76010.76990.79860.76190.68000.76720.76470.7727
LBD0.58510.84620.75600.76340.79700.84940.89670.62880.89330.83240.33670.77890.84040.8295
MS0.70760.72310.73380.70670.51630.67200.64860.67140.53500.50960.54400.57310.67360.7177
CTC0.61260.44170.63810.63010.32360.64810.63460.65570.59320.84030.82630.64820.56950.5399
AVG0.81690.81500.82820.81470.76300.84370.84810.85910.83760.85460.76900.84730.84910.8523
CSIQAWGN0.93440.89740.94710.93800.95420.92620.96760.96370.96740.95750.93870.96310.96650.9680
JPEG0.90080.95430.96310.96600.96140.96520.96510.96150.96700.97030.95250.96700.96890.9685
JP2K0.93070.96050.96820.96820.97520.96840.97170.96920.97460.96710.96290.97720.97780.9777
AGPN0.93150.89240.93300.90570.95680.92330.95020.96360.94790.95090.91140.95190.95160.9525
GB0.93590.96080.97110.97810.96810.97280.97120.96790.97470.97440.95830.97670.97880.9777
CTC0.88610.79250.95280.95400.92100.94200.90370.95050.95090.93450.94780.95300.93240.9354
AVG0.91990.90970.95590.95170.95610.94970.95490.96270.96380.95910.94530.96480.96270.9633
LIVEJP2K0.95060.97620.98020.97910.97750.98190.98230.97000.98250.97380.97020.97270.98150.9813
JPEG0.93610.95980.96260.96020.94640.96250.96070.95340.96130.95680.94690.95460.96150.9614
AWGN0.86430.98010.98450.98070.99040.97980.98470.98810.98890.98990.82420.98650.98870.9900
GB0.03590.95170.97330.98380.96920.98320.97510.97030.97280.98260.84530.97820.98040.9804
FF0.93060.96430.96900.96740.97480.97070.96580.96440.97230.97780.79290.96660.97630.9767
AVG0.74350.96640.97390.97420.97170.97560.97370.96920.97560.97620.87590.97170.97770.9780
– For each row, the first, second and third-ranked performances are highlighted respectively in blue, red and black colors. – The acronyms for distortions are defined in Table 3, AVG refers to average performance over all noises in a database.
Table 7. Running time comparison of the IQA models.
Table 7. Running time comparison of the IQA models.
IQARunning Time (ms)Images per Second
PSNR3.47288.49
SSIM7.47133.88
MS-SSIM25.7138.90
IW-SSIM157.856.34
MAD696.001.44
FSIM115.418.66
GMSD3.80263.05
VSI64.1215.60
MCSD6.34157.70
VIF560.191.79
NQM106.599.38
SRSIM10.2997.20
VSP10.7593.03
CEQI (Proposed)15.2765.51

Share and Cite

MDPI and ACS Style

Layek, M.A.; Uddin, A.F.M.S.; Le, T.P.; Chung, T.; Huh, E.-N. Center-Emphasized Visual Saliency and a Contrast-Based Full Reference Image Quality Index. Symmetry 2019, 11, 296. https://doi.org/10.3390/sym11030296

AMA Style

Layek MA, Uddin AFMS, Le TP, Chung T, Huh E-N. Center-Emphasized Visual Saliency and a Contrast-Based Full Reference Image Quality Index. Symmetry. 2019; 11(3):296. https://doi.org/10.3390/sym11030296

Chicago/Turabian Style

Layek, Md. Abu, A. F. M. Shahab Uddin, Tuyen P. Le, TaeChoong Chung, and Eui-Nam Huh. 2019. "Center-Emphasized Visual Saliency and a Contrast-Based Full Reference Image Quality Index" Symmetry 11, no. 3: 296. https://doi.org/10.3390/sym11030296

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop