ISSN ONLINE(2278-8875) PRINT (2320-3765)

All submissions of the EM system will be redirected to Online Manuscript Submission System. Authors are requested to submit articles directly to Online Manuscript Submission System of respective journal.

Performance Evaluation of Blind Image Quality Assessment on Deconvolving Images

R.Likhitha1, K.Ramesh2
  1. M.Tech Student, Dept. Of Electronics & Communication Engineering, Sree Vidyanikethan Engineering College, Tirupati, AP, India
  2. M.Tech, Assistant Professor, Dept. Of Electronics & Communication Engineering, Sree Vidyanikethan Engineering College, Tirupati, AP, India
Related article at Pubmed, Scholar Google

Visit for more related articles at International Journal of Advanced Research in Electrical, Electronics and Instrumentation Engineering

Abstract

We develop an efficient general-purpose blind/no-reference image quality assessment (IQA) algorithm using a natural scene statistics (NSS) model of discrete cosine transform (DCT) coefficients. The algorithm is computationally appealing, given the availability of platforms optimized for DCT computation. The approach relies on a simple Bayesian inference model to predict image quality scores given certain extracted features. The features are based on an NSS model of the image DCT coefficients. The estimated parameters of the model are utilized to form features that are indicative of perceptual quality. These features are used in a simple Bayesian inference approach to predict quality scores. The resulting algorithm, which we name BLIINDS-II, requires minimal training and adopts a simple probabilistic model for score prediction. Given the extracted features from a test image, the quality score that maximizes the probability of the empirically determined inference model is chosen as the predicted quality score of that image. When tested on the LIVE IQA database, BLIINDS-II is shown to correlate highly with human judgments of quality, at a level that is competitive with the popular SSIM index.

INTRODUCTION

The ubiquity of transmitted digital visual information in daily and professional life, and the broad range of applications that rely on it, such as personal digital assistants, high-definition televisions, internet video streaming, and video on demand, necessitate the means to evaluate the visual quality of this information. The various stages of the pipeline through which an image passes can introduce distortions to the image, beginning with its capture until its consumption by a viewer. The acquisition, digitization, compression, storage, transmission, and display processes all introduce modifications to the original image.
Only recently did full-reference image quality assessment (FR-IQA) methods reach a satisfactory level of performance, as demonstrated by high correlations with human subjective judgments of visual quality. SSIM, MS-SSIM, VSNR, VIF index, and the divisive normalization-based indices in and are examples of successful FR-IQA algorithms. These methods require the availability of a reference signal against which to compare the test signal. In many applications, however, the reference signal is not available to perform a comparison against. This strictly limits the application domain of FR-IQA algorithms and points to the need for reliable blind/NR-IQA algorithms. However, no NR-IQA algorithm has been proven consistently reliable in performance. While some FR-IQA algorithms are reliable enough to be deployed in standards, generic NR-IQA algorithms have been regarded as having a long way to go before reaching similar useful levels of performance. The problem of blindly assessing the visual quality of images, in the absence of a reference, and without assuming a single distortion type, requires dispensing with older ideas of quality such as fidelity, similarity, and metric comparison. Presently, NR-IQA algorithms generally follow one of three trends: 1) distortion-specific approaches. These employ a specific distortion model to drive an objective algorithm to predict a subjective quality score. These algorithms quantify one or more distortions such as blockiness, blur, or ringing and score the image accordingly; 2) training based approaches: these train a model to predict the image quality score based on a number of features extracted from the image ; and 3) natural scene statistics (NSS) approaches: these rely on the hypothesis that images of the natural world (i.e., distortion-free images) occupy a small subspace of the space of all possible images and seek to find a distance between the test image and the subspace of natural images To perform the JPEG coding, an image (in colour or grey scales) is first subdivided into blocks of 8x8 pixels. The Discrete Cosine Transform (DCT) is then performed on each block. This generates 64 coefficients which are then quantised to reduce their magnitude. The coefficients are then reordered into a onedimensional array in a zigzag manner before further entropy encoding. The compression is achieved in two stages; the first is during quantisation and the second during the entropy coding process.JPEG decoding is the reverse process of coding.

EXISTING METHOD

The alternating direction method of multipliers (ADMM), originally proposed in the 1970’s, emerged recently as a flexible and efficient tool for several imaging inverse problems, such as denoising , deblurring, inpainting, reconstruction, motion segmentation, to mention only a few classical problems. ADMM-based approaches make use of variable splitting, which allows a straightforward treatment of various priors/regularizers, such as those based on frames or on total-variation (TV), as well as the seamless inclusion of several types of constraints. ADMM is closely related to other techniques, namely the socalled Bregman and split Bregman methods and Douglas-Rachford splitting. Several ADMM-based algorithms for imaging inverse problems require, at each iteration, solving a linear system (equivalently, inverting a matrix). As illustrated in Figure 1, all these standard BCs are quite unnatural, as they do not correspond to any realistic imaging system, being motivated merely by computational convenience. Namely, assuming a periodic boundary condition has the advantage of allowing a very fast implementation of the convolution as a point-wise multiplication in the DFT

Fig. 1. Illustration of the (unnatural) assumptions underlying the periodic,reflexive, and zero boundary conditions.
domain, efficiently implementable using the FFT. However, in real problems, there is no reason for the external (unobserved) pixels to follow periodic (or any other) boundary conditions. A well known consequence of this mismatch is a degradation of the deconvolved images, such as the appearance of ringing artifacts emanating from the boundaries. These artifacts can be reduced by pre-processing the image to reduce the spurious discontinuities at the boundaries, created by the (wrong) periodicity assumption; this is what is done by the “edgetaper” function in the MATLAB Image Processing Toolbox. In a more sophisticated version of this idea that was recently proposed, the observed image is extrapolated to create a larger image with smooth BCs `
image
image

PROPOSED METHOD

Our approach relies on the IQA algorithm learning how the NSS model parameters vary across different perceptual levels of image distortion. The algorithm is trained using features derived directly from a generalized parametric statistical model of natural image DCT coefficients against various perceptual levels of image distortion. The learning model is then used to predict perceptual image quality scores. Unlike much of the prior work on image/video quality assessment (QA) .We make little direct use of specific perceptual models such as area cortical decompositions , masking and motion perception. Yet we consider our approach as perceptually relevant since the NSS models reflect statistical properties of the world that drive perceptual functions of the HVS. This is a consequence of the belief that the HVS is adapted to the statistics of its visual natural environment. In other words, models of natural scenes embody characteristics of the HVS, which is hypothesized to be evolutionally adapted to models conforming to natural scenes. HVS characteristics that are intrinsic to, or that can be incorporated into NSS models include: 1) visual sensitivity to structural information; 2) perceptual masking ; 3) visual sensitivity to directional information ; 4) multi scale spatial visual processing ; and 5) intolerance to flagrantly visible visual distortions . In the following sections we explain how one or more of these HVS properties are embedded in the model. The generalized Gaussian model has recently been used as afeature in a NSS-based RR-IQA algorithm and in a simpletwo-stage NR-IQA algorithm .
The univariate generalized Gaussian density is given by
image
image
where μ is the mean, γ is the shape parameter, and α and β are the normalizing and scale parameters given by
image
where σ is the standard deviation, and _ denotes the gamma function given by
image
This family of distributions includes the Gaussian distribution (β = 2) and the Laplacian distribution (β = 1). As β → ∞, the distribution converges to a uniform distribution.The GGD at varying levels of the shape parameter γ. A variety of parameter estimation methods have been proposed for this model. The multivariate version of the generalized Gaussian density is given by
image
Visual images are subjected to local spatial frequency decompositions in the visual cortex. Likewise, in our IQA model, feature extraction is performed in the local frequency (DCT) domain. The main motivation behind feature extraction in the DCT domain is the observation that the statistics of DCT coefficients change with the degree and type of image distortion. Another advantage is computational convenience: optimized DCT-specific platforms , and fast algorithms for DCT computation,can ease computation. For instance, DCTs can be computed efficiently by variable-change transforms from computationally efficient fast Fourier transform algorithms. Many image and video compression algorithms are based on block-based DCT transforms (JPEG, MPEG2, H263, and H264 that relies on a variation of the DCT). Consequently, the model-based method could be applied to already-computed coefficients, resulting in even greater computational efficiency. Finally, and perhaps most importantly, it is possible to define simple and naturally defined model-based DCT features that capture perceptually relevant image and distortion characteristics in a natural and convenient manner.We illustrate one instance of how the statistics of DCT coefficients changes as an image becomes distorted, which shows the DCT coefficient histograms of a distortionfree image and a Gaussian blur distorted image ,respectively. The differences in observed DCT coefficient distributions between distorted and non distorted images are exploited in the design of features of visual quality score prediction.

EXPERIMENTAL RESULTS

image
Fig.1.Debluring output
image
Fig.2.Deconvolution output
image
Fig.3.Debluring output

image

Fig.4.BIAQ Deconvolution Image

Table for comparing Improved SNR values for different algorithms with present method
image
image
(a)-(b) Improved SNR values

CONCLUSION

The main limitation of these types of “learning based” algorithms is that they require training to learn the prediction parameters (i.e., they suffer regression limitations). Consequently, if these algorithms are trained on a subset (of all possible) image distortions, then these algorithms are expected to perform well on the distortions they have encountered during training, or on distortions that affect images in a similar manner to the ones encountered during training. We leave the design of effective no-reference methods that are completely non reliant on training as challenging future work.
There are significant design differences between DIIVINE and BLIINDS-II. DIIVINE uses a dense complex representation of images in the wavelet domain and extracts a large number of features to train two stages of the algorithm: 1) a nonlinear SVM training for classification and 2) a nonlinear SRV training for regression within each class. Given M assumed distortions, DIIVINE requires M distortion-specific quality assessment engines to be trained and applied. Hence, DIIVINE does not directly accomplish multi distortion QA. Instead, it computes a probability There are significant design differences between DIIVINE and BLIINDS-II. DIIVINE uses a dense complex representation of images in the In addition, the DIIVINE index and the BLIINDS index target essentially different application domains. The DIIVINE index, by a twostage strategy, enables the identification of distortions afflicting the image. This is not only valuable for accomplishing directed quality assessment but also for iden SAAD.: BLIND IMAGE QUALITY ASSESSMENT 3351 tifying image distortions to be repaired. This is accomplished at considerable computational expense using a much larger feature set and sophisticated learning mechanisms. By comparison, the BLIINDS index is designed to achieve the speed and performance required by a quality assessment algorithm operating in a high speed video network. It accomplishes this via a simple onestage QA process using a small number of NSS features that are easily computed from a small (subsampled) number of fast DCT coefficients, using a very simple probabilistic classifier. In the future, we envision NSS-based QA algorithms that use spatiotemporal features for video-QA and NSS-depth features for stereo-QA. These may efficiently operate in the DCT domain like BLIINDS-II.

References