No reference image quality assessment for JPEG2000 based on spatial features

https://doi.org/10.1016/j.image.2008.03.005Get rights and content

Abstract

Perceptual image quality evaluation has become an important issue, due to increasing transmission of multimedia contents over the Internet and 3G mobile networks. Most of the no reference perceptual image quality evaluations traditionally attempted to quantify the predefined artifacts of the coded images. Under the assumption that human visual perception is very sensitive to edge information of an image and any kinds of artifacts create pixel distortion, we propose a new approach for designing a no reference image quality evaluation model for JPEG2000 images in this paper, which uses pixel distortions and edge information. Subjective experiment results on the images are used to train and test the model, which has achieved good quality prediction performance.

Introduction

It is an ever increasing requirement to send more multimedia data over tighter bandwidth which has been driven to develop advanced compression technology. Due to the advanced development of different image compression techniques and processing systems, there is a very big concern about the levels of image quality both for providers and users in many image processing applications from compression to printing. Obviously, digital images suffer a wide variety of distortion in these applications and perceptual quality of the images are degraded. Therefore, perceptual image quality measurement is an important problem. Though the subjective test is considered to be the most accurate method since it reflects human perception, it is time consuming and expensive. Furthermore, it cannot be done in real time. As a result, developing objective image quality evaluation methods are getting more attention nowadays. There are three types of methods that are used for objective image quality evaluation: full-reference (FR), reduced-reference (RR) and no-reference (NR). In the FR method, a reference/original image is required to assess the quality of the distorted image. Therefore, it is highly desirable to develop a quality assessment method that does not require full access to the reference images. In the RR method, some extracted features of the reference/original image are required to assess the quality. However, in many practical applications, the reference image is not available and an NR quality assessment approach is desirable.

The most widely used objective image quality/distortion metrics are peak signal-to-noise ratio (PSNR) and mean squared error (MSE), but they are widely criticized, among others things, for not correlation well with perceived quality measurement. In the past, a great deal of effort has been made to develop new objective image/video quality metrics that incorporate perceptual quality measures by considering human visual system (HVS) characteristics [1], [2], [3], [4], [5], [6]. Most of the proposed image quality assessment approaches require the original image as a reference.

Nevertheless, human beings do not need to have access to the reference image to make judgements regarding quality. Human observers can easily assess the quality of distorted images without using any reference image. By contrast, designing objective NR quality measurement algorithms is a very difficult task. This is mainly due to the limited understanding of the HVS, and it is believed that effective NR quality assessment is feasible only when prior knowledge about the image distortion types is available. Although only a limited number of methods have been proposed in the literatures for objective NR quality assessment, this topic has attracted a great deal of attention recently [7], [8], [9], [10], [11], [12], [13], [14], [15], [16], [17], [18], [19]. Since the predominant mode for image and video coding and transmission is using block-based video compression algorithms, blind measurement of the blocking artifact has been the main emphasis of NR quality assessment researches [7], [8], [9], [10], [11], [12], [13]. Blockiness, activity and segmentation-based measure are explained either in the spatial domain [7], [8], [9], [10], [11] or in the frequency domain [12], [13].

However, the above described methods would obviously fail for any other distortion types, such as ringing or blurring introduced by the JPEG2000 image compression algorithm, or the H.264 video compression algorithm. Some researchers have attempted to quantify the blurring and ringing artifacts without reference. In [14], a visible ringing measure (VRM) is proposed that captures the ringing artifact around strong edges. The algorithm is based on constructing an image mask that exposes only those parts of the image that are in the vicinity of strong edges, and the ringing measure is considered to be the pixel intensity variance around the edges in the masked image. Still, the measure was not compared to the human score of quality. In [15], [16] an NR blur metric is proposed based on measuring average edge transition widths, and this blur measure was used to predict the quality of JPEG2000 compressed images. In [17], an NR algorithm is proposed based on natural scene statistics. In [18], a principal component analysis is performed on edge points, beforehand classified as distorted or not, in order to measure both blurring and ringing effects, as well as the combination of spatial ringing and blurring measures, which are also presented in [19].

All of the proposed NR perceptual image quality assessment algorithms are implemented according to the predefined specific artifacts of specific coders [7], [8], [9], [10], [11], [12], [13], [14], [15], [16], [17], [18], [19]. A lot of NR quality evaluations have already focus on measuring the blocking artifacts, specially for JPEG images with quite sufficient correlation with the subjective scores. However, very few NR evaluations have been performed for JPEG2000 images and these evaluations’ generalization ability and performances are not widely verified and well matched to the subjective scores. Whereas, nowadays, the JPEG2000 coder is getting more attention compared to the JPEG coder, due to its high coding performance, although JPEG was previously the standard coder for still image. The JPEG2000 coder served better in many image processing applications, such as digital cameras, 3G mobile phones, video streaming, printers, scanners, high quality frame-based video recording, nonlinear video editing, storage, etc. Specifically, motion JPEG2000 is the leading digital film standard currently supported by digital cinema initiatives for the storage, distribution and exhibition of motion pictures.

In this research, we propose a new method for NR quality evaluation of JPEG2000 images, irrespective of any predefined specific artifacts based on pixel distortions and edge information measure. This type of quality assessment is used to assess image quality and produces results comparable to those of subjective scores. The subjective experiment results on our database (JPEG2000 color images) were used to train and test the model and it achieved a sufficient quality prediction performance. The other database was also used to verify the model's performance. We report that the performance of the model is sufficient and reliable.

Section snippets

Our database [22]

We conducted subjective experiments on 24 bits/pixel RGB color images on our database. There were 98 images of size 768×512 in the database of JPEG2000. Out of all, 14 were reference images that are shown in Fig. 1. The rest of the images were JPEG2000 coded. Six compression ratios (CR: 12, 24, 32, 48, 72 and 96) were selected for the JPEG2000 encoder [20]. Single stimulus (SS) adjectival categorical judgement method was used in these subjective experiments. Prior to participating the session

Proposed model

Many researches have already established that the main function of the HVS is to extract structural or edge information from the viewing field, and the HVS is highly adapted for this purpose [4], [5]. Under the assumption that human visual perception is very sensitive to edge information, natural image signals are highly structured, specifically the samples of the signals have strong dependencies between each other, especially when they are close in space. Therefore, any kind of artifacts

Results

In order to verify our proposed model performance against other quality assessment algorithms, we want to consider a general purpose FR model and application specific NR models. These models include MSSIM (general purpose, FR) [6], Sheikh et al. (JPEG2000, NR) [17], and Marziliano et al. (JPEG2000, NR) [16]. Although such comparison is unfair to one method or another in different aspects, it provides a useful indication about the relative performance of the proposed model.

To the best of our

Conclusions

In this paper, we proposed a no-reference image quality assessment model irrespective of any predefined specific artifacts of JPEG2000 images. We claimed that any kinds of artifacts create pixel distortions and human visual perception is very sensitive to edge information. Therefore, we presented a new approach of image quality assessment model of JPEG2000 based on pixel distortions and edge information. The proposed model had been given good agreement with the MOS. Although the approach is

Acknowledgments

The authors would like to thank Dr. H.R. Sheikh for supplying the LIVE Quality Assessment Database (http://live.ece.utexas.edu/research/quality). The authors would also like to thank Prof. Murat Kunt and Prof. Touradj Ebrahimi for their higher motivation about the image quality evaluation methodology when he stayed at EPFL for visiting researcher supported by SNSF & JSPS.

References (29)

  • F. Pan et al.

    A locally adaptive algorithm for measuring blocking artifacts in images and videos

    Signal Process.: Image Commun.

    (2004)
  • P. Marziliano et al.

    Perceptual blur and ringing metrics: applications to JPEG2000

    Signal Process.: Image Commun.

    (2004)
  • Z. Wang et al.

    A universal image quality index

    IEEE Signal Process. Lett.

    (March 2002)
  • A.M. Eskicioglu et al.

    Image quality measures and their performance

    IEEE Trans. Commun.

    (December 1995)
  • B. Girod

    What's wrong with mean-square error

  • Z. Wang, Rate scalable foveated image and video communications, Ph.D. Thesis, Department of ECE, The University of...
  • Z. Wang et al.

    Why is image quality assessment so difficult?

  • Z. Wang et al.

    Image quality assessment: from error visibility to structural similarity

    IEEE Trans. Image Process.

    (April 2004)
  • H.R. Wu et al.

    A generalized block-edge impairment metric for video coding

    IEEE Signal Process. Lett.

    (November 1997)
  • Z. Wang et al.

    No-Reference perceptual quality assessment of JPEG compressed images

  • Z.M. Parvez Sazzad et al.

    Image quality assessment models for JPEG and JPEG2000 compressed color images

  • Y. Horita et al.

    Segmentation and local features based image quality evaluation

  • Z. Wang et al.

    Blind measurement of blocking artifacts in images

  • A.C. Bovik et al.

    DCT-domain blind measurement of blocking artifacts in DCT-coded images

  • Cited by (141)

    • A modulation method that can improve the performance of LED multi-spectral imaging

      2023, Spectrochimica Acta - Part A: Molecular and Biomolecular Spectroscopy
    • Medical image fusion based on improved multi-scale morphology gradient-weighted local energy and visual saliency map

      2022, Biomedical Signal Processing and Control
      Citation Excerpt :

      Subjectively, the fusion images obtained by the new algorithm proposed in this study are superior to those of other methods in terms of visual contrast and image details. We also used six evaluation indexes (entropy (EN) [2,4,11,13–15,50], average gradient (AG) [11,50], peak signal-to-noise ratio (PSNR) [9,12,35–37], Pearson linear correlation coefficient(CC) [51], corresponding mutual information (MI) [17,37,47], edge reservation(QAB/F) [5,50], Total information loss(LAB/F)and NAB/F [52]to evaluate the fusion results of the new method and the other methods described above. EN, AG, and PSNR are common indexes used to evaluate image quality, whereas CC, MI, QAB/F, LAB/F and NAB/F are the evaluation index for image fusion.

    • Objective quality assessment of synthesized images by local variation measurement

      2021, Signal Processing: Image Communication
      Citation Excerpt :

      There are also some SSIM variants, e.g., information content weighting SSIM (IW-SSIM) [13], multi-scale SSIM (MS-SSIM) [14] and structural similarity weighting-SSIM (SW-SSIM) [15]. These FR-IQA methods [16–18] are tailor for the natural images with the distortions of blur [19,20], blockiness [21], ringing [22] and etc. Nevertheless, the quality research on synthesized images is less activate.

    View all citing articles on Scopus
    View full text