• Corpus ID: 207870151

Domain-Aware No-Reference Image Quality Assessment

  title={Domain-Aware No-Reference Image Quality Assessment},
  author={Weihao Xia and Yujiu Yang and Jing-Hao Xue and Jing Xiao},
No-reference image quality assessment (NR-IQA) is a fundamental yet challenging task in low-level computer vision. It is to predict the perceptual quality of an image with unknown distortion. Its difficulty is particularly pronounced as the corresponding reference for assessment is typically absent. Various mechanisms to extract features ranging from natural scene statistics to deep features have been leveraged to boost the NR-IQA performance. However, these methods treat images of different… 

Figures and Tables from this paper



No-Reference Image Quality Assessment in the Spatial Domain

Despite its simplicity, it is able to show that BRISQUE is statistically better than the full-reference peak signal-to-noise ratio and the structural similarity index, and is highly competitive with respect to all present-day distortion-generic NR IQA algorithms.

Hallucinated-IQA: No-Reference Image Quality Assessment via Adversarial Learning

A hallucination-guided quality regression network is proposed to address the issue of no-reference image quality assessment, and significantly outperforms all the previous state-of-the-art methods by large margins.


This paper proposes a full reference framework named Mask Gated Convolutional Network (MGCN) for evaluating the image quality score and identifying distortions simultaneously, which achieves the state-of-the-art performance on LIVE2, TID2008 and TID2013 datasets.

Blind Image Quality Assessment: From Natural Scene Statistics to Perceptual Quality

DIIVINE is capable of assessing the quality of a distorted image across multiple distortion categories, as against most NR IQA algorithms that are distortion-specific in nature, and is statistically superior to the often used measure of peak signal-to-noise ratio (PSNR) and statistically equivalent to the popular structural similarity index (SSIM).

Blind Quality Assessment of Camera Images Based on Low-Level and High-Level Statistical Features

The proposed IQA approach extracts the low-level and high-level statistical features, which can capture the quality degradations effectively and design a neural network to integrate all the features and convert them to the final quality score.

Unsupervised feature learning framework for no-reference image quality assessment

This paper uses raw image patches extracted from a set of unlabeled images to learn a dictionary in an unsupervised manner and uses soft-assignment coding with max pooling to obtain effective image representations for quality estimation.

Deep Objective Quality Assessment Driven Single Image Super-Resolution

Experimental results prove that the proposed SISR method achieves better performance than other methods both qualitatively and quantitatively in most of the cases.

Waterloo Exploration Database: New Challenges for Image Quality Assessment Models

This work establishes a large-scale database named the Waterloo Exploration Database, which in its current state contains 4744 pristine natural images and 94 880 distorted images created from them, and presents three alternative test criteria to evaluate the performance of IQA models, namely, the pristine/distorted image discriminability test, the listwise ranking consistency test, and the pairwise preference consistency test.

FSIM: A Feature Similarity Index for Image Quality Assessment

A novel feature similarity (FSIM) index for full reference IQA is proposed based on the fact that human visual system (HVS) understands an image mainly according to its low-level features.

A Statistical Evaluation of Recent Full Reference Image Quality Assessment Algorithms

This paper presents results of an extensive subjective quality assessment study in which a total of 779 distorted images were evaluated by about two dozen human subjects and is the largest subjective image quality study in the literature in terms of number of images, distortion types, and number of human judgments per image.