The student-t mixture as a natural image patch prior with application to image compression
@article{Oord2014TheSM, title={The student-t mixture as a natural image patch prior with application to image compression}, author={A{\"a}ron van den Oord and Benjamin Schrauwen}, journal={J. Mach. Learn. Res.}, year={2014}, volume={15}, pages={2061-2086} }
Recent results have shown that Gaussian mixture models (GMMs) are remarkably good at density modeling of natural image patches, especially given their simplicity. In terms of log likelihood on real-valued data they are comparable with the best performing techniques published, easily outperforming more advanced ones, such as deep belief networks. They can be applied to various image processing tasks, such as image denoising, deblurring and inpainting, where they improve on other generic prior…
Figures and Tables from this paper
34 Citations
Factoring Variations in Natural Images with Deep Gaussian Mixture Models
- Computer ScienceNIPS
- 2014
This paper proposes a new scalable deep generative model for images, called the Deep Gaussian Mixture Model, that is a straightforward but powerful generalization of GMMs to multiple layers, and shows that deeper GMM architectures generalize better than more shallow ones.
Single Image Super-Resolution via Multiple Mixture Prior Models
- Computer ScienceIEEE Transactions on Image Processing
- 2018
This paper utilizes an effective way with mixture prior models to transform the large nonlinear feature space of LR images into a group of linear subspaces in the training phase, and finds that student-t prior shows stronger performance than the well-known Gaussian prior.
Image restoration with generalized Gaussian mixture model patch priors
- Computer ScienceArXiv
- 2018
It is shown that a generalized Gaussian mixture model (GGMM) captures the underlying distribution of patches better than a GMM and is consistently a better fit for modeling image patch distribution and performs better in image denoising task.
Image Transformer
- Computer ScienceICML
- 2018
This work generalizes a recently proposed model architecture based on self-attention, the Transformer, to a sequence modeling formulation of image generation with a tractable likelihood, and significantly increases the size of images the model can process in practice, despite maintaining significantly larger receptive fields per layer than typical convolutional neural networks.
Multi-channel expected patch log likelihood for color image denoising
- Computer ScienceNeurocomputing
- 2019
MMPM for Single Image Super-Resolution
- Engineering
- 2019
The MMPM (Multiple Mixture Prior Model) is learning based SR method which is used for converting Low Resolution (LR) image to High Resolution (HR) image. Other time or space intensive SR methods have…
Conditional Image Generation with PixelCNN Decoders
- Computer ScienceNIPS
- 2016
The gated convolutional layers in the proposed model improve the log-likelihood of PixelCNN to match the state-of-the-art performance of PixelRNN on ImageNet, with greatly reduced computational cost.
Generative Image Modeling Using Spatial LSTMs
- Computer ScienceNIPS
- 2015
This work introduces a recurrent image model based on multidimensional long short-term memory units which is particularly suited for image modeling due to their spatial structure and outperforms the state of the art in quantitative comparisons on several image datasets and produces promising results when used for texture synthesis and inpainting.
Lossy Image Compression with Compressive Autoencoders
- Computer ScienceICLR
- 2017
It is shown that minimal changes to the loss are sufficient to train deep autoencoders competitive with JPEG 2000 and outperforming recently proposed approaches based on RNNs, and furthermore computationally efficient thanks to a sub-pixel architecture, which makes it suitable for high-resolution images.
Gaussian mixture model learning based image denoising method with adaptive regularization parameters
- Computer ScienceMultimedia Tools and Applications
- 2016
This paper attempts to incorporate the gradient fidelity term with the Gaussian mixture model learning based image denoising method to preserve more fine structures of images and constructs an adaptive regularization parameter selection scheme by combing the image gradient with the local entropy of the image.
References
SHOWING 1-10 OF 40 REFERENCES
Non-local sparse models for image restoration
- Computer Science2009 IEEE 12th International Conference on Computer Vision
- 2009
Experimental results in image denoising and demosaicking tasks with synthetic and real noise show that the proposed method outperforms the state of the art, making it possible to effectively restore raw images from digital cameras at a reasonable speed and memory cost.
From learning models of natural image patches to whole image restoration
- Computer Science2011 International Conference on Computer Vision
- 2011
A generic framework which allows for whole image restoration using any patch based prior for which a MAP (or approximate MAP) estimate can be calculated is proposed and a generic, surprisingly simple Gaussian Mixture prior is presented, learned from a set of natural images.
"Natural Images, Gaussian Mixtures and Dead Leaves"
- Environmental ScienceNIPS
- 2012
An in depth analysis of what such a GMM model learns from natural images as a function of number of mixture components — including covariance structure, contrast variation and intricate structures such as textures, boundaries and more.
Fields of Experts: a framework for learning image priors
- Computer Science2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05)
- 2005
A framework for learning generic, expressive image priors that capture the statistics of natural scenes and can be used for a variety of machine vision tasks, developed using a Products-of-Experts framework.
Image Denoising Via Sparse and Redundant Representations Over Learned Dictionaries
- Computer ScienceIEEE Transactions on Image Processing
- 2006
This work addresses the image denoising problem, where zero-mean white and homogeneous Gaussian additive noise is to be removed from a given image, and uses the K-SVD algorithm to obtain a dictionary that describes the image content effectively.
What makes a good model of natural images?
- Computer Science2007 IEEE Conference on Computer Vision and Pattern Recognition
- 2007
A tractable lower and upper bounds on the partition function of models based on filter outputs and efficient learning algorithms that do not require any sampling are presented and applied to previous models shows that the nonintuitive features learned are not an artifact of the learning process but rather are capturing robust properties of natural images.
Lloyd clustering of Gauss mixture models for image compression and classification
- Computer ScienceSignal Process. Image Commun.
- 2005
Mixtures of Conditional Gaussian Scale Mixtures Applied to Multiscale Image Representations
- Computer SciencePloS one
- 2012
It is demonstrated that this probabilistic model based on mixtures of Gaussian scale mixtures and a simple multiscale representation yields the best performance reported to date when evaluated with respect to the cross-entropy rate, a measure tightly linked to the average log-likelihood.
Learning a piecewise linear transform coding scheme for images
- Computer ScienceInternational Conference on Graphic and Image Processing
- 2013
A piecewise linear transform coding scheme based on Vector Quantization is proposed for the use of GMM's for image compression, demonstrating that the proposed techniques outperform JPEG, with results comparable to JPEG2000 for a broad class of images.
Image Compression Using Sparse Representations and the Iteration-Tuned and Aligned Dictionary
- Computer ScienceIEEE Journal of Selected Topics in Signal Processing
- 2011
It is established experimentally that the ITAD structure results in lower-complexity representations that enjoy greater sparsity when compared to other recent dictionary structures, and a global rate-distortion criterion is proposed that distributes the code bits across the various image blocks.