This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Joint Image Compression and Denoising via Latent-Space Scalability

Saeed Ranjbar Alvar, Mateen Ulhaq, Hyomin Choi and Ivan V. Bajić The authors are with the School of Engineering Science, Simon Fraser University, Burnaby, BC, Canada.
Abstract

When it comes to image compression in digital cameras, denoising is traditionally performed prior to compression. However, there are applications where image noise may be necessary to demonstrate the trustworthiness of the image, such as court evidence and image forensics. This means that noise itself needs to be coded, in addition to the clean image itself. In this paper, we present a learning-based image compression framework where image denoising and compression are performed jointly. The latent space of the image codec is organized in a scalable manner such that the clean image can be decoded from a subset of the latent space (the base layer), while the noisy image is decoded from the full latent space at a higher rate. Using a subset of the latent space for the denoised image allows denoising to be carried out at a lower rate. Besides providing a scalable representation of the noisy input image, performing denoising jointly with compression makes intuitive sense because noise is hard to compress; hence, compressibility is one of the criteria that may help distinguish noise from the signal. The proposed codec is compared against established compression and denoising benchmarks, and the experiments reveal considerable bitrate savings compared to a cascade combination of a state-of-the-art codec and a state-of-the-art denoiser.

I Keywords:

image denoising, image compression, deep learning, multi-task compression, scalable coding

II Introduction

Images obtained from digital imaging sensors are degraded by the noise generated due to many factors such as lighting of the scene, sensors, shutter speed, etc. In practice, noticeable noise is often encountered in low-light conditions, as illustrated in the Smartphone Image Denoising Dataset (SIDD) [1]. In a typical image processing pipeline, noise in the captured image is attenuated or removed before compressing the image. The noise removed in the pre-processing stage cannot be restored, and the compressed image does not carry information about the original noise. While it is a desirable feature not to have noise in the stored image for the majority of applications, the captured noise may carry useful information for certain applications, such as court evidence, image forensics, and artistic intent. For such applications, the noise needs to be preserved in the compressed image. In fact, compressed-domain denoising together with techniques to preserve the noise is part of the recent JPEG AI call for proposals [2]. The major drawback of encoding the noise is that it significantly increases the bitrate required for storing and transferring the images. As an example, it is known that independent and identically distributed (iid) Gaussian source, which is a common noise model, has the worst rate-distortion performance among all the sources with the same variance [3]. Another issue is that when the clean (denoised) image is needed, the denoising should be applied to the reconstructed noisy images. The additional denoising step may increase the run time and the complexity of the pipeline.

To overcome the mentioned drawbacks of encoding the noisy image and performing denoising in cascade, we present a scalable multi-task image compression framework that performs compression and denoising jointly. We borrow the terminology from scalable video coding [4], where the input video is encoded into a scalable representation consisting of a base layer and one or more enhancement layers, which enables reconstructing various representations of the original video - different resolutions and/or frame rates and/or qualities. In the proposed Joint Image Compression and Denoising (JICD) framework, the encoder maps the noisy input to a latent representation that is partitioned into a base layer and an enhancement layer. The base layer contains the information about the clean image, while the enhancement layer contains information about noise. When the denoised image is needed, only the base layer needs to be encoded (and decoded), thereby avoiding noise coding. The enhancement layer is encoded only when the noisy input reconstruction is needed.

The scalable design of the system provides several advantages. Since only a subset of latent features is encoded for the denoised image, the bitrate is reduced compared to using the entire latent space. Another advantage is that the noise is not completely removed from the latent features, only separated from the features corresponding to the denoised image. Therefore, when the noisy input reconstruction is needed, the enhancement features are used in addition to the base features to decode the noisy input. The multi-task nature of the framework means that compression and denoising are trained jointly, and it also allows us to obtain both reconstructed noisy input and the corresponding denoised image in single forward pass, which reduces the complexity compared to the cascade implementation of compression and denoising. In fact, our results demonstrate that such a system provides improved performance – better denoising accuracy at the same bitrate – compared to a cascade combination of a state-of-the-art codec and a state-of-the-art denoiser.

The novel contributions of this paper are as follows:

  • We develop JICD, the first multi-task image coding framework that supports both image denoising and noisy image reconstruction.

  • JICD employs latent space scalability, such that the information about the clean image is mapped to a subset of the latent space (base layer) while noise information is mapped to the remainder (enhancement layer).

  • Unlike many methods in the literature, which are either developed for a particular type of noise and/or require some noise parameter(s) in order to operate properly, the proposed JICD is capable of handling unseen noise.

The remainder of the paper is organized as follows. Section III briefly describes prior work related to compression, denoising, and joint compression and denoising. Section IV discusses the preliminaries related to learning-based multi-task image compression. Section V presents the proposed method. Section VI describes the experiments and analyzes the experimental results. Finally, Section VII presents concluding remarks.

III Related Works

The proposed JICD framework is a multi-task image codec that performs image compression and denoising jointly. In this section, we briefly discuss the most relevant works related to image denoising (Section III-A), learning-based image compression (Section III-B), and multi-task image compression including joint compression and denoising (Section III-C).

III-A Image Denoising

State-of-the-art classical image denoising methods are based on Non-local Self Similarity (NSS). In these methods, repetitive local patterns in a noisy image are used to capture signal and noise characteristics, and perform denoising. In BM3D [5], similar patches are first found by block matching. Then, they are stacked to form a 3D block. Finally, transform-domain collaborative filtering is applied to obtain the clean patch. [6] used adaptive filtering to improve BM3D. WNNM [7] performs denoisng by applying low rank matrix approximation to the stacked noisy patches. In [8], a patch group based NSS prior learning scheme to learn explicit NSS models from natural images is proposed. The denoising method in [9] used NSS priors in both the degraded images and the external clean images to perform denoising. CBM3D [10] and MCWNNM [11] are the extensions of BM3D and WNNM, respectively, created to handle color images.

More recently, (deep) learning-based denoising methods have gained popularity and surpassed the performance of classical methods.  [12] used a multi-layer perceptron (MLP) to achieve denoising results comparable to the state-of-the-art classic method. Among the learning-based denoisers, DnCNN [13] was the first Convolutional Neural Network (CNN) to perform blind Gaussian denoising. FFDNet [14] improved upon DnCNN by proposing a fast and flexible denoising CNN that could handle different noise levels with a single model. In [15], noise estimation subnetwork is added prior to the CNN-based denoiser to get an accurate estimate of the noise level in the real-world noisy photographs. A Generative Adversarial Networks (GAN)-based denoising method is proposed in [16]. The mentioned works are supervised methods where clean reference image is needed for training. In [17, 18], self-supervised denoising methods are proposed.

III-B Learning-based Image Compression

In recent years, there has been an increasing interest in the development of learning-based image codecs. Some of the early works [19, 20, 21] were based on Recurrent Neural Networks (RNNs), whose purpose was to model spatial dependence of pixels in an image. More recently, the focus has shifted to Convolutional Neural Network (CNN)-based autoencoders. [22] introduced Generalized Divisive Normalization (GDN) as a key component of the nonlinear transform in the encoder. The image codec based on GDN was improved by introducing a hyperprior to capture spatial dependencies and take advantage of statistical redundancy in the entropy model [23]. To further improve the coding gains, discretized Gaussian mixture likelihoods are used in [24] to parameterize the distributions of latent codes. Most recently, this approach has been extended using advanced latent-space context modelling [25] to achieve even better performance.

Most state-of-the-art learning-based image codding approaches [23, 24, 25] train different models for different bitrates, by changing the Lagrange multiplier that trades-off rate and distortion. Such approach is meant to explore the potential of learning-based compression, rather than be used in practice as is. There has also been a considerable amount of work on variable-rate learning-based compression [19, 26, 27, 28, 29], where a single model is able to produce multiple rate-distortion points. However, in terms of rate-distortion performance, “fixed-rate” approaches such as [24, 25] currently seem to have an advantage over variable-rate ones.

III-C Multi-Task Image Compression

The mentioned learning-based codec are single-task models, where the task is the reconstruction of the input image, just like with conventional codecs. However, the real power of learning-based codecs is their ability to be trained for multiple tasks, for example, image processing or computer vision tasks, besides the usual input reconstruction. In fact, the goal of JPEG AI standardization is to develop such a coding framework that could support multiple tasks from a common compressed representation [30].

[31] proposed a scalable multi-task model with multiple segments in the latent space to handle computer vision tasks in addition to input reconstruction. The concept was based on latent-space scalability [32], where the latent space is partitioned in a scalable manner, from tasks that require less information to tasks that require more information. Our JICD framework is also based on latent-space scalability [32]. However, unlike these earlier works, the latent space is organized such that it supports image denoising from the base layer and noisy input reconstruction from the full latent space. In other words, the tasks are different compared to these earlier works.

Recently, [33] and [34] developed joint image compression and denoising pipelines built upon learning-based image codecs, where the pipeline is trained to take the input noisy image, compress it, and decode a denoised image. However, with these approaches, it is not possible to reconstruct the original noisy image, hence they are not multi-task models. Our proposed JICD performs the denoising task in its base layer, but it keeps the noise information in the enhancement layer, thereby also enabling noisy input reconstruction if needed.

IV Prelimineries

In thinking about how to construct a learning-based system that can produce both the denoised image and reconstruct the noisy image, it is useful to consider the processing pipeline in which noisy image is first compressed, then decoded, and then denoising is applied to obtain the denoised image. Let 𝐗n\mathbf{X}_{n} be the noisy input image. If such an image is input to a learning-based codec [22, 23, 35, 24], encoding would proceed in three steps:

𝒴}𝒳\ϕ\mathbfcal{Y}=g_{a}(\mathbf{X}_{n};\phi) (1)
𝒴^=Q(𝒴\widehat{\mathbfcal{Y}}=Q(\mathbfcal{Y})\\ (2)
𝐁=AE(𝒴^)\mathbf{B}=A_{E}(\widehat{\mathbfcal{Y}}) (3)

where gag_{a} is the analysis transform, ϕ\phi represents the parameters of gag_{a}, QQ is the quantization function, and 𝐁\mathbf{B} is the bitstream obtained by applying the arithmetic encoder AEA_{E} to 𝒴^\widehat{\mathbfcal{Y}}.

The noisy input image is reconstructed at the decoder by applying the entropy decoding and synthesis transform to the encoded bitstream as:

𝒴^=AD(𝐁)\widehat{\mathbfcal{Y}}=A_{D}(\mathbf{B}) (4)
𝐗^n=gs(𝒴^;θ)\widehat{\mathbf{X}}_{n}=g_{s}(\widehat{\mathbfcal{Y}};\theta) (5)

where ADA_{D} is the entropy decoder, gsg_{s} and θ\theta are the synthesis transform and its parameters, respectively. Then the denoised image can be obtained by applying a denoiser to the reconstructed noisy input as:

𝐗^=F(𝐗^n,ψ)\widehat{\mathbf{X}}={F}(\widehat{\mathbf{X}}_{n},\psi) (6)

where FF and ψ\psi are the denoiser and its parameters, respectively, and 𝐗^\widehat{\mathbf{X}} is the denoised image.

This processing pipeline forms a Markov chain 𝐗n𝒴^𝐗^n𝐗^\mathbf{X}_{n}\to\widehat{\mathbfcal{Y}}\to\widehat{\mathbf{X}}_{n}\to\widehat{\mathbf{X}}. Applying the data processing inequality (DPI) [3] to this Markov chain, we get

I(𝒴^;𝐗^n)I(𝒴^;𝐗^),I(\widehat{\mathbfcal{Y}};\widehat{\mathbf{X}}_{n})\geq I(\widehat{\mathbfcal{Y}};\widehat{\mathbf{X}}), (7)

where I(;)I(\cdot\,;\,\cdot) is the mutual information [3] between two random quantities. Based on (7) we can conclude that latent representation 𝒴^\widehat{\mathbfcal{Y}} carries less information about the denoised image 𝐗^\widehat{\mathbf{X}} than it does about the noisy reconstructed image 𝐗^n\widehat{\mathbf{X}}_{n}. Moreover, because 𝐗^\widehat{\mathbf{X}} is obtained from 𝐗^n\widehat{\mathbf{X}}_{n}, the information that 𝒴^\widehat{\mathbfcal{Y}} carries about 𝐗^\widehat{\mathbf{X}} is a subset of the information that it carries about 𝐗^n\widehat{\mathbf{X}}_{n}. This motivates us to structure the latent representation 𝒴^\widehat{\mathbfcal{Y}} in such a way that only a part of it (the base layer) is used to reconstruct the denoised image 𝐗^\widehat{\mathbf{X}}, while the whole of 𝒴^\widehat{\mathbfcal{Y}} (base+enhancement) is used to reconstruct the noisy image 𝐗^n\widehat{\mathbf{X}}_{n}.

V Proposed method

Refer to caption
Figure 1: The proposed joint compression and denoising (JICD) framework. AE/AD represent arithmetic encoder and decoder, respectively. CmC_{m} and EP stand for the context model and entropy parameters, respectively. Q represents the quantizer, which is simple rounding to the nearest integer. The architecture of the individual building blocks is the same as in [24, 32, 31], but they have been retrained to support different tasks, as explained in the text.
Refer to caption

(a) Analysis Transform, /2: stride=2

Refer to caption

(b) Hyper Synthesis

Refer to caption

(c) Synthesis Transform, *2: up-sampling by the factor of 2

Refer to caption

(d) Context model Cm,jC_{m,j} and EPjEP_{j}, |\mathbfcal{H}_{j} is the output of Hyper Synthesis

Figure 2: The architecture of the components inside the proposed JICD framework.

The proposed joint image compression and denoising (JICD) framework consists of an encoder and two task-specific decoders, as illustrated in Fig. 1. The architecture of the blocks that make up the encoder and two decoders in Fig. 1 is shown in Fig. 2. Note that the architecture of the individual building blocks (Analysis Transform, Synthesis Transform, etc.) is the same as in [24, 32, 31], but these blocks have been retrained to support a scalable latent representation for joint compression and denoising. Specifically, compared to [24], our encoder is trained to produce a scalable latent space that enables both denoising and noisy input reconstruction. Compared to [32, 31], our system is trained to support different tasks, and correspondingly the structure of the latent space and the training procedure is different. Details of individual components are described below.

V-A Encoder

The encoder employs an analysis transform to obtain a high fidelity latent-space representation for the input image. In addition, the encoder has blocks to efficiently encode the obtained latent-space tensor. The encoder’s analysis transform is borrowed from [24] due to its high compression efficiency. In addition to the analysis transform, we also adopted the entropy parameter (EP) module, the context model (CTX) for arithmetic encoder/decoder (AE/AD), synthesis transform and hyper analysis/synthesis without attention layers from [24].

The analysis transform converts the input image 𝐗\mathbf{X} into 𝒴𝒩××𝒞\mathbfcal{Y}\in\mathbb{R}^{N\times M\times C}, with C=192C=192 as in [35, 24]. Unlike [35, 24], the latent representation 𝒴\mathbfcal{Y} is split into two separate sub-latents 𝒴𝒴𝒴\mathbfcal{Y}=\mathbfcal{Y}_{1}\cup\mathbfcal{Y}_{2}, 𝒴𝒴\mathbfcal{Y}_{1}\cap\mathbfcal{Y}_{2}=\emptyset, where 𝒴\mathbfcal{Y}_{1} is the base layer containing ii channels, 𝒴{𝒴𝒴𝒴}\mathbfcal{Y}_{1}=\{\mathbf{Y}_{1},\mathbf{Y}_{2},...\mathbf{Y}_{i}\}, and 𝒴\mathbfcal{Y}_{2} is the enhancement layer containing CiC-i channels, 𝒴{𝒴𝒴𝒴𝒞}\mathbfcal{Y}_{2}=\{\mathbf{Y}_{i+1},\mathbf{Y}_{i+2},...,\mathbf{Y}_{C}\}. This allows the latent representation to be used efficiently for multiple purposes, namely denoising (from 𝒴\mathbfcal{Y}_{1}) and noisy input reconstruction (from 𝒴𝒴\mathbfcal{Y}_{1}\cup\mathbfcal{Y}_{2}). Since denoising requires only 𝒴\mathbfcal{Y}_{1}, it can be accomplished at a lower bitrate compared to decoding the full latent space. The sub-latents are then quantized to produce 𝒴^1\widehat{\mathbfcal{Y}}_{1} and 𝒴^2\widehat{\mathbfcal{Y}}_{2}, respectively, and then coded using their respective context models to produce two independently-decodable bitstreams, as discussed in [31, 32]. The side bitstream shown in Fig. 1 is considered to be a part of the base layer and its rate is included in bitrate calculations for the base layer bitstream in the experiments.

V-B Decoder

Two task-specific decoders are constructed: one for denoised image decoding and one for noisy input image reconstruction. The hyperpriors used in both decoders are reconstructed from the side bitstream which, as mentioned above, is considered to be a part of the base layer. Quantized base representation 𝒴^1\widehat{\mathbfcal{Y}}_{1} is reconstructed in the base decoder by decoding the base bitstream, and used to produce the denoised image 𝐗^\widehat{\mathbf{X}}. Unlike [32, 31], where the base layer was dedicated to object detection/segmentation, our decoder does not require latent space transformation from 𝒴^1\widehat{\mathbfcal{Y}}_{1} into another latent space; the synthesis transform (Fig. 1) produces the denoised image 𝐗^\widehat{\mathbf{X}} directly from 𝒴^1\widehat{\mathbfcal{Y}}_{1}. Quantized enhancement representation 𝒴^2\widehat{\mathbfcal{Y}}_{2} is decoded only when noisy input reconstruction is needed. The reconstructed noisy input image 𝐗^n\widehat{\mathbf{X}}_{n} is produced by the second decoder using 𝒴^=𝒴^1𝒴^2\widehat{\mathbfcal{Y}}=\widehat{\mathbfcal{Y}}_{1}\cup\widehat{\mathbfcal{Y}}_{2}.

Although not pursued in this work, it is worth mentioning that the proposed JICD framework can be extended to perform various computer vision tasks as well, such as image classification or object detection. These tasks typically require clean images, so one can think of the processing pipeline described by the following Markov chain: 𝐗n𝒴^1𝐗^T\mathbf{X}_{n}\to\widehat{\mathbfcal{Y}}_{1}\to\widehat{\mathbf{X}}\to T, where TT is the output of a computer vision task, for example a class label or object bounding boxes. Applying the DPI to this Markov chain we have

I(𝒴^1;𝐗^)I(𝒴^1;T),I(\widehat{\mathbfcal{Y}}_{1};\widehat{\mathbf{X}})\geq I(\widehat{\mathbfcal{Y}}_{1};T), (8)

which implies that a subset of information from 𝒴^1\widehat{\mathbfcal{Y}}_{1} is sufficient to produce TT. Hence, if such tasks are required, the encoder’s latent space can be further partitioned by splitting 𝒴^1\widehat{\mathbfcal{Y}}_{1}, in a manner similar to [32, 31], to support such tasks at an even lower bitrate than our base layer.

V-C Training

The model is trained end-to-end with a rate-distortion Lagrangian loss function in the form of:

=R+λD,\mathcal{L}=R+\lambda\cdot D, (9)

where RR is an estimate of rate rate, DD is the total distortion of both tasks, and λ\lambda is the Lagrange multiplier. The estimated rate is affected by latent and hyper-priors as in [35],

R=𝔼xpx[log2py^(y^)]latent+𝔼xpx[log2pz^(z^)]hyper-priors,R=\underbrace{\mathbb{E}_{x\sim p_{x}}\left[-\textup{log}_{2}p_{\hat{y}}(\hat{y})\right]}_{\textup{latent}}+\underbrace{\mathbb{E}_{x\sim p_{x}}\left[-\textup{log}_{2}p_{\hat{z}}(\hat{z})\right]}_{\textup{hyper-priors}}, (10)

where xx denotes input data, y^\hat{y} is the quantized latent data and z^\hat{z} is the quantized hyper-prior. Total distortion DD is computed as the weighted average of image denoising distortion and noisy input reconstruction distortion:

D=(1w)MSE(𝐗,𝐗^)+wMSE(𝐗n,𝐗^n),D=(1-w)\cdot\textup{MSE}\left(\mathbf{X},\widehat{\mathbf{X}}\right)+w\cdot\textup{MSE}\left(\mathbf{X}_{n},\widehat{\mathbf{X}}_{n}\right), (11)

where ww is the trade-off factor to adjust the importance of the tasks. The automatic differentiation [36] ensures that the gradients from DD flow through the corresponding parameters without further modification to the back-propagation algorithm.

VI Experimental Results

VI-A Network Training

The proposed multi-task model is trained from scratch using the randomly cropped 256×256256\times 256 patches from the CLIC dataset [37]. The noisy images are obtained using additive white Gaussian noise (AWGN) with three noise levels σ={15,25,50}\sigma=\{15,25,50\}, clipping the resulting values to [0, 255] and quantizing the clipped values to mimic how noisy images are stored in practice. The batch size is set to 16. Training is ran for 300 epochs using the Adam optimizer with initial learning rate of 1×1041\times 10^{-4}. The learning rate is reduced by factor of 0.50.5 when the training loss plateaus. We trained 6 different models by changing the value of λ\lambda in (9). The list of different values for λ\lambda is shown in Table I. For all the models we used w=0.05w=0.05 in (11). We trained the model for the first rate point (lowest λ\lambda) from scratch. However, for the remaining rate points we fine-tune the model starting from the previous rate point’s weights

Model Index 1 2 3 4 5 6
λ\lambda 0.0035 0.0067 0.013 0.025 0.0483 0.09
TABLE I: λ\lambda values used for training various models. Higher λ\lambda leads to higher qualities and higher bitrates.

We trained models under two different settings. In the first setting, a given model is trained for each noise level. For this case, the number of enhancement channels CiC-i is chosen according to the strength of the noise. For stronger noise, we allocate more channels to the enhancement layer, so that it can capture enough information to reconstruct the noise. The number of enhancement channels is reduced as the noise gets weaker. Specifically, the number of enhancement channels is empirically set to 32, 12, and 2 for σ=50\sigma=50, σ=25\sigma=25, and σ=15\sigma=15, respectively. The second training setting is to train a single model with different noise levels σ{50,25,15}\sigma\in\{50,25,15\} simultaneously, and use the final trained model to perform denoising for all noise levels. This is beneficial when the noise level information is not given. In this model, we used 180 base channels and 12 enhancement channels. σ\sigma at each training iteration is uniformly chosen from {50,25,15}\{50,25,15\}.

VI-B Data

To evaluate the performance of the proposed JICD framework, four color image datasets are used: 1) CBSD68 [38], 2) Kodak24 [39], 3) McMaster [40] and 4) JPEG AI testset [30], which is used in the JPEG AI exploration experiments. The mentioned datasets contain 68, 24, 18, and 16 images, respectively. The resolution of the images in the Kodak24 and McMaster dataset is fixed to 500×500500\times 500. CBS68 dataset contains the lowest-resolution images among the four datasets, with the height and width of images ranging between 321 and 481. The images in the JPEG AI testset are high-resolution images with the height varying between 872 and 2456 pixels and width varying between 1336 and 3680 pixels. The results are reported for two sets of noisy images. In the first set, we added synthesized AWGN to the testing images with three noise levels: σ={15,25,50}\sigma=\{15,25,50\} and tested the results with the quantized noisy images. In the second set, we used the synthesized noise obtained from the noise simulator in [41], which was also used to generate the final test images for the denoising tasks in the ongoing JPEG AI standardization. This type of noise was not used during the training of the proposed JICD framework. Hence, the goal of testing with this second set of images is to evaluate how well the proposed JICD generalizes to the noise that is not seen during the training.

VI-C Baselines

The denoising performance of the proposed JICD framework is compared against well-established baselines: CBM3D [10] and FFDNet [14]. CBM3D is a NSS-based denoising method, and FFDNet belongs to the learning-based denoising category. FFDNet was trained using AWGN with different noise levels during the training. At inference time, FFDNet needs the variance of the noise as input. FFDNet-clip [14] is a version of FFDNet that is trained with quantized noisy images. Since our focus is on practical settings with quantized noisy images, we used FFDNet-clip as a baseline in the experiments. We also tested the DRUNet denoiser [42], which is one of the latest state-of-the-art denoisers. DRUNet assumes that the noise is not quantized, and when tested with quantized noise, it performs worse than FFDNet-clip. As a result, we did not include it in the experiments.

Two baselines are established by applying CBM3D and FFDNet-clip directly on noisy images, without compression. However, to assess the interaction of compression and denoising, we establish one more baseline. In this third baseline, the noisy image is first compressed using the end-to-end image compression model from [24] (the “Cheng model”) with an implementation from CompressAI [43], and then decoded. Then FFDNet-clip is used to denoise the decoded noisy image. We call this cascade denoising approach as Cheng+FFDNet-clip. It is worth mentioning that Cheng+FFDNet-clip, similar to the proposed JICD framework, is able to obtain both the reconstructed noisy images and denoised images, hence it could be considered a multi-task approach.

VI-D Experiments on AWGN Removal

We evaluate the baselines and the proposed JICD method using the quantized noisy images obtained using AWGN with three noise levels, σ{15,25,50}\sigma\in\{15,25,50\}. The test results with the strongest noise (σ=50\sigma=50) across the four datasets (CBSD68, Kodak24, McMaster, and JPEG AI) are shown Fig. 3 in terms of rate vs. Peak Signal-to-Noise Ratio (PSNR) and in Fig. 4 in terms of rate vs. Structural Similarity Index Measure (SSIM). The horizontal lines in the figure correspond to applying CBM3D and FFDNet-clip to the raw (uncompressed) noisy images. The blue curve shows the results for Cheng+FFDNet-clip. The six points on this curve correspond to the six Cheng models from CompressAI [43]. For JICD, two curves are shown. The orange curve shows the results obtained from the models trained for σ=50\sigma=50 with 160 base feature channels and 32 enhancement channels. The yellow curve corresponds to the results obtained using the model that was trained with variable σ\sigma values and has 180 base and 12 enhancement channels. The six points on the orange and yellow curves correspond to the six JICD models we trained with λ\lambda values shown in Table I.

Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 3: Denoising rate-PSNR curves for σ=50\sigma=50: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI
Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 4: Denoising rate-SSIM curves for σ=50\sigma=50: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI

As seen in Fig. 3, for σ=50\sigma=50, the quality of the images denoised by CBM3D is considerably lower compared to those obtained using FFDNet-clip. It was shown in [14] that CBM3D and FFDNet-clip achieve comparable performance for non-quantized noisy images. Our results show that CBM3D’s performance is degraded when the noise deviates (due to clipping and quantization) from the assumed model, at least at high noise levels.

The comparison of the results obtained by JICD and Cheng+FFDNet-clip reveal that JICD is able to reduce the bitrate substantially while achieving the same denoising performance as Cheng+FFDNet-clip. This is due to the fact that the Cheng model allocates the entire latent representation to noisy input reconstruction, whereas the proposed method uses a subset of the latent features to perform denoising. The results of JICD trained with variable σ\sigma are also shown in the curves. Since the number of base channels is larger in this model compared to the model trained for σ=50\sigma=50, its denoising performance is improved.

To summarize the differences between the performance-rate curves, we compute Bjøntegaard Delta-rate (BD-rate) [44]. The BD-rate of the proposed JICD compared to Cheng+FFDNet-clip on the four datasets is given in the first two rows of Table II for PSNR, and Table III for SSIM. It can be seen that the proposed method achieves up to 80.2% BD-rate savings compared to Cheng+FFDNet-clip. Both JICD and Cheng+FFDNet-clip denoising methods outperform CBMD3D for all the tested rate points at σ=50\sigma=50. Using the proposed JICD method, we are able to denoise images at a quality close to what FFDNet-clip achieves on raw images, and at the same time compress the input.

Noise type Model CBSD68 Kodak24 McMaster JPEG AI
AWGN σ=50\sigma=50 σ=50\sigma=50 69.28-69.28% 72.91-72.91% 72.69-72.69% 74.45-74.45%
variable σ\sigma 70.66-70.66% 77.27-77.27% 76.55-76.55% 80.20-80.20%
AWGN σ=25\sigma=25 σ=25\sigma=25 30.58-30.58% 41.00-41.00% 33.18-33.18% 45.13-45.13%
variable σ\sigma 30.28-30.28% 42.61-42.61% 33.52-33.52% 45.77-45.77%
AWGN σ=15\sigma=15 σ=15\sigma=15 1.071.07% 11.99-11.99% 4.95-4.95% 15.82-15.82%
variable σ\sigma 8.008.00% 2.99-2.99% 9.229.22% 5.78-5.78%
Practical noise
simulator
variable σ\sigma 23.25-23.25% 33.83-33.83% 21.51-21.51% 23.42-23.42%
TABLE II: The PSNR-based BD-rate of the proposed JICD compared to Cheng+FFDNet-clip on the image denoising task.
Noise type Model CBSD68 Kodak24 McMaster JPEG AI
AWGN σ=50\sigma=50 σ=50\sigma=50 63.34-63.34% 72.08-72.08% 72.29-72.29% 72.89-72.89%
variable σ\sigma 64.09-64.09% 73.64-73.64% 75.57-75.57% 76.43-76.43%
AWGN σ=25\sigma=25 σ=25\sigma=25 24.14-24.14% 38.99-38.99% 53.40-53.40% 43.11-43.11%
variable σ\sigma 24.45-24.45% 39.86-39.86% 52.21-52.21% 43.97-43.97%
AWGN σ=15\sigma=15 σ=15\sigma=15 4.524.52% 11.85-11.85% 21.57-21.57% 15.32-15.32%
variable σ\sigma 9.149.14% 5.96-5.96% 8.60-8.60% 8.78-8.78%
Practical noise
simulator
variable σ\sigma 15.83-15.83% 27.66-27.66% 28.18-28.18% 37.72-37.72%
TABLE III: The SSIM-based BD-rate of the proposed JICD compared to Cheng+FFDNet-clip on the image denoising task.

We repeat the denoising experiment for σ=25\sigma=25, and the results are shown in Fig. 5 for PSNR and Fig. 6 for SSIM. As seen in the figures, the gap between the CBM3D and FFDNet-clip performance is now reduced, and the compression-based methods now outperform CBM3D only at the higher rates. The gap between the curves corresponding to JICD and Cheng+FFDNet-clip is also reduced. However, JICD still achieves a considerable BD-rate saving compared to Cheng+FFDNet-clip, as shown in the third row of Table II and Table III. JICD trained with variable σ\sigma has slightly better PSNR performance compared to the noise-specific model on three datasets, and a slightly worse performance (by 0.3%) on the low-resolution CBSD68 dataset.

Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 5: Denoising rate-PSNR curves for σ=25\sigma=25: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI
Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 6: Denoising rate-SSIM curves for σ=25\sigma=25: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI

At the lowest noise level (σ=15\sigma=15), the gap between CBM3D and FFDNet-clip shrinks further. It can be seen in the denoised rate-PSNR curves in Fig. 7 and rate-SSIM curves in Fig. 8 that when the noise is weak, applying denoising to the raw images achieves high PSNR, and the compression-based methods cannot outperform either CBM3D, or FFDNet-clip at the tested rates. The gap between JICD and Cheng+FFDNet-clip curves is also reduced compared to the higher noise levels. This can also be be seen from the BD-rates in the fourth row of Table II and Table III. JICD trained for σ=15\sigma=15 outperforms Cheng+FFDNet-clip on three datasets, but it suffers a 1% (4.5% for SSIM) loss on the low-resolution CBSD68.

As seen above, the performance of the proposed JICD framework is lower on the low-resolution CBSD68 dataset than on other datasets. The reason is the following. The processing pipeline USED in JICD expects the input dimensions to be multiples of 64. For images whose dimensions do not satisfy this requirement, the input is padded up to the nearest multiple of 64. At low resolutions, the padded area may be somewhat large in relation to the original image, which causes noticeable performance degradation. At high resolutions, the padded area is insignificant compared to the original image, and the impact on JICD’s performance is correspondingly smaller. It is worth mentioning that for σ=15\sigma=15, the JICD trained with variable σ\sigma has a weaker denoising performance compared to the model trained specifically for σ=15\sigma=15. This is because the number of base channels in the variable-σ\sigma model (180) is smaller than the number of base channels in the noise-specific model (190). At low noise levels, fewer channels are needed to hold noise information, which means the number of base channels could be higher. Hence, the structure chosen for the noise-specific model is better suited for this case. However, we show in the next subsection that the variable-σ\sigma model is more useful when the noise parameters are not known.

Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 7: Denoising rate-PSNR curves for σ=15\sigma=15: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI
Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 8: Denoising rate-SSIM curves for σ=15\sigma=15: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI

VI-E Experiments on Unseen Noise Removal

VI-E1 Image Compression and Denoising

The proposed JICD denoiser and the baselines are also tested with the noise that was not used in the training. The purpose of this experiment is to evaluate how well the denoisers are able to handle unseen noise. To generate unseen noise, we used the noise simulator from [41]. This noise simulator, which we subsequently refer to as ‘practical noise simulator,’ was created by fitting the Poissonian-Gaussian noise model [45] to the noise from the Smartphone Image Denoising Dataset (SIDD) [1]. It is worth mentioning that this noise simulator is used in the evaluation of the image denoising task in JPEG AI standardization.

For this experiment we use the JICD model trained with variable σ\sigma. One advantage of this model is that, unlike some of the baselines, it does not require any additional input or noise information, besides the noisy image. On the other hand, FFDNet needs σ\sigma to perform denoising. In the experiment, the σ\sigma is estimated for each image by computing the standard deviation of the difference between the noisy test image and the corresponding clean image.

Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 9: Denoising rate-PSNR curves for the unseen noise: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI
Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 10: Denoising rate-SSIM curves for the unseen noise: (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG AI

The denoising rate-PSNR and rate-SSIM curves are illustrated in Fig. 9 and Fig. 10, respectively. Since the variance of the noise obtained from the practical noise simulator is not large, the PSNR range of the denoised images is close to that observed in the AWGN experiments with σ=15\sigma=15 and σ=25\sigma=25. The results indicate that JICD achieves better denoising performance compared to Cheng+FFDNet-clip across all four datasets. Moreover, at higher bitrates (1 bpp and above), JICD outperforms CBM3D applied to uncompressed noisy images. BD-rate results are summarized in the last row of Table II and Table III. It is seen in the table that JICD achieves 1515-30%30\% gain over Cheng+FFDNer-clip across the four datasets.

A visual example comparing the denoised images obtained from JICD and Cheng+FFDNet-clip encoded at similar bitrates is shown in Fig. 11. As seen in the figure, JICD preserves more details compared to Cheng+FFDNet-clip. In addition, the colors inside the white circle are reproduced closer to the ground truth with JICD compared to the image produced by Cheng+FFDNet-clip.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 11: An example of denoised images. Top to Bottom: noisy image, clean image, Denoised: Cheng+FFDNet-clip (bpp=0.57), Denoised: proposed (bpp=0.55). Images in the right column show the red square in the left images

VI-E2 Noisy Image Reconstruction

Besides denoising, the proposed JICD framework is also able to reconstruct the noisy input image when enhancement features are decoded together with base features. While the main focus of this work was on denoising (and the majority of experiments devoted to that goal), for completeness we also evaluate the noisy image reconstruction performance using unseen noise. We compare the noisy input reconstruction performance of JICD against [24], i.e., the compression model used earlier in the Cheng+FFDNet-clip baseline. The PSNR between the noisy input and the reconstructed noisy images is shown against bitrate in Fig. 12, while Fig. 13 shows SSIM vs. bitrate. As illustrated in Fig. 12, our JICD achieves better noisy input reconstruction compared to [24] in most cases. BD-rate results corresponding to Fig. 12 and Fig. 13 are given in Table IV and Table V, respectively. As the numbers in the Table IV indicate, the proposed JICD achieves noticeable BD-rate savings on three of the four test datasets; the only exception is, again, the low-resolution CBSD68 dataset, where the loss is mainly concentrated at higher bitrates. It is worth noting that, since our proposed method is trained using the MSE loss, it performs better in terms of PSNR than SSIM. Overall, the proposed JICD framework achieves gains on both denoising and compression tasks compared to Cheng+FFDNet-clip and [24] models.

Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 12: The rate-PSNR curves for noisy input reconstruction. (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG-AI
Refer to caption

(a)

Refer to caption

(b)

Refer to caption

(c)

Refer to caption

(d)

Figure 13: The rate-SSIM curves for noisy input reconstruction. (a) CBSD68, (b) Kodak24, (c) McMaster, (d) JPEG-AI
Noise type Model CBSD68 Kodak24 McMaster JPEG AI
Practical noise
simulator
variable σ\sigma 5.505.50% 11.74-11.74% 3.97-3.97% 13.49-13.49%
TABLE IV: The PSNR-based BD-rate of the proposed JICD compared to the Cheng model on noisy input reconstruction.
Noise type Model CBSD68 Kodak24 McMaster JPEG AI
Practical noise
simulator
variable σ\sigma 22.5822.58% 1.901.90% 4.054.05% 0.580.58%
TABLE V: The SSIM-based BD-rate of the proposed JICD compared to the Cheng model on noisy input reconstruction.

VII Conclusion

In this work, we presented a joint image compression and denoising framework. The proposed framework is a scalable multi-task image compression model based on the latent-space scalability. The base features are used to perform the denoising and the enhancement features are used when the noisy input reconstruction is needed. Extensive experiments show that the proposed framework achieves significant BD-rate savings up to 80.20% across different dataset compared to the cascade compression and denoising method. The experimental results also indicate that the proposed method achieves improved results for the unseen noise for both denoising and noisy input reconstruction tasks.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Author Contributions

SRA and IVB contributed to conception and design of the study. HC developed the initial code. MU and SRA contributed to further code development and optimization. SRA wrote the first draft of the manuscript and worked with IVB on the revisions.

Funding

Funding for this work was provided by the Natural Sciences and Engineering Research Council (NSERC) of Canada under the grants RGPIN-2021-02485 and RGPAS-2021-00038, and by Huawei Technologies.

References

  • [1] A. Abdelhamed, S. Lin, and M. S. Brown, “A high-quality denoising dataset for smartphone cameras,” in Proc. CVPR’18, Jun. 2018, pp. 1692–1700.
  • [2] ISO/IEC and ITU-T, “Final call for proposals for JPEG AI,” ISO/IEC JTC 1/SC29/WG1 N100095, 2022.
  • [3] T. M. Cover and J. A. Thomas, Elements of Information Theory, 2nd ed.   Wiley, 2006.
  • [4] H. Schwarz, D. Marpe, and T. Wiegand, “Overview of the scalable video coding extension of the h.264/avc standard,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 17, no. 9, pp. 1103–1120, 2007.
  • [5] K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Image denoising by sparse 3-d transform-domain collaborative filtering,” IEEE Trans. Image Process., vol. 16, no. 8, pp. 2080–2095, 2007.
  • [6] A. A. Yahya, J. Tan, B. Su, M. Hu, Y. Wang, K. Liu, and A. N. Hadi, “Bm3d image denoising algorithm based on an adaptive filtering,” Multimedia Tools and Applications, vol. 79, no. 27, pp. 20 391–20 427, 2020.
  • [7] S. Gu, L. Zhang, W. Zuo, and X. Feng, “Weighted nuclear norm minimization with application to image denoising,” in Proc. IEEE CVPR’14, 2014, pp. 2862–2869.
  • [8] J. Xu, L. Zhang, W. Zuo, D. Zhang, and X. Feng, “Patch group based nonlocal self-similarity prior learning for image denoising,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 244–252.
  • [9] Z. Zha, X. Yuan, B. Wen, J. Zhang, J. Zhou, and C. Zhu, “Simultaneous nonlocal self-similarity prior for image denoising,” in 2019 IEEE International Conference on Image Processing (ICIP).   IEEE, 2019, pp. 1119–1123.
  • [10] K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Color image denoising via sparse 3d collaborative filtering with grouping constraint in luminance-chrominance space,” in Proc. IEEE ICIP’07, vol. 1, 2007, pp. I – 313–I – 316.
  • [11] J. Xu, L. Zhang, D. Zhang, and X. Feng, “Multi-channel weighted nuclear norm minimization for real color image denoising,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 1096–1104.
  • [12] H. C. Burger, C. J. Schuler, and S. Harmeling, “Image denoising: Can plain neural networks compete with bm3d?” in 2012 IEEE conference on computer vision and pattern recognition.   IEEE, 2012, pp. 2392–2399.
  • [13] K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, “Beyond a gaussian denoiser: Residual learning of deep CNN for image denoising,” IEEE Trans. Image Process., vol. 26, no. 7, pp. 3142–3155, 2017.
  • [14] K. Zhang, W. Zuo, and L. Zhang, “FFDNet: toward a fast and flexible solution for cnn-based image denoising,” IEEE Trans. Image Process., vol. 27, no. 9, pp. 4608–4622, 2018.
  • [15] S. Guo, Z. Yan, K. Zhang, W. Zuo, and L. Zhang, “Toward convolutional blind denoising of real photographs,” in Proc. IEEE CVPR’19, 2019, pp. 1712–1722.
  • [16] J. Chen, J. Chen, H. Chao, and M. Yang, “Image blind denoising with generative adversarial network based noise modeling,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3155–3164.
  • [17] S. Laine, T. Karras, J. Lehtinen, and T. Aila, “High-quality self-supervised deep image denoising,” Advances in Neural Information Processing Systems, vol. 32, 2019.
  • [18] Y. Quan, M. Chen, T. Pang, and H. Ji, “Self2self with dropout: Learning self-supervised denoising from single image,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1890–1898.
  • [19] G. Toderici, S. M. O’Malley, S. J. Hwang, D. Vincent, D. Minnen, S. Baluja, M. Covell, and R. Sukthankar, “Variable rate image compression with recurrent neural networks,” in ICLR, 2016.
  • [20] D. Minnen, G. Toderici, M. Covell, T. Chinen, N. Johnston, J. Shor, S. J. Hwang, D. Vincent, and S. Singh, “Spatially adaptive image compression using a tiled deep network,” in Proc. IEEE ICIP, 2017, pp. 2796–2800.
  • [21] N. Johnston, D. Vincent, D. Minnen, M. Covell, S. Singh, T. Chinen, S. Jin Hwang, J. Shor, and G. Toderici, “Improved lossy image compression with priming and spatially adaptive bit rates for recurrent networks,” in Proc. IEEE/CVF CVPR, 2018, pp. 4385–4393.
  • [22] J. Ballé, V. Laparra, and E. Simoncelli, “End-to-end optimized image compression,” in Proc. ICLR’17, 2017.
  • [23] J. Ballé, D. Minnen, S. Singh, S. J. Hwang, and N. Johnston, “Variational image compression with a scale hyperprior,” in Proc. ICLR’18, 2018.
  • [24] Z. Cheng, H. Sun, M. Takeuchi, and J. Katto, “Learned image compression with discretized gaussian mixture likelihoods and attention modules,” in Proc. IEEE CVPR’20, 2020, pp. 7936–7945.
  • [25] Z. Guo, Z. Zhang, R. Feng, and Z. Chen, “Causal contextual prediction for learned image compression,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 4, pp. 2329–2341, 2022.
  • [26] Y. Choi, M. El-Khamy, and J. Lee, “Variable rate deep image compression with a conditional autoencoder,” in Proc. IEEE/CVF ICCV, 2019, pp. 3146–3154.
  • [27] F. Yang, L. Herranz, J. v. d. Weijer, J. A. I. Guitián, A. M. López, and M. G. Mozerov, “Variable rate deep image compression with modulated autoencoder,” IEEE Signal Processing Letters, vol. 27, pp. 331–335, 2020.
  • [28] D. Sebai, “Multi-rate deep semantic image compression with quantized modulated autoencoder,” in Proc. IEEE MMSP, 2021, pp. 1–6.
  • [29] S. Yin, C. Li, Y. Bao, Y. Liang, F. Meng, and W. Liu, “Universal efficient variable-rate neural image compression,” in Proc. IEEE ICASSP, 2022, pp. 2025–2029.
  • [30] ISO/IEC and ITU-T, “JPEG AI use cases and requirements,” ISO/IEC JTC 1/SC29/WG1 N100094, 2022.
  • [31] H. Choi and I. V. Bajić, “Scalable image coding for humans and machines,” IEEE Trans. Image Process., vol. 31, pp. 2739–2754, Mar. 2022.
  • [32] ——, “Latent-space scalability for multi-task collaborative intelligence,” in Proc. IEEE ICIP, Sep. 2021, pp. 3562–3566.
  • [33] M. Testolina, E. Upenik, and T. Ebrahimi, “Towards image denoising in the latent space of learning-based compression,” in Applications of Digital Image Processing XLIV, vol. 11842.   SPIE, 2021, pp. 412–422.
  • [34] V. Alves de Oliveira, M. Chabert, T. Oberlin, C. Poulliat, M. Bruno, C. Latry, M. Carlavan, S. Henrot, F. Falzon, and R. Camarero, “Satellite image compression and denoising with neural networks,” IEEE Geoscience and Remote Sensing Letters, vol. 19, pp. 1–5, 2022.
  • [35] D. Minnen, J. Ballé, and G. D. Toderici, “Joint autoregressive and hierarchical priors for learned image compression,” Advances in Neural Information Processing Systems, vol. 31, pp. 10 771–10 780, 2018.
  • [36] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Kopf, E. Yang, Z. DeVito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala, “Pytorch: An imperative style, high-performance deep learning library,” in Advances in Neural Information Processing Systems 32.   Curran Associates, Inc., 2019, pp. 8024–8035. [Online]. Available: http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
  • [37] CLIC, “Challenge on learned image compression (CLIC),” [Online]: http://www.compression.cc/, 2019.
  • [38] D. R. Martin, C. C. Fowlkes, D. Tal, and J. Malik, “A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” in Proc. IEEE ICCV’01, 2001, pp. 416–425.
  • [39] R. Franzen, “Kodak lossless true color image suite,” source: http://r0k. us/graphics/kodak, vol. 4, no. 2, 1999.
  • [40] L. Zhang, X. Wu, A. Buades, and X. Li, “Color demosaicking by local directional interpolation and nonlocal adaptive thresholding,” J. Electronic Imaging, vol. 20, no. 2, p. 023016, 2011.
  • [41] S. Ranjbar Alvar and I. V. Bajić, “Practical noise simulation for RGB images,” arXiv preprint arXiv:2201.12773, 2022.
  • [42] K. Zhang, Y. Li, W. Zuo, L. Zhang, L. Van Gool, and R. Timofte, “Plug-and-play image restoration with deep denoiser prior,” IEEE Trans. Pattern Anal. Mach. Intell., 2021.
  • [43] J. Bégaint, F. Racapé, S. Feltman, and A. Pushparaja, “Compressai: a pytorch library and evaluation platform for end-to-end compression research,” arXiv preprint arXiv:2011.03029, 2020.
  • [44] G. Bjøntegaard, “VCEG-M33: Calculation of average PSNR differences between RD-curves,” in Video Coding Experts Group (VCEG).   ITU –Telecommunications Standardization, Apr. 2001.
  • [45] A. Foi, M. Trimeche, V. Katkovnik, and K. Egiazarian, “Practical Poissonian-Gaussian noise modeling and fitting for single-image raw-data,” IEEE Trans. Image Process., vol. 17, no. 10, pp. 1737–1754, Oct. 2008.