![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Our method uses fast-to-compute high-resolution auxiliary features to support super resolution of Monte Carlo rendering results. Our method generates high-quality visual details than both off-the-shelf super resolution methods and MSSPL, a dedicated super resolution method for Monte Carlo renderings that uses both more high-resolution auxiliary features and the corresponding high resolution low sample rendering result.
Auxiliary Features-Guided Super Resolution for Monte Carlo Rendering
Abstract
This paper investigates super resolution to reduce the number of pixels to render and thus speed up Monte Carlo rendering algorithms. While great progress has been made to super resolution technologies, it is essentially an ill-posed problem and cannot recover high-frequency details in renderings. To address this problem, we exploit high-resolution auxiliary features to guide super resolution of low-resolution renderings. These high-resolution auxiliary features can be quickly rendered by a rendering engine and at the same time provide valuable high-frequency details to assist super resolution. To this end, we develop a cross-modality Transformer network that consists of an auxiliary feature branch and a low-resolution rendering branch. These two branches are designed to fuse high-resolution auxiliary features with the corresponding low-resolution rendering. Furthermore, we design residual densely-connected Swin Transformer groups to learn to extract representative features to enable high-quality super-resolution. Our experiments show that our auxiliary features-guided super-resolution method outperforms both super-resolution methods and Monte Carlo denoising methods in producing high-quality renderings.
{CCSXML}<ccs2012> <concept> <concept_id>10010147.10010371.10010372.10010374</concept_id> <concept_desc>Computing methodologies Ray tracing</concept_desc> <concept_significance>500</concept_significance> </concept> </ccs2012>
\ccsdesc[500]Computing methodologies Ray tracing \printccsdesc
keywords:
Super resolution, Fast-to-compute auxiliary features, Transformer, Monte Carlo rendering1 Introduction
Monte Carlo rendering algorithms are now widely used to generate photo realistic computer graphics images for applications such as visual effects, video games, and computer animations. These algorithms generate a pixel’s color by integrating over all the light paths arriving at a single point [CPC84]. To rendering a high-quality image, a large number of rays need to be cast for each pixel, which makes Monte Carol rendering a slow process.
A great amount of effort has been devoted to speeding up Monte Carlo rendering. The core idea is to reduce the number of rays for each pixel. For instance, numerous denoising algorithms are now available to reconstruct a high-quality image from a rendering produced at a low sampling rate. Such Monte Carlo denoising algorithms often use auxiliary features generated by a rendering algorithm to help denoise the noisy rendering result. The recent deep neural network-based denoising algorithms can now generate very high-quality images at a fairly low sampling rate [BVM*17, CKS*17, KKR18, GLA*19].
Monte Carlo rendering can also be sped up by reducing the number of pixels to render. For example, pixels from the frames that have already been rendered can be warped to generate frames in-between existing frames to increase the frame rate [BDM*21] or to generate future frames to reduce latency [GFL*21]. Another approach is to only render one pixel for a block of neighboring pixels to further reduce the total number of pixels to render. This can be implemented by first rendering a low-resolution image and then applying super resolution to increase its resolution [XNC*20, HLM*21]. As super resolution is a fundamentally ill-posed problem, it alone often cannot recover high-frequency details from only the low-resolution rendering. To address this problem, Hou et al.render a high-resolution rendering with a low sampling rate and use that together with the high-resolution auxiliary features to help super resolve the low-resolution rendering rendered at a high sampling rate. While this method produces a high-quality result, it needs to render the high-resolution image at a low sampling rate, which still takes a considerable amount of time [HLM*21].
Can we only use the fast-to-obtain high-resolution auxiliary features without the high-resolution-low-sample rendering to effectively assist super resolution of the corresponding low-resolution rendering? If so, we can further speed up Monte Carlo rendering. We are encouraged by the recent work on neural frame synthesis that showed fast-to-obtain auxiliary features of the target frames can greatly help interpolate or extrapolate the target frames [BDM*21, GFL*21]. On the other hand, Hou et al.showed that using a wide range of auxiliary features and the high-resolution-low-sample rendering help super resolution more than only using a subset of auxiliary features within their own deep neural network-based super resolution framework [HLM*21]. Therefore, if we only use a small number of fast-to-compute auxiliary features, we need to have a better super resolution method.
This paper presents a Cross-modality Residual Densely-connected Swin Transformer (XRDS) for super resolution of a Monte Carlo rendering guided by its auxiliary features. For the seek of speed, we only use two auxiliary features: albedo and normal. To effectively use these features, we design a super resolution network based on Swin Transformer that recently has been shown powerful for a wide variety of computer vision tasks. Our Transformer network has two branches, one for the low resolution rendering and the other for the auxiliary features. Such two branches are designed to perform cross-modality fusion to effectively use auxiliary features to assist super resolution of the low-resolution rendering. While the auxiliary feature branch consists of convolutional blocks, the branch for the low-resolution rendering consists of a sequence of residual densely-connected Swin Transformer blocks to extract effective features. The features from the two branches are combined together using a cross-modality fusion module and are finally used to generate the high-resolution high-quality rendering.
This paper contributes to Monte Carlo rendering as follows. First, we present the first super resolution approach to Monte Carlo rendering that only uses fast-to-compute high-resolution auxiliary features to enable high-quality upsampling of a low-resolution rendering. Second, we design a dedicated Cross-modality Swin Transformer-based super resolution network that can learn to effectively combine high-resolution auxiliary features with the corresponding low-resolution rendering to generate the final high-resolution high-quality image. Third, our experiments show that our method outperforms super-resolution and denoising methods in producing high-quality renderings.
2 Related Work

This section briefly discusses relevant work to our paper, including Monte Carlo denoising, super resolution, and vision Transformers.
Monte Carlo Denoising. Monte Carlo rendering algorithms need numerous samples per pixel to generate a high-quality rendering [CPC84, Kaj86]. With insufficient samples, the rendering results suffer from noise. To address this problem, many Monte Carlo denoising methods have been developed to reconstruct high-quality renderings from only a small number of samples. Traditional methods reconstruct renderings in a similar way to general image denoising methods by designing specific denoising kernels based on image variance or geometric features or directly regress the final result [DSHL10, JC95, LSK*07, LR90, McC99, RW94, SIMP06, SD12, XP05]. Zwicker et al.provided a good survey on this [ZJL*15]. Alternatively, adaptive sampling algorithms can be used to reduce the overall sample numbers for the whole image [BM98, ETH*09, Jen01, LWC12, MA06, MMMG16, ODR09, RKZ11, WABG06, WRC88].
Recently, deep neural network-based denoising methods have shown impressive performance. These methods learn to reconstruct high-quality renderings from small number of samples. In their seminal work, Kalantari et al.estimated optimal filter parameters using a multi-layer perceptron neural network [KBS15]. Bako et al.estimated spatially adaptive kernels for denoising in a convolutional manner [BVM*17]. Vogels et al.extended the concept of kernel prediction methods to temporal denoising [VRM*18]. With asymmetric loss functions, their method could produce high-quality results for a sequence of frames. Chaitanya et al.developed a recurrent autoencoder to denoise a sequence of frames while maintaining temporal stability [CKS*17]. Xu et al.developed an adversarial approach to Monte Carlo rendering denoising that can greatly reduce artifacts such as blurs and unfaithful details from denoising results [XZW*19]. Gharbi et al.developed a kernel splatting network that reconstructs the final image by splatting samples to pixels according to the estimated splatting kernels [GLA*19]. Munkberg et al.proposed to filter auxiliary layers of individual samples [MH20]. Their method works well on outliers and complex visibility. Hasselgren et al.proposed a neural spatial-temporal sampling method for Monte Carlo video denoising [HMS*20]. Their method first estimates the sampling map from the temporal reprojection and auxiliary features and then denoises the resulting image generated using the sampling map to produce high-quality results. Zheng et al.proposed an ensemble denoising technique that learns to combine multiple denoiser together [ZZXY21]. Yu et al.designed a transformer-based neural network for Monte Carlo denoising [YNL*21]. Their network consists multi-scale feature extractor and a self-attention module and achieved promising results. Unlike these denoising methods, our method explores an orthogonal approach that speeds up Monte Carlo rendering by reducing the number of pixels to render via super-resolution.
Super resolution. Super resolution is a classic problem in computer vision. It aims to reconstruct a high-resolution image from the low-resolution input. Recently, the state of the art of super resolution research has been advanced significantly due to the use of deep neural networks [DLHT14, AKS18, HSU19, HWG18, KKM16, LSK*17, LTT*19, LWLS19, XMS19, ZZZ18, ZWLQ19]. Specifically, in their seminal work, Dong et al.trained a three-layer fully convolutional neural network for single image super resolution [DLHT14]. Since that, a large variety of super resolution deep neural networks have been invented by leveraging carefully designed network architectures, including residual blocks [HZRS16], densely connected blocks [HLVW17, ZTK*18], channel attention blocks [HSS18, ZLL*18], transformers [LCS*21], and others.
Super resolution has recently been explored to speed up rendering. Xiao et al.designed a super resolution network that takes both the low-resolution rendering and the corresponding low-resolution auxiliary features as input and outputs a high-resolution frame. They leveraged neighboring frames to further improve super resolution quality [XNC*20]. While their method was designed for a rasterization-based renderering engine, in principal, it can be applied to Monte Carlo rendering. Thomas et al.combined super resolution and Monte Carlo denoising for videos. Their network takes a low-resolution rendering as well as a warped previous frame as input and produces a high-resolution frame [TLP*22]. However, super resolution is essentially an ill-posed problem and cannot fully recover missing high-frequency visual details from only the low-resolution input. To address this problem, Hou et al.developed a super resolution approach based on multiple-resolution sampling. Their method first renders a low-resolution image at a high-sampling rate and a high-resolution image at a low sampling rate. Their method then exploits the high-resolution noisy image to recover high frequency visual details [HLM*21]. While their method generates high-quality renderings, it needs to render the high resolution noisy image and auxiliary features, which takes a considerable amount of time. Different from the above methods, our method obtains high-frequency information from only fast-to-compute high-resolution auxiliary features as inspired by recent interpolation and extrapolation methods that use fast-to-compute auxiliary features of the target frames to help generate the target frames [GFL*21, BDM*21].
Vision transformer. Transformer was initially designed for natural language process tasks [VSP*17]. Due to its self-attention mechanism, it can efficiently capture the long-term information from the input. Recently, transformer networks have attracted considerable attention in the computer vision community and achieved great success in various computer vision tasks, including image recognition [LLC*21], object detection [CMS*20], semantic segmentation [ZLZ*21], and image restoration [LCS*21]. Dosovitskiy et al.developed the first transformer network for image recognition [DBK*20]. They split the input image into image patches and then feed these image patches as tokens to the transformer network. Chen et al.presented an image processing transformer for various restoration problems and demonstrated that pretraining on large datasets could greatly improve the capability of a transformer network for low-level computer vision tasks [CWG*21]. Liang et al.developed SwinIR for image restoration. Their network adapted the Swin Transformer [LLC*21] as their backbone and achieved promising results [LCS*21]. However, transformer for the Monte-Carlo denoising is less explored. Inspired by the success of these vision transformer networks, we are the first to design a dedicated cross-modality transformer network for super resolution of Monte Carlo renderings that can effectively leverage fast-to-compute high-resolution auxiliary features to recover high-frequency visual details when upsampling a low-resolution rendering.
3 Our Method
This paper proposes a super resolution method guided by the fast-to-compute auxiliary features to speed up the Monte Carlo rendering. Our method takes a low resolution rendering and its high-resolution fast-to-compute auxiliary features as input, and outputs the corresponding high-quality high-resolution result . The high-resolution auxiliary features provide the essential high-frequency information for the super-resolution.
Different from the previous work [HLM*21], which leverages a wide range of auxiliary features, our method only employs the auxiliary features that can be computed very fast [BDM*21], including albedo and normal. On the one hand, although our method doesn’t leverage the shading layers, albedo and normal could provide a lot of high-frequency information, e.g., the texture of the material, which is essential for super-resolution. As we will show, it can help improve the super-resolution results. On the other hand, albedo and normal can be computed fast [BDM*21]. It not only reduces the rendering time but also enables us to render these high-resolution layers at a relatively higher sampling rate, which typically contains fewer artifacts, such as aliasing.

We design a cross-modality transformer network to effectively fuse two categories of visual input, namely the low-resolution rendering and its corresponding high-resolution auxiliary features, to recover visual details. Figure 1 shows the architecture of our network. It contains two parallel branches, one for the low-resolution rendering and the other for the corresponding high-resolution auxiliary features.
Auxiliary feature branch. The auxiliary feature branch takes auxiliary features as inputs, which provide essential high frequency visual details. As discussed above, we select albedo and normal, which are relatively fast to acquire. Since this branch processes high-resolution input, we design a shallow architecture for the sake of memory and speed. As shown in Figure 1, we employ a convolutional layer and residual blocks (RB) [HZRS16] in a sequence to get the features
(1) |
where indicates the convolution operation. indicates the operation of a residual block. In our experiments, we set the channels as 32 for the auxiliary feature branch.
We then obtain the downsampled features with a group of deshuffle layers [HLM*21], which is able to downscale the feature while keeping the high frequency information.
(2) |
where indicates the deshuffle layer.
Low resolution rendering branch. Following the recent works on image super resolution [ZLL*18, ZZZ18, LCS*21], we first adopt a convolutional layer with 64 channels to get the shallow feature from the low resolution rendering .
(3) |
We feed the resulting feature to a sequence of cross-modality residual densely-connected Swin Transformer groups (XDG).
(4) |
where indicates the XDG module. indicates the number of XDG. We choose in our experiments. XDG is designed to fuse the auxiliary features and the low-resolution rendering features . It consists of a cross-modality module (XM) and a sequence of residual densely-connected Swin Transformer blocks (RDST). Specifically, XM is designed to fuse the local information from the low-resolution rendering and the high frequency information from the auxiliary features, while the RDST sequence learns more dedicated representations for super resolution from them.
Cross-modality module (XM). Inspired by the success of Swin Transformer [LLC*21, LCS*21] and Transformer Decoder [GLDZ22], we design XM based on Swin Transformer, which can efficiently model the long-range dependency. Figure 2 shows the architecture of XM. It takes features from the low-resolution rendering branch and features from the auxiliary feature branch as input and outputs the fused feature . It consists of Layer Norm layers (LN), a Window-based Multi-head Self-Attention layer (W-MSA), a Window-based Multi-head Cross Attention layer (W-MCA), and a Multi-Layer Perception layer (MLP). The key idea behind XM is to combine the features from the low-resolution rendering branch with the features from the high-resolution auxiliary branch using cross-attention, creating a more comprehensive representation for super resolution. The process starts by extracting intermediate features from , which serve as the "query" . From , which holds high-resolution information, the "key" and "value" are extracted. Then, the cross-attention is calculated following [VSP*17] and combined with to generate . Finally, an MLP layer is used to integrate the features from the low-resolution branch and the cross-attention.

Methods | ||||||
---|---|---|---|---|---|---|
PSNR | RelMSE | PSNR | RelMSE | PSNR | RelMSE | |
Bicubic | 30.57 | 0.0141 | 25.39 | 0.0858 | 22.36 | 0.2473 |
EDSR | 32.01 | 0.0079 | 30.70 | 0.0119 | 27.97 | 0.0241 |
RCAN | 32.03 | 0.0084 | 30.73 | 0.0117 | 27.92 | 0.0253 |
SwinIR | 31.05 | 0.0118 | 30.78 | 0.0152 | 28.04 | 0.0364 |
MSSPL | 38.40 | 0.0015 | 34.27 | 0.0039 | 31.08 | 0.0079 |
Ours | 42.48 | 0.0007 | 37.45 | 0.0021 | 31.94 | 0.0076 |
Residual Densely-connected Swin Transformer block (RDST). As shown in Figure 1, we feed the fused feature from XM to a sequence of residual densely-connected Swin Transformer blocks (RDST),
(5) |
where indicates the RDST block. We also use a short skip connection to combine the shallow feature with the deep feature
(6) |
We design RDST by combining the ideas of the Residual Densely-connected Network (RDN) [ZTK*18] and Swin Transformer [LLC*21]. We are specifically inspired by SwinIR [LCS*21] that explores Swin Transformers for image restoration tasks. It replaces traditional convolutional layers with Swin layers in residual blocks, allowing for the learning of more descriptive features and delivering impressive results. Taking inspiration from RDN [ZTK*18], we introduce RDST, where the convolution layers in densely-connected blocks are replaced with Swin layers. As shown in Figure 3, RDST consists of a sequence of densely-connected Swin Transformer blocks and a local feature fusion block. For the densely-connected Swin Transformer blocks, we shift the windows. We also use a local skip connection to fuse the features from the shallow layer.
Upscale. We adopt the pixel shuffle layer [SCH*16] to upscale the dense feature to a high resolution feature. We also uses a convolutional layer with 3 channels to predict the final high resolution image .
(7) |
where indicates the operation of the pixel shuffle layer.
Training details. We adopt the robust loss to handle the prediction with a high dynamic range image [HLM*21].
(8) |
where indicates the ground truth image. indicates the number of pixels. indicates the robust factor, which is set to 0.1.
We implement our network in PyTorch. We train our super resolution network on examples of size . We select Adam [KB14] with , as the optimizer. The learning rate is set to 0.0001. We train the network for 400 epochs with a mini-batch size of 16 for our super resolution models, and we fine-tune our other models using the pretrained weights. It takes about one week to train a single model using 4 Nvidia A40 GPUs. We adopt the BCR dataset [HLM*21] as the training dataset. BCR dataset contains 2449 images from 1463 scenes rendered by Blender Cycles. Following MSSPL [HLM*21], we use 2126 images from 1283 scenes for training, 193 images from 76 scenes for validation, and 130 images from 104 scenes for testing.

4 Experiments
We evaluate our network by quantitatively and qualitatively comparing them with state-of-the-art image super resolution methods and the Monte Carlo denoising methods on the BCR dataset [HLM*21] and the Gharbi dataset [GLA*19]. We also conduct the ablation study to examine our method. Following [HLM*21], we adopt Relative Mean Square Error (RelMSE) and PSNR to evaluate our methods in the scene linear color space and the sRGB space, respectively. Please refer to the supplementary material for an interactive demo that provides more results.
4.1 Comparison with Super Resolution Methods
We compare our method with state-of-the-art super-resolution methods, including EDSR [LSK*17], RCAN [ZLL*18], and SwinIR [LCS*21], a recent transformer-based approach, as well as the multiple sampling-based super resolution method MSSPL [HLM*21]. We obtained the results of compared methods either from the authors [HLM*21] or from finetuning the official models [LSK*17, ZLL*18, LCS*21] on the BCR dataset.
As shown in Table 1, our method outperforms super-resolution methods. This improvement can be largely attributed to the use of high-resolution auxiliary features to capture high-frequency visual details. For this experiment, we use the ground truth auxiliary features as they are fast to acquire. We also vary the number of samples used to generate these features in order to examine their effect on our method. As shown in Figure 4, while having more samples to generate these auxiliary features benefits our method, the features generated with only one sample per pixel allow our method to outperform the standard super-resolution methods.
MSSPL takes both the low-resolution rendering and the high resolution noisy rendering as well as a wide variety of high resolution auxiliary features as input [HLM*21]. In this test, the high resolution rendering and features are rendered with one sample per pixel. As shown in Figure 4 and Table 1, our method, when only using albedo and normal as auxiliary features obtained with one sample per pixel, can achieve 35.16 dB which is higher than MSSPL (34.27 dB) for the task, even though our method takes much less input information from the high resolution input.
Speed and memory. Table 2 reports the speeds and the peak memory of the above methods. As our method is based on the Transformer, our method is slower than CNN-based methods, including EDSR [LSK*17], RCAN [ZLL*18] and MSSPL [HLM*21]. This is consistent with many other works that Transformer tends to be slower than CNN [DBK*20, TCD*21, LCS*21]. Compared to Transformer-based method SwinIR, our method is slightly faster. We also compare the peak memory to produce a image in Table 2. Our method uses less peak memory than EDSR and MSSPL but more memory than RCAN and SwinIR.
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
![]() |
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
4.2 Comparison with Denoising Methods
Scale | EDSR | RCAN | SwinIR | MSSPL | Ours | |
---|---|---|---|---|---|---|
Runtime(ms) | 503.96 | 280.51 | 1149.25 | 125.24 | 1009.08 | |
Peak memory(MB) | 2493.9 | 672.1 | 806.0 | 739.70 | 941.3 | |
Peak memory(MB) | 2375.6 | 621.3 | 659.0 | 1010.1 | 803.8 | |
Peak memory(MB) | 2359.7 | 615.4 | 608.4 | 1008.0 | 783.4 |
Method | 2spp | 4spp | 8spp | |||
---|---|---|---|---|---|---|
PSNR | RelMSE | PSNR | RelMSE | PSNR | RelMSE | |
Input | 18.12 | 0.2953 | 21.51 | 0.1400 | 24.75 | 0.0646 |
KPCN | 25.87 | 0.0390 | 27.31 | 0.0299 | 28.11 | 0.0276 |
KPCN-ft | 31.03 | 0.0078 | 33.69 | 0.0043 | 35.83 | 0.0026 |
Bitterli | 26.67 | 0.0293 | 27.22 | 0.0252 | 27.45 | 0.0226 |
Gharbi | 30.73 | 0.0068 | 31.61 | 0.0057 | 32.29 | 0.0050 |
MSSPL | 33.27 | 0.0044 | 35.15 | 0.0027 | 36.74 | 0.0019 |
MSSPL | 33.94 | 0.0039 | 35.21 | 0.0028 | 36.31 | 0.0022 |
MSSPL | 31.37 | 0.0075 | 32.35 | 0.0057 | 33.14 | 0.0049 |
AdvMC-ft | 30.33 | - | 32.30 | - | 33.69 | - |
MCSA-ft | 32.68 | 0.0049 | 34.81 | 0.0031 | 36.61 | 0.0021 |
Ours | (1 - 1) | ( 2 - 2) | (4 - 4) | |||
31.04 | 0.0078 | 34.67 | 0.0030 | 36.62 | 0.0020 | |
Ours | (4 - 1) | ( 8 - 2) | (16 - 4) | |||
34.12 | 0.0035 | 35.49 | 0.0026 | 37.09 | 0.0018 | |
Ours | (16 - 1) | (32 - 2) | (64 - 4) | |||
34.08 | 0.0046 | 35.06 | 0.0034 | 35.77 | 0.0029 | |
Ours | (64 - 1) | (128 - 2) | (250 - 4) | |||
31.10 | 0.0095 | 31.36 | 0.0093 | 31.62 | 0.0093 |
Method | 4 spp | 8 spp | 16 spp | |||
---|---|---|---|---|---|---|
PSNR | RelMSE | PSNR | RelMSE | PSNR | RelMSE | |
Input | 19.58 | 17.5358 | 21.91 | 7.5682 | 24.17 | 11.2189 |
Sen | 28.23 | 1.0484 | 28.00 | 0.5744 | 27.64 | 0.3396 |
Rousselle | 30.01 | 1.9407 | 32.32 | 1.9660 | 34.36 | 1.9446 |
Kalantari | 31.33 | 1.5573 | 33.00 | 1.6635 | 34.43 | 1.8021 |
Bitterli | 28.98 | 1.1024 | 30.92 | 0.9297 | 32.40 | 0.9640 |
KPCN | 29.75 | 1.0616 | 30.56 | 7.0774 | 31.00 | 20.2309 |
KPCN-ft | 29.86 | 0.5004 | 31.66 | 0.8616 | 33.39 | 0.2981 |
Gharbi | 33.11 | 0.0486 | 34.45 | 0.0385 | 35.36 | 0.0318 |
MSSPL | 34.02 | 1.5025 | 35.30 | 1.4902 | 36.43 | 1.4748 |
MSSPL | 33.94 | 5.5586 | 35.22 | 5.6781 | 35.97 | 5.7436 |
MSSPL | 31.56 | 3.7228 | 32.60 | 4.2300 | 33.22 | 4.5045 |
Ours | (2 - 2) | (4 - 4) | (8 - 8) | |||
27.41 | 0.3438 | 30.39 | 0.3092 | 32.88 | 0.3062 | |
Ours | (8 - 2) | (16 - 4) | (32 - 8) | |||
34.29 | 2.2587 | 35.47 | 1.5480 | 36.37 | 1.5417 | |
Ours | (32 - 2) | (64 - 4) | (128 - 8) | |||
34.26 | 20.7861 | 35.12 | 29.0364 | 35.52 | 28.1264 | |
Ours | (128 - 2) | (16 - 8) | (32 - 16) | |||
31.57 | 1.3474 | 31.26 | 1.1718 | 31.51 | 1.0940 |
Method | Buffer | 2spp | 4spp | 8spp | |||
---|---|---|---|---|---|---|---|
PSNR | RelMSE | PSNR | RelMSE | PSNR | RelMSE | ||
MSSPL | Full | (16 - 1) | (32 - 2) | (64 - 4) | |||
33.94 | 0.0039 | 35.21 | 0.0028 | 36.31 | 0.0022 | ||
MSSPL | Fast | (16 - 1) | (32 - 2) | (64 - 4) | |||
32.32 | 0.0079 | 33.42 | 0.0060 | 33.96 | 0.0050 | ||
Ours | Full | (16 - 1) | (32 - 2) | (64 - 4) | |||
34.84 | 0.0033 | 36.01 | 0.0024 | 37.22 | 0.0018 | ||
Ours | Fast | (16 - 1) | (32 - 2) | (64 - 4) | |||
34.08 | 0.0046 | 35.06 | 0.0034 | 35.77 | 0.0029 |
Auxiliary Layer | None | Normal | Albedo | Normal + Albedo |
---|---|---|---|---|
PSNR | 30.49 | 34.85 | 36.42 | 37.45 |
RelMSE | 0.0141 | 0.0042 | 0.0030 | 0.0021 |
Method | AdvMC-ft | MCSA | Ours |
---|---|---|---|
PSNR | 27.96 | 30.01 | 34.12 |
LPIPS | 0.320 | 0.202 | 0.090 |
We compare our methods to the state-of-the-art Monte Carlo denoising methods, including Sen [SD12], Rousselle [RKZ11], Kalantari [KBS15], Bitterli [BRM*16], KPCN[BVM*17], Gharbi [GLA*19], MSSPL [HLM*21], AdvMC [XZW*19], and MCSA [YNL*21]. Table 3 and Table 4 report results on the BCR dataset [HLM*21] and the Gharbi dataset [GLA*19], respectively. We obtain the results of the comparing methods either from their authors [HLM*21] or from their project websites [GLA*19]. MSSPL [HLM*21] was trained on the BCR dataset. For KPCN [BVM*17], AdvMC [XZW*19], and MCSA [YNL*21], we finetuned their official models on the BCR dataset using their official training scripts. For our model, We trained a distinct model for each scale and sampling count.
As most denoising methods do not take high-resolution auxiliary features as input, we follow MSSPL [HLM*21] to compute the average spp for our method and MSSPL as , where indicates the scale. and indicate the sampling rates for the low-resolution and high-resolution inputs, respectively. In our case, we take the sampling rates for the auxiliary features as . We would like to note that this measurement of spp is unfair to our method, as our method only uses high-resolution albedo and normal features which takes much less time than rendering all the shading layers to obtain the high-resolution rendering as done in MSSPL.
As shown in Table 3, our method generates better results than the state-of-the-art methods on the BCR dataset [HLM*21]. Ours model wins 0.18dB, 0.28dB, and 0.35 dB in terms of PSNR on 2spp, 4sppp, and 8spp, respectively.
We also conduct our experiments on scale. On the one hand, with , our method produces worse results than MSSPL because MSSPL uses the high-resolution RGB image as input that is not available to our method. While the high-resolution RGB input to MSSPL is rendered at a low sampling rate, it still provides useful information. As shown in the existing literature on Monte Carlo denoising, even the rendering result at 1 spp can be denoised to a reasonable quality. At such a high upsampling rate of , super-resolution is very difficult. On the other hand, in practice, given a target overall spp rate, our method can select an optimal combination of (spp rate, super-resolution scale) that outperforms MSSPL and other methods, as shown in Table 3. In practice, will not be used for rendering by either MSSPL or our method to achieve an overall target spp as it produces the worst results among alternative combinations of spp rate and super-resolution scale.
Figure 6 shows the visual comparisons. Our results are more visually plausible. Briefly, instead of working in the pixel color space that can potentially cause the color fidelity problem, our method fuses the low-resolution RGB and high-resolution feature maps in the feature space and learns to fuse them into correct colors, thus alleviating the color ambiguities/artifacts at fine details. For example, In Figure 7, the wall of our results is less noisy and more accurate than the results from other methods that are either blurred or inconsistent with the ground truth. In the second example, our method produces high-frequency geometric details in the wine basket area that well differentiates the mesh color and the background color.
Table 4 reports the comparison on the Gharbi dataset [GLA*19]. Following MSSPL [GLA*19], we directly test our models pretrained on the BCR dataset without fine-tuning as the training set of the Gharbi dataset is not available. Our model wins 0.27dB and 0.17dB in terms of PSNR on 4spp and 8spp, respectively. When the spp is 16, our PSNR is slightly lower than MSSPL [HLM*21]. We would like point out our method takes less high-resolution information than MSSPL. Our input high-resolution auxiliary features only include the albedo and normal, while MSSPL also takes all the shading layers as inputs. When the high resolution input is rendered at a high spp, the shading layers can contribute a lot of high frequency information. Similar to the findings in MSSPL [HLM*21], our results on RelMSE are heavily affected by a small number of pixels with abnormal large errors. Excluding these abnormal pixels can greatly improve our scores on RelMSE. As shown in Figure 7, our method produces high-quality results with much fewer artifacts when compared to the ground truth.
4.3 Discussions
Auxiliary features sampling rates. As discussed above and shown in Figure 4, using more samples to generate the auxiliary features help our method generate better super resolution results. However, even using one sample per pixel to generate the auxiliary features can already enable our method to significantly outperform standard super resolution methods. Moreover, when we use 16 samples to generate these features, our results are already very close to the results that use the features generated using 4000 samples per pixel denoted as in the figure.
Input layers of auxiliary features. We examine how our method works with different auxiliary feature layers. The upsampling scale is set to . We use 4000 spp for and . As shown in Table 7, both albedo and normal can improve the results significantly, as they can provide the essential high frequency visual details for super resolution. The performance of our network can be further improved if we take both of them as inputs. These findings are consistent with previous denoising methods [BVM*17, GLA*19] where intermediate layers can improve the final results.
Ablation study w.r.t. MSSPL [HLM*21]. We evaluated the performance of both our method and MSSPL [HLM*21] using fast-to-compute auxiliary features as well as full auxiliary features. In the experiments, the upsampling scale is set to . As shown in Table 5, both our network and MSSPL benefit from using the full auxiliary features due to the richer high-resolution information they provide. However, our method with fast-to-compute layers still outperforms MSSPL with full auxiliary layers, which demonstrates the effectiveness of our network architecture.
Network | RDB | RSTB | RDST | RDST + XM |
---|---|---|---|---|
PSNR | 35.56 | 36.63 | 37.27 | 37.45 |
RelMSE | 0.0034 | 0.0098 | 0.0022 | 0.0021 |
Method | 2spp | 4spp | 8spp | |||
---|---|---|---|---|---|---|
PSNR | LPIPS | PSNR | LPIPS | PSNR | LPIPS | |
AdvMC-ft | 30.33 | 0.209 | 32.30 | 0.155 | 33.69 | 0.126 |
MCSA-ft | 32.68 | 0.108 | 34.81 | 0.080 | 36.61 | 0.068 |
Ours | 34.12 | 0.090 | 35.49 | 0.070 | 37.09 | 0.057 |
RDST Num | PSNR | RelMSE | Flops(T) | Macs(G) | Params(M) |
---|---|---|---|---|---|
5 | 34.08 | 0.0046 | 1.45 | 723.68 | 9.36 |
3 | 33.47 | 0.0056 | 1.04 | 519.13 | 6.21 |
1 | 32.60 | 0.0091 | 0.63 | 314.59 | 3.06 |
XDG Num | PSNR | RelMSE | Flops(T) | Macs(G) | Params(M) |
---|---|---|---|---|---|
3 | 34.08 | 0.0046 | 1.45 | 723.68 | 9.36 |
2 | 33.19 | 0.0066 | 1.02 | 507.86 | 6.42 |
1 | 32.30 | 0.1193 | 0.59 | 292.04 | 3.49 |
Network Effectiveness. We examine how our network architecture works by comparing to AdvMC [XZW*19] and MCSA [YNL*21]. Specifically, we feed high-resolution 1-spp RGB and 1-spp auxiliary buffers to AdvMC and MCSA and fine tune them on the BCR dataset. In this experiment, our method takes 4-spp low-resolution RGB (, effectively the same sampling rate as 1 spp at the high resolution) and 1-spp high-resolution auxiliary buffers. Table 7 shows our method outperforms these methods, which demonstrates the effectiveness of our transformer-based network architecture.
Network architecture components. We examine the effect of the network architecture. The upsampling scale is set to . In this test, we remove XM modules and replace our RDST with the state-of-the-art blocks, including RDB from RRN [ZTK*18] and RSTB from SwinIR [LCS*21]. As shown in Table 8, our RDST can greatly improve the results. These improvements can be attributed to the strong generalization capability of RDST. Besides, XM modules can further improve the results.
Number of RDST blocks. We examine how our network architecture works with different RDST blocks in each XDG block on the BCR dataset [HLM*21]. In this test, the upsampling scale is set to x4. To check the impact of RDST, we set the XDG number as 3, and we investigated our results across different RDST numbers of each XDG block, including 1, 3, and 5. Besides, we also measure the flops, macs, and parameters for a single 1024x1024 image [RRRH20]. As shown in Table 10, decreasing the number of RDST blocks accelerates the network but compromises performance.
Number of XDG blocks. Similar to RDST, we investigate our results across different XDG numbers, including 1, 2, and 3. The upsampling scale is set to x4 and the number of RDST of each XDG block is set to 5. As the results reported in Table 11, reducing the number of XDG blocks accelerates the network but also compromises performance.
Our robust loss vs SMAPE loss [Mea86]. Our robust loss is used based on our observations that there are a very small number of pixels with abnormally large intensity values in our dataset, mostly due to the firefly artifacts. These pixels will often incur very large errors during training and thus compromise the performance of our model. We use the robust loss to reduce these undesirable impacts of these pixels as this robust loss will limit the maximal loss value to 1 no matter how large the pixel error is. We compared these two loss functions. In our experiments, the upsampling factor is set to 4, and we set the sampling rate to (16 - 1). Models trained with the SMAPE loss showed slightly worse results: 33.96 v.s.34.12 in PSNR, and 0.0046 v.s.0.0035 in RelMSE.
Super resolution scale. We investigate our results across multiple scales, including , , , and . Among them, scales and exhibit weaker performance compared to and . When comparing scales and , takes less peak memory and is faster than , but leads to better quality. To make a fair comparison, we maintain a consistent average sampling rate across different scales. Consequently, the low-resolution input of our model is rendered at a much lower average sampling rate than that of our model. This makes the resulting input RGB image to our model very noisy for and thus comprises the final quality of Ours , as reported in the 2-spp column of Table 3. In the 4-spp column of the same table, the difference between Ours and Ours is less significant as in this setting, the average sampling rate of Ours is reasonably higher and provides more information for our model to synthesize higher-quality results.
In addition, we used the same training pipeline for our model as we did for other scales, keeping the number of epochs consistent across all scales. However, due to the high memory requirement to train the model, we have to set a smaller mini-batch size. This would also potentially impact the performance, but we believe that this is not as significant as the first reason we discussed above.
Perceptual quality. We examine the perceptual quality of our results using the LPIPS metric [ZIE*18]. Table 7 and Table 9 present the results for AdvMC [XZW*19], MCSA [YNL*21], and our method. Our approach outperforms the others in terms of both PSNR and LPIPS, thereby demonstrating its ability to generate images with high perceptual quality.
5 Limitations and Future Work
The fusion for the high-reflection parts is challenging. Our method produces high-frequency visual details by two means: 1) train a neural network to learn to recover high-frequency information from low-resolution input and 2) use high-frequency information from the high-resolution albedo and normal maps. Our neural network can learn to produce visual details for many examples. However, super resolution from a low-resolution input alone is necessarily an ill-posed problem. In the high-reflection parts of the scene, such as the example shown in Figure 8, when the high-resolution normal and albedo map cannot, by its nature, provide high-frequency details in those regions, our method may fail.
Compared to CNN-based methods, our method is slow. However, compared to another Transformer-based method [YNL*21], our method uses less peak memory (0.89Gb vs 30.56Gb) and is faster (1.0s vs 2.5s) when producing a 1024x1024 image using an Nvidia A40. Research on fast transformers has been advancing quickly recently. Patro et al. [PA23] offer an extensive review of efficient vision transformers. Through the advancement of effective token mixing strategies and efficient MLP layers, vision transformers can be significantly accelerated [LWZ*22, GHW*22, YPL*22]. For example, both CMT [GHW*22] and WaveViT [YPL*22] outperform EfficientNet [TL19] while maintaining a lower computational complexity. Moreover, several transformer hardware accelerators have been introduced to expedite Transformer networks, such as SwiftTron [MDC*23]. We believe that our method can benefit from the quick advance of research on Transformer.
In this paper, we specifically explored albedo and normal as quick-to-compute auxiliary features. However, we acknowledge that other auxiliary features, such as a whitted ray-traced layer, could offer valuable high-frequency information and be generated fast. Incorporating such a layer can potentially improve the performance of our method. Unfortunately, the BCR dataset doesn’t contain such layers. We plan to explore this in our future research.
Source
GT
LR
Albedo
Normal
Ours
6 Conclusion
This paper explored high-resolution fast-to-compute auxiliary features to guide super resolution of Monte Carlo renderings. We developed a dedicated cross-modality Transformer network to fuse high-resolution fast-to-compute auxiliary features with the corresponding low-resolution rendering. We designed a Transformer-based cross-modality module to fuse the features from two modalities. We also developed a Residual Densely-connected Swin Transformer block to learn more representative features. Experimental results indicate that our proposed method surpasses existing state-of-the-art super-resolution and denoising techniques in producing high-quality images.
References
- [AKS18] Namhyuk Ahn, Byungkon Kang and Kyung-Ah Sohn “Fast, accurate, and lightweight super-resolution with cascading residual network” In Proceedings of the European Conference on Computer Vision, 2018, pp. 252–268
- [BDM*21] Karlis Martins Briedis et al. “Neural frame interpolation for rendered content” In ACM Transactions on Graphics (TOG) 40.6 ACM New York, NY, USA, 2021, pp. 1–13
- [BM98] Mark R Bolin and Gary W Meyer “A perceptually based adaptive sampling algorithm” In Proceedings of the 25th annual conference on Computer graphics and interactive techniques ACM, 1998, pp. 299–309
- [BRM*16] Benedikt Bitterli et al. “Nonlinearly Weighted First-order Regression for Denoising Monte Carlo Renderings” In Computer Graphics Forum 35.4, 2016, pp. 107–117 Wiley Online Library
- [BVM*17] Steve Bako et al. “Kernel-predicting convolutional networks for denoising Monte Carlo renderings” In ACM Transactions on Graphics 36.4 ACM, 2017, pp. 97
- [CKS*17] Chakravarty R Alla Chaitanya et al. “Interactive reconstruction of Monte Carlo image sequences using a recurrent denoising autoencoder” In ACM Transactions on Graphics 36.4 ACM, 2017, pp. 98
- [CMS*20] Nicolas Carion et al. “End-to-end object detection with transformers” In European conference on computer vision, 2020, pp. 213–229 Springer
- [CPC84] Robert L Cook, Thomas Porter and Loren Carpenter “Distributed ray tracing” In ACM SIGGRAPH computer graphics 18.3, 1984, pp. 137–145
- [CWG*21] Hanting Chen et al. “Pre-Trained Image Processing Transformer”, 2021 arXiv:2012.00364 [cs.CV]
- [DBK*20] Alexey Dosovitskiy et al. “An image is worth 16x16 words: Transformers for image recognition at scale” In arXiv preprint arXiv:2010.11929, 2020
- [DLHT14] Chao Dong, Chen Change Loy, Kaiming He and Xiaoou Tang “Learning a deep convolutional network for image super-resolution” In European conference on computer vision, 2014, pp. 184–199 Springer
- [DSHL10] Holger Dammertz, Daniel Sewtz, Johannes Hanika and Hendrik Lensch “Edge-avoiding À-Trous wavelet transform for fast global illumination filtering” In Proceedings of the Conference on High Performance Graphics, 2010, pp. 67–75 Eurographics Association
- [ETH*09] Kevin Egan et al. “Frequency analysis and sheared reconstruction for rendering motion blur” In ACM Transactions on Graphics 28.3, 2009, pp. 93
- [GFL*21] Jie Guo et al. “ExtraNet: real-time extrapolated rendering for low-latency temporal supersampling” In ACM Transactions on Graphics (TOG) 40.6 ACM New York, NY, USA, 2021, pp. 1–16
- [GHW*22] Jianyuan Guo et al. “Cmt: Convolutional neural networks meet vision transformers” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 12175–12185
- [GLA*19] Michaël Gharbi et al. “Sample-based Monte Carlo denoising using a kernel-splatting network” In ACM Transactions on Graphics 38.4 ACM New York, NY, USA, 2019, pp. 1–12
- [GLDZ22] Zhicheng Geng, Luming Liang, Tianyu Ding and Ilya Zharkov “RSTT: Real-time Spatial Temporal Transformer for Space-Time Video Super-Resolution” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022
- [HLM*21] Qiqi Hou et al. “Fast Monte Carlo Rendering via Multi-Resolution Sampling” In Graphics Interface Conference, 2021, pp. 25:1–9
- [HLVW17] Gao Huang, Zhuang Liu, Laurens Van Der Maaten and Kilian Q Weinberger “Densely connected convolutional networks” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4700–4708
- [HMS*20] Jon Hasselgren et al. “Neural temporal adaptive sampling and denoising” In Computer Graphics Forum 39.2, 2020, pp. 147–155 Wiley Online Library
- [HSS18] Jie Hu, Li Shen and Gang Sun “Squeeze-and-excitation networks” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 7132–7141
- [HSU19] Muhammad Haris, Gregory Shakhnarovich and Norimichi Ukita “Recurrent back-projection network for video super-resolution” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 3897–3906
- [HWG18] Zheng Hui, Xiumei Wang and Xinbo Gao “Fast and accurate single image super-resolution via information distillation network” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 723–731
- [HZRS16] Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
- [JC95] Henrik Wann Jensen and Niels Jørgen Christensen “Optimizing path tracing using noise reduction filters” Václav Skala-UNION Agency, 1995
- [Jen01] Henrik Wann Jensen “Realistic image synthesis using photon mapping” AK Peters/CRC Press, 2001
- [Kaj86] James T Kajiya “The rendering equation” In ACM SIGGRAPH computer graphics 20.4 ACM, 1986, pp. 143–150
- [KB14] Diederik P Kingma and Jimmy Ba “Adam: A method for stochastic optimization” In arXiv preprint arXiv:1412.6980, 2014
- [KBS15] Nima Khademi Kalantari, Steve Bako and Pradeep Sen “A machine learning approach for filtering Monte Carlo noise.” In ACM Trans. Graph. 34.4, 2015, pp. 122–1
- [KKM16] Jiwon Kim, Jung Kwon Lee and Kyoung Mu Lee “Deeply-recursive convolutional network for image super-resolution” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1637–1645
- [KKR18] Alexandr Kuznetsov, Nima Khademi Kalantari and Ravi Ramamoorthi “Deep adaptive sampling for low sample count rendering” In Computer Graphics Forum 37.4, 2018, pp. 35–44
- [LCS*21] Jingyun Liang et al. “Swinir: Image restoration using swin transformer” In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 1833–1844
- [LLC*21] Ze Liu et al. “Swin transformer: Hierarchical vision transformer using shifted windows” In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 10012–10022
- [LR90] Mark E Lee and Richard A Redner “A note on the use of nonlinear filtering in computer graphics” In IEEE Computer Graphics and Applications 10.3 IEEE, 1990, pp. 23–29
- [LSK*07] Samuli Laine et al. “Incremental instant radiosity for real-time indirect illumination” In Proceedings of the 18th Eurographics conference on Rendering Techniques, 2007, pp. 277–286 Eurographics Association
- [LSK*17] Bee Lim et al. “Enhanced deep residual networks for single image super-resolution” In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2017, pp. 136–144
- [LTT*19] Yawei Li et al. “3D appearance super-resolution with deep learning” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 9671–9680
- [LWC12] Tzu-Mao Li, Yu-Ting Wu and Yung-Yu Chuang “SURE-based optimization for adaptive sampling and reconstruction” In ACM Transactions on Graphics 31.6 ACM, 2012, pp. 194
- [LWLS19] Zhi-Song Liu, Li-Wen Wang, Chu-Tak Li and Wan-Chi Siu “Hierarchical back projection network for image super-resolution” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2019
- [LWZ*22] Kunchang Li et al. “Uniformer: Unifying convolution and self-attention for visual recognition” In arXiv preprint arXiv:2201.09450, 2022
- [MA06] Mark Meyer and John Anderson “Statistical acceleration for animated global illumination” In ACM Transactions on Graphics 25.3 ACM, 2006, pp. 1075–1080
- [McC99] Michael D McCool “Anisotropic diffusion for Monte Carlo noise reduction” In ACM Transactions on Graphics 18.2 ACM, 1999, pp. 171–194
- [MDC*23] Alberto Marchisio et al. “SwiftTron: An Efficient Hardware Accelerator for Quantized Transformers” In arXiv preprint arXiv:2304.03986, 2023
- [Mea86] Nigel Meade “Long range forecasting: From crystal ball to computer” In Journal of the Operational Research Society 37.5 Taylor & Francis, 1986, pp. 533–535
- [MH20] Jacob Munkberg and Jon Hasselgren “Neural denoising with layer embeddings” In Computer Graphics Forum 39.4, 2020, pp. 1–12 Wiley Online Library
- [MMMG16] Bochang Moon, Steven McDonagh, Kenny Mitchell and Markus Gross “Adaptive polynomial rendering” In ACM Transactions on Graphics 35.4 ACM, 2016, pp. 40
- [ODR09] Ryan S Overbeck, Craig Donner and Ravi Ramamoorthi “Adaptive wavelet rendering.” In ACM Trans. Graph. 28.5, 2009, pp. 140
- [RKZ11] Fabrice Rousselle, Claude Knaus and Matthias Zwicker “Adaptive sampling and reconstruction using greedy error minimization” In ACM Transactions on Graphics 30.6 ACM, 2011, pp. 159
- [RRRH20] Jeff Rasley, Samyam Rajbhandari, Olatunji Ruwase and Yuxiong He “Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters” In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2020, pp. 3505–3506
- [RW94] Holly E Rushmeier and Gregory J Ward “Energy preserving non-linear filters” In Proceedings of the 21st annual conference on Computer graphics and interactive techniques ACM, 1994, pp. 131–138
- [SCH*16] Wenzhe Shi et al. “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1874–1883
- [SD12] Pradeep Sen and Soheil Darabi “On filtering the noise from the random parameters in Monte Carlo rendering.” In ACM Trans. Graph. 31.3, 2012, pp. 18–1
- [SIMP06] Benjamin Segovia, Jean Claude Iehl, Richard Mitanchey and Bernard Péroche “Non-interleaved deferred shading of interleaved sample patterns” In Graphics Hardware, 2006, pp. 53–60
- [TCD*21] Hugo Touvron et al. “Training data-efficient image transformers distillation through attention” In International Conference on Machine Learning 139, 2021, pp. 10347–10357
- [TL19] Mingxing Tan and Quoc Le “Efficientnet: Rethinking model scaling for convolutional neural networks” In International conference on machine learning, 2019, pp. 6105–6114 PMLR
- [TLP*22] Manu Mathew Thomas et al. “Temporally Stable Real-Time Joint Neural Denoising and Supersampling” In Proceedings of the ACM on Computer Graphics and Interactive Techniques 5.3 ACM New York, NY, USA, 2022, pp. 1–22
- [VRM*18] Thijs Vogels et al. “Denoising with kernel prediction and asymmetric loss functions” In ACM Transactions on Graphics (TOG) 37.4 ACM New York, NY, USA, 2018, pp. 1–15
- [VSP*17] Ashish Vaswani et al. “Attention is all you need” In Advances in neural information processing systems 30, 2017
- [WABG06] Bruce Walter, Adam Arbree, Kavita Bala and Donald P Greenberg “Multidimensional lightcuts” In ACM Transactions on graphics 25.3 ACM, 2006, pp. 1081–1088
- [WRC88] Gregory J Ward, Francis M Rubinstein and Robert D Clear “A ray tracing solution for diffuse interreflection” In ACM SIGGRAPH Computer Graphics 22.4 ACM, 1988, pp. 85–92
- [XMS19] Xiangyu Xu, Yongrui Ma and Wenxiu Sun “Towards real scene super-resolution with raw images” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 1723–1731
- [XNC*20] Lei Xiao et al. “Neural supersampling for real-time rendering” In ACM Transactions on Graphics (TOG) 39.4 ACM New York, NY, USA, 2020, pp. 142–1
- [XP05] Ruifeng Xu and Sumanta N Pattanaik “A novel Monte Carlo noise reduction operator” In IEEE Computer Graphics and Applications 25.2 IEEE, 2005, pp. 31–35
- [XZW*19] Bing Xu et al. “Adversarial Monte Carlo denoising with conditioned auxiliary feature modulation.” In ACM Trans. Graph. 38.6, 2019, pp. 224–1
- [YNL*21] Jiaqi Yu et al. “Monte Carlo denoising via auxiliary feature guided self-attention.” In ACM Trans. Graph. 40.6, 2021, pp. 273–1
- [YPL*22] Ting Yao et al. “Wave-vit: Unifying wavelet and transformers for visual representation learning” In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXV, 2022, pp. 328–345 Springer
- [ZIE*18] Richard Zhang et al. “The Unreasonable Effectiveness of Deep Features as a Perceptual Metric” In CVPR, 2018
- [ZJL*15] Matthias Zwicker et al. “Recent advances in adaptive sampling and reconstruction for Monte Carlo rendering” In Computer Graphics Forum 34.2, 2015, pp. 667–681
- [ZLL*18] Yulun Zhang et al. “Image super-resolution using very deep residual channel attention networks” In Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 286–301
- [ZLZ*21] Sixiao Zheng et al. “Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 6881–6890
- [ZTK*18] Yulun Zhang et al. “Residual dense network for image super-resolution” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 2472–2481
- [ZWLQ19] Zhifei Zhang, Zhaowen Wang, Zhe Lin and Hairong Qi “Image super-resolution by neural texture transfer” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 7982–7991
- [ZZXY21] Shaokun Zheng, Fengshi Zheng, Kun Xu and Ling-Qi Yan “Ensemble denoising for Monte Carlo renderings” In ACM Transactions on Graphics (TOG) 40.6 ACM New York, NY, USA, 2021, pp. 1–17
- [ZZZ18] Kai Zhang, Wangmeng Zuo and Lei Zhang “Learning a single convolutional super-resolution network for multiple degradations” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 3262–3271