This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: 1The University of Texas at Austin
11email: {zhiwenfan,yifanjiang97,peihaowang,xinyu.gong,dejia,atlaswang}@utexas.edu

Unified Implicit Neural Stylization

Zhiwen Fan 11 ⋆ ⋆    Yifan Jiang 11 ⋆ ⋆    Peihao Wang 11 ⋆ ⋆   
Xinyu Gong1
   Dejia Xu1    Zhangyang Wang
11
Abstract

Representing visual signals by implicit neural representation (INR) has prevailed among many vision tasks. Its potential for editing/processing given signals remains less explored. This work explores a new intriguing direction: training a stylized implicit representation, using a generalized approach that can apply to various 2D and 3D scenarios. We conduct a pilot study on a variety of INRs, including 2D coordinate-based representation, signed distance function, and neural radiance field. Our solution is a Unified Implicit Neural Stylization framework, dubbed INS. In contrary to vanilla INR, INS decouples the ordinary implicit function into a style implicit module and a content implicit module, in order to separately encode the representations from the style image and input scenes. An amalgamation module is then applied to aggregate these information and synthesize the stylized output. To regularize the geometry in 3D scenes, we propose a novel self-distillation geometry consistency loss which preserves the geometry fidelity of the stylized scenes. Comprehensive experiments are conducted on multiple task settings, including fitting images using MLPs, stylization for implicit surfaces and sylized novel view synthesis using neural radiance. We further demonstrate that the learned representation is continuous not only spatially but also style-wise, leading to effortlessly interpolating between different styles and generating images with new mixed styles. Please refer to the video on our project page for more view synthesis results: https://zhiwenfan.github.io/INS.

1 Introduction

Implicit Neural Representation (INR) has gained remarkable popularity \star Equal contribution in representing concise signal representation in computer vision and computer graphics [70, 53, 58, 49, 82]. As an alternative to discrete grid-based signal representation, implicit representation is able to parameterize modern signals as samples of a continuous manifold, using multi-layer perceptions (MLP) to map between coordinates and signal values. Several seminal works [53, 70, 83] have verified the effectiveness of INR in representing image, video, and audio. Followups further apply INR to more challenging tasks including novel-view synthesis [53, 6, 7, 86], 3D-aware generative model [90, 10, 9], and inverse problem [16, 74].

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
SIREN [70] SIREN+INS Style Image SDF [83] SDF+INS Style Image
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
NeRF [53] Style Image NeRF+INS NeRF+INS NeRF+INS NeRF+INS
Figure 1: Representative visual examples generated by the proposed method. We show stylized results on three different types of implicit representation, including 2D coordinate-based mapping function (SIREN [70]), Signed Distance Function(SDF [58]), and Neural Radiance Field (NeRF [53]).

While implicit neural representation reveals multiple advantages compared to conventional discrete signals, a general question of curiosity might be: which and how modern visual signal processing approaches/tasks designed for discrete signals can also be applied to continuous representations? Research pursuing this answer has been conducted on implicit neural representation since its origin. Chen et al. [17] apply a local implicit function to image super-resolution and they observe that INR can surpass bilinear and nearest upsampling. Sun et al. [74] demonstrate the effectiveness of INR in the context of sparse-view X-ray CT. Dupont et al. [23] propose to store the weights of a neural implicit function instead of pixel values, which surprisingly outperforms JPEG compression format. [87] further demonstrates superior video compression using similar ideas.

We investigate a novel setting: to yield visually pleasing stylized examples under various 2D and 3D scenarios, using a generalized approach leveraging implicit neural representations. Note that, training a stylized implicit neural representation still faces many hurdles. On one hand, the aforementioned works mostly have the access to dense measurements or at least sparse clean data, which enables training an implicit neural network under the supervision of target signal. In contrast to those tasks/approaches, current image stylization mechanisms are mostly conducted in an unsupervised manner, due to the absence of stylized ground truth data. Consequently, it is still unknown whether coordinate-based MLP can be optimized without accessing corresponding ground truth signals. On the other hand, existing style images are mostly based on 2D scenes, which raises obstacles when being considered as the appearance of 3D implicit representation. Prior art [19] attempted on marrying stylization with one specific type of Implicit Neural Representation, the neural radiance field (NeRF) [53]. Nevertheless, it still captures the statistics of style information by a series of pre-trained convolution-based hypernetwork to generate model weights, rather than a direct implicitly encoding stylization. As indicated by recent literature [46], training a robust hypernetwork requires a large amount of training samples, while novel-view synthesis tasks commonly hold no more than hundreds of views, potentially jeopardizing the synthesized visual quality.

To conquer the aforementioned fragility, we propose a Unified Implicit Neural Stylization framework, coined as INS. Different from the vanilla implicit function which is built upon a single MLP network, the proposed framework divides an ordinary implicit neural representation to multiple individual components. Concretely speaking, we introduce a Style Implicit Module to the ordinary implicit representation, and coin the later one as Content Implicit Module in our framework. During the training process, the stylized information and content scene are encoded as one continuous representation, and then fused by another Amalgamation Module. To further regularize the geometry of given scenes, we utilize an additional self-distilled geometry consistency loss on top of the rendered density, for the stylization of NeRF. Eventually, INS is able to render view-consistent stylized scenes from novel views, with visually impressive texture details: a few examples are shown in Figure. 1.

Our contributions are outlined below:

  • We propose INS, a unified implicit neural stylization framework, consists of a style implicit module, a content implicit module, and an amalgamation module, which enables us to synthesize promising stylized scenes under multiple 2D and 3D implicit representations.

  • We conduct comprehensive experiments on several popular implicit representation frameworks in this novel stylization setting, including 2D coordinate-based framework (SIREN [70]), Neural Radiance Field (NeRF [53]), and Signed Distance Functions (SDF [58]). The rendering results are found to be more consistent, in both shape and style details, from different views.

  • We further demonstrate that INS is able to learn representations that are continuous not only with regard to spatial placements (including views), but also in the style space. This leads to effortlessly interpolating between different styles and generating images rendered by the new mixed styles.

2 Related Works

2.1 Implicit Function

Recent research has exhibited the potential of Implicit Neural Representation (INR) to replace traditional discrete signals with continuous functions parameterized by multilayer perceptrons (MLP), in computer vision and graphics [75, 71]. The coordinate-based neural representations [18, 49, 50] have become a popular representation for various tasks such as representing image/video [70, 23, 87], 3D reconstruction [27, 4, 57, 18, 29, 49, 56, 58, 62, 64], and 3D-aware generative modelling [10, 21, 30, 33, 48, 55, 65, 90]. Analogously, as this representation is differentiable, prior works apply coordinate-based MLPs to many inverse problems in computational photography [16, 12, 72, 3, 66, 77] and scientific computing [45, 32, 89].

2.2 Implicit 3D Scene Representation

Traditional 3D reconstruction methods utilizes discrete representations such as point cloud [2], meshes [43, 67], multi-plane images [52], depth maps [81, 31] and voxel grids [44]. Recently, INR has also prevailed among 3D scene representation tasks, which simply adopt an MLP that maps from any continuous input 3D coordinate to the geometry of the scene, including signed distance function (SDF) [58, 41, 5, 50, 29, 73, 40], 3D occupancy network [49, 18], and so on. In addition to representing shape, INR has also been extended to encode object appearance. Among them, Neural Radiance Field (NeRF [53]) is one of the most effective coordinate-based neural representations for photo-realistic view synthesis that represents a scene as a field of particles. Draw inspiration from the preliminary success made by NeRF, a lot of following works further improve and extend it to wider application [6, 7, 8, 24, 47, 84, 51, 61, 13, 34, 6, 7, 61, 60, 78]. Different from the grid-based approaches, training a stylized implicit representation can not access ground truth signals, which further amplifies the difficulty of optimizing the implicit neural representation.

2.3 Stylization

Traditionally, image stylization is formulated as a painterly rendering process through stroke prediction [88, 85]. The first neural style transfer method, proposed by Gatys et al.  [26], builds an iterative framework to optimize the input image in order to minimize the content and style loss defined by a pre-trained VGG network. Due to the frustratingly large cost of training time, a number of follow-ups further explore how to design a feed-forward deep neural networks [42, 76], which obtain real-time performance without sacrificing too much style information. Recently, several works extend it to video stylization [63, 37, 11, 14] and 3D environment [54, 36, 19, 28, 35].

Ha-NeRF [15] is proposed for recovering a realistic NeRF at a different time of day from a group of tourism images, with a CNN to encode the appearance latent code. The most related NeRF-based stylization works: Style3D [19] and StylizedNeRF [38] still require a CNN-based hypernetwork or decoder to generate the stylized parameters for neural radiance field. In comparison, our proposed INS framework can work for more general implicit representations beyond neural radiance field, and can also be extended to encoding multiple styles. Experiments demonstrate that INS generate more faithful stylization on NeRF compared with Style3D [19].

Refer to caption
Figure 2: The main pipeline of unified implicit neural stylization (INS) framework and its components. We took SDF with the proposed INS, for example, it inputs with implicit coordinates along with ray directions and style embeddings. Style Implicit Module (SIM) and Content Implicit Module (CIM) are used to extract conditional implicit style features and implicit scene features. Amalgamate Module (AM) is applied to fuse features in the two spaces, generating styliezed density and color intensity of each 3D point. An implicit rendering step is applied on the top of AM (i.e., Volume rendering for NeRF, surface rendering for SDF) to render the pixel intensity.

3 Preliminary

This section introduces the relevant background on several implicit representations and volumetric radiance representations, including image fitting [70], neural radiance field [53] and signed distance function [58].

Implicit Image Fitting:

The most prototypical example of neural implicit representation is image regression [70, 75]. Consider fitting a function f:23f:{}^{2}\rightarrow{}^{3} that encodes the pixel array of a given image \mathcal{I} into a continuous representation. Function f(𝒙)f(\boldsymbol{x}) takes pixel coordinates 𝒙2\boldsymbol{x}\in{}^{2} as the inputs, and outputs the corresponding RGB colors 𝒄3\boldsymbol{c}\in{}^{3}. Parameterizing ff with a multi-layer perception networks (MLPs), it can be optimized by the mean-squared error (MSE) loss function =𝔼𝒙()f(𝒙)𝒄22\mathcal{L}=\operatorname{\mathbb{E}}_{\boldsymbol{x}\sim\operatorname{\mathbb{P}}(\mathcal{I})}\lVert f(\boldsymbol{x})-\boldsymbol{c}\rVert_{2}^{2}, where ()\operatorname{\mathbb{P}}(\mathcal{I}) is a probability measure support in image lattice \mathcal{I}.

Neural Radiance Field:

In contrast to point-wisely regression of implicit fields, NeRF [53] proposes to reconstruct a radiance field by inversing a differentiable rendering equation from captured images. Specifically, NeRF learns an MLP f:(𝒙,𝜽)(𝒄,σ)f:(\boldsymbol{x},\boldsymbol{\theta})\mapsto(\boldsymbol{c},\sigma) with parameters 𝚯\boldsymbol{\Theta}, where 𝒙\boldsymbol{x} is the spatial coordinate in 3D space and 𝜽\boldsymbol{\theta} represents the view directions [π,π]2\in[-\pi,\pi]^{2}. The output 𝒄3\boldsymbol{c}\in{}^{3} indicates the predicted color of the sampled point, σ+\sigma\in{}_{+} signifies its density value. The pixel color intensity can be obtained using volume rendering [22] by ray tracing, integrating the predicted color and density along the ray. To render a pixel on the image plane, NeRF casts a ray 𝒓=(𝒐,𝒅,𝜽)\boldsymbol{r}=(\boldsymbol{o},\boldsymbol{d},\boldsymbol{\theta}) through the pixel and accumulate the color and density of KK point samples along the view direction in the 3D space. The pixel color intensity can be estimated:

𝑪(𝒓|𝚯)=k=1KTk(1exp(σkΔtk))𝒄k,\displaystyle\boldsymbol{C}(\boldsymbol{r}|\boldsymbol{\Theta})=\sum_{k=1}^{K}T_{k}(1-\exp(-\sigma_{k}\Delta t_{k}))\boldsymbol{c}_{k}, (1)

where (𝒄k,σk)=f(𝒙k,𝜽)(\boldsymbol{c}_{k},\sigma_{k})=f(\boldsymbol{x}_{k},\boldsymbol{\theta}), 𝒙k=𝒐+tk𝒅\boldsymbol{x}_{k}=\boldsymbol{o}+t_{k}\boldsymbol{d}, tkt_{k} are the marching distance of sampled points, and Tk=exp(l=1k1σlΔtl)T_{k}=\exp(-\sum_{l=1}^{k-1}\sigma_{l}\Delta t_{l}) is known as the transmittance to model occlusion. Δtk=tk+1tk\Delta t_{k}=t_{k+1}-t_{k} indicates the distance of sampled point in 3D space. With this approximated rendering pipeline, the model weights are optimized by minimizing the L2L_{2} distance between rendered ray colors 𝑪(𝒓)\boldsymbol{C}(\boldsymbol{r}) and captured pixel colors 𝑪^(𝒓)\widehat{\boldsymbol{C}}(\boldsymbol{r}) as follows:

𝚯=argmin𝚯𝔼𝒓()𝑪(𝒓|𝚯)𝑪^(𝒓)22,\displaystyle\boldsymbol{\Theta}^{*}=\operatorname*{arg\,min}_{\boldsymbol{\Theta}}\operatorname{\mathbb{E}}_{\boldsymbol{r}\sim\operatorname{\mathbb{P}}(\mathcal{R})}\left\lVert\boldsymbol{C}(\boldsymbol{r}|\boldsymbol{\Theta})-\widehat{\boldsymbol{C}}(\boldsymbol{r})\right\rVert_{2}^{2}, (2)

where \mathcal{R} is a collection of rays cast from all pixels in the training set, and ()\operatorname{\mathbb{P}}(\mathcal{R}) defines a distribution over it.

Implicit Surface Representation:

Signed Distance Function (SDF) [20] f:3f:{}^{3}\rightarrow\real is an implicit representation of 3D geometries. SDF specifies each spatial point with the signed distance to the implicit iso-surface, where the sign indicates whether the point is inside or outside the object. Recent works of [59, 69] propose to employ MLPs to represent this continuous field via direct supervision using point clouds. To optimize a textured SDF from multi-view images like NeRF [51], Yariv et al.[83] proposes a neural rendering pipeline, named IDR, which enables rendering images from an SDF. With this framework, one can indirectly supervise SDF using its multi-view projections. Suppose given a camera pose, we can cast rays 𝒓=(𝒐,𝒗)\boldsymbol{r}=(\boldsymbol{o},\boldsymbol{v}) through each pixel to trace an intersected point with the surface:

𝒙^=𝒐+t0𝒗𝒗f0𝒗0f(𝒐+t0𝒗),\displaystyle\boldsymbol{\hat{x}}=\boldsymbol{o}+t_{0}\boldsymbol{v}-\frac{\boldsymbol{v}}{\nabla f_{0}\cdot\boldsymbol{v}_{0}}f(\boldsymbol{o}+t_{0}\boldsymbol{v}), (3)

where t0t_{0}, 𝒗0\boldsymbol{v}_{0} and f0f_{0} are initial states when performing ray tracing (see [83]). After obtaining the intersection 𝒙^\boldsymbol{\hat{x}} of ray and surface, IDR also lets the SDF network ff output an appearance embedding 𝜸^\boldsymbol{\hat{\gamma}}, and computes the normal 𝒏^=f(𝒙^)\boldsymbol{\hat{n}}=\nabla f(\boldsymbol{\hat{x}}). Then the ray color can be rendered by another rendering MLP conditioned on both point coordinate 𝒙^\boldsymbol{\hat{x}} and normal 𝒏^\boldsymbol{\hat{n}}:

𝑪(𝒓|𝚯)=r(𝒙^,𝒏^,𝒅,𝜸^).\displaystyle\boldsymbol{C}(\boldsymbol{r}\lvert\boldsymbol{\Theta})=r(\boldsymbol{\hat{x}},\boldsymbol{\hat{n}},\boldsymbol{d},\boldsymbol{\hat{\gamma}}). (4)

Similar to NeRF [51], ff and rr are simultaneously optimized by photometric loss between captured image pixels and rendered rays (see Equation 2).

4 Method

We next illustrate the main pipeline of Unified Implicit Neural Stylization (INS). INS consists of a Style Implicit Module (SIM) to transform the input style embedding into implicit style representations, a Content Implicit Module (CIM) to map the input coordinates into implicit scene representations, and an Amalgamate Module (AM) which amalgamates the two representation to predict RGB intensity. To preserve the geometry fidelity while generating the stylized texture of rendered views, a self-distilled geometry consistency regularization is applied upon the INS framework.

4.1 Implicit Style and Content Representation

Generating the stylized images 𝒀\boldsymbol{Y} can be formulated as an energy minimization problem [26]. It consists of a content loss and a style loss, defined under a pre-trained VGG network [68]. We build upon the prior work and thus propose our implicit stylization framework for SIREN, SDF and NeRF.

Content Representation

The content loss in 2D images stylization pipeline [26] content\mathcal{L}_{\mathrm{content}} is defined as:

content(𝑪,𝒀)=1Ci,jWi,jHi,j(i,j)𝒥Fi,j(𝑪)Fi,j(𝒀)F2,\displaystyle\mathcal{L}_{\mathrm{content}}(\boldsymbol{C},\boldsymbol{Y})=\frac{1}{C_{i,j}W_{i,j}H_{i,j}}\sum_{(i,j)\in\mathcal{J}}\left\lVert F_{i,j}(\boldsymbol{C})-F_{i,j}(\boldsymbol{Y})\right\rVert_{F}^{2}, (5)

where 𝑪\boldsymbol{C} denotes the content ground truth image and 𝒀\boldsymbol{Y} denotes synthesized output, Fi,jF_{i,j} denotes the feature map extracted from a VGG-16 model pre-trained on ImageNet, ii represents its ii-th max pooling, and jj represents its jj-th convolutional layer after ii-th max pooling layer. Ci,jC_{i,j}, Wi,jW_{i,j} and Hi,jH_{i,j} are the dimensions of the extracted feature maps. We adapt the content loss to the intermediate layer of INS pipeline to preserve the content of the predicted color image patch, we choose ii = 2, jj = 2 by default.

Style Representation

To extract representation of the stylized information,  [68] introduces a different feature space to capture texture information [25]. Similar to the content loss, the feature space is built upon the filter response in multiple layers of a pre-trained VGG network. By capturing the correlations of the filter responses expressed by the Gram matrix 𝑮i,jCi,j×Ci,j\boldsymbol{G}_{i,j}\in{}^{C_{i,j}\times C_{i,j}} between the style image 𝑺\boldsymbol{S} and the synthesized image 𝒀\boldsymbol{Y}, multi-scale representations can be obtained to capture the texture information from the style image and endow such texture on the stylized image. Here, we define our style loss style\mathcal{L}_{\mathrm{style}} using the same layers of VGG-16 with [26] on the top of the prediction of implicit neural representations and the given style image:

style(𝑺,𝒀)=(i,j)𝒥𝑮i,j(𝑺)𝑮i,j(𝒀)F2,\displaystyle\mathcal{L}_{\mathrm{style}}(\boldsymbol{S},\boldsymbol{Y})=\sum_{(i,j)\in\mathcal{J}}\left\lVert\boldsymbol{G}_{i,j}(\boldsymbol{S})-\boldsymbol{G}_{i,j}(\boldsymbol{Y})\right\rVert_{F}^{2}, (6)
where [𝑮i,j]c,c(𝒀)=1Ci,jWi,jHi,jk=1H×WFi,j(𝒀)c,kFi,j(𝒀)c,k,\displaystyle\text{where }[\boldsymbol{G}_{i,j}]_{c,c^{\prime}}(\boldsymbol{Y})=\frac{1}{C_{i,j}W_{i,j}H_{i,j}}\sum_{k=1}^{H\times W}F_{i,j}(\boldsymbol{Y})_{c,k}F_{i,j}(\boldsymbol{Y})_{c^{\prime},k}, (7)

where 𝒥\mathcal{J} are the indices of selected feature maps. In practice, we choose 𝒥={(1,2),(2,2),(3,3),(4,3)}\mathcal{J}=\{(1,2),(2,2),(3,3),(4,3)\} in our experiments.

Conditional INS Stylization

Conditional encoding has been widely applied in convolutional networks [39, 80]. Similarly, we propose the conditional implicit representation by input with style conditioned embeddings using and extracting style-dependent features to render stylized color and density, which is shown in Figure 2. In training, we prepare nn style images with an nn dimensional one-hot style-condition vector. A mini-batch is constructed with the combinations of one content training patch and all candidate style images. The one-hot vector is fed into SIM to extract the ww-dimensional style features, they are then concatenated with the implicit representations output from CIM. The following layers of AM take the two features, aggregate them to render the pixel intensity and scene geometry along the rays. A pre-trained VGG [68] is appended on the top of the INS pipeline to apply implicit style and content constraints during training. During the inference stage, we discard the VGG network, the INS framework becomes a pure MLPs-based network.

Refer to caption
Figure 3: The proposed self-distilled geometry consistency. The CIM weights in the grey box are from a pre-trained NeRF and are kept fixed since then. During fine-tuning, the output density of the fixed CIM serves as geometry constrains for the stylized density from the output of AM.

4.2 Geometry Consistency for Neural Radiance Field

NeRF [53] learns reasonable 3D geometry inherently due to the particular design of implicit radiance field and the supervision from multiple views. However, the INS framework is expected to integrate style statistics from 2D image into 3D radiance field, where no multi-view style images accessible during training process. To specialize INS for neural radiance fields, we propose to regularize INS with proper geometry constraint to produce faithful shape and appearance. As the ground truth of target geometry is unavailable in most novel view synthesis benchmarks, we seek help from the self-distillation framework [79]. Concretely speaking, we first train the content implicit module (CIM) only to obtain a clean geometry σ1\sigma_{1}, following the vanilla NeRF training pipeline. After that, we copy the trained weight of CIM and keep that fixed (as shown in the grey block in Figure 3). In the next, we turn to optimize the whole INS framework (SIM, CIM and AM) with implicit stylization constrains. Meanwhile, a self-distilled geometry constraint between the original geometry σ1\sigma_{1} produced by the fixed CIM weight and the final stylized geometry σ2\sigma_{2} reconstructed by the implicit neural stylization framework. The objective of self-distilled geometry consistency loss is formulated as geo=|σ1σ2|\mathcal{L}_{\mathrm{geo}}=\lvert\sigma_{1}-\sigma_{2}\rvert, where we adopt the mean-absolute error for the densities of each sampled point.

Sampling-Stride Ray Sampling

Neural Radiance Field casts a number of rays (typically not adjacent) from camera origin, intersecting the pixel, into the volume and accumulating the color based on density along the ray.

While our model input with rays intersected with an image patch of size 𝒫K×K\mathcal{P}\in K\times K, predicting the stylized patch 𝒫K×K\mathcal{P^{\prime}}\in K\times K with its texture closed to the given style images. 2D style transfer methods [11, 37] typically crop the patch larger than 256×256256\times 256. However, it is too expensive for the neural radiance field as it queries the MLPs more than 256×256×N256\times 256\times N times of the MLP for each step [53], where NN indicates sampled points number along each ray.

Similar to [65, 48], we adopt a Sampling-Stride Ray Sampling strategy to enlarge the receptive field of the sampled patch to capture a more global context. The illustration of the ray sampling can be found in our supplementary materials, where a sampling stride larger than 1 result in a large receptive field while keeping computational cost fixed.

4.3 Optimization

Let 𝑭(𝑻)\boldsymbol{F}(\boldsymbol{T}) denote an INS model parameterized by 𝚯\boldsymbol{\Theta}, which synthesizes an image (patch) from the view specified by the camera pose 𝑻4×4\boldsymbol{T}\in{}^{4\times 4} by marching and rendering the rays for all the pixels on the image (patch). Given multi-view images and the corresponding camera parameters 𝒯={𝑪i,𝑻i}i=1N\mathcal{T}=\{\boldsymbol{C}_{i},\boldsymbol{T}_{i}\}_{i=1}^{N}, as well as a set of style images 𝒮={𝑺i}i=1M\mathcal{S}=\{\boldsymbol{S}_{i}\}_{i=1}^{M}, we train the INS using a combination of losses including reconstruction loss recon\mathcal{L}_{\mathrm{recon}}, geometry consistency loss geo\mathcal{L}_{\mathrm{geo}}, content loss content\mathcal{L}_{\mathrm{content}} and style loss style\mathcal{L}_{\mathrm{style}}:

total(𝚯|𝒯,𝒮,𝚯vgg)=𝔼(𝑪,𝑻)(𝒯)[λ0recon(𝑭(𝑻),𝑪)+λ1geo(𝑭(𝑻))]\displaystyle\mathcal{L}_{\mathrm{total}}(\boldsymbol{\Theta}|\mathcal{T},\mathcal{S},\boldsymbol{\Theta}_{\mathrm{vgg}})=\operatorname{\mathbb{E}}_{(\boldsymbol{C},\boldsymbol{T})\sim\operatorname{\mathbb{P}}(\mathcal{T})}\left[\lambda_{0}\mathcal{L}_{\mathrm{recon}}(\boldsymbol{F}(\boldsymbol{T}),\boldsymbol{C})+\lambda_{1}\mathcal{L}_{\mathrm{geo}}(\boldsymbol{F}(\boldsymbol{T}))\right] (8)
+𝔼𝑺(𝒮)𝔼(𝑪,𝑻)(𝒯)[λ2content(𝑭(𝑻),𝑪|𝚯vgg)+λ3style(𝑭(𝑻),𝑺|𝚯vgg)],\displaystyle+\operatorname{\mathbb{E}}_{\boldsymbol{S}\sim\operatorname{\mathbb{P}}(\mathcal{S})}\operatorname{\mathbb{E}}_{(\boldsymbol{C},\boldsymbol{T})\sim\operatorname{\mathbb{P}}(\mathcal{T})}\left[\lambda_{2}\mathcal{L}_{\mathrm{content}}(\boldsymbol{F}(\boldsymbol{T}),\boldsymbol{C}|\boldsymbol{\Theta}_{\mathrm{vgg}})+\lambda_{3}\mathcal{L}_{\mathrm{style}}(\boldsymbol{F}(\boldsymbol{T}),\boldsymbol{S}|\boldsymbol{\Theta}_{\mathrm{vgg}})\right],

where λ0\lambda_{0}, λ1\lambda_{1}, λ2\lambda_{2}, λ3\lambda_{3} control the strength of each loss term, 𝚯vgg\boldsymbol{\Theta}_{\mathrm{vgg}} denotes the parameters of the VGG network, and ()\operatorname{\mathbb{P}}(\cdot) defines a distribution over a support.

5 Experiments

5.1 Stylization on SIREN MLPs

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
SIREN SIREN+INS Style Image SIREN+INS Style Image
Figure 4: Visual examples generated by applying INS to SIREN [70]. Given an input image used for fitting and style images, SIREN+INS can express the style statistics via an implicit manner(i.e., MLPs).

As one representative example, we apply INS on fitting an image via SIREN [71] MLPs. We reuse the original SIREN framework [70] as CIM and follow its training recipes to fit images of 512×\times512 pixels. Besides that, we also incorporate the SIM and AM on SIREN. A pre-trained VGG-16 network is appended on the output to provide style and content supervisions during training. As seen in Figure 4, the proposed framework successfully in representing the images with the given style statistics in an implicit way.

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Style Image INS(view1) INS(view2) INS(view3)
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Style Image INS(view1) INS(view2) INS(view3)
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Style Image INS(view1) INS(view2) INS(view3)
Figure 5: Multi-view examples rendered by applying INS to the neural radiance field. The rendered scenes and objects show consistent stylized texture under different views.

5.2 Novel View Synthesis with NeRF

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Ours(view1)  [42](view1) Style3D [19](view1) Adain [37](view1)
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Style Image Ours(view2)  [42](view2) Style3D [19](view2) Adain [37](view2)
Figure 6: Qualitative comparisons. We compare INS with several methods on novel view synthesis datasets. Three scenes with different style images are demonstrated, we can see our proposed INS method preserves the learned geometry better (e.g. with clean background) and capture global context while achieving the desired styles.

Experimental Settings

We train our INS framework on NeRF-Synthetic [53] dataset and Local Light Field Fusion(LLFF) [52] dataset. NeRF-Synthetic consists of complex scenes with 360-degree views, where each scene has a central object with 100 inward-facing cameras distributed randomly on the upper hemisphere. Both rendered images and ground truth meshes are provided in NeRF-Synthetic dataset. LLFF dataset consists of forward-facing scenes, with fewer images. We implement INS on the same architecture and training strategy with the original NeRF [53]. λ1\lambda_{1}, λ2\lambda_{2} and λ3\lambda_{3} are set as zero in the first 150k iterations and then set as 1e6, 1 and 1e8 in the following 50k iterations. The self-distilled density supervision depicted in Figure 3 is generated from the CIM with 150k iterations pre-training. Adam optimizer is adopted with learning rates of 0.0005. Hyper-parameters are carefully tuned via grid searches and the best configuration is applied to all experiments. All experiments are trained on one NVIDIA RTX A6000 GPU. We retrain Style3D [19] in NeRF-Synthetic and LLFF datasets using their provided code and setting. We train all methods using the same number of style images for fair comparisons.

Results

In Figure 5, we can see INS generates faithful and view-consistent results for new viewpoints, with rich textures across scenes and styles. We further compare INS with three state-of-the art methods, including 3D neural stylization [19], image-based stylization methods [37, 42]. As is shown in Figure 6, We can see that stylizations from image-based methods produce noisy and view-inconsistent stylization as they transfer styles based on a single image. Style3D [19] generates blur results as it still relies on convolution networks (a.k.a. hypernetwork) to generate the MLP weights for the subsequent volume rendering. Our proposed implicit neural stylization method is trained to preserve correct scene geometry as well as capture global context, generating better view-consistent stylizations.

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Style Color w/ SS Color w/o SS Depth w/ SS Depth w/o SS
Figure 7: Qualitative evaluation on the effectiveness of Sampling Stride technique (SS). We can see INS with a larger receptive field generate more satisfying stylization.

5.3 Stylization on Signed Distance Function

Experimental Settings

DeepSDF [59] only learns the 3D geometry from given inputs. Later work IDR [83] extend it to reconstruct both 3D surface and appearance. We follow IDR [83] to implement the implicit neural stylization framework. To encode style statistic onto IDR, we project the learned textured SDF into multi-view images and implement our style loss on the rendered results. In the experiments, we picked 2 scenes from the DTU dataset [1], where each scene consists of 50 to 100 images and object masks captured from different angles. Similar to NeRF, we pre-train the IDR model for chosen scenes by minimizing the loss between the ground truth image and the rendered result. Then both the SDF network and rendering network are jointly optimized the proposed framework from projected views. Note that due to IDR’s architectural design, we are no longer able to impose self-distilled geometry consistency loss. Instead, we employ content/style loss in the masked region, which is similar to [83]. Besides, we observe that the SDF representation is more sensitive to parameter variations. To maintain intact geometries, we adjust the learning rate for the SDF network to 101110^{-11} times smaller than the rendering network.

Results

As are shown in Figure 8, the visualizations of two-view SDF representation demonstrate that both the learned appearance and geometry have deformed to fit the given style statistics.

Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Style Image SDF(view1) SDF(view2)
Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Style Image SDF(view1) SDF(view2)
Figure 8: Visualization results of applying INS to the Signed Distance Function. Given multi-view color images and style images, the proposed INS can learn the style statistics for the disentangled geometry and appearance.

5.4 Conditional Style Interpolation

Training with style-conditioned one-hot embedding, we can interpolate between style images to mix multiple styles with arbitrary weights. Specifically, we train INS on NeRF with two style images along with a two-dimensional one-hot vector as conditional code. After training, we mix the two style statistics by using a weighted two-dimensional vector. As is shown in Figure 9, the synthesized results can smoothly transfer from the first style to the second style when we linearly mix the two style embeddings at inference time.

Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image [1, 0, …, 0] [0.9, 0.1, …, 0] [0.8, 0.2, …, 0] [0.7, 0.3, …, 0] [0.6, 0.4, …, 0]
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Style Images [0.5, 0.5, …, 0] [0.4, 0.6, …, 0] [0.3, 0.7, …, 0] [0.2, 0.8, …, 0] [0, 1, …, 0]
Figure 9: Visualization of the mixture of two styles using different mixture weights. We can see the style smoothly transfers from one style to another.
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Color Image Color w/ GC Color w/ GC Depth w/ GC Depth w/ GC
Refer to caption Refer to caption Refer to caption Refer to caption Refer to caption
Style Image Color w/o GC Color w/o GC Depth w/o GC Depth w/o GC
Figure 10: Qualitative results of the self-distilled geometry consistency(“GC”). Both the rendered images and depth maps are shown for validation.

5.5 Ablation Study

Effect of Updating NeRF’s Geometry Field

The geometry modification in the density branch enables a more flexible stylization, by stylizing shape tweaks on the object surface. As shown in Figure. 11, only updating the color branch easily collapses to the low-level color transformation instead of the painterly texture transformation, violating our original goal. This phenomenon is also addressed in the previous 3D mesh stylization [43], where they explicitly model the 3D shape by deforming the mesh vertexes.

Refer to caption
Figure 11: Ablation study of optimizing style-conditioned density and color. By updating the density, it can render richer textures than only updating the color branch.

Effect of Self-distilled Geometry Consistency

To evaluate the effectiveness of the proposed geometry consistency regularizer, we visualize the front and back viewpoints of the synthesized color images and depth maps. As is shown in Figure 10, the proposed self-distilled geometry consistency learns a good trade-off between stylization and clean geometry.

Should INS Learned with Larger Receptive Field?

To investigate the effect by using the Sampling Stride (SS) Ray Sampling strategy, we conduct a comparison of ray sampling with and without sampling stride for NeRF stylization. For a fair comparison, we set the ray number as 64×\times64 in both settings. INS with sampling stride covers the content resolution of (64×\timess) ×\times (64×\timess) where ss indicates sampling strides depicted in Section 4.2 and here we set ss=4. Figure 7 shows that INS with SS achieves significantly better visual results, as it results in a higher receptive field in perceiving content statistics.

6 Conclusions

In this work, we present a Unified Implicit Neural Stylization framework (INS) to stylize complex 2D/3D scenes using implicit function. We conduct a pilot study on different types of implicit representations, including 2D coordinate-based mapping function, Neural Radiance Field, and Signed Distance Function. Comprehensive experiments demonstrate that the proposed method yields photo-realistic images/videos with visually consistent stylized textures. One limitation of our work lies in the training efficiency issue, similar to most implicit representation, rendering a style scene requires several hours of training, precluding on-device training. Addressing this issue could become a future direction.

References

  • [1] Aanæs, H., Jensen, R.R., Vogiatzis, G., Tola, E., Dahl, A.B.: Large-scale data for multiple-view stereopsis. International Journal of Computer Vision 120(2), 153–168 (2016)
  • [2] Aliev, K.A., Sevastopolsky, A., Kolos, M., Ulyanov, D., Lempitsky, V.: Neural point-based graphics. In: European Conference on Computer Vision. pp. 696–712. Springer (2020)
  • [3] Attal, B., Laidlaw, E., Gokaslan, A., Kim, C., Richardt, C., Tompkin, J., O’Toole, M.: Törf: Time-of-flight radiance fields for dynamic scene view synthesis. Advances in neural information processing systems 34 (2021)
  • [4] Atzmon, M., Haim, N., Yariv, L., Israelov, O., Maron, H., Lipman, Y.: Controlling neural level sets. Advances in Neural Information Processing Systems 32 (2019)
  • [5] Atzmon, M., Lipman, Y.: Sal: Sign agnostic learning of shapes from raw data. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2565–2574 (2020)
  • [6] Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 5855–5864 (2021)
  • [7] Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Mip-nerf 360: Unbounded anti-aliased neural radiance fields. arXiv preprint arXiv:2111.12077 (2021)
  • [8] Bergman, A., Kellnhofer, P., Wetzstein, G.: Fast training of neural lumigraph representations using meta learning. Advances in Neural Information Processing Systems 34 (2021)
  • [9] Chan, E.R., Lin, C.Z., Chan, M.A., Nagano, K., Pan, B., De Mello, S., Gallo, O., Guibas, L., Tremblay, J., Khamis, S., et al.: Efficient geometry-aware 3d generative adversarial networks. arXiv preprint arXiv:2112.07945 (2021)
  • [10] Chan, E.R., Monteiro, M., Kellnhofer, P., Wu, J., Wetzstein, G.: pi-gan: Periodic implicit generative adversarial networks for 3d-aware image synthesis. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 5799–5809 (2021)
  • [11] Chen, D., Liao, J., Yuan, L., Yu, N., Hua, G.: Coherent online video style transfer. In: Proceedings of the IEEE International Conference on Computer Vision. pp. 1105–1114 (2017)
  • [12] Chen, H., He, B., Wang, H., Ren, Y., Lim, S.N., Shrivastava, A.: Nerv: Neural representations for videos. Advances in Neural Information Processing Systems 34 (2021)
  • [13] Chen, T., Wang, P., Fan, Z., Wang, Z.: Aug-nerf: Training stronger neural radiance fields with triple-level physically-grounded augmentations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 15191–15202 (2022)
  • [14] Chen, X., Zhang, Y., Wang, Y., Shu, H., Xu, C., Xu, C.: Optical flow distillation: Towards efficient and stable video style transfer. In: European Conference on Computer Vision. pp. 614–630. Springer (2020)
  • [15] Chen, X., Zhang, Q., Li, X., Chen, Y., Feng, Y., Wang, X., Wang, J.: Hallucinated neural radiance fields in the wild. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 12943–12952 (2022)
  • [16] Chen, Y., Liu, S., Wang, X.: Learning continuous image representation with local implicit image function. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 8628–8638 (2021)
  • [17] Chen, Y., Liu, S., Wang, X.: Learning continuous image representation with local implicit image function. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 8628–8638 (2021)
  • [18] Chen, Z., Zhang, H.: Learning implicit fields for generative shape modeling. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 5939–5948 (2019)
  • [19] Chiang, P.Z., Tsai, M.S., Tseng, H.Y., Lai, W.S., Chiu, W.C.: Stylizing 3d scene via implicit representation and hypernetwork. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. pp. 1475–1484 (2022)
  • [20] Curless, B., Levoy, M.: A volumetric method for building complex models from range images. In: Proceedings of the 23rd annual conference on Computer graphics and interactive techniques. pp. 303–312 (1996)
  • [21] DeVries, T., Bautista, M.A., Srivastava, N., Taylor, G.W., Susskind, J.M.: Unconstrained scene generation with locally conditioned radiance fields. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 14304–14313 (2021)
  • [22] Drebin, R.A., Carpenter, L., Hanrahan, P.: Volume rendering. ACM Siggraph Computer Graphics 22(4), 65–74 (1988)
  • [23] Dupont, E., Goliński, A., Alizadeh, M., Teh, Y.W., Doucet, A.: Coin: Compression with implicit neural representations. arXiv preprint arXiv:2103.03123 (2021)
  • [24] Gao, C., Shih, Y., Lai, W.S., Liang, C.K., Huang, J.B.: Portrait neural radiance fields from a single image. arXiv preprint arXiv:2012.05903 (2020)
  • [25] Gatys, L., Ecker, A.S., Bethge, M.: Texture synthesis using convolutional neural networks. Advances in neural information processing systems 28 (2015)
  • [26] Gatys, L.A., Ecker, A.S., Bethge, M.: A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576 (2015)
  • [27] Genova, K., Cole, F., Vlasic, D., Sarna, A., Freeman, W.T., Funkhouser, T.: Learning shape templates with structured implicit functions. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 7154–7164 (2019)
  • [28] Gong, X., Huang, H., Ma, L., Shen, F., Liu, W., Zhang, T.: Neural stereoscopic image style transfer. In: Proceedings of the European Conference on Computer Vision (ECCV) (September 2018)
  • [29] Gropp, A., Yariv, L., Haim, N., Atzmon, M., Lipman, Y.: Implicit geometric regularization for learning shapes. arXiv preprint arXiv:2002.10099 (2020)
  • [30] Gu, J., Liu, L., Wang, P., Theobalt, C.: Stylenerf: A style-based 3d-aware generator for high-resolution image synthesis. arXiv preprint arXiv:2110.08985 (2021)
  • [31] Gu, X., Fan, Z., Zhu, S., Dai, Z., Tan, F., Tan, P.: Cascade cost volume for high-resolution multi-view stereo and stereo matching. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 2495–2504 (2020)
  • [32] Han, J., Jentzen, A., Weinan, E.: Solving high-dimensional partial differential equations using deep learning. Proceedings of the National Academy of Sciences 115(34), 8505–8510 (2018)
  • [33] Hao, Z., Mallya, A., Belongie, S., Liu, M.Y.: Gancraft: Unsupervised 3d neural rendering of minecraft worlds. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 14072–14082 (2021)
  • [34] Hedman, P., Srinivasan, P.P., Mildenhall, B., Barron, J.T., Debevec, P.: Baking neural radiance fields for real-time view synthesis. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 5875–5884 (2021)
  • [35] Höllein, L., Johnson, J., Nießner, M.: Stylemesh: Style transfer for indoor 3d scene reconstructions. arXiv preprint arXiv:2112.01530 (2021)
  • [36] Huang, H.P., Tseng, H.Y., Saini, S., Singh, M., Yang, M.H.: Learning to stylize novel views. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 13869–13878 (2021)
  • [37] Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE international conference on computer vision. pp. 1501–1510 (2017)
  • [38] Huang, Y.H., He, Y., Yuan, Y.J., Lai, Y.K., Gao, L.: Stylizednerf: consistent 3d scene stylization as stylized nerf via 2d-3d mutual learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 18342–18352 (2022)
  • [39] Iizuka, S., Simo-Serra, E., Ishikawa, H.: Let there be color! joint end-to-end learning of global and local image priors for automatic image colorization with simultaneous classification. ACM Transactions on Graphics (ToG) 35(4), 1–11 (2016)
  • [40] Jiang, C., Sud, A., Makadia, A., Huang, J., Nießner, M., Funkhouser, T., et al.: Local implicit grid representations for 3d scenes. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 6001–6010 (2020)
  • [41] Jiang, Y., Ji, D., Han, Z., Zwicker, M.: Sdfdiff: Differentiable rendering of signed distance fields for 3d shape optimization. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1251–1261 (2020)
  • [42] Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: European conference on computer vision. pp. 694–711. Springer (2016)
  • [43] Kato, H., Ushiku, Y., Harada, T.: Neural 3d mesh renderer. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 3907–3916 (2018)
  • [44] Kutulakos, K.N., Seitz, S.M.: A theory of shape by space carving. International journal of computer vision 38(3), 199–218 (2000)
  • [45] Li, Z., Kovachki, N., Azizzadenesheli, K., Liu, B., Bhattacharya, K., Stuart, A., Anandkumar, A.: Fourier neural operator for parametric partial differential equations. arXiv preprint arXiv:2010.08895 (2020)
  • [46] Lorraine, J., Duvenaud, D.: Stochastic hyperparameter optimization through hypernetworks. arXiv preprint arXiv:1802.09419 (2018)
  • [47] Martin-Brualla, R., Radwan, N., Sajjadi, M.S., Barron, J.T., Dosovitskiy, A., Duckworth, D.: Nerf in the wild: Neural radiance fields for unconstrained photo collections. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 7210–7219 (2021)
  • [48] Meng, Q., Chen, A., Luo, H., Wu, M., Su, H., Xu, L., He, X., Yu, J.: Gnerf: Gan-based neural radiance field without posed camera. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 6351–6361 (2021)
  • [49] Mescheder, L., Oechsle, M., Niemeyer, M., Nowozin, S., Geiger, A.: Occupancy networks: Learning 3d reconstruction in function space. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 4460–4470 (2019)
  • [50] Michalkiewicz, M., Pontes, J.K., Jack, D., Baktashmotlagh, M., Eriksson, A.: Implicit surface representations as layers in neural networks. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 4743–4752 (2019)
  • [51] Mildenhall, B., Hedman, P., Martin-Brualla, R., Srinivasan, P., Barron, J.T.: Nerf in the dark: High dynamic range view synthesis from noisy raw images. arXiv preprint arXiv:2111.13679 (2021)
  • [52] Mildenhall, B., Srinivasan, P.P., Ortiz-Cayon, R., Kalantari, N.K., Ramamoorthi, R., Ng, R., Kar, A.: Local light field fusion: Practical view synthesis with prescriptive sampling guidelines. ACM Transactions on Graphics (TOG) 38(4), 1–14 (2019)
  • [53] Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: Representing scenes as neural radiance fields for view synthesis. In: European conference on computer vision. pp. 405–421. Springer (2020)
  • [54] Mu, F., Wang, J., Wu, Y., Li, Y.: 3d photo stylization: Learning to generate stylized novel views from a single image. arXiv preprint arXiv:2112.00169 (2021)
  • [55] Niemeyer, M., Geiger, A.: Giraffe: Representing scenes as compositional generative neural feature fields. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 11453–11464 (2021)
  • [56] Niemeyer, M., Mescheder, L., Oechsle, M., Geiger, A.: Occupancy flow: 4d reconstruction by learning particle dynamics. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 5379–5389 (2019)
  • [57] Oechsle, M., Mescheder, L., Niemeyer, M., Strauss, T., Geiger, A.: Texture fields: Learning texture representations in function space. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 4531–4540 (2019)
  • [58] Park, J.J., Florence, P., Straub, J., Newcombe, R., Lovegrove, S.: Deepsdf: Learning continuous signed distance functions for shape representation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 165–174 (2019)
  • [59] Park, J.J., Florence, P., Straub, J., Newcombe, R., Lovegrove, S.: Deepsdf: Learning continuous signed distance functions for shape representation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 165–174 (2019)
  • [60] Park, K., Sinha, U., Barron, J.T., Bouaziz, S., Goldman, D.B., Seitz, S.M., Martin-Brualla, R.: Nerfies: Deformable neural radiance fields. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 5865–5874 (2021)
  • [61] Park, K., Sinha, U., Hedman, P., Barron, J.T., Bouaziz, S., Goldman, D.B., Martin-Brualla, R., Seitz, S.M.: Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields. arXiv preprint arXiv:2106.13228 (2021)
  • [62] Peng, S., Niemeyer, M., Mescheder, L., Pollefeys, M., Geiger, A.: Convolutional occupancy networks. In: European Conference on Computer Vision. pp. 523–540. Springer (2020)
  • [63] Ruder, M., Dosovitskiy, A., Brox, T.: Artistic style transfer for videos. In: German conference on pattern recognition. pp. 26–36. Springer (2016)
  • [64] Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 2304–2314 (2019)
  • [65] Schwarz, K., Liao, Y., Niemeyer, M., Geiger, A.: Graf: Generative radiance fields for 3d-aware image synthesis. Advances in Neural Information Processing Systems 33, 20154–20166 (2020)
  • [66] Shen, S., Wang, Z., Liu, P., Pan, Z., Li, R., Gao, T., Li, S., Yu, J.: Non-line-of-sight imaging via neural transient fields. IEEE Transactions on Pattern Analysis and Machine Intelligence (2021)
  • [67] Shrestha, R., Fan, Z., Su, Q., Dai, Z., Zhu, S., Tan, P.: Meshmvs: Multi-view stereo guided mesh reconstruction. In: 2021 International Conference on 3D Vision (3DV). pp. 1290–1300. IEEE (2021)
  • [68] Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
  • [69] Sitzmann, V., Chan, E., Tucker, R., Snavely, N., Wetzstein, G.: Metasdf: Meta-learning signed distance functions. Advances in Neural Information Processing Systems 33, 10136–10147 (2020)
  • [70] Sitzmann, V., Martel, J., Bergman, A., Lindell, D., Wetzstein, G.: Implicit neural representations with periodic activation functions. Advances in Neural Information Processing Systems 33, 7462–7473 (2020)
  • [71] Sitzmann, V., Martel, J., Bergman, A., Lindell, D., Wetzstein, G.: Implicit neural representations with periodic activation functions. Advances in Neural Information Processing Systems 33, 7462–7473 (2020)
  • [72] Sitzmann, V., Rezchikov, S., Freeman, W.T., Tenenbaum, J.B., Durand, F.: Light field networks: Neural scene representations with single-evaluation rendering. arXiv preprint arXiv:2106.02634 (2021)
  • [73] Sitzmann, V., Zollhöfer, M., Wetzstein, G.: Scene representation networks: Continuous 3d-structure-aware neural scene representations. Advances in Neural Information Processing Systems 32 (2019)
  • [74] Sun, Y., Liu, J., Xie, M., Wohlberg, B., Kamilov, U.S.: Coil: Coordinate-based internal learning for imaging inverse problems. arXiv preprint arXiv:2102.05181 (2021)
  • [75] Tancik, M., Srinivasan, P., Mildenhall, B., Fridovich-Keil, S., Raghavan, N., Singhal, U., Ramamoorthi, R., Barron, J., Ng, R.: Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems 33, 7537–7547 (2020)
  • [76] Ulyanov, D., Lebedev, V., Vedaldi, A., Lempitsky, V.S.: Texture networks: Feed-forward synthesis of textures and stylized images. In: ICML. vol. 1, p. 4 (2016)
  • [77] Wang, P., Fan, Z., Chen, T., Wang, Z.: Neural implicit dictionary learning via mixture-of-expert training. In: International Conference on Machine Learning. pp. 22613–22624. PMLR (2022)
  • [78] Xu, D., Jiang, Y., Wang, P., Fan, Z., Shi, H., Wang, Z.: Sinnerf: Training neural radiance fields on complex scenes from a single image. arXiv preprint arXiv:2204.00928 (2022)
  • [79] Xu, Y., Qiu, X., Zhou, L., Huang, X.: Improving bert fine-tuning via self-ensemble and self-distillation. arXiv preprint arXiv:2002.10345 (2020)
  • [80] Yanai, K., Tanno, R.: Conditional fast style transfer network. In: Proceedings of the 2017 ACM on International Conference on Multimedia Retrieval. pp. 434–437 (2017)
  • [81] Yao, Y., Luo, Z., Li, S., Fang, T., Quan, L.: Mvsnet: Depth inference for unstructured multi-view stereo. In: Proceedings of the European conference on computer vision (ECCV). pp. 767–783 (2018)
  • [82] Yariv, L., Kasten, Y., Moran, D., Galun, M., Atzmon, M., Ronen, B., Lipman, Y.: Multiview neural surface reconstruction by disentangling geometry and appearance. Advances in Neural Information Processing Systems 33, 2492–2502 (2020)
  • [83] Yariv, L., Kasten, Y., Moran, D., Galun, M., Atzmon, M., Ronen, B., Lipman, Y.: Multiview neural surface reconstruction by disentangling geometry and appearance. Advances in Neural Information Processing Systems 33, 2492–2502 (2020)
  • [84] Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelnerf: Neural radiance fields from one or few images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 4578–4587 (2021)
  • [85] Zeng, K., Zhao, M., Xiong, C., Zhu, S.C.: From image parsing to painterly rendering. ACM Trans. Graph. 29(1),  2–1 (2009)
  • [86] Zhang, K., Riegler, G., Snavely, N., Koltun, V.: Nerf++: Analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492 (2020)
  • [87] Zhang, Y., van Rozendaal, T., Brehmer, J., Nagel, M., Cohen, T.: Implicit neural video compression. arXiv preprint arXiv:2112.11312 (2021)
  • [88] Zhao, M., Zhu, S.C.: Customizing painterly rendering styles using stroke processes. In: Proceedings of the ACM SIGGRAPH/Eurographics Symposium on non-photorealistic animation and rendering. pp. 137–146 (2011)
  • [89] Zhong, E.D., Bepler, T., Berger, B., Davis, J.H.: Cryodrgn: reconstruction of heterogeneous cryo-em structures using neural networks. Nature Methods 18(2), 176–185 (2021)
  • [90] Zhou, P., Xie, L., Ni, B., Tian, Q.: Cips-3d: A 3d-aware generator of gans based on conditionally-independent pixel synthesis. arXiv preprint arXiv:2110.09788 (2021)