This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Neural Operator: Graph Kernel Network
for Partial Differential Equations

Zongyi Li, Nikola Kovachki, Kamyar Azizzadenesheli, Burigede Liu,
Kaushik Bhattacharya, Andrew Stuart, Anima Anandkumar
Abstract

The classical development of neural networks has been primarily for mappings between a finite-dimensional Euclidean space and a set of classes, or between two finite-dimensional Euclidean spaces. The purpose of this work is to generalize neural networks so that they can learn mappings between infinite-dimensional spaces (operators). The key innovation in our work is that a single set of network parameters, within a carefully designed network architecture, may be used to describe mappings between infinite-dimensional spaces and between different finite-dimensional approximations of those spaces. We formulate approximation of the infinite-dimensional mapping by composing nonlinear activation functions and a class of integral operators. The kernel integration is computed by message passing on graph networks. This approach has substantial practical consequences which we will illustrate in the context of mappings between input data to partial differential equations (PDEs) and their solutions. In this context, such learned networks can generalize among different approximation methods for the PDE (such as finite difference or finite element methods) and among approximations corresponding to different underlying levels of resolution and discretization. Experiments confirm that the proposed graph kernel network does have the desired properties and show competitive performance compared to the state of the art solvers.

1 INTRODUCTION

There are numerous applications in which it is desirable to learn a mapping between Banach spaces. In particular, either the input or the output space, or both, may be infinite-dimensional. The possibility of learning such mappings opens up a new class of problems in the design of neural networks, with widespread potential applicability. New ideas are required to build on traditional neural networks which are mappings from finite-dimensional Euclidean spaces into classes, or into another finite-dimensional Euclidean space. We study the development of neural networks in the setting in which the input and output spaces comprise real-valued functions defined on subsets in d.\mathbb{R}^{d}.

1.1 Literature Review And Context

We formulate a new class of neural networks, which are defined to map between spaces of functions on a bounded open set DD in d.\mathbb{R}^{d}. Such neural networks, once trained, have the important property that they are discretization invariant, sharing the same network parameters between different discretizations. In contrast, standard neural network architectures depend heavily on the discretization and have difficulty in generalizing between different grid representations. Our methodology has an underlying Nyström approximation formulation [Nyström, 1930] which links different grids to a single set of network parameters. We illustrate the new conceptual class of neural networks within the context of partial differential equations, and the mapping between input data (in the form of a function) and output data (the function which solves the PDE). Both supervised and semisupervised settings are considered.

In PDE applications, the defining equations are often local, whilst the solution operator has non-local effects which, nonetheless, decay. Such non-local effects can be described by integral operators with graph approximations of Nyström type [Belongie et al., 2002] providing a consistent way of connecting different grid or data structures arising in computational methods. For this reason, graph networks hold great potential for the solution operators of PDEs, which is the point of departure for our work.

Partial differential equations (PDEs).

A wide range of important engineering and physical problems are governed by PDEs. Over the past few decades, significant progress has been made on formulating [Gurtin, 1982] and solving [Johnson, 2012] the governing PDEs in many scientific fields from micro-scale problems (e.g., quantum and molecular dynamics) to macro-scale applications (e.g., civil and marine engineering). Despite the success in the application of PDEs to solve real-life problems, two significant challenges remain. First, identifying/formulating the underlying PDEs appropriate for the modeling of a specific problem usually requires extensive prior knowledge in the corresponding field which is then combined with universal conservation laws to design a predictive model; for example, modeling the deformation and fracture of solid structures requires detailed knowledge on the relationship between stress and strain in the constituent material. For complicated systems such as living cells, acquiring such knowledge is often elusive and formulating the governing PDE for these systems remains prohibitive; the possibility of learning such knowledge from data may revolutionize such fields. Second, solving complicated non-linear PDE systems (such as those arising in turbulence and plasticity) is computationally demanding; again the possibility of using instances of data from such computations to design fast approximate solvers holds great potential. In both these challenges, if neural networks are to play a role in exploiting the increasing volume of available data, then there is a need to formulate them so that they are well-adapted to mappings between function spaces.

We first outline two major neural network based approaches for PDEs. We consider PDEs of the form

(au)(x)=f(x),xDu(x)=0,xD,\displaystyle\begin{split}(\mathcal{L}_{a}u)(x)&=f(x),\qquad x\in D\ \\ u(x)&=0,\qquad\quad\>\>x\in\partial D,\end{split} (1)

with solution u:Du:D\to\mathbb{R}, and parameter a:Da:D\to\mathbb{R} entering the definition of a\mathcal{L}_{a}. The domain DD is discretized into KK points (see Section 2) and NN training pairs of coefficient functions and (approximate) solution functions {aj,uj}j=1N\{a_{j},u_{j}\}_{j=1}^{N} are used to design a neural network. The first approach parametrizes the solution operator as a deep convolutional neural network between finite Euclidean space :K×ΘK\mathcal{F}:\mathbb{R}^{K}\times\Theta\to\mathbb{R}^{K} [Guo et al., 2016, Zhu and Zabaras, 2018, Adler and Oktem, 2017, Bhatnagar et al., 2019]. Such an approach is, by definition, not mesh independent and will need modifications to the architecture for different resolution and discretization of KK in order to achieve consistent error (if at all possible). We demonstrate this issue in section 4 using the architecture of [Zhu and Zabaras, 2018] which was designed for the solution of (3) on a uniform 64×6464\times 64 mesh. Furthermore, this approach is limited to the discretization size and geometry of the training data hence it is not possible to query solutions at new points in the domain. In contrast we show, for our method, both invariance of the error to grid resolution, and the ability to transfer the solution between meshes in section 4.

The second approach directly parameterizes the solution uu as a neural network :D×Θ\mathcal{F}:D\times\Theta\to\mathbb{R} [E and Yu, 2018, Raissi et al., 2019, Bar and Sochen, 2019]. This approach is, of course, mesh independent since the solution is defined on the physical domain. However, the parametric dependence is accounted for in a mesh-dependent fashion. Indeed, for any given new equation with new coefficient function aa, one would need to train a new neural network a\mathcal{F}_{a}. Such an approach closely resembles classical methods such as finite elements, replacing the linear span of a finite set of local basis functions with the space of neural networks. This approach suffers from the same computational issue as the classical methods: one needs to solve an optimization problem for every new parameter. Furthermore, the approach is limited to a setting in which the underlying PDE is known; purely data-driven learning of a map between spaces of functions is not possible. The methodology we introduce circumvents these issues.

Our methodology most closely resembles the classical reduced basis method [DeVore, 2014] or the method of [Cohen and DeVore, 2015]. Along with the contemporaneous work [Bhattacharya et al., 2020], our method, to the best of our knowledge, is the first practical deep learning method that is designed to learn maps between infinite-dimensional spaces. It remedies the mesh-dependent nature of the approach in [Guo et al., 2016, Zhu and Zabaras, 2018, Adler and Oktem, 2017, Bhatnagar et al., 2019] by producing a quality of approximation that is invariant to the resolution of the function and it has the ability to transfer solutions between meshes. Moreover it need only be trained once on the equations set {aj,uj}j=1N\{a_{j},u_{j}\}_{j=1}^{N}; then, obtaining a solution for a new aμa\sim\mu, only requires a forward pass of the network, alleviating the major computational issues incurred in [E and Yu, 2018, Raissi et al., 2019, Herrmann et al., 2020, Bar and Sochen, 2019]. Lastly, our method requires no knowledge of the underlying PDE; the true map \mathcal{F}^{\dagger} can be treated as a black-box, perhaps trained on experimental data or on the output of a costly computer simulation, not necessarily a PDE.

Graph neural networks.

Graph neural network (GNNs), a class of neural networks that apply on graph-structured data, have recently been developed and seen a variety of applications. Graph networks incorporate an array of techniques such as graph convolution, edge convolution, attention, and graph pooling [Kipf and Welling, 2016, Hamilton et al., 2017, Gilmer et al., 2017, Veličković et al., 2017, Murphy et al., 2018]. GNNs have also been applied to the modeling of physical phenomena such as molecules [Chen et al., 2019] and rigid body systems [Battaglia et al., 2018], as these problems exhibit a natural graph interpretation: the particles are the nodes and the interactions are the edges.

The work [Alet et al., 2019] performed an initial study that employs graph networks on the problem of learning solutions to Poisson’s equation among other physical applications. They propose an encoder-decoder setting, constructing graphs in the latent space and utilizing message passing between the encoder and decoder. However, their model uses a nearest neighbor structure that is unable to capture non-local dependencies as the mesh size is increased. In contrast, we directly construct a graph in which the nodes are located on the spatial domain of the output function. Through message passing, we are then able to directly learn the kernel of the network which approximates the PDE solution. When querying a new location, we simply add a new node to our spatial graph and connect it to the existing nodes, avoiding interpolation error by leveraging the power of the Nyström extension for integral operators.

Refer to caption

Graph kernel network for the solution of (3). It can be trained on a small resolution and will generalize to a large one. Error is point-wise squared error.

Figure 1: Train on 16×1616\times 16, test on 241×241241\times 241

Continuous neural networks.

The concept of defining neural networks in infinite-dimensional spaces is a central problem that long been studied [Williams, 1996, Neal, 1996, Roux and Bengio, 2007, Globerson and Livni, 2016, Guss, 2016]. The general idea is to take the infinite-width limit which yields a non-parametric method and has connections to Gaussian Process Regression [Neal, 1996, Matthews et al., 2018, Garriga-Alonso et al., 2018], leading to the introduction of deep Gaussian processes [Damianou and Lawrence, 2013, Dunlop et al., 2018]. Thus far, such methods have not yielded efficient numerical algorithms that can parallel the success of convolutional or recurrent neural networks in finite dimensions in the setting of mappings between function spaces. Another idea is to simply define a sequence of compositions where each layer is a map between infinite-dimensional spaces with a finite-dimensional parametric dependence. This is the approach we take in this work, going a step further by sharing parameters between each layer.

1.2 Contributions

We introduce the concept of Neural Operator and instantiate it through graph kernel networks, a novel deep neural network method to learn the mapping between infinite-dimensional spaces of functions defined on bounded open subsets of d\mathbb{R}^{d}.

  • Unlike existing methods, our approach is demonstrably able to learn the mapping between function spaces, and is invariant to different approximations and grids, as demonstrated in Figure 1.

  • We deploy the Nyström extension to connect the neural network on function space to families of GNNs on arbitrary, possibly unstructured, grids.

  • We demonstrate that the Neural Operator approach has competitive approximation accuracy to classical and deep learning methods.

  • Numerical results also show that the Neural Operator needs only to be trained on a few samples in order to generalize to the whole class of problems.

  • We show the ability to perform semi-supervised learning, learning from data at only a few points and then generalizing to the whole domain.

These concepts are illustrated in the context of a family of elliptic PDEs prototypical of a number of problems arising throughout the sciences and engineering.

2 PROBLEM SETTING

Our goal is to learn a mapping between two infinite dimensional spaces by using a finite collection of observations of input-output pairs from this mapping: supervised learning. Let 𝒜\mathcal{A} and 𝒰\mathcal{U} be separable Banach spaces and :𝒜𝒰\mathcal{F}^{\dagger}:\mathcal{A}\to\mathcal{U} a (typically) non-linear map. Suppose we have observations {aj,uj}j=1N\{a_{j},u_{j}\}_{j=1}^{N} where ajμa_{j}\sim\mu is an i.i.d. sequence from the probability measure μ\mu supported on 𝒜\mathcal{A} and uj=(aj)u_{j}=\mathcal{F}^{\dagger}(a_{j}) is possibly corrupted with noise. We aim to build an approximation of \mathcal{F}^{\dagger} by constructing a parametric map

:𝒜×Θ𝒰\mathcal{F}:\mathcal{A}\times\Theta\to\mathcal{U} (2)

for some finite-dimensional parameter space Θ\Theta and then choosing θΘ\theta^{\dagger}\in\Theta so that (,θ)\mathcal{F}(\cdot,\theta^{\dagger})\approx\mathcal{F}^{\dagger}.

This is a natural framework for learning in infinite-dimensions as one could define a cost functional C:𝒰×𝒰C:\mathcal{U}\times\mathcal{U}\to\mathbb{R} and seek a minimizer of the problem

minθΘ𝔼aμ[C((a,θ),(a))]\min_{\theta\in\Theta}\mathbb{E}_{a\sim\mu}[C(\mathcal{F}(a,\theta),\mathcal{F}^{\dagger}(a))]

which directly parallels the classical finite-dimensional setting [Vapnik, 1998]. Showing the existence of minimizers, in the infinite-dimensional setting, remains a challenging open problem. We will approach this problem in the test-train setting in which empirical approximations to the cost are used. We conceptualize our methodology in the infinite-dimensional setting. This means that all finite-dimensional approximations can share a common set of network parameters which are defined in the (approximation-free) infinite-dimensional setting. To be concrete we will consider infinite-dimensional spaces which are Banach spaces of real-valued functions defined on a bounded open set in d\mathbb{R}^{d}. We then consider mappings \mathcal{F}^{\dagger} which take input functions to a PDE and map them to solutions of the PDE, both input and solutions being real-valued functions on d\mathbb{R}^{d}.

A common instantiation of the preceding problem is the approximation of the second order elliptic PDE

(a(x)u(x))=f(x),xDu(x)=0,xD\displaystyle\begin{split}-\nabla\cdot(a(x)\nabla u(x))&=f(x),\quad x\in D\\ u(x)&=0,\qquad\>\>x\in\partial D\end{split} (3)

for some bounded, open set DdD\subset\mathbb{R}^{d} and a fixed function fL2(D;)f\in L^{2}(D;\mathbb{R}). This equation is prototypical of PDEs arising in numerous applications including hydrology [Bear and Corapcioglu, 2012] and elasticity [Antman, 2005]. For a given a𝒜=L(D;+)L2(D;+)a\in\mathcal{A}=L^{\infty}(D;\mathbb{R}^{+})\cap L^{2}(D;\mathbb{R}^{+}), equation (3) has a unique weak solution u𝒰=H01(D;)u\in\mathcal{U}=H_{0}^{1}(D;\mathbb{R}) [Evans, 2010] and therefore we can define the solution operator \mathcal{F}^{\dagger} as the map aua\mapsto u. Note that while the PDE (3) is linear, the solution operator \mathcal{F}^{\dagger} is not.

Since our data aja_{j} and uju_{j} are , in general, functions, to work with them numerically, we assume access only to point-wise evaluations. To illustrate this, we will continue with the example of the preceding paragraph. To this end let PK={x1,,xK}DP_{K}=\{x_{1},\dots,x_{K}\}\subset D be a KK-point discretization of the domain DD and assume we have observations aj|PK,uj|PKKa_{j}|_{P_{K}},u_{j}|_{P_{K}}\in\mathbb{R}^{K}, for a finite collection of input-output pairs indexed by jj. In the next section, we propose a kernel inspired graph neural network architecture which, while trained on the discretized data, can produce an answer u(x)u(x) for any xDx\in D given a new input aμa\sim\mu. That is to say that our approach is independent of the discretization PKP_{K} and therefore a true function space method; we verify this claim numerically by showing invariance of the error as KK\to\infty. Such a property is highly desirable as it allows a transfer of solutions between different grid geometries and discretization sizes.

We note that, while the application of our methodology is based on having point-wise evaluations of the function, it is not limited by it. One may, for example, represent a function numerically as a finite set of truncated basis coefficients. Invariance of the representation would then be with respect to the size of this set. Our methodology can, in principle, be modified to accommodate this scenario through a suitably chosen architecture. We do not pursue this direction in the current work.

3 GRAPH KERNEL NETWORK

We propose a graph kernel neural network for the solution of the problem outlined in section 2. A table of notations is included in Appendix A.1. As a guiding principle for our architecture, we take the following example. Let a\mathcal{L}_{a} be a differential operator depending on a parameter a𝒜a\in\mathcal{A} and consider the PDE

(au)(x)=f(x),xDu(x)=0,xD\displaystyle\begin{split}(\mathcal{L}_{a}u)(x)&=f(x),\qquad x\in D\ \\ u(x)&=0,\qquad\quad\>\>x\in\partial D\end{split} (4)

for a bounded, open set DdD\subset\mathbb{R}^{d} and some fixed function ff living in an appropriate function space determined by the structure of a\mathcal{L}_{a}. The elliptic operator a=div(a)\mathcal{L}_{a}\cdot=-\text{div}(a\nabla\cdot) from equation (3) is an example. Under fairly general conditions on a\mathcal{L}_{a} [Evans, 2010], we may define the Green’s function G:D×DG:D\times D\to\mathbb{R} as the unique solution to the problem

aG(x,)=δx\mathcal{L}_{a}G(x,\cdot)=\delta_{x}

where δx\delta_{x} is the delta measure on d\mathbb{R}^{d} centered at xx. Note that GG will depend on the parameter aa thus we will henceforth denote it as GaG_{a}. The solution to (4) can then be represented as

u(x)=DGa(x,y)f(y)𝑑y.u(x)=\int_{D}G_{a}(x,y)f(y)\>dy. (5)

This is easily seen via the formal computation

(au)(x)\displaystyle(\mathcal{L}_{a}u)(x) =D(aG(x,))(y)f(y)𝑑y\displaystyle=\int_{D}(\mathcal{L}_{a}G(x,\cdot))(y)f(y)\>dy
=Dδx(y)f(y)𝑑y\displaystyle=\int_{D}\delta_{x}(y)f(y)\>dy
=f(x)\displaystyle=f(x)

Generally the Green’s function is continuous at points xyx\neq y, for example, when a\mathcal{L}_{a} is uniformly elliptic [Gilbarg and Trudinger, 2015], hence it is natural to model it via a neural network. Guided by the representation (5), we propose the following iterative architecture for t=0,,T1t=0,\dots,T-1.

vt+1(x)=σ(Wvt(x)+Dκϕ(x,y,a(x),a(y))vt(y)νx(dy))\displaystyle\begin{split}v_{t+1}(x)=\sigma\bigg{(}Wv_{t}(x)\!\!+\!\!\int_{D}\kappa_{\phi}(x,y,a(x),a(y))v_{t}(y)\>\nu_{x}(dy)\bigg{)}\!\!\!\!\end{split} (6)

where σ:\sigma:\mathbb{R}\to\mathbb{R} is a fixed function applied element-wise, νx\nu_{x} is a fixed Borel measure for each xDx\in D. The matrix Wn×nW\in\mathbb{R}^{n\times n}, together with the parameters ϕ\phi entering kernel κϕ:2(d+1)n×n\kappa_{\phi}:\mathbb{R}^{2(d+1)}\to\mathbb{R}^{n\times n}, are to be learned from data. We model κϕ\kappa_{\phi} as a neural network mapping 2(d+1)\mathbb{R}^{2(d+1)} to n×n.\mathbb{R}^{n\times n}.

Discretization of the continuum picture may be viewed as replacing Borel measure νx\nu_{x} by an empirical approximation based on the KK grid points being used. In this setting we may view κϕ\kappa_{\phi} as a K×KK\times K kernel block matrix, where each entry κϕ(x,y)\kappa_{\phi}(x,y) is itself a n×nn\times n matrix. Each block shares the same set of network parameters. This is the key to making a method that shares common parameters independent of the discretization used.

Finally, we observe that, although we have focussed on neural networks mapping aa to uu, generalizations are possible, such as mapping ff to uu, or having non-zero boundary data gg on D\partial D and mapping gg to uu. More generally one can consider the mapping from (a,f,g)(a,f,g) into uu and use similar ideas. Indeed to illustrate ideas we will consider the mapping from ff to uu below (which is linear and for which an analytic solution is known) before moving on to study the (nonlinear) mapping from aa to uu.

Example 1: Poisson equation.

We consider a simplification of the foregoing in which we study the map from ff to uu. To this end we set v0(x)=f(x)v_{0}(x)=f(x), T=1T=1, n=1n=1, σ(x)=x\sigma(x)=x, W=w=0W=w=0, and νx(dy)=dy\nu_{x}(dy)=dy (the Lebesgue measure) in (6). We then obtain the representation (5) with the Green’s function GaG_{a} parameterized by the neural network κϕ\kappa_{\phi} with explicit dependence on a(x)a(x), a(y)a(y). Now consider the setting where D=[0,1]D=[0,1] and a(x)1a(x)\equiv 1, so that (3) reduces to the 1-dimensional Poisson equation with explicitly computable Green’s function. Indeed,

G(x,y)=12(x+y|yx|)xy.G(x,y)=\frac{1}{2}\left(x+y-|y-x|\right)-xy.

Note that although the map fuf\mapsto u is, in function space, linear, the Green’s function itself is not linear in either argument. Figure 2 shows κϕ\kappa_{\phi} after training with N=2048N=2048 samples fjμ=𝒩(0,(Δ+I)1)f_{j}\sim\mu=\mathcal{N}(0,(-\Delta+I)^{-1}) with periodic boundary conditions on the operator Δ+I-\Delta+I. (samples from this measure can be easily implemented by means of a random Fourier series – Karhunen-Loeve – see [Lord et al., 2014]).

Refer to caption
Refer to caption

Proof of concept: graph kernel network on 11 dimensional Poisson equation; comparison of learned and truth kernel.

Figure 2: Kernel for one-dimensional Green’s function

Example 2: 2D Poisson equation.

We further demonstrate the power of the graph kernel network by extending the Poisson equation studied in example 1 to the two dimensional (2D) case, where we approximate the map f(x)u(x)f(x)\mapsto u(x) where xD=[0,1]×[0,1]x\in D=[0,1]\times[0,1]. We consider two approaches: (i)(i) graph kernel network to approximate the 2D Green’s function G(x,y)G(x,y) and (ii)(ii) dense neural network with f(x)f(x) as input and u(x)u(x) as output such that the mapping f(x)u(x)f(x)\mapsto u(x) is directly approximated.

The two neural networks are trained with the same training sets of different sizes ranging from 1 to 100 samples and tested with 1000 test samples. The relative l2l_{2} test errors as a function of training samples are illustrated in Figure 7. We observe that the Graph kernel network approximates the map with a minimum (5)(\approx 5) number of training samples while still possessing a smaller test error comparing to that of a dense neural network trained with 100 samples. Therefore, the Graph kernel network potentially significantly reduces the number of required training samples to approximate the mapping. This property is especially important in practice as obtaining a huge number of training data for certain engineering/physics problems is always prohibitive.

The reason for the Graph kernel network to have such strong approximation power is because that it is able to learn the truth Green’s function for the Poisson equation, as already demonstrated in the 1D case. We shall demonstrate further in Appendix A.4 that the Graph kernel networks are apt to capture the truth kernel for the 2D Poisson equation.

Refer to caption
Figure 3: Comparison between the performance of graph kernel network and the dense neural network on the approximation of (3). Plots of relative l2l_{2} test approximation errors versus the number of training samples for the Graph kernel network and dense neural networks when approximating the problem (3).

Algorithmic framework.

The initialization v0(x)v_{0}(x) to our network (6) can be viewed as the initial guess we make for the solution u(x)u(x) as well as any other dependence we want to make explicit. A natural choice is to start with the coefficient a(x)a(x) itself as well as the position in physical space xx. This (d+1)(d+1)-dimensional vector field is then lifted to a nn-dimensional vector field, an operation which we may view as the first layer of the overarching neural network. This is then used as an initialization to the kernel neural network, which is iterated TT times. In the final layer, we project back to the scalar field of interest with another neural network layer.

Due to the smoothing effect of the inverse elliptic operator in (3) with respect to the input data aa (and indeed ff when we consider this as input), we augment the initialization (x,a(x))(x,a(x)) with a Gaussian smoothed version of the coefficients aϵ(x)a_{\epsilon}(x), together with their gradient aϵ(x)\nabla a_{\epsilon}(x). Thus we initialize with a 2(d+1)2(d+1)-dimensional vector field. Throughout this paper the Gaussian smoothing is performed with a centred isotropic Gaussian with variance 5.5. The Borel measure νx\nu_{x} is chosen to be the Lebesgue measure supported on a ball at xx of radius rr. Thus we have

v0(x)=P(x,a(x),aϵ(x),aϵ(x))+p\displaystyle v_{0}(x)=P(x,a(x),a_{\epsilon}(x),\nabla a_{\epsilon}(x))+p (7)
vt+1(x)=σ(Wvt(x)+B(x,r)κϕ(x,y,a(x),a(y))vt(y)dy)\displaystyle\begin{split}v_{t+1}(x)=\sigma\Big{(}Wv_{t}(x)+\int_{B(x,r)}\!\!\!\kappa_{\phi}\big{(}x,y,a(x),a(y)\big{)}v_{t}(y)\>\mathrm{d}y\Big{)}\end{split} (8)
u(x)=QvT(x)+q\displaystyle u(x)=Qv_{T}(x)+q (9)

where Pn×2(d+1)P\in\mathbb{R}^{n\times 2(d+1)}, pnp\in\mathbb{R}^{n}, vt(x)nv_{t}(x)\in\mathbb{R}^{n} and Q1×nQ\in\mathbb{R}^{1\times n}, qq\in\mathbb{R}. The integration in (8) is approximated by a Monte Carlo sum via a message passing graph network with edge weights (x,y,a(x),a(y))(x,y,a(x),a(y)). The choice of measure νx(dy)=𝟙B(x,r)dy\nu_{x}(\mathrm{d}y)=\mathds{1}_{B(x,r)}\mathrm{d}y is two-fold: 1) it allows for more efficient computation and 2) it exploits the decay property of the Green’s function. Note that if more information is known about the true kernel, it can be added into this measure. For example, if we know the true kernel has a Gaussian structure, we can define νx(dy)=𝟙B(x,r)ρx(y)dy\nu_{x}(\mathrm{d}y)=\mathds{1}_{B(x,r)}\rho_{x}(y)\mathrm{d}y where ρx(y)\rho_{x}(y) is a Gaussian density. Then κϕ\kappa_{\phi} will need to learn a much less complicated function. We however do not pursue this direction in the current line of work.

Message passing graph networks.

Message passing graph networks comprise a standard architecture employing edge features [Gilmer et al., 2017]. If we properly construct the graph on the spatial domain DD of the PDE, the kernel integration can be viewed as an aggregation of messages. Given node features vt(x)nv_{t}(x)\in\mathbb{R}^{n}, edge features e(x,y)nee(x,y)\in\mathbb{R}^{n_{e}}, and a graph GG, the message passing neural network with averaging aggregation is

vt+1(x)=σ(Wvt(x)+1|N(x)|yN(x)κϕ(e(x,y))vt(y))v_{t+1}(x)=\sigma\Bigl{(}Wv_{t}(x)+\frac{1}{|N(x)|}\!\!\sum_{y\in N(x)}\!\!\!\kappa_{\phi}\big{(}e(x,y)\big{)}v_{t}(y)\Bigr{)}\!\! (10)

where Wn×nW\!\!\in\!\mathbb{R}^{n\times n}, N(x)N(x) is the neighborhood of xx according to the graph, κϕ(e(x,y))\kappa_{\phi}\big{(}e(x,y)\big{)} is a neural network taking as input edge features and as output a matrix in n×n\mathbb{R}^{n\times n}. Relating to (8), e(x,y)=(x,y,a(x),a(y))2(d+1)e(x,y)\!=\!(x,y,a(x),a(y))\!\in\!\mathbb{R}^{2(d+1)}.

Graph construction.

To use the message passing framework (10), we need to design a graph that connects the physical domain DD of the PDE. The nodes are chosen to be the KK discretized spatial locations. Here we work on a standard uniform mesh, but there are many other possibilities such as finite-element triangulations and random points at which data is acquired. The edge connectivity is then chosen according to the integration measure in (8), namely Lebesgue restricted to a ball. Each node xdx\in\mathbb{R}^{d} is connected to nodes which lie within B(x,r)B(x,r), defining the neighborhood set N(x)N(x). Then for each neighbor yN(x)y\in N(x), we assign the edge weight e(x,y)=(x,y,a(x),a(y))e(x,y)=(x,y,a(x),a(y)). Equation (10) can then be viewed as a Monte Carlo approximation of (8). This local structure allows for more efficient computation while remaining invariant to mesh-refinement. Indeed, since the radius parameter rr is chosen in physical space, the size of the set N(x)N(x) grows as the discretization size KK grows. This is a key feature that makes our methodology mesh-independent.

Nyström approximation of the kernel.

While the aforementioned graph structure severely reduces the computational overhead of integrating over the entire domain DD (corresponding to a fully-connected graph), the number of edges still scales like 𝒪(K2)\mathcal{O}(K^{2}). To overcome this, we employ a random Nyström-type approximation of the kernel. In particular, we uniformly sample mKm\ll K nodes from the original graph, constructing a new random sub-graph. This process is repeated ll\in\mathbb{N} times, yielding ll random sub-graphs each with mm nodes. This can be thought of as a way of reducing the variance in the estimator. We use these sub-graphs when evaluating (10) during training, leading to the more favorable scaling 𝒪(lm2)\mathcal{O}(lm^{2}). Indeed, numerically we find that l=4l=4 and m=200m=200 is sufficient even when K=4212=177,241K=421^{2}=177,241. In the evaluation phase, when we want the solution on a particular mesh geometry, we simply partition the mesh into sub-graphs each with mm nodes and evaluate each separately.

We will now highlight the quality of this kernel approximation in a RHKS setting. A real Reproducing Kernel Hilbert Space (RKHS) (,,,)(\mathcal{H},\langle\cdot,\cdot\rangle,\|\cdot\|) is a Hilbert space of functions f:Df:D\to\mathbb{R} where point-wise evaluation is a continuous linear functional, i.e. |f(x)|Cf|f(x)|\leq C\|f\| for some constant C0C\geq 0, independent of xx. For every RHKS, there exists a unique, symmetric, positive definite kernel κ:D×D\kappa:D\times D\to\mathbb{R}, which gives the representation f(x)=f,κ(,x)f(x)=\langle f,\kappa(\cdot,x)\rangle. Let T:T:\mathcal{H}\to\mathcal{H} be a linear operator on \mathcal{H} acting via the kernel

Tf=B(,r)κ(,y)f(y)ν(dy).Tf=\int_{B(\cdot,r)}\kappa(\cdot,y)f(y)\mathrm{\nu}(dy).

Let Tm:T_{m}:\mathcal{H}\to\mathcal{H} be its mm-point empirical approximation

Tm=B(,r)κ(,y)f(y)νm(dy),\displaystyle T_{m}=\int_{B(\cdot,r)}\kappa(\cdot,y)f(y)\mathrm{\nu}_{m}(dy),
νm(dy)=1mk=1mδyk(dy),\displaystyle\nu_{m}(dy)=\frac{1}{m}\sum_{k=1}^{m}\delta_{y_{k}}(dy),

so that

Tmf=1mk=1mκ(,yk)f(yk).T_{m}f=\frac{1}{m}\sum_{k=1}^{m}\kappa(\cdot,y_{k})f(y_{k}).

The error of this approximation achieves the Monte Carlo rate O(m1/2)O(m^{-1/2}):

Proposition 1.

Suppose 𝔼yν[κ(,y)4]<\mathbb{E}_{y\sim\nu}[\kappa(\cdot,y)^{4}]<\infty then there exists a constant C0C\geq 0 such that

𝔼TTmHSCm\mathbb{E}\|T-T_{m}\|_{HS}\leq\frac{C}{\sqrt{m}}

where HS\|\cdot\|_{HS} denotes the Hilbert-Schmidt norm on operators acting on \mathcal{H}.

For proof see Appendix A.2. With stricter assumptions similar results can also be proven with high probability [Rosasco et al., 2010].

Lemma 2.

(cite Belkin) TT and TKT_{K} are Hilbert-Schmidt. Furthermore, with probability greater than 12eτ1-2e^{\tau}

TTKHS22kτK\|T-T_{K}\|_{HS}\leq\frac{2\sqrt{2}k\sqrt{\tau}}{\sqrt{K}}

where k=supxDκ(x,x).k=\sup_{x\in D}{\kappa(x,x)}.

We note that, in our algorithm, κ:D×Dn×n\kappa:D\times D\to\mathbb{R}^{n\times n} whereas the preceding results are proven only in the setting n=1n=1; nonetheless they provide useful intuition regarding the approximations used in our methodology.

4 EXPERIMENTS

In this section we illustrate the claimed properties of our methodology, and compare it to existing approaches in the literature. All experimental results concern the mapping aua\mapsto u defined by (3) with D=[0,1]2D=[0,1]^{2}. Coefficients are generated according to aμa\sim\mu where μ=ψ#𝒩(0,(Δ+9I)2)\mu=\psi_{\#}\mathcal{N}(0,(-\Delta+9I)^{-2}) with a Neumann boundry condition on the operator Δ+9I-\Delta+9I. The mapping ψ:\psi:\mathbb{R}\to\mathbb{R} takes the value 12 on the positive part of the real line and 3 on the negative. Hence the coefficients are piecewise constant with random geometry and a fixed contrast of 4. Such constructions are prototypical of physical properties such as permeability in sub-surface flows and material microstructures in elasticity. Solutions uu are obtained by using a second-order finite difference scheme on a 241×241241\times 241 grid. Different resolutions are downsampled from this dataset.

To be concrete we set the dimension of representation nn (i.e. the width of graph network) to be 6464, the number of iterations TT to be 66, σ\sigma to be the ReLU function, and the inner kernel network κ\kappa to be a 33-layer feed-forward network with widths (6,512,1024,n2)(6,512,1024,n^{2}) and ReLU activation. We use the Adam optimizer with a learning rate 1e41e-4 and train for 200200 epochs with respect to the absolute mean squared error on the normalized data unless otherwise stated. These chosen hyperparameters are not optimized and could be adapted to improve performance. We employ the message passing network from the standard Pytorch graph network library Torch-geometric [Fey and Lenssen, 2019]. All the test errors are relative L2(D)L^{2}(D) errors on the original data. A table of notations for the hyper-parameters can be found in Appendix A.1. The code and data can be found at https://github.com/wumming/graph-pde.

4.1 Supervised Setting

First we consider the supervised scenario that we are given NN training pairs {aj,uj}1N\{a_{j},u_{j}\}_{1}^{N}, where each aja_{j} and uju_{j} are provided on a s×ss\times s grid (K=s2K=s^{2}).

Generalization of resolutions on full grids.

To examine the generalization property, we train the graph kernel network on resolution s×ss\times s and test on another resolution s×ss^{\prime}\times s^{\prime}. We fix the radius to be r=0.10r=0.10, train on N=100N=100 equation pairs and test on 4040 equation pairs.

Table 1: Comparing resolutions on full grids
Resolutions s=16s^{\prime}=16 s=31s^{\prime}=31 s=61s^{\prime}=61
s=16s=16 0.05250.0525 0.05910.0591 0.05850.0585
s=31s=31 0.07870.0787 0.05380.0538 0.05880.0588

r=0.10r=0.10, N=100N=100, relative l2l_{2} test error

As shown in Table 1, for each row, the test errors at different resolutions remain on the same scale, illustrating the desired design feature that graph kernel networks can train on one resolution and generalize to another resolution. The test errors on the diagonal (s=s=16s=s^{\prime}=16 and s=s=31s=s^{\prime}=31) are the smallest, which means the network has the best performance when the training grid and the test grid are the same. Interestingly, for the second row, when training on s=31s=31, it is easier to general to s=61s^{\prime}=61 than to s=16s^{\prime}=16. This is because when generalizing to a larger grid, the support of the kernel becomes large which does not hurt the performance. But when generalizing to a smaller grid, part of the support of the kernel is lost, which causes the kernel to be inaccurate.

Expressiveness and overfitting.

We compare the training error and test error with a different number of training pairs NN to see if the kernel network can learn the kernel structure even with a small amount of data. We study the expressiveness of the kernel network, examining how it overfits. We fix r=0.10r=0.10 on the s=s=31s=s^{\prime}=31 grid and train with N=10,100,1000N=10,100,1000 whilst employing 20002000, 500500, 100100 epochs respectively.

Table 2: Comparing number of training pairs
Training Size Training Error Test Error
N=10N=10 0.00890.0089 0.09310.0931
N=100N=100 0.01830.0183 0.04780.0478
N=1000N=1000 0.02550.0255 0.03450.0345

20002000, 500500, 100100 epochs respectively.

We see from Table 2 that the kernel network already achieves a reasonable result when N=10N=10, and the accuracy is competitive when N=100N=100. In all three cases, the test error is larger than the training error suggesting that the kernel network has enough expressiveness to overfit the training set. This overfitting is not severe as the training error will not be pushed to zero even for N=10N=10, after 20002000 epochs.

4.2 Semi-Supervised Setting

In the semi-supervised setting, we are only given mm nodes sampled from a s×ss\times s grid for each training pair, and want to evaluate on mm^{\prime} nodes sampled from an s×ss^{\prime}\times s^{\prime} grid for each test pair. To be concrete, we set the number of sampled nodes m=m=200m=m^{\prime}=200. For each training pair, we sample twice l=2l=2; for each test pair, we sample once l=1l^{\prime}=1. We train on N=100N=100 pairs and test on N=100N^{\prime}=100 pairs. The radius for both training and testing is set to r=r=0.25r=r^{\prime}=0.25.

Generalization of resolutions on sampled grids.

Similar to the first experiments, we train the graph kernel network with nodes sampled from the s×ss\times s resolution and test on nodes sampled from the s×ss^{\prime}\times s^{\prime} resolution.

Table 3: Generalization of resolutions on sampled Grids
Resolutions s=61s^{\prime}=61 s=121s^{\prime}=121 s=241s^{\prime}=241
s=16s=16 0.07170.0717 0.07680.0768 0.07240.0724
s=31s=31 0.07260.0726 0.07100.0710 0.07220.0722
s=61s=61 0.06870.0687 0.07280.0728 0.07230.0723
s=121s=121 0.06870.0687 0.06640.0664 0.06850.0685
s=241s=241 0.06490.0649 0.06580.0658 0.06490.0649

N=100N=100, m=m=200m=m^{\prime}=200, r=r=0.25r=r^{\prime}=0.25, l=2l=2

As shown in Table 3, for each row, the test errors on different resolutions are about the same, which means the graph kernel network can also generalize in the semi-supervised setting. Comparing the rows, large training resolutions ss tend to have a smaller error. When sampled from a finer grid, there are more edges because the support of the kernel is larger on the finer grid. Still, the performance is best when s=ss=s^{\prime}.

The number of examples vs the times of sampling.

Increasing the number of times we sample, ll, reduces the error from the Nyström approximation. By comparing different ll we determine which value will be sufficient. When we sample ll times for each equation, we get a total of NlNl sampled training pairs, Table 4.

Table 4: Number of training pairs and sampling
l=1l=1 l=2l=2 l=4l=4 l=8l=8
N=10N=10 0.12590.1259 0.10690.1069 0.09670.0967 0.10260.1026
N=100N=100 0.07860.0786 0.06870.0687 0.06900.0690 0.06210.0621
N=1000N=1000 0.06040.0604 0.05790.0579 0.05400.0540 0.04830.0483

s=241s=241, m=m=200m=m^{\prime}=200, r=r=0.25r=r^{\prime}=0.25

Table 4 indicates that the larger ll the better, but l=2l=2 already gives a reasonable performance. Moreover, the same order of sampled training pairs, (N=100,l=8)(N=100,l=8), and (N=1000,l=1)(N=1000,l=1), result in a similar performance. It implies that in a low training data regime, increasing ll improves the performance.

Different number of nodes in training and testing.

To further examine the Nyström approximation, we compare different numbers of node samples m,mm,m^{\prime} for both training and testing.

Table 5: Number of nodes in the training and testing
m=100m^{\prime}=100 m=200m^{\prime}=200 m=400m^{\prime}=400 m=800m^{\prime}=800
m=100m=100 0.08710.0871 0.07160.0716 0.06620.0662 0.06090.0609
m=200m=200 0.09720.0972 0.07340.0734 0.06060.0606 0.05620.0562
m=400m=400 0.09910.0991 0.06990.0699 0.05600.0560 0.05060.0506
m=800m=800 0.10840.1084 0.07510.0751 0.05730.0573 0.04780.0478

s=121,r=r=0.15,l=5s=121,r=r^{\prime}=0.15,l=5

As can be seen from Table 5, in general the larger mm and mm^{\prime} the better. For each row, fixing mm, the larger mm^{\prime} the better. But for each column, when fixing mm^{\prime}, increasing mm may not lead to better performance. This is again due to the fact that when learning on a larger grid, the kernel network learns a kernel with larger support. When evaluating on a smaller grid, the learned kernel will be truncated to have small support, leading to an increased error. In general, m=mm=m^{\prime} will be the best choice.

The radius and the number of nodes.

The computation and storage of graph networks directly scale with the number of edges. In this experiment, we want to study the trade-off between the number of nodes mm and the radius rr when fixing the number of edges.

Table 6: The radius and the number of nodes
m=100m=100 m=200m=200 m=400m=400
r=0.05r=0.05 0.110(176)0.110(176) 0.109(666)0.109(666) 0.099(3354)0.099(3354)
r=0.15r=0.15 0.086(512)0.086(512) 0.070(2770)0.070(2770) 0.053(14086)0.053(14086)
r=0.40r=0.40 0.064(1596)0.064(1596) 0.051(9728)0.051(9728) 0.040(55919)0.040(55919)
r=1.00r=1.00 0.059(9756)0.059(9756) 0.048(38690)0.048(38690) -

Error (Edges), s=121s=121, l=5l=5, m=mm^{\prime}=m

Table 6 shows the test error with the number of edges for different rr and mm. In general, the more edges, the better. For a fixed number of edges, the performance depends more on the radius rr than on the number of nodes mm. In other words, the error of truncating the kernel locally is larger than the error from the Nyström approximation. It would be better to use larger rr with smaller mm.

Inner Kernel Network κ\kappa.

To find the best network structure of κ\kappa, we compare different combinations of width and depth. We consider three cases of κ\kappa: 1. a 22-layer feed-forward network with widths (6,width,n2)(6,width,n^{2}), 2. a 33-layer feed-forward network with widths (6,width/2,width,n2)(6,width/2,width,n^{2}), and 3. a 55-layer feed-forward network with widths (6,width/4,width/2,width,width,n2)(6,width/4,width/2,width,width,n^{2}), all with ReLU activation and learning rate 1e41e-4.

Table 7: Comparing the width and depth for the inner kernel network κ\kappa
depth =2=2 depth =3=3 depth =5=5
width = 6464 0.06850.0685 0.06950.0695 0.07700.0770
width = 128128 0.06300.0630 0.06330.0633 0.07020.0702
width = 256256 0.06170.0617 0.06100.0610 0.06880.0688
width = 10241024 0.06410.0641 0.05910.0591 0.06080.0608
width = 40964096 0.29340.2934 0.06900.0690 0.06380.0638

s=241s=241, m=200m=200, r=0.25r=0.25

As shown in Table 7, have wider and deeper network increase the expensiveness of the kernel network. The diagonal combinations width=256,depth=2width=256,depth=2, width=1024,depth=3width=1024,depth=3, and width=4096,depth=5width=4096,depth=5 have better test error. Notice the wide but shallow network width=4096,depth=2width=4096,depth=2 has very bad performance. Both its training and testing error once decreased to 0.090.09 around 1010th epochs, but then blow off. The 1e41e-4 learning rate is probably too high for this combination. In general, depth of 33 with with of 256256 is a good combination of Dracy Equation dataset.

4.3 Comparison with Different Benchmarks

In the following section, we compare the Graph Kernel Network with different benchmarks on a larger dataset of N=1024N=1024 training pairs computed on a 421×421421\times 421 grid. The network is trained and evaluated on the same full grid. The results are presented in Table 8.

  • NN is a simple point-wise feedforward neural network. It is mesh-free, but performs badly due to lack of neighbor information.

  • FCN is the state of the art neural network method based on Fully Convolution Network [Zhu and Zabaras, 2018]. It has a dominating performance for small grids s=61s=61. But fully convolution networks are mesh-dependent and therefore their error grows when moving to a larger grid.

  • PCA+NN is an instantiation of the methodology proposed in [Bhattacharya et al., 2020]: using PCA as an autoencoder on both the input and output data and interpolating the latent spaces with a neural network. The method provably obtains mesh-independent error and can learn purely from data, however the solution can only be evaluated on the same mesh as the training data.

  • RBM is the classical Reduced Basis Method (using a PCA basis), which is widely used in applications and provably obtains mesh-independent error [DeVore, 2014]. It has the best performance but the solutions can only be evaluated on the same mesh as the training data and one needs knowledge of the PDE to employ it.

  • GKN stands for our graph kernel network with r=0.25r=0.25 and m=300m=300. It enjoys competitive performance against all other methods while being able to generalize to different mesh geometries. Some figures arising from GKN are included in Appendix A.3.

Table 8: Error of different methods
Networks s=85s=85 s=141s=141 s=211s=211 s=421s=421
NN 0.17160.1716 0.17160.1716 0.17160.1716 0.17160.1716
FCN 0.02530.0253 0.04930.0493 0.07270.0727 0.10970.1097
PCA+NN 0.02990.0299 0.02980.0298 0.02980.0298 0.02990.0299
RBM 0.02440.0244 0.02510.0251 0.02550.0255 0.02590.0259
GKN 0.03460.0346 0.03320.0332 0.03420.0342 0.03690.0369

5 DISCUSSION AND FUTURE WORK

We have introduced the concept of Neural Operator and instantiated it through graph kernel networks designed to approximate mappings between function spaces. They are constructed to be mesh-free and our numerical experiments demonstrate that they have the desired property of being able to train and generalize on different meshes. This is because the networks learn the mapping between infinite-dimensional function spaces, which can then be shared with approximations at various levels of discretization. A further advantage is that data may be incorporated on unstructured grids, using the Nyström approximation. We demonstrate that our method can achieve competitive performance with other mesh-free approaches developed in the numerical analysis community, and beats state-of-the-art neural network approaches on large grids, which are mesh-dependent. The methods developed in the numerical analysis community are less flexible than the approach we introduce here, relying heavily on the variational structure of divergence form elliptic PDEs. Our new mesh-free method has many applications. It has the potential to be a faster solver that learns from only sparse observations in physical space. It is the only method that can work in the semi-supervised scenario when we only have measurements on some parts of the grid. It is also the only method that can transfer between different geometries. For example, when computing the flow dynamic of many different airfoils, we can construct different graphs and train together. When learning from irregular grids and querying new locations, our method does not require any interpolation, avoid subsequently interpolation error.

Disadvantage.

Graph kernel network’s runtime and storage scale with the number of edges E=O(K2)E=O(K^{2}). While other mesh-dependent methods such as PCA+NN and RBM require only O(K)O(K). This is somewhat inevitable, because to learn the continuous function or the kernel, we need to capture pairwise information between every two nodes, which is O(K2)O(K^{2}); when the discretization is fixed, one just needs to capture the point-wise information, which is O(K)O(K). Therefore training and evaluating the whole grid is costly when the grid is large. On the other hand, subsampling to ameliorate cost loses information in the data, and causes errors which make our method less competitive than PCA+NN and RBM.

Future works.

To deal with the above problem, we propose that ideas such as multi-grid and fast multipole methods [Gholami et al., 2016] may be combined with our approach to reduce complexity. In particular, a multi-grid approach will construct multi-graphs corresponding to different resolutions so that, within each graph, nodes only connect to their nearest neighbors. The number of edges then scale as O(K)O(K) instead of O(K2)O(K^{2}). The error terms from Nyström approximation and local truncation can be avoided. Another direction is to extend the framework for time-dependent PDEs. Since the graph kernel network is itself an iterative solver with the time step tt, it is natural to frame it as an RNN that each time step corresponds to a time step of the PDEs.

Acknowledgements

Z. Li gratefully acknowledges the financial support from the Kortschak Scholars Program. K. Azizzadenesheli is supported in part by Raytheon and Amazon Web Service. A. Anandkumar is supported in part by Bren endowed chair, DARPA PAIHR00111890035, LwLL grants, Raytheon, Microsoft, Google, Adobe faculty fellowships, and DE Logi grant. K. Bhattacharya, N. B. Kovachki, B. Liu and A. M. Stuart gratefully acknowledge the financial support of the Amy research Laboratory through the Cooperative Agreement Number W911NF-12-0022. Research was sponsored by the Army Research Laboratory and was accomplished under Cooperative Agreement Number W911NF-12-2-0022. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the Army Research Laboratory or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright notation herein.

References

  • [Adler and Oktem, 2017] Adler, J. and Oktem, O. (2017). Solving ill-posed inverse problems using iterative deep neural networks. Inverse Problems.
  • [Alet et al., 2019] Alet, F., Jeewajee, A. K., Villalonga, M. B., Rodriguez, A., Lozano-Perez, T., and Kaelbling, L. (2019). Graph element networks: adaptive, structured computation and memory. In 36th International Conference on Machine Learning. PMLR.
  • [Antman, 2005] Antman, S. S. (2005). Problems In Nonlinear Elasticity. Springer.
  • [Bar and Sochen, 2019] Bar, L. and Sochen, N. (2019). Unsupervised deep learning algorithm for pde-based forward and inverse problems. arXiv preprint arXiv:1904.05417.
  • [Battaglia et al., 2018] Battaglia, P. W., Hamrick, J. B., Bapst, V., Sanchez-Gonzalez, A., Zambaldi, V., Malinowski, M., Tacchetti, A., Raposo, D., Santoro, A., Faulkner, R., et al. (2018). Relational inductive biases, deep learning, and graph networks. arXiv preprint arXiv:1806.01261.
  • [Bear and Corapcioglu, 2012] Bear, J. and Corapcioglu, M. Y. (2012). Fundamentals of transport phenomena in porous media. Springer Science & Business Media.
  • [Belongie et al., 2002] Belongie, S., Fowlkes, C., Chung, F., and Malik, J. (2002). Spectral partitioning with indefinite kernels using the nyström extension. In European conference on computer vision. Springer.
  • [Bhatnagar et al., 2019] Bhatnagar, S., Afshar, Y., Pan, S., Duraisamy, K., and Kaushik, S. (2019). Prediction of aerodynamic flow fields using convolutional neural networks. Computational Mechanics, pages 1–21.
  • [Bhattacharya et al., 2020] Bhattacharya, K., Kovachki, N. B., and Stuart, A. M. (2020). Model reduction and neural networks for parametric pde(s). preprint.
  • [Chen et al., 2019] Chen, C., Ye, W., Zuo, Y., Zheng, C., and Ong, S. P. (2019). Graph networks as a universal machine learning framework for molecules and crystals. Chemistry of Materials, 31(9):3564–3572.
  • [Cohen and DeVore, 2015] Cohen, A. and DeVore, R. (2015). Approximation of high-dimensional parametric pdes. Acta Numerica.
  • [Damianou and Lawrence, 2013] Damianou, A. and Lawrence, N. (2013). Deep gaussian processes. In Artificial Intelligence and Statistics, pages 207–215.
  • [DeVore, 2014] DeVore, R. A. (2014). Chapter 3: The Theoretical Foundation of Reduced Basis Methods.
  • [Dunlop et al., 2018] Dunlop, M. M., Girolami, M. A., Stuart, A. M., and Teckentrup, A. L. (2018). How deep are deep gaussian processes? The Journal of Machine Learning Research, 19(1):2100–2145.
  • [E and Yu, 2018] E, W. and Yu, B. (2018). The deep ritz method: A deep learning-based numerical algorithm for solving variational problems. Communications in Mathematics and Statistics.
  • [Evans, 2010] Evans, L. C. (2010). Partial Differential Equations, volume 19. American Mathematical Soc.
  • [Fey and Lenssen, 2019] Fey, M. and Lenssen, J. E. (2019). Fast graph representation learning with PyTorch Geometric. In ICLR Workshop on Representation Learning.
  • [Garriga-Alonso et al., 2018] Garriga-Alonso, A., Rasmussen, C. E., and Aitchison, L. (2018). Deep Convolutional Networks as shallow Gaussian Processes. arXiv e-prints, page arXiv:1808.05587.
  • [Gholami et al., 2016] Gholami, A., Malhotra, D., Sundar, H., and Biros, G. (2016). Fft, fmm, or multigrid? a comparative study of state-of-the-art poisson solvers for uniform and nonuniform grids in the unit cube. SIAM Journal on Scientific Computing.
  • [Gilbarg and Trudinger, 2015] Gilbarg, D. and Trudinger, N. S. (2015). Elliptic partial differential equations of second order. springer.
  • [Gilmer et al., 2017] Gilmer, J., Schoenholz, S. S., Riley, P. F., Vinyals, O., and Dahl, G. E. (2017). Neural message passing for quantum chemistry. In Proceedings of the 34th International Conference on Machine Learning.
  • [Globerson and Livni, 2016] Globerson, A. and Livni, R. (2016). Learning infinite-layer networks: Beyond the kernel trick. CoRR, abs/1606.05316.
  • [Guo et al., 2016] Guo, X., Li, W., and Iorio, F. (2016). Convolutional neural networks for steady flow approximation. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.
  • [Gurtin, 1982] Gurtin, M. E. (1982). An introduction to continuum mechanics. Academic press.
  • [Guss, 2016] Guss, W. H. (2016). Deep Function Machines: Generalized Neural Networks for Topological Layer Expression. arXiv e-prints, page arXiv:1612.04799.
  • [Hamilton et al., 2017] Hamilton, W., Ying, Z., and Leskovec, J. (2017). Inductive representation learning on large graphs. In Advances in neural information processing systems, pages 1024–1034.
  • [Herrmann et al., 2020] Herrmann, L., Schwab, C., and Zech, J. (2020). Deep relu neural network expression rates for data-to-qoi maps in bayesian pde inversion.
  • [Johnson, 2012] Johnson, C. (2012). Numerical solution of partial differential equations by the finite element method. Courier Corporation.
  • [Kipf and Welling, 2016] Kipf, T. N. and Welling, M. (2016). Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907.
  • [Lord et al., 2014] Lord, G. J., Powell, C. E., and Shardlow, T. (2014). An introduction to computational stochastic PDEs, volume 50. Cambridge University Press.
  • [Matthews et al., 2018] Matthews, A. G. d. G., Rowland, M., Hron, J., Turner, R. E., and Ghahramani, Z. (2018). Gaussian Process Behaviour in Wide Deep Neural Networks.
  • [Murphy et al., 2018] Murphy, R. L., Srinivasan, B., Rao, V., and Ribeiro, B. (2018). Janossy pooling: Learning deep permutation-invariant functions for variable-size inputs. arXiv preprint arXiv:1811.01900.
  • [Neal, 1996] Neal, R. M. (1996). Bayesian Learning for Neural Networks. Springer-Verlag.
  • [Nyström, 1930] Nyström, E. J. (1930). Über die praktische auflösung von integralgleichungen mit anwendungen auf randwertaufgaben. Acta Mathematica.
  • [Raissi et al., 2019] Raissi, M., Perdikaris, P., and Karniadakis, G. E. (2019). Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational Physics, 378:686–707.
  • [Rosasco et al., 2010] Rosasco, L., Belkin, M., and Vito, E. D. (2010). On learning with integral operators. J. Mach. Learn. Res., 11:905–934.
  • [Roux and Bengio, 2007] Roux, N. L. and Bengio, Y. (2007). Continuous neural networks. In Meila, M. and Shen, X., editors, Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics.
  • [Vapnik, 1998] Vapnik, V. N. (1998). Statistical Learning Theory. Wiley-Interscience.
  • [Veličković et al., 2017] Veličković, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., and Bengio, Y. (2017). Graph attention networks.
  • [Williams, 1996] Williams, C. K. I. (1996). Computing with infinite networks. In Proceedings of the 9th International Conference on Neural Information Processing Systems, Cambridge, MA, USA. MIT Press.
  • [Zhu and Zabaras, 2018] Zhu, Y. and Zabaras, N. (2018). Bayesian deep convolutional encoder–decoder networks for surrogate modeling and uncertainty quantification. Journal of Computational Physics.

Appendix A Appendix

A.1 Table of Notations

Table 9: Table of notations
Notation Meaning
PDE
a𝒜a\in\mathcal{A} The input coefficient functions
u𝒰u\in\mathcal{U} The target solution functions
DdD\subset\mathbb{R}^{d} The spatial domain for the PDE
xDx\in D Points in the the spatial domain
PKP_{K} KK-point discretization of DD
:𝒜𝒰\mathcal{F}:\mathcal{A}\to\mathcal{U} The operator mapping the coefficients to the solutions
μ\mu A probability measure where aja_{j} sampled from.
Graph Kernel Networks
κ:2(d+1)n×n\kappa:\mathbb{R}^{2(d+1)}\to\mathbb{R}^{n\times n} The kernel maps (x,y,a(x),a(y))(x,y,a(x),a(y)) to a n×nn\times n matrix
ϕ\phi The parameters of the kernel network κ\kappa
t=0,,Tt=0,\ldots,T The time steps
σ\sigma The activation function
v(x)nv(x)\in\mathbb{R}^{n} The neural network representation of u(x)u(x)
ν\nu A Borel measure for xx
aϵa_{\epsilon} Gaussian smoothing of aa
aϵ\nabla a_{\epsilon} The gradients of aϵa_{\epsilon}
W,P,Q,p,qW,P,Q,p,q Learnable parameters of the networks.
Hyperparameters
NN The number of training pairs
NN^{\prime} The number of testing pairs
ss The underlying resolution of training points
ss^{\prime} The underlying resolution of testing points
K=s2K=s^{2} Total number of nodes in the grid
mm The number of sampled nodes per training pair
mm^{\prime} The number of sampled nodes per testing pair
ll The times of re-sampling per training pair
ll^{\prime} The times of re-sampling per testing pair
rr The radius of the ball of kernel integration for training
rr^{\prime} The radius of the ball of kernel integration for testing

A.2 Nyström Approximation

Proof of Proposition 1.

Let {yj}j=1m\{y_{j}\}_{j=1}^{m} be an i.i.d. sequence with yjνy_{j}\sim\nu. Define κy=κ(,y)\kappa_{y}=\kappa(\cdot,y) for any yDy\in D. Notice that by the reproducing property,

𝔼yν[𝟙B(,r)(κyκy)f]\displaystyle\mathbb{E}_{y\sim\nu}[\mathds{1}_{B(\cdot,r)}(\kappa_{y}\otimes\kappa_{y})f] =𝔼yν[𝟙B(,r)κy,fκy]\displaystyle=\mathbb{E}_{y\sim\nu}[\mathds{1}_{B(\cdot,r)}\langle\kappa_{y},f\rangle\kappa_{y}]
=D𝟙B(,r)κy,fκyν(dy)\displaystyle=\int_{D}\mathds{1}_{B(\cdot,r)}\langle\kappa_{y},f\rangle\kappa_{y}\>\nu(dy)
=D𝟙B(,r)κ(,y)f(y)ν(dy)\displaystyle=\int_{D}\mathds{1}_{B(\cdot,r)}\kappa(\cdot,y)f(y)\>\nu(dy)
=B(,r)κ(,y)f(y)ν(dy)\displaystyle=\int_{B(\cdot,r)}\kappa(\cdot,y)f(y)\>\nu(dy)

hence

T=𝔼yν[𝟙B(,r)(κyκy)]T=\mathbb{E}_{y\sim\nu}[\mathds{1}_{B(\cdot,r)}(\kappa_{y}\otimes\kappa_{y})]

and similarly

Tm=1mj=1m𝟙(yjB(,r))(κyjκyj).T_{m}=\frac{1}{m}\sum_{j=1}^{m}\mathds{1}(y_{j}\in B(\cdot,r))(\kappa_{y_{j}}\otimes\kappa_{y_{j}}).

Define T(j):=𝟙(yjB(,r))(κyjκyj)T^{(j)}:=\mathds{1}(y_{j}\in B(\cdot,r))(\kappa_{y_{j}}\otimes\kappa_{y_{j}}) for any j{1,,m}j\in\{1,\dots,m\} and Ty:=𝟙B(,r)(κyκy)T_{y}:=\mathds{1}_{B(\cdot,r)}(\kappa_{y}\otimes\kappa_{y}) for any yDy\in D, noting that 𝔼yν[Ty]=T\mathbb{E}_{y\sim\nu}[T_{y}]=T and 𝔼[T(j)]=T\mathbb{E}[T^{(j)}]=T. Further we note that

𝔼uνTyHS2𝔼yνκy4<\mathbb{E}_{u\sim\nu}\|T_{y}\|^{2}_{HS}\leq\mathbb{E}_{y\sim\nu}\|\kappa_{y}\|^{4}<\infty

and, by Jensen’s inequality,

THS2𝔼yνTyHS2<\|T\|^{2}_{HS}\leq\mathbb{E}_{y\sim\nu}\|T_{y}\|^{2}_{HS}<\infty

hence TT is Hilbert-Schmidt (as is TmT_{m} since it has finite rank). We now compute,

𝔼TmTHS2\displaystyle\mathbb{E}\|T_{m}-T\|_{HS}^{2} =𝔼1mj=1mT(j)THS2\displaystyle=\mathbb{E}\|\frac{1}{m}\sum_{j=1}^{m}T^{(j)}-T\|^{2}_{HS}
=𝔼1mj=1mT(j)HS221mj=1m𝔼[T(j)],THS+THS2\displaystyle=\mathbb{E}\|\frac{1}{m}\sum_{j=1}^{m}T^{(j)}\|^{2}_{HS}-2\langle\frac{1}{m}\sum_{j=1}^{m}\mathbb{E}[T^{(j)}],T\rangle_{HS}+\|T\|^{2}_{HS}
=𝔼1mj=1mT(j)HS2THS2\displaystyle=\mathbb{E}\|\frac{1}{m}\sum_{j=1}^{m}T^{(j)}\|^{2}_{HS}-\|T\|^{2}_{HS}
=1m𝔼yνTyHS2+1m2j=1mkjm𝔼[T(j)],𝔼[T(k)]HSTHS2\displaystyle=\frac{1}{m}\mathbb{E}_{y\sim\nu}\|T_{y}\|_{HS}^{2}+\frac{1}{m^{2}}\sum_{j=1}^{m}\sum_{k\neq j}^{m}\langle\mathbb{E}[T^{(j)}],\mathbb{E}[T^{(k)}]\rangle_{HS}-\|T\|^{2}_{HS}
=1m𝔼yνTyHS2+m2mm2THS2THS2\displaystyle=\frac{1}{m}\mathbb{E}_{y\sim\nu}\|T_{y}\|_{HS}^{2}+\frac{m^{2}-m}{m^{2}}\|T\|^{2}_{HS}-\|T\|^{2}_{HS}
=1m(𝔼yνTyHS2THS2)\displaystyle=\frac{1}{m}\big{(}\mathbb{E}_{y\sim\nu}\|T_{y}\|_{HS}^{2}-\|T\|^{2}_{HS}\big{)}
=1m𝔼yνTyTHS2.\displaystyle=\frac{1}{m}\mathbb{E}_{y\sim\nu}\|T_{y}-T\|^{2}_{HS}.

Setting C2=𝔼yνTyTHS2C^{2}=\mathbb{E}_{y\sim\nu}\|T_{y}-T\|^{2}_{HS}, we now have

𝔼TmTHS2=C2m.\mathbb{E}\|T_{m}-T\|_{HS}^{2}=\frac{C^{2}}{m}.

Applying Jensen’s inequality to the convex function xx2x\mapsto x^{2} gives

𝔼TmTHSCm.\mathbb{E}\|T_{m}-T\|_{HS}\leq\frac{C}{\sqrt{m}}.

A.3 Figures of Table 8

Figure 4, 5, and 6 show graph kernel network’s performance for the first 55 testing examples with s=141,211,421s=141,211,421 respectively. As can be seen in the figures, most of the error occurs around the singularity of the coefficient aa (where the two colors cross).

łRefer to caption

Figure 4: s=141s=141
Refer to caption
Figure 5: s=211s=211
Refer to caption
Figure 6: s=421s=421

A.4 Approximating the 2D Green’s function

Refer to caption
Figure 7: Contour plot of the learned and truth kernel of Poisson equation on a unit disk. The contour plot shows the Green’s function of G(0,0;ρ~,θ~)G(0,0;\widetilde{\rho},\widetilde{\theta})

In Section 3 it has been shown that the graph kernel network is capable of approximating the truth kernel of 1D Poisson equation. Here we further demonstrate the approximation power of the graph kernel network for the 2D Poisson equation by showing that the graph kernel net work is able to capture its truth Green’s function.

Although a closed-form expression for general Green’s function of arbitrary 2D domain does not exist, it does exist when the partial differential equation is defined on a unit disk. Hence consider the 2D Poisson’s equation as introduced in Section 3 with xx defined on a unit disk D:x12+x221D:x_{1}^{2}+x_{2}^{2}\leq 1. Recall the corresponding Green’s function in this case can be expressed in the polar coordinate as:

Gt(ρ,θ;ρ~,θ~)=14πlnρ~2+ρ22ρρ~(cos(θ~)θ)ρ~2ρ2+12ρρ~(cos(θ~)θ).G_{t}(\rho,\theta;\widetilde{\rho},\widetilde{\theta})=\frac{1}{4\pi}\ln\frac{\widetilde{\rho}^{2}+\rho^{2}-2\rho\widetilde{\rho}(\cos(\widetilde{\theta})-\theta)}{\widetilde{\rho}^{2}\rho^{2}+1-2\rho\widetilde{\rho}(\cos(\widetilde{\theta})-\theta)}.

We first apply the graph kernel network to approximate the mapping f(x)u(x)f(x)\mapsto u(x), and then compare the learned kernel GG to the truth kernel GtG_{t}. Figure 7 shows contour plots of the truth and learned Green’s function evaluated at G(0,0;ρ~,θ~)G(0,0;\widetilde{\rho},\widetilde{\theta}). Again the graph kernel network is capable of learning the true Green’s function, albeit the approximating error becomes larger near the singularity at zero. As a consequence, the exceptional generalization power of the graph neural network seen in Fig. 3 is explained and expected.