This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Multi-Step Embed to Control: A Novel Deep Learning-based Approach for Surrogate Modelling in Reservoir Simulation

Jungang Chen, Eduardo Gildin, John Killough
Harold Vance Department of Petroleum Engineering, Texas A&M University, USA
{jungangc, egildin, jkillough}@tamu.edu
Abstract

Reduced-order models, also known as proxy model or surrogate model, are approximate models that are less computational expensive as opposed to fully descriptive models. With the integration of machine learning, these models have garnered increasing research interests recently. However, many existing reduced-order modeling methods, such as embed to control (E2C) and embed to control and observe (E2CO), fall short in long-term predictions due to the accumulation of prediction errors over time. This issue arises partly from the one-step prediction framework inherent in E2C and E2CO architectures. This paper introduces a deep learning-based surrogate model, referred as multi-step embed-to-control model, for the construction of proxy models with improved long-term prediction performance. Unlike E2C and E2CO, the proposed network considers multiple forward transitions in the latent space at a time using Koopman operator, allowing the model to incorporate a sequence of state snapshots during training phrases. Additionally, the loss function of this novel approach has been redesigned to accommodate these multiple transitions and to respect the underlying physical principles. To validate the efficacy of the proposed method, the developed framework was implemented within two-phase (oil and water) reservoir model under a waterflooding scheme. Comparative analysis demonstrate that the proposed model significantly outperforms the conventional E2C model in long-term simulation scenarios. Notably, there was a substantial reduction in temporal errors in the prediction of saturation profiles and a decent improvement in pressure forecasting accuracy.

1 Introduction

Reservoir simulation is a computational process used to model multi-physics behaviors of fluids within the subsurface reservoir over a specified spatiotemporal domain. It primarily focuses on utilizing mathematical models to predict the flow of oil, gas and water through porous media. Recent efforts have expanded to include thermos-hydro-mechanical interactions and applications in geothermal energy, CO2 sequestration and hydrogen storage. However, full scale simulations, especially those modeling multi-physics, are computationally intensive and time-consuming, making them impractical for real-time decision-making, optimization, uncertainty quantification, and frequent data assimilation.

To address these challenges, reduced-order models have been introduced to simplify those complex simulations, enabling faster and more efficient decision making while maintaining essential accuracy. Traditional reduced-order modeling (ROM) techniques, such as those using Proper Orthogonal Decomposition (POD), seek to project the original high-dimensional space xx to a lower-dimensional latent state space zz using the POD basis Φ\Phi derived from collected data snapshots, e.g. z=Φxz=\Phi x [20, 18]. Despite the effectiveness, their reliance on direct source code access could hinder their practical applications. Alternatively, deep learning-based ROMs offer a non-intrusive approach that can seamlessly integrate with existing commercial simulators [8, 5]. Moreover, traditional ROMs assume that the dynamics of the system can be accurately captured in a linear subspace of the full state space. This assumption can limit its effectiveness in handling systems with highly nonlinear behavior. On the other hand, deep learning-based ROMs leverage neural networks to account for nonlinear dynamics.

With the development of machine learning techniques and their application to subsurface problems [11, 25, 26, 12], various machine learning-based surrogate models have been proposed for different settings [19, 27, 2]. However, the embed to control (E2C) model and its variants are particularly suitable for scenarios with time-varying well settings, such as bottom-hole pressure and injection rates, where well-control optimization is involved. The original Embed-to-Control (E2C) is a variational autoencoder-based deep learning model for stochastic optimal control [22]. It has been modified to a deterministic form for surrogate modeling in reservoir simulation [8, 5]. This model consists of an encoder network that projects high-dimensional state spaces (saturation, pressure, etc.) into a lower latent space, a transition model that governs the evolution of this latent space from current time to the next, and a decoder that reconstructs states of next time from latent representation of corresponding time step. However, the conventional E2C model, primarily designed for short-term forecasting due to its reliance on adjacent state pairs for training, encounters limitations in precise long-term simulation scenarios.

In this paper, we propose a novel deep learning-based ROMs that is more accurate in terms of long-term predictions. Similar to the variational autoencoder structure in E2C, the newly proposed model also employs an encoder that projects original high-dimensional state to a lower-dimensional latent representation and a decoder that reconstruct the state using predicted latent state. However, unlike the locally linear transition model in E2C, we use Koopman operator to perform multiple forward transitions in latent space simultaneously, which allows the network to make multi-step state predictions by decoding predicted latent representations. Notably, the Koopman operator is also approximated by neural networks, so the overall network is trained in an unsupervised manner. Furthermore, the loss function has been modified to account for the multi-step prediction loss. We note these modifications improve the long-term prediction accuracy. This enhancement can be primarily attributed to the integration of multiple transitions within the model’s framework, which effectively mitigates the accumulation of error along the trajectory. By improving the accuracy of long-term predictions of states and observations, the model can facilitate the development of more reliable well rate/pressure control strategies for longer reservoir production period [4].

This paper is organized as follows: we begin with a review of deep generative models and show their relationship with reduced-order modeling. We then highlight the limitations of previous models in long-term simulations and introduce our model which is more accurate for long-term predictions. Following that, we present and compare our results with previous models. Finally, we conclude this work and discuss future research directions.

2 Problem Statement

Instead of comparing deep-learning ROMs with traditional ROM techniques, in this section, we present the deep learning-based ROMs within the framework of deep generative models for sequential data. The rational behind is that given a sequence of control inputs and the initial state, what would be the most probable resulting state trajectory? Mathematically, it can be formulated as p(x1:T|u1:T,xo)p(x_{1:T}|u_{1:T},x_{o}). Once p(x1:T|u1:T,xo)p(x_{1:T}|u_{1:T},x_{o}) is obtained through deep learning, the trained model can generate the most probable trajectory x1:Tx_{1:T} for any given testing sequence of control inputs u1:Tu_{1:T} and initial state x0x_{0}. Although the generation process is stochastic, it can be easily converted to deterministic form, as in reservoir simulation case, by modifying the encoding and decoding network architectures. The model is called ’reduced order’ because it leverages the encoder-decoder structure to compress the original state into a latent space and reconstruct original states using predicted latent states.

2.1 Variational autoencoder (VAE)

Variational autoencoder is a type of deep generative model that employs an encoder-decoder structure to generate new data samples similar to input data it has been trained on [9, 3, 13, 6]. VAE-based generative models often consists of two processes: generative process and inference process. The generative process assumes that the observed data x is generated from latent variable z which has a simpler distribution (e.g. standard Gaussian). Once z has been sampled from the prior distribution p(z)p(z), the new data can be generated from the conditional probability pθ(x|z)p_{\theta}(x|z) which is parameterized by a neural network with parameter θ\theta. The inference process involves calculation of the posterior distribution pθ(z|x)=pθ(x|z)p(z)p(x)p_{\theta}(z|x)=\frac{p_{\theta}(x|z)p(z)}{p(x)}. However, directly computing the posterior distribution is intractable because it requires integration over latent variable z. To tackle this issue, VAEs adopts variational inference to approximate the posterior distribution using a distribution which is parameterized by neural network, expressed as qφ(z|x)q_{\varphi}(z|x).

To approximate the true posterior, the VAE is trained to maximize the evidence lower bound (ELBO), which reads,

(θ,φ,x)=𝔼qφ(z|x)[log(pθ(x|z))]DKL(qφ(z|x))p(z)),\mathcal{L}(\theta,\varphi,x)=\mathbb{E}_{q_{\varphi}(z|x)}[\log(p_{\theta}(x|z))]-D_{KL}(q_{\varphi}(z|x))\parallel p(z)), (1)

The first term is referred as reconstruction, which measures how well the data can be reconstructed through the encoder-decoder process. The second term is called regularization term, which enforces the approximate posterior qφ(z|x)q_{\varphi}(z|x) to be close to the prior p(z)p(z).

Despite their effectiveness, VAEs are not designed for sequential data [7]. In fact, VAE assumes the input data being independent and identically distributed (i.i.d.), which hinders its application to sequential data. Furthermore, VAEs cannot handle system controls, therefore it is not suitable for dynamical systems where both system inputs and state transitions need to be considered.

2.2 Embed to Control (E2C)

The original Embed-to-Control (E2C) is a variational autoencoder-based deep learning model for stochastic optimal control of dynamical systems [22]. Similar to VAE, the E2C has an encoder-decoder structure. Instead of modeling data likelihood p(x)p(x), the E2C models the pair likelihood p(xt,xt+1|ut)p(x_{t},x_{t+1}|u_{t}), where xt+1x_{t+1} is a result from xtx_{t} by applying utu_{t}. E2C assumes xtx_{t} and xt+1x_{t+1} can be generated from latent variable ztz_{t} and zt+1z_{t+1} respectively. To consider state transitions, it enforces zt+1z_{t+1} to be both the embedding of xt+1x_{t+1} and the evolution of ztz_{t} by applying utu_{t}. The latent dynamics in general is expressed as:

z^t+1=fz(zt,ut),\hat{z}_{t+1}=f_{z}(z_{t},u_{t}), (2)

Where fzf_{z} signifies the latent dynamics operator that evolves the latent state based on current state ztz_{t} and control input utu_{t}. Accordingly, the E2C aims to maximize the following evidence lower bound (ELBO) [1]:

(θ,φ,ω,xt,xt+1)=𝔼qφ(zt|xt),qφ(zt+1|xt+1)[log(pθ(xt|zt))+log(pθ(xt+1|zt+1))]\displaystyle\mathcal{L}(\theta,\varphi,\omega,x_{t},x_{t+1})=\mathbb{E}_{q_{\varphi}(z_{t}|x_{t}),q_{\varphi}(z_{t+1}|x_{t+1})}[\log(p_{\theta}(x_{t}|z_{t}))+\log(p_{\theta}(x_{t+1}|z_{t+1}))] (3)
DKL(qφ(zt|xt))p(zt))DKL(qω(z^t+1|zt,ut))qφ(zt+1|xt+1)),\displaystyle-D_{KL}(q_{\varphi}(z_{t}|x_{t}))\parallel p(z_{t}))-D_{KL}(q_{\omega}(\hat{z}_{t+1}|z_{t},u_{t}))\parallel q_{\varphi}(z_{t+1}|x_{t+1})),

Where qω(z^t+1|zt,ut)q_{\omega}(\hat{z}_{t+1}|z_{t},u_{t}) represents the generative dynamics of equation 3 which is parameterized by neural network with weights ω\omega. In practice, it’s usually replaced by a locally linear formulation. Likewise, the first term in equation 3 represents the reconstruction term which measures how well xtx_{t} and xt+1x_{t+1} can be reconstructed, the second and last term means that not only it enforces the posterior qφ(zt|xt)q_{\varphi}(z_{t}|x_{t}) to be like the prior p(zt)p(z_{t}), it also enforces the result z^t+1\hat{z}_{t+1} from the latent dynamics to be close to the embedding zt+1z_{t+1}.

[8] has modified the above formulation to a deterministic version for surrogate modelling of reservoir simulation applications. This model consists of an encoder network that projects high-dimensional states into a lower latent space, a transition model that governs the evolution of this latent space from current time to the next, and a decoder that reconstructs states of next time from latent representation of corresponding time step. From a mathematical standpoint, the encoder-transition-decoder architecture, is detailed as follows:

Encoder: zt=Qφ(xt)z_{t}=Q_{\varphi}(x_{t})

Transition: z^t+1=Fω1(zt,ut)\hat{z}_{t+1}=F_{\omega_{1}}(z_{t},u_{t})

Decoder: x^t+1=Pθ(z^t+1)\hat{x}_{t+1}=P_{\theta}(\hat{z}_{t+1})

where QφQ_{\varphi} and PθP_{\theta} denote the deterministic form of qφq_{\varphi}and pθp_{\theta} , and Fω1F_{\omega_{1}} represents the deterministic latent state dynamics. A graphical representation of E2C is presented in Figure 1.

Refer to caption
Figure 1: Graphical model of E2C.

In practice, the latent dynamics is usually replaced by a locally linear form:

z^t+1=A(zt)zt+B(zt)ut,\hat{z}_{t+1}=A(z_{t})*z_{t}+B(z_{t})*u_{t}, (4)

Although effective, E2C is not designed for long trajectory prediction, as indicated by equation 1, E2C aims to maximize the pair marginal likelihood. In other words, for a given xtx_{t} and ztz_{t}, E2C tries to predict the most probable xt+1x_{t+1}, which referred as one-step prediction. For long-term trajectory prediction, the one-step prediction will naturally lead to error accumulation problem.

3 Methods

Unlike E2C, the newly proposed model performs multi-step forward prediction through a deep Koopman operator. Instead of relying on single-step transition in the latent space and utilizing adjacent states for training, the new framework employs a multiple transition strategy. This approach incorporates multiple state snapshots and control data to facilitate the learning of a more refined latent space, thereby significantly improving the accuracy of long-term predictions. The framework of the newly proposed models is presented in Figure 2 below:

Refer to caption
Figure 2: Framework of proposed multi-step E2C model. The latent space is further lifted by Koopman operator to a space where the state transition is linear. The network can be equipped with multiple state snapshots for training and therefore improving the long-term prediction accuracy.

In this proposed approach, it is assumed that a sequence of states xtx_{t}, xt+1x_{t+1}, …, xt+kx_{t+k} can be generated by corresponding latent states ztz_{t}, zt+1z_{t+1}, …, zt+kz_{t+k} with the decoder structure. The key distinction of our method lies in how the latent space is propagated. [21] attempted to use the E2C approach but found that special treatment was needed to maintain stability. Unlike the locally linear form used in E2C, where locally A and B matrices are applied to transitions (by locally we meant that A and B matrices in 4 change with different latent states), our model leverages the Koopman operator to project the latent space into a higher-dimensional space where multi-step transitions become globally linear, as in 7. As a result, our model, referred to multi-step embed to control (MSE2C), models the likelihood p(xt,xt+1,xt+k|ut,ut+1,ut+k1)p(x_{t},x_{t+1},...x_{t+k}|u_{t},u_{t+1},...u_{t+k-1}), where kk is the forward step. The resulting loss of the whole framework total\mathcal{L}_{total} is written as follows(the details are presented in next section):

total=rec+reg+koop\displaystyle\mathcal{L}_{total}=\mathcal{L}_{rec}+\mathcal{L}_{reg}+\mathcal{L}_{koop} (5)

3.1 Deep Koopman Operator

The Koopman operator is a powerful tool to analyze nonlinear dynamical systems by mapping them into an infinite-dimensional framework [10]. The basic concept of Koopman operator is to shift the focus from the state space to the space of observable functions where linear operator can be applied. Practical implementations, such as dynamic mode decomposition (DMD) and its variant extended DMD, seek to approximate the Koopman operator using data from the system. For more details regarding Koopman operators, we refer to [15, 23, 16]. Recent advancements involve using deep learning to automatically learn the appropriate observables and Koopman operator representations from data [14, 17](Morton, J et al., 2019; Shi, H et al., 2022). [24] proposed a data-driven approach to model and control autonomous vehicle with Koopman operators approximated by neural network. However, directly applying Koopman operators to original states may not be effective for high-dimensional systems. In this paper, instead of applying Koopman operators directly to the original states, we lift the latent space to observables using neural networks, which we believe is more effective for high-dimensional systems. A draft structure is presented in Figure 3.

Refer to caption
Figure 3: The structure of deep Koopman operator, where green box represents dense layers.

Again, let z^t+1=fz(zt,ut)\hat{z}_{t+1}=f_{z}(z_{t},u_{t}) be the nonlinear transition dynamics of the E2C. The deep Koopman operator tried to lift the latent space representation ztz_{t} to a space where the following operation is linear:

Φe,t+1=AΦe,t+But\Phi_{e,t+1}=A\Phi_{e,t}+Bu_{t} (6)

Where Φe,t=[zt,lt]T\Phi_{e,t}=[z_{t},l_{t}]^{T}, and ltl_{t} denotes the output of last layer of the encoder in Figure 3, which reads as lt=enc(zt)l_{t}=enc(z_{t}). The matrix AA and BB (we slightly abuse the notation here and note that these are different with what in the locally linear form), are defined as the approximate Koopman operator with finite dimension. For brevity, equation 6 can be rearranged as Φe,t+1=KΨt\Phi_{e,t+1}=K\Psi_{t}. K=[A,B]K=[A,B] is the Koopman operator approximated by neural network weights, and Ψt=[Φe,t,ut]T\Psi_{t}=[\Phi_{e,t},u_{t}]^{T}.

With approximate Koopman operator, multi-step forward prediction can be performed in the lifted space, which follows:

Φe,t+p=AΦe,t+p1+But+p1(p=1,2,,k)\Phi_{e,t+p}=A\Phi_{e,t+p-1}+Bu_{t+p-1}\qquad(p=1,2,\dots,k) (7)

To learn the Koopman operator, the error of the state evolution in the lifted space and those from observations needs to be minimized. To this end, the following k-step loss is minimized:

koop=1ki=1kΦ^e,t+iK[i]Ψt2,\displaystyle\mathcal{L}_{koop}=\frac{1}{k}\sum_{i=1}^{k}||\hat{\Phi}_{e,t+i}-K^{[i]}\Psi_{t}||^{2}, (8)

Where Φ^e,t+i=[Qφ(xt+i),enc(Qφ(xt+i))]T\hat{\Phi}_{e,t+i}=[Q_{\varphi}(x_{t+i}),enc(Q_{\varphi}(x_{t+i}))]^{T} are mapped from observations, and K[i]ΨtK^{[i]}\Psi_{t} are predicted from the Koopman operator.

3.2 Multi-step E2C training

In this section, we outline the loss functions used to train the multi-step E2C network. The loss consists of 3 components: reconstruction loss that enforces the predicted states close to the true states, regularization loss to ensure the encoded latent states are also results of k-step prediction from Koopman operator, and the k-step loss that uses to approximate the Koopman operator. The first two losses are similar to the VAE loss in equation 1 and E2C loss in equation 3, but note that in this context it’s different sign as we minimize the loss instead of maximizing likelihood. The k-step loss follows the definition in equation 8. In summary, the loss terms are defined as:

rec=i=1k𝔼qφ(zt+1|xt+1)[log(pθ(xt+1|zt=1))],\displaystyle\mathcal{L}_{rec}=-\sum_{i=1}^{k}\mathbb{E}_{q_{\varphi}(z_{t+1}|x_{t+1})}[\log(p_{\theta}(x_{t+1}|z_{t=1}))], (9)
reg=DKL(qφ(z|x))p(z))+i=1kDKL(kω(z^t+i|zt,ut)qφ(zt+i|xt+i))\displaystyle\mathcal{L}_{reg}=D_{KL}(q_{\varphi}(z|x))\parallel p(z))+\sum_{i=1}^{k}D_{KL}(k_{\omega}(\hat{z}_{t+i}|z_{t},u_{t})\parallel q_{\varphi}(z_{t+i}|x_{t+i})) (10)
koop=1ki=1kΦ^e,t+iK[i]Ψt2,\displaystyle\mathcal{L}_{koop}=\frac{1}{k}\sum_{i=1}^{k}||\hat{\Phi}_{e,t+i}-K^{[i]}\Psi_{t}||^{2}, (11)

where kω(z^t+i|zt,ut)k_{\omega}(\hat{z}_{t+i}|z_{t},u_{t}) in reg\mathcal{L}_{reg} denotes the latent state predictions from Koopman operator.To sum up, we present the training tuple and training loss of different methods in Table 1. It can be observed that one obvious advantage of our approach is that more data considered for training, which significantly reduces the error accumulation along the simulation trajectory.

The training process consists of 400 epochs with a batch size of 6. The Adam optimizer is used for training, with a learning rate set at 0.01. The data is split into training and testing sets with a ratio of 4:1.

Table 1: Comparison of training tuple and training loss for different methods
Methods Training tuple Training losses111We abuse the loss term here.rec\mathcal{L}_{rec} and reg\mathcal{L}_{reg} represent the reconstruction and regularization loss of each unique approach.
VAE {xx} {rec\mathcal{L}_{rec},reg\mathcal{L}_{reg}}
Embed to Control (E2C) {xtx_{t},utu_{t},xt+1x_{t+1}} {rec\mathcal{L}_{rec}, reg\mathcal{L}_{reg}}
Multi-step E2C {xtx_{t},utu_{t},xt+1x_{t+1},ut+1u_{t+1},…,xt+kx_{t+k}} {rec\mathcal{L}_{rec}, reg\mathcal{L}_{reg}, koop\mathcal{L}_{koop} }

4 Results

The proposed framework is applied within a water flooding case where water is injected into an oil reservoir. The model consists of 64 × 64 × 1 grid blocks, each of which is of the same size and has dimension of 65.6 ft × 65.6 ft × 65.6 ft in x, y and z directions. We simulated scenarios involving the water injection using 4 injector wells and the extraction of oil using 5 producer wells. The well locations and permeability map are plotted in Figure 4. This model runs for 2000 days with well controls, including water injection rates and bottom hole pressures (BHPs), changing every 100 days.

Refer to caption
Figure 4: Permeability map and well locations.

In this experiment, a total of 600 simulation runs are performed during the offline stage to gather data. Each simulation run, also referred as one trajectory, consists of 21 state (pressure, saturation) snapshots and 20 well controls (BHP, injection rate). For a k-step model, we can collect 21k+121-k+1 sets of data per trajectory, resulting in a total of 600(21k+1)600*(21-k+1) data samples for all simulation runs.

Three approaches, including the E2C, multistep E2C with k=5k=5, and multistep E2C with k=7k=7, are compared to validate the efficacy of proposed approach.

4.1 Comparison of E2C and Multi-step E2C

Figure 5 and 6 compares the water saturation and fluid pressure predicted by E2C, multi-step E2C with a forward prediction step of 5 (k=5) and multi-step E2C with a forward prediction step of 7 (k=7), against those obtained from numerical simulations.

Refer to caption
Figure 5: Comparison of E2C, multi-step E2C (k=5) and multi-step E2C (k=7) predictions on water saturation evolution at different days.
Refer to caption
Figure 6: Comparison of E2C, multi-step E2C (k=5) and multi-step E2C (k=7) predictions on fluid pressure evolution at different days.

The top row of Figure 5 shows the water saturation at different times from numerical simulation. The second row of Figure 5 depicts the absolute error of water saturations between the numerical simulation and those predicted by E2C at corresponding timesteps. The third row and last row present the absolute error of water saturation predictions between the numerical simulation and multi-step E2C (k=5) and multistep E2C (k=7) respectively. It can be observed that both multi-step E2C methods perform better than the vanilla E2C at every time steps.

Similarly, we present the fluid pressure and its absolute error map from different approaches at difference time in Figure 6. The top row represents the fluid pressures from numerical simulator, while the second row presents the corresponding absolute error map predicted by E2C. The last two rows depict the absolute error map predicted by multistep E2C (k=5) and multistep E2C (k=7) respectively. It can be also observed that the multi-step E2C achieves significantly better long-term accuracy for predicting fluid pressures.

Finally, to demonstrate the effectiveness of the proposed multi-step E2C across various testing scenarios, Figure 7 presents a boxplot that shows the absolute error of 100 testing cases over time for different approaches. As the plot illustrates, the E2C approach consistently exhibits the highest absolute error across timesteps, indicating it is less accurate than the other methods. In contrast, the multi-step E2C approaches with k=5 and k=7 show similar accuracy, with the multi-step E2C (k=5) performing slightly better. This slight performance difference can be attributed to the reduction in training samples when the forward timesteps are increased(as presented in 8), making the network more challenging to train effectively.

Refer to caption
Figure 7: Boxplot of absolute error along with timesteps for different approaches.
Refer to caption
Figure 8: Collected samples decreases as forward step kk increases.

5 Conclusions

This study introduces a novel deep learning-based reduced-order model designed for efficient long-term predictions in reservoir simulations. The proposed multi-step E2C method incorporates multiple forward transitions in the latent space using the Koopman operator, significantly enhancing the accuracy of long-term predictions by mitigating error accumulation typically observed in conventional single-step E2C models. The efficacy of the proposed method is validated using a two-phase (oil and water) reservoir model under a water flooding scheme. Comparative analysis highlights that the multi-step E2C model outperforms conventional E2C in both water saturation and fluid pressure predictions over extended simulation periods. By improving the accuracy of long-term predictions, the proposed model can facilitate more reliable well rate and pressure control strategies, ultimately contributing to more effective reservoir management and decision-making processes. In conclusion, the multi-step E2C model offers a robust and computationally efficient solution for long-term reservoir simulation and optimization, addressing key limitations of existing reduced-order modeling techniques.

Acknowledgments

The authors would like to thank Texas A&M High Performance Research Computing for advanced computing resources.

References

  • Banijamali et al. [2018] E. Banijamali, R. Shu, H. Bui, A. Ghodsi, et al. Robust locally-linear controllable embedding. In International Conference on Artificial Intelligence and Statistics, pages 1751–1759. PMLR, 2018.
  • Chen et al. [2023a] J. Chen, E. Gildin, and J. E. Killough. Physics-informed convolutional recurrent surrogate model for reservoir simulation with well controls. arXiv preprint arXiv:2305.09056, 2023a.
  • Chen et al. [2023b] J. Chen, C.-K. Huang, J. F. Delgado, and S. Misra. Generating subsurface earth models using discrete representation learning and deep autoregressive network. Computational Geosciences, 27(6):955–974, 2023b.
  • Chen et al. [2024] J. Chen, E. Gildin, and J. E. Killough. Optimization of pressure management strategies for geological co2 sequestration using surrogate model-based reinforcement learning. arXiv preprint arXiv:2403.07360, 2024.
  • Coutinho et al. [2021] E. J. R. Coutinho, M. Dall’Aqua, and E. Gildin. Physics-aware deep-learning-based proxy reservoir simulation model equipped with state and well output prediction. Frontiers in Applied Mathematics and Statistics, 7:651178, 2021.
  • Ding et al. [2024] Z. Ding, P. Li, Q. Yang, and S. Li. Enhance image-to-image generation with llava prompt and negative prompt. arXiv preprint arXiv:2406.01956, 2024.
  • Girin et al. [2020] L. Girin, S. Leglaive, X. Bie, J. Diard, T. Hueber, and X. Alameda-Pineda. Dynamical variational autoencoders: A comprehensive review. arXiv preprint arXiv:2008.12595, 2020.
  • Jin et al. [2020] Z. L. Jin, Y. Liu, and L. J. Durlofsky. Deep-learning-based surrogate model for reservoir simulation with time-varying well controls. Journal of Petroleum Science and Engineering, 192:107273, 2020.
  • Kingma and Welling [2013] D. P. Kingma and M. Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
  • Koopman [1931] B. O. Koopman. Hamiltonian systems and transformation in hilbert space. Proceedings of the National Academy of Sciences, 17(5):315–318, 1931.
  • Li et al. [2023] C. Li, B. Yan, R. Kou, and S. Gao. Rapid inference of reservoir permeability from inversion of travel time data under a fast marching method based deep learning framework. In SPE EuropEC-Europe Energy Conference featured at the 84th EAGE Annual Conference & Exhibition. OnePetro, 2023.
  • Li et al. [2018] P. Li, Y. Lin, and E. Schultz-Fellenz. Contextual hourglass network for semantic segmentation of high resolution aerial imagery. arXiv preprint arXiv:1810.12813, 2018.
  • Misra et al. [2023] S. Misra, J. Chen, Y. Falola, P. Churilova, C.-K. Huang, and J. Delgado. Massive geomodel compression and rapid geomodel generation using advanced autoencoders and autoregressive neural networks. In SPE EuropEC-Europe Energy Conference featured at the 84th EAGE Annual Conference & Exhibition. OnePetro, 2023.
  • Morton et al. [2019] J. Morton, F. D. Witherden, and M. J. Kochenderfer. Deep variational koopman models: Inferring koopman observations for uncertainty-aware dynamics modeling and control. arXiv preprint arXiv:1902.09742, 2019.
  • Proctor et al. [2018] J. L. Proctor, S. L. Brunton, and J. N. Kutz. Generalizing koopman theory to allow for inputs and control. SIAM Journal on Applied Dynamical Systems, 17(1):909–930, 2018.
  • Schmid [2010] P. J. Schmid. Dynamic mode decomposition of numerical and experimental data. Journal of fluid mechanics, 656:5–28, 2010.
  • Shi and Meng [2022] H. Shi and M. Q.-H. Meng. Deep koopman operator with control for nonlinear systems. IEEE Robotics and Automation Letters, 7(3):7700–7707, 2022.
  • Tan et al. [2019] X. Tan, E. Gildin, H. Florez, S. Trehan, Y. Yang, and N. Hoda. Trajectory-based deim (tdeim) model reduction applied to reservoir simulation. Computational Geosciences, 23:35–53, 2019.
  • Tang et al. [2022] M. Tang, X. Ju, and L. J. Durlofsky. Deep-learning-based coupled flow-geomechanics surrogate model for co2 sequestration. International Journal of Greenhouse Gas Control, 118:103692, 2022.
  • Trehan and Durlofsky [2016] S. Trehan and L. J. Durlofsky. Trajectory piecewise quadratic reduced-order model for subsurface flow, with application to pde-constrained optimization. Journal of Computational Physics, 326:446–473, 2016.
  • Tytarenko [2022] A. Tytarenko. Multi-step prediction in linearized latent state spaces for representation learning. arXiv preprint arXiv:2209.01127, 2022.
  • Watter et al. [2015] M. Watter, J. Springenberg, J. Boedecker, and M. Riedmiller. Embed to control: A locally linear latent dynamics model for control from raw images. Advances in neural information processing systems, 28, 2015.
  • Williams et al. [2015] M. O. Williams, I. G. Kevrekidis, and C. W. Rowley. A data–driven approximation of the koopman operator: Extending dynamic mode decomposition. Journal of Nonlinear Science, 25:1307–1346, 2015.
  • Xiao et al. [2022] Y. Xiao, X. Zhang, X. Xu, X. Liu, and J. Liu. Deep neural networks with koopman operators for modeling and control of autonomous vehicles. IEEE transactions on intelligent vehicles, 8(1):135–146, 2022.
  • Yan et al. [2023] B. Yan, C. Li, Z. Tariq, and K. Zhang. Estimation of heterogeneous permeability using pressure derivative data through an inversion neural network inspired by the fast marching method. Geoenergy Science and Engineering, 228:211982, 2023.
  • Zhang et al. [2023] C. Zhang, Y. Pan, W. Ma, and R. Bi. Application of combined statistics, machine learning and game theory approach for shale gas production performance evaluation: Insights of frac-hit timing and severity. In SPE/AAPG/SEG Unconventional Resources Technology Conference, page D021S042R001. URTEC, 2023.
  • Zheng et al. [2024] F. Zheng, B. Jha, and B. Jafarpour. A deep learning-based surrogate model for rapid assessment of geomechanical risks in geologic co2 storage. In SPE/AAPG/SEG Carbon, Capture, Utilization, and Storage Conference and Exhibition, page D031S022R001. SPE, 2024.