This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\SetWatermarkText

Draft \SetWatermarkLightness0.85 \SetWatermarkScale1.3

GotFlow3D: Recurrent Graph Optimal Transport for Learning 3D Flow Motion in Particle Tracking

Jiaming Liang, Chao Xu and Shengze Cai J. Liang and C. Xu are affiliated with the College of Control Science & Engineering, Zhejiang University (ZJU), Hangzhou, 310027 China, the State Key Laboratory of Industrial Control Technology, the National Engineering Center for Industrial Automation, the Institute of Cyber-Systems & Control at ZJU, and ZJU Huzhou Institute.
S. Cai is with the Institute of Cyber-Systems & Control, College of Control Science & Engineering, Zhejiang University, Hangzhou, 310027 China.
Corresponding authors: shengze_cai@zju.edu.edu, cxu@zju.edu.cn
Abstract

Flow visualization technologies such as particle tracking velocimetry (PTV) are broadly used in understanding the all-pervasiveness three-dimensional (3D) turbulent flow from nature and industrial processes. Despite the advances in 3D acquisition techniques, the developed motion estimation algorithms in particle tracking remain great challenges of large particle displacements, dense particle distributions and high computational cost. By introducing a novel deep neural network based on recurrent Graph Optimal Transport, called GotFlow3D, we present an end-to-end solution to learn the 3D fluid flow motion from double-frame particle sets. The proposed network constructs two graphs in the geometric and feature space and further enriches the original particle representations with the fused intrinsic and extrinsic features learnt from a graph neural network. The extracted deep features are subsequently utilized to make optimal transport plans indicating the correspondences of particle pairs, which are then iteratively and adaptively retrieved to guide the recurrent flow learning. Experimental evaluations, including assessments on numerical experiments and validations on real-world experiments, demonstrate that the proposed GotFlow3D achieves state-of-the-art performance against both recently-developed scene flow learners and particle tracking algorithms, with impressive accuracy, robustness and generalization ability, which can provide deeper insight into the complex dynamics of broad physical and biological systems.

Index Terms:
Graph Neural Networks, Recurrent Neural Networks, Optimal Transport, Fluid Motion Estimation, Point Clouds.

1 Introduction

Visualization and measurement of the turbulent flow can reveal dynamic behaviors of great complexity in fluid mechanics, biological locomotion and soft materials, which has long been recognized by Leonardo da Vinci and carried out to track fine particles in turbulent flows of the aortic heart valve in the 1400s [1]. Among the diverse flow visualization technologies, particle tracking velocimetry (PTV) is commonly used for quantitative and non-intrusive measurement of global velocity field, that can provide a deeper perception into the complex flow phenomena [2]. In particular, tracking the suspended particles embedded within a fluid medium (e.g., liquid, gas or other material with continuous deformation) in PTV has been playing a significant role in elucidating natural phenomena formation processes [3, 4], hydrodynamics of biological locomotion [5], tissue-scale or intracellular flow in organisms [6, 7, 8, 9], biological collective phenomena or transport processes [10, 11], dynamic mechanisms of liquids [12] as well as atomic rearrangement motion in solids [13].

The PTV algorithms rely on the detection and tracking of individual particles between consecutive particle images, resulting in Lagrangian velocity vectors that can considerably represent the local motions of the fluid flow. In general, the particles are detected and localized with sub-pixel spatial coordinates in different ways depending on various experimental setups (e.g., defocusing PTV [14] or traction force microscopy [15]). Compared to other developed flow analysis techniques, e.g., particle image velocimetry (PIV) [16], PTV shows the great superiority for providing flow fields with substantially increased spatial resolution as well as improved accuracy by avoiding bias errors due to the spatial averaging in PIV [2]. Moreover, due to the individual particle tracking manner, PTV demonstrates better performance in measuring the flows with strong velocity gradients and inhomogeneity of particle spatial distribution due to Saffman effect [16], especially in a three-dimensional (3D) spatial domain.

Despite the aforementioned advantages, the ever-growing PTV demands in scientific researches and engineering applications pose great challenges to classical particle tracking algorithms, which need to be further generalized to tricky flow scenarios involving diverse motion patterns, large dynamic velocity ranges, dense particle distribution, etc. In addition, the classical approaches commonly involve complicated iterative algorithms with adjustable parameters or introduce additional temporal smoothness from multiple frames [17, 2], which are generally time-consuming in real experiments. To overcome these limitations, we incorporate deep neural networks developed for 3D point clouds to solve the fundamental problem of learning the 3D flow motion from double-frame particle sets in PTV experiments.

Deep learning in fluid mechanics has attracted increasing attention and developed rapidly over the past few years [18]. In the community of flow measurement, PIV has thrived due to the development of deep neural networks, especially convolutional neural networks (CNN) [19, 20, 21, 22]. However, due to the difficulty of applying CNN-based approaches to cope with the unstructured grid particle data, PTV has until recently benefited few from deep learning and conversely still remains room for improvement. A recent work on learning-based PTV has been proposed by [23], where a recurrent neural network is applied to deal with the particle linking issue. However, multiple frames of particle trajectories are required as input and a pre-trained model is specialized for one flow case. On the other hand, pioneered by the recent PointNet-based backbone [24], numerous approaches have been put forwards to estimate scene flow (mostly rigid motion) from unstructured 3D point clouds [25, 26, 27, 28]. Inspired by the scene flow learning network, a preliminary investigation has been conducted to estimate 2D flow in PTV analysis, which shows a comparable performance with classical state-of-the-art PTV approaches on accuracy, robustness and efficiency [29]. In this work, we present how to extract spatial geometric information from the basic PTV data (merely 3D particle positions of two frames) with graph neural network (GNN) and generalize the scene flow to complex non-rigid flow motion (more common in physical phenomena) using our proposed optimal-transport guided RNN framework. The designed deep learning model is called GotFlow3D. The learnt flow by GotFlow3D can directly indicates the physical dynamics or be implemented as a-priori knowledge of the flow field into predictor-corrector schemes [2] to boost the particle tracking. Our propose method is a new paradigm in PTV technology to recover the complex particle motion, which plays a significant role for further dynamic behavior analysis in physical and biological systems.

Refer to caption
Figure 1: Overview of GotFlow3D for learning flow motion from particle sets. (a) Setup and the flowchart of GotFlow3D in particle tracking experiments. Double-frame particle distributions are obtained by tomographic configuration and then processed by GotFlow3D to output a 3D vector field. (b) Illustration of the synthetic dataset generation. (c) Architecture of the proposed network, GotFlow3D, which consists of 3 key components: (1) graph-based feature extractor, (2) optimal transport, (3) recurrent GRU. (d) Visualization of SDGNN, marked in purple box in (c), which fuses the static spatial graph and the dynamic feature graph to extract the geometric information of particle sets. (e) Illustration of the adaptive point-voxel optimal transport (AdaPT) plan retrieval in the correspondence searching. The first raw of (e) depicts the searching of dynamically reduced nearest neighbors (red balls) of the investigated point (the yellow ball) in the point-based transport plan retrieval, while the bottom one represents the adaptively-deformable multi-scale cuboids in the voxel-based retrieval.

2 Methods

2.1 Problem Setup

To better understand the principles of the proposed model, we first describe the formulation of the complex flow learning problem. The whole framework is shown in Figure 1(a). The proposed model consumes, as inputs, two unorganized consecutive 3D particle sets 𝒫\mathcal{P}, 𝒬\mathcal{Q} and produces the dense vector field 𝒱\mathcal{V}. The particle sets can be presented as the source set 𝒫={piDin}i=1n1\mathcal{P}=\{p_{i}\in\mathbb{R}^{D_{in}}\}_{i=1}^{n_{1}} with n1n_{1} particles and the target set 𝒬={qjDin}j=1n2\mathcal{Q}=\{q_{j}\in\mathbb{R}^{D_{in}}\}_{j=1}^{n_{2}} with n2n_{2} particles, where each particle pip_{i} (or qjq_{j}) suspends in various flow media and faithfully follows the media flow motion. Note that the attributes of each particle pip_{i} (or qjq_{j}) may contain the 3D coordinate xi3x_{i}\in\mathbb{R}^{3} together with some additional state features ficf_{i}\in\mathbb{R}^{c} (e.g., intensity or color information). To enhance the generalization for various flow measurement tasks, we only consider the 3D coordinate (i.e., Din=3D_{in}=3) in this work.

Each element viv_{i} in 𝒱={vi3}i=1n1\mathcal{V}=\{v_{i}\in\mathbb{R}^{3}\}_{i=1}^{n_{1}} represents the predicted flow motion corresponding to each particle pip_{i}. Given the source set 𝒫\mathcal{P} and the target set 𝒬\mathcal{Q}, the iteratively updated flow 𝒱t\mathcal{V}_{t} (in the tt-th iteration) gradually translates the source 𝒫\mathcal{P} to 𝒫t\mathcal{P}^{\prime}_{t}, which is expected to approach the target 𝒬\mathcal{Q}. Before the iterative process, we first employ the GNNs to extract the geometric pointwise feature 𝒢\mathcal{F}_{\mathcal{G}}, which is adopted to assign the optimal transport plan 𝒯\mathcal{T}. Then, the recurrent neural network updates the flow by adaptively seeking the neighbors of 𝒫t\mathcal{P}^{\prime}_{t} in 𝒬\mathcal{Q} and following the guidance of the corresponding optimal transport plan.

To promote the learning of complex non-rigid fluid flow beyond scene flow, the properties of the flow motion in fluids should be taken into consideration. For examples, multi-scale eddies and chaotic property changes frequently exist in the complex fluid flow, which poses a great challenge to flow estimators. The challenge encourages an emphasis on both the extraction of large-scale motion fields (e.g., laminar flow) and the refinement of small-scale structures (e.g., vortex). In addition, the spatial local structures assembled with neighboring particles show approximately motion invariance and geometric similarities due to the local smooth fluid motion. All these properties are considered while designing the deep neural network.

2.2 GotFlow3D

This section introduces the proposed method used to learn the complex flow motion from two consecutive 3D particle sets. As shown in Figure 1(c), this model is designed to extract the geometric patterns of the 3D particle sets in the flow media, make the optimal transport plan that encodes the point correspondences, and iteratively undate the flow motion by adaptively retrieving the corresponding transport plan between the two particle sets. In the following, we present the architecture of GotFlow3D, which is constructed by three essential modules performing (1) graph-based feature extraction; (2) adaptive optimal transport plan lookup; and (3) iterative flow updating. Eventually, we also describe the implementation and training details of GotFlow3D.

2.2.1 Graph-based Feature Extraction

To ensure an accurate flow estimation, it is of great importance to capture reliable pointwise features, which are supposed to be discriminative enough for exact correspondences searching. Considering that the 3D coordinate is the only available input, the key point is to learn pointwise features representing the local geometry information. To achieve the effective feature exploration on the unstructured and irregular input format, we construct two kinds of geometric graphs on the particle sets and employ the static-dynamic fusion GNN (SDGNN) to derive the deep features of each particle. Specifically, we construct the static graph 𝒢s\mathcal{G}^{s} on 3D spatial space and the dynamic graph 𝒢d\mathcal{G}^{d} on high-dimensional feature space.

We first define the static graph 𝒢s=(𝒱er,s)\mathcal{G}^{s}=(\mathcal{V}_{er},\mathcal{E}^{s}), where the vertices 𝒱er={1,2,,n1(n2)}\mathcal{V}_{er}=\{1,2,...,n_{1}(n_{2})\} are composed of the entire given particle set 𝒫\mathcal{P} (or 𝒬\mathcal{Q}). The edges s|𝒱er|×|𝒱er|\mathcal{E}^{s}\subseteq\lvert\mathcal{V}_{er}\rvert\times\lvert\mathcal{V}_{er}\rvert are yielded by connecting each particle (vertex) pip_{i} to its kk-nearest spatial neighbors pik𝒩ks(pi)\forall p_{i_{k}}\in\mathcal{N}_{k}^{s}(p_{i}) with the 3D Euclidean distance metric. The edges combining pip_{i} and its neighbors form local geometric structures and further generate the associated vertex feature f~i=[pikpi;rik;θik;φik]\tilde{f}_{i}=[p_{i_{k}}-p_{i};r_{i_{k}};\theta_{i_{k}};\varphi_{i_{k}}], f~ik×6\tilde{f}_{i}\in\mathbb{R}^{k\times 6}, where pikpip_{i_{k}}-p_{i} denotes the 3D directional vector. In addition, rik,θikr_{i_{k}},\theta_{i_{k}} and φik\varphi_{i_{k}} represent the radial distance, azimuthal angle and polar angle in spherical coordinates, respectively, which supplement the relative spatial relationship of the neighboring vertices from another perspective. Such spherical representations further provide rotation invariant correspondences, which are advantageous in the estimation of complex motions and larger deformations [30]. The static graph for a particle set needs to be computed only once before training and is kept for the following retrieval.

As shown in Figure 1(d), given 𝒢s\mathcal{G}^{s} and 𝒫\mathcal{P}, we employ the SDGNN to encode the local geometric characteristics to generate the distinctive feature descriptor for each particle. We denote the extracted pointwise feature of vertex pip_{i} at the β\beta-th layer as giβDβg_{i}^{\beta}\in\mathbb{R}^{D_{\beta}}. In the deep hierarchical SDGNN, graph convolution operations are performed on the local geometric structure of each vertex pip_{i} to map the vertex feature giβ1g_{i}^{\beta-1} of the previous layer to a new feature giβg_{i}^{\beta} in the subsequent layer. Inspired by the PointNet++ framework [31], we propose a GeoSetConv layer to further embed the vertex features of the static graph to a high-dimensional feature space to better characterize the local geometric structures:

giβs=maxj𝒩ks(pi){ϕβ(f~i,gjβ1sgiβ1s)},\displaystyle{{}^{s}g_{i}^{\beta}}=\mathop{\max}_{j\in\mathcal{N}_{k}^{s}(p_{i})}\ \left\{\phi^{\beta}\left(\tilde{f}_{i},\;{{}^{s}g_{j}^{\beta-1}}-{{}^{s}g_{i}^{\beta-1}}\right)\right\}, (1)

here giβs{{}^{s}g_{i}^{\beta}} denotes the feature extracted from the static graph at the β\beta-th layer, and 𝒩sk(pi)\mathcal{N}_{s}^{k}(p_{i}) depicts the kk-nearest neighbors of vertex pip_{i} in the static graph. In addition, ϕβ\phi^{\beta} is the β\beta-th GeoSetConv layer, which is implemented as multi-layer perceptrons (MLPs). And max\max denotes the max-pooling operation, which aggregates features over the kk neighbors. The gi0s{}^{s}g_{i}^{0} is initialized by the 3D coordinate of pip_{i} in the first GeoSetConv layer.

To further enrich the feature representations describing the detailed topological information, we explore the feature learning on the dynamic feature graph, as EdgeConv did in [32]. It is the first attempt at applying the dynamic graph CNN to extract features in the flow learning task. The following experiments also demonstrate that the introduction of the dynamic feature graph contributes to the improvement of the accuracy. The dynamic graph 𝒢d\mathcal{G}^{d} is constructed and updated by aggregating the kk-nearest neighbors on the high-dimensional feature space generated at each layer. Specifically, for each layer, we build the dynamic graph 𝒢d=(𝒱er,d)\mathcal{G}^{d}=(\mathcal{V}_{er},\mathcal{E}^{d}), where 𝒢d\mathcal{G}^{d} shares the same vertices as that of 𝒢s\mathcal{G}^{s} and the edges d\mathcal{E}^{d} link feature level neighbors to form local structures in the feature manifold. Then, we formulate the update of the vertex feature giβd{{}^{d}g_{i}^{\beta}} of pip_{i} on the dynamic graph at the β\beta-th layer as

giβd=maxj𝒩kf(pi){φβ(gjβ1d,gjβ1dgiβ1d)},\displaystyle{{}^{d}g_{i}^{\beta}}=\mathop{\max}_{j\in\mathcal{N}_{k}^{f}(p_{i})}\ \left\{\varphi^{\beta}\left({{}^{d}g_{j}^{\beta-1}},\;{{}^{d}g_{j}^{\beta-1}}-{{}^{d}g_{i}^{\beta-1}}\right)\right\}, (2)

here 𝒩kf(pi)\mathcal{N}_{k}^{f}(p_{i}) depicts the index of the kk-nearest neighbors of pip_{i} in the feature space. And φβ\varphi^{\beta} denotes the β\beta-th EdgeConv layer, which is also implemented as MLPs. Different from the static graph, whose receptive field is limited in the local neighborhood and hinders the acquirement of long-term dependencies, the dynamic graph enlarges the receptive field to the whole input set. The dynamic graph learns to construct a special receptive field by assigning more attention to the neighbors with similar feature (flow) properties.

Then, the proposed hierarchical SDGNN incorporates the extrinsic geometric topology and the intrinsic feature correspondences by fusing the learnt features from the two types of graphs. More precisely, we further concatenate the hierarchical features of these graphs from different layers of the GNN. Finally, several MLPs with a skip connection are employed to derive the final deep pointwise feature

𝒢=ψ(β(gβs,gβd)),\displaystyle\mathcal{F}_{\mathcal{G}}={\psi}\left(\mathop{\oplus}_{\beta}\left({{}^{s}g^{\beta}},{{}^{d}g^{\beta}}\right)\right), (3)

where ψ{\psi} depicts the MLPs and β(,)\mathop{\oplus}\limits_{\beta}(\cdot,\cdot) denotes the concatenation of features from different graphs and different layers. The extracted pointwise feature is adopted for the following optimal transport planning to infer the particle correspondences via the feature similarity. Furthermore, we employ another feature encoder to extract the context feature 𝒞\mathcal{F}_{\mathcal{C}} of 𝒫\mathcal{P}, which provides additional context information to the flow update. Such context feature encoder simply consists of several SetConv layers from PointNet++, which only operates on the spatial static graph.

2.2.2 Adaptive Point-Voxel Optimal Transport

Optimal transport. Optimal transport theory [33] guides the transportation and allocation between two distributions. In this work, we regard the correspondences searching of two particle sets as the transportation and allocation of the particle mass, where each particle pip_{i} (qjq_{j}) is set to have a mass of n11n_{1}^{-1} (n21n_{2}^{-1}). Then, the cost has to be paid to transport each unit of mass from pip_{i} to qjq_{j}. Therefore, the key issue of optimal transport is to seek a transport plan transforming a source particle distribution 𝒫\mathcal{P} to a target particle distribution 𝒬\mathcal{Q} whilst minimizing the transport cost [34]. In more detail, the computation of a transport plan 𝒯+n1×n2\mathcal{T}\in\mathbb{R}_{+}^{n_{1}\times n_{2}} between two particle sets can be formulated as

𝒯=argminU+n×ni,j=1CijUij\displaystyle\mathcal{T}=\mathop{\arg\min}\limits_{U\in\mathbb{R}_{+}^{n\times n}}\sum_{i,j=1}C_{ij}U_{ij} (4)
s.t.U1n1=𝝁p,UT1n2=𝝁q,\displaystyle s.t.\quad U\textbf{1}_{n_{1}}=\bm{\mu}_{p},\ U^{T}\textbf{1}_{n_{2}}=\bm{\mu}_{q},

where CijC_{ij} is the transport cost from particle pi𝒫p_{i}\in\mathcal{P} to qj𝒬q_{j}\in\mathcal{Q}, and UijU_{ij} denotes the assigned quantity of mass to be transported from pip_{i} to qjq_{j}. In addition, we define 𝝁p=1n11n1\bm{\mu}_{p}=\frac{1}{n_{1}}\textbf{1}_{n_{1}} and 𝝁q=1n21n2\bm{\mu}_{q}=\frac{1}{n_{2}}\textbf{1}_{n_{2}} in the equality constraints. The constraint manifests that the mass of each particle pip_{i} from the source distribution is supposed to be completely transported to the target and each particle qjq_{j} of the target distribution is expected to be filled up with the transported mass.

However, due to the measurement noise in real experiments such as 3D reconstruction errors and object occlusions, the map reflecting the correspondences between two particle sets is neither injective nor surjective. Therefore, the mass constraints cannot hold ideally. To cope with this problem, it is preferred to apply the constraint relaxation to solve the optimal transport problem. We follow the formulation proposed by FLOT [26] to employ the Sinkhorn algorithm [35] to efficiently optimize the transport plan with an entropic regularization and a mass regularization:

𝒯=argminU+n×n\displaystyle\mathcal{T}=\mathop{\arg\min}\limits_{U\in\mathbb{R}_{+}^{n\times n}} [i,j=1NCijUij+ϵUij(logUij1)]\displaystyle\left[\sum_{i,j=1}^{N}C_{ij}U_{ij}+\epsilon U_{ij}(\log U_{ij}-1)\right] (5)
+λKL(U1n,𝝁p)+λKL(UT1n,𝝁q),\displaystyle+\lambda KL(U\textbf{1}_{n},\bm{\mu}_{p})+\lambda KL(U^{T}\textbf{1}_{n},\bm{\mu}_{q}),

here KLKL represents the KLKL-divergence. In addition, ϵ\epsilon and λ\lambda denote the entropic and the mass regularization parameter, which are also learnt through the network training. A small ϵ\epsilon leads to a sparse transport plan, while a high λ\lambda tends to reinforce the effect of the mass constraint. Then, we define the transport cost between all pairs as the dissimilarity of the extracted pointwise feature via the cosine similarity:

Cij=1𝒢(pi)T𝒢(qj)𝒢(pi)2𝒢(qj)2,\displaystyle C_{ij}=1-\frac{\mathcal{F}_{\mathcal{G}}(p_{i})^{T}\mathcal{F}_{\mathcal{G}}(q_{j})}{\|\mathcal{F}_{\mathcal{G}}(p_{i})\|_{2}\|\mathcal{F}_{\mathcal{G}}(q_{j})\|_{2}}, (6)

where 2\|\cdot\|_{2} denotes the l2l_{2}-norm. A small cost between two particles generally corresponds to a high plan, which implies a latent possibility of particle correspondences. The transport plan is constructed only once and is kept available for the following retrieval, which conduces to avoid numerous repeat matrix calculations.

Adaptive point-voxel optimal transport. In this paper, we propose a novel approach to adaptively retrieve the optimal transport plan for the iterative flow updating. For each iteration tt, given the currently estimated flow 𝒱t\mathcal{V}_{t}, the source 𝒫\mathcal{P} is translated to 𝒫t\mathcal{P}^{\prime}_{t}, where Pt=P+𝒱tP^{\prime}_{t}=P+\mathcal{V}_{t}. The estimated flow is initialized at 𝒱0=0\mathcal{V}_{0}=0 and accumulated by the previous flow updates as

𝒱t=l=1tΔfl,\mathcal{V}_{t}=\sum_{l=1}^{t}\Delta f_{l}, (7)

where ll denotes the index of the iteration. Next, we can consult the transport plan between 𝒫t\mathcal{P}^{\prime}_{t} and 𝒬\mathcal{Q} to guide the flow update Δft+1\Delta f_{t+1} of the next iteration. More specifically, the consultation is adopted by searching the neighbors of the warped source 𝒫t\mathcal{P}^{\prime}_{t} in the target 𝒬\mathcal{Q} and retrieving the corresponding transport plan 𝒯t\mathcal{T}_{t}. This iterative transport plan consultation contributes to lowering the difficulty of long range motion estimation and promote the refinement of small scale motion.

An essential step in the above-mentioned algorithm is: how to efficiently retrieve abundance and reliable information (𝒯t\mathcal{T}_{t}) from the full transport plan (𝒯\mathcal{T}) through the neighborhood relation reasoning of the unstructured particle set. To address this issue, the proposed approach further generates the adaptive point-voxel optimal transport (AdaPT) plan 𝒯t\mathcal{T}^{*}_{t} during each consultation based on the original point-voxel operation [28]. More precisely, we integrate the captured point-based transport plan 𝒯p\mathcal{T}^{*}_{p} and the voxel-based transport plan 𝒯v\mathcal{T}^{*}_{v} to offer the long-range and small-scale correspondence of particle pairs in the consultation:

𝒯t=𝒯p(Pt,Q)+𝒯v(Pt,Q).\displaystyle\mathcal{T}^{*}_{t}=\mathcal{T}^{*}_{p}(P^{\prime}_{t},Q)+\mathcal{T}^{*}_{v}(P^{\prime}_{t},Q). (8)

First, we present the generation of the point-based transport plan 𝒯p\mathcal{T}^{*}_{p} using a dynamic ktk_{t}-nearest neighbors searching, where the number of the located neighbors are gradually reduced as the iteration progressed. Such construction adaptively narrows the search region to refine the fine-grained correspondence along with the increasing of iterations. From the deep point of view, the learning of the subsequent smaller residual flow amounts to cope with particles with relatively larger spatial distance. Accordingly, we dynamically reduce the searched neighboring particles to adapt to sparsely seeded regions, in which small number of neighbors perform better as noted in [30]. The computation of 𝒯p\mathcal{T}^{*}_{p} and the attenuation of ktk_{t} are formulated as:

𝒯p=maxj𝒩Qkt(Pt)Φ1(𝒯(Qj),QjPt),\displaystyle\mathcal{T}^{*}_{p}=\mathop{\max}_{j\in\mathcal{N}_{Q}^{k_{t}}(P^{\prime}_{t})}\Phi_{1}(\mathcal{T}(Q_{j}),\ Q_{j}-P^{\prime}_{t}), (9)
kt=k0ηt,k_{t}=k_{0}-\eta t, (10)

here 𝒩Qkt(Pt)\mathcal{N}_{Q}^{k_{t}}(P^{\prime}_{t}) denotes the index set of the ktk_{t}-nearest neighbors of PtP^{\prime}_{t} in QQ, and 𝒯(Qj)\mathcal{T}(Q_{j}) is the corresponding retrieved transport plan between QjQ_{j} and PtP^{\prime}_{t}. In addition, we further combine the retrieved 𝒯(Qj)\mathcal{T}(Q_{j}) and the relative spatial relation QjPtQ_{j}-P^{\prime}_{t} of the corresponding neighbors to deep encode the final plan 𝒯p\mathcal{T}^{*}_{p}, where the local spatial relation plays a similar role as the geometric attention weight. The deep encoding is performed by the MLPs Φ1\Phi_{1} with a max\max operation over the ktk_{t} dimension. Furthermore, k0k_{0} is the initial value of ktk_{t}, while η\eta is set to a constant value.

To further cope with the large-scale motion embedded in the long-range correspondence of transport plan, we innovatively construct adaptively deformable multi-scale cuboids, which compartmentalize the geometric space containing particles into pyramid voxels to yield the voxel-based transport plan 𝒯v\mathcal{T}^{*}_{v}. Specifically, each constructed cuboid is centered at each particle of PtP^{\prime}_{t} and collects the enveloped neighbors in QQ. As depicted in Figure 1(e), the cuboid is composed of d3d^{3} identical sub-cuboids, whose length (e1\vec{e}_{1}), width (e2\vec{e}_{2}) and height (e3\vec{e}_{3}) vectors are adaptively defined according to the several recent flow updates as:

{e1=(18𝒮(|Δfl|)+18cosΔfl,Δfl1+1)rl1,e2=(18𝒮(|Δfl1|)+0.9)rl2,e3=0.9rl3,\displaystyle\begin{cases}\vec{e}_{1}=(\frac{1}{8}\mathcal{S}(\lvert\Delta f_{l}\rvert)+\frac{1}{8}\cos\left\langle\Delta f_{l},\Delta f_{l-1}\right\rangle+1)r\vec{l}_{1}\vspace{1ex},\\ \vec{e}_{2}=(\frac{1}{8}\mathcal{S}(\lvert\Delta f_{l-1}\rvert)+0.9)r\vec{l}_{2}\vspace{1ex},\\ \vec{e}_{3}=0.9r\vec{l}_{3},\end{cases} (11)

where 𝒮(|Δfl|)\mathcal{S}(\lvert\Delta f_{l}\rvert) and 𝒮(|Δfl1|)\mathcal{S}(\lvert\Delta f_{l-1}\rvert) denote the normalized length of the current updated flow and the last updated flow over all particles in PtP^{\prime}_{t}, respectively. Such normalization encourages a longer-range correspondence search when large variations appear on the flow. And cosΔfl,Δfl1cos\left\langle\Delta f_{l},\Delta f_{l-1}\right\rangle calculates the cosine similarity between Δfl\Delta f_{l} and Δfl1\Delta f_{l-1}. Such similarity calculation also reinforces a longer-range correspondence search when recent flow updates keep consistent but prompts a smaller-range correspondence search when they remain inconsistent. Moreover, l1\vec{l}_{1}, l2\vec{l}_{2} and l3\vec{l}_{3} are unit orthogonal basis vectors of the three principal orientations used to guide the search direction of the correspondence, which are established as:

{l1=Δfln,l2=Δfl1Δfl1,ΔflΔfl,ΔflΔfln,l3=l1l2n,\displaystyle\begin{cases}\vec{l}_{1}=\|\Delta f_{l}\|_{n}\vspace{1ex},\\ \vec{l}_{2}=\|\Delta f_{l-1}-\frac{\left\langle\Delta f_{l-1},\Delta f_{l}\right\rangle}{\left\langle\Delta f_{l},\Delta f_{l}\right\rangle}\Delta f_{l}\|_{n}\vspace{1ex},\\ \vec{l}_{3}=\|\vec{l}_{1}\otimes\vec{l}_{2}\|_{n},\end{cases} (12)

where ,\left\langle\cdot,\cdot\right\rangle denotes the inner product of two vectors, while \otimes represents the outer product; n\|\cdot\|_{n} means that the vector is normalized to unit length. Furthermore, we define a basic scale parameter rr, which controls the size deformation of the cuboid:

r=2s1dmα,\displaystyle r=2^{s-1}d_{m}\alpha, (13)

where dmd_{m} denotes the mean distance between each particle and its kdk_{d}-nearest neighbors in the input set, which is utilized to obtain the local particle concentration information. And α\alpha is a learnable scale parameter optimized with the network training, which contributes to an adaptive process of particle sets with various particle densities. In addition, ss is the index of the pyramid iteration, which gradually doubles the length size of the sub-cuboid to retrieve the multi-scale voxel-based plan. The adaptive variation of the multi-scale cuboid size rr makes it possible to cover farther desired particles.

As illustrated above, the constructed cuboid presents the powerful deformability to adapt to current flow updates and input particle densities, while showing puissant capabilities to capture long-range correspondences with the multi-scale strategy. Then, the cuboid partitions the enveloped neighboring particles in QQ into d3d^{3} sub-cuboids, which indicate the relative direction of the neighbours to the central particle in PtP^{\prime}_{t}. Furthermore, the neighboring particles and the corresponding transport plan in each sub-cuboid are aggregated and averaged for the final calculation of 𝒯v\mathcal{T}^{*}_{v}:

𝒯v=Φ2(s(b(𝒯¯(𝒩Qbs(Pt))))),\displaystyle\mathcal{T}^{*}_{v}=\Phi_{2}\left(\mathop{\uplus}_{s}\left(\mathop{\uplus}_{b}\left(\overline{\mathcal{T}}\left({}^{s}\mathcal{N}_{Q}^{b}\left(P^{\prime}_{t}\right)\right)\right)\right)\right), (14)

here 𝒩Qbs(Pt){}^{s}\mathcal{N}_{Q}^{b}\left(P^{\prime}_{t}\right) denotes the neighboring particles of PtP^{\prime}_{t} in QQ enveloped in the bb-th sub-cuboid for the ss-th pyramid iteration. And 𝒯¯\overline{\mathcal{T}} is the average over the transport plan of particles in the bb-th sub-cuboid. Moreover, s\mathop{\uplus}_{s} and b\mathop{\uplus}_{b} denote the concatenation over all pyramid iterations and all sub-cuboids, respectively. There followed another MLPs Φ2\Phi_{2}, which is used to implement the similar deep encoding for the final plan 𝒯v\mathcal{T}^{*}_{v}.

GRU flow refinement. Inspired by RAFT [36], we finally employ a recurrent GRU to estimate the flow update Δft+1\Delta f_{t+1}, which mimics the iterative strategy in the optimization algorithm. The applied GRU consumes the integrated AdaPT plan 𝒯t\mathcal{T}^{*}_{t}, the current flow estimation 𝒱t\mathcal{V}_{t}, the context feature 𝒞\mathcal{F}_{\mathcal{C}} and the hidden state ht1h_{t-1} transmitted from the last iteration, while yielding a new hidden state hth_{t} encoding the future flow information:

zt=σ(Conv1d([ht1,xt],Wz)),z_{t}=\sigma(Conv_{1d}([h_{t-1},x_{t}],W_{z})), (15)
rt=σ(Conv1d([ht1,xt],Wr)),r_{t}=\sigma(Conv_{1d}([h_{t-1},x_{t}],W_{r})), (16)
ht~=tanh(Conv1d([rtht1,xt],Wh)),\tilde{h_{t}}=\tanh(Conv_{1d}([r_{t}\odot h_{t-1},x_{t}],W_{h})), (17)
ht=(1zt)ht1+ztht~,h_{t}=(1-z_{t})\odot h_{t-1}+z_{t}\odot\tilde{h_{t}}, (18)

where xtx_{t} denotes the concatenated feature of 𝒯t\mathcal{T}^{*}_{t}, 𝒱t\mathcal{V}_{t} and 𝒞\mathcal{F}_{\mathcal{C}}. And \odot is the element-wise product. In addition, Wz(r,h)W_{z(r,h)} denotes the weights of convolutional layers in different steps. Then, the hth_{t} is passed through a SetConv layer and several convolutional layers to predict the flow update Δft+1\Delta f_{t+1}.

2.2.3 Implementation Details

In this framework, there are some predefined parameters in the graph construction and the AdaPT plan retrieval process. The number of the searched neighboring particles used to construct the static graph and the dynamic graph is set as k=32k=32. For the AdaPT plan retrieval, the dynamic ktk_{t} is initialized as k0=32k_{0}=32, while η\eta is fixed to 2. And when the iteration exceeds 8, the ktk_{t} is fixed to 16. The resolution of the cuboid dd and the number of the pyramid iteration ss are both set as 3. In addition, the kdk_{d} is set as 3 to calculate the mean local particle distance. Moreover, the scale parameter α\alpha is initialized as 0.5 and has a special learning rate, which is 0.1 times the global learning rate, to enhance the training stability. The dimension of the final pointwise feature extracted by the hierarchical GNN is set as 128. Furthermore, the ϵ\epsilon and λ\lambda used to control the two regularizations are initialized as 0 and 1, respectively. And the flow is iteratively updated for 8 iterations.

The network training are performed with supervised loss using the entire estimated flow sequence {𝒱1e,𝒱Te}\{\mathcal{V}^{e}_{1},...\mathcal{V}^{e}_{T}\}:

=1Ni=1Nt=1Tγt𝒱te𝒱gt,\mathcal{L}=\frac{1}{N}\sum_{i=1}^{N}\sum_{t=1}^{T}\gamma_{t}\|\mathcal{V}^{e}_{t}-\mathcal{V}^{gt}\|, (19)

where 𝒱te\mathcal{V}^{e}_{t} denotes the flow estimation of the tt-th iteration, while 𝒱gt\mathcal{V}^{gt} is the ground truth flow. In addition, \|\cdot\| denotes the l1l_{1} distance, and TT is the total number of flow update iterations. Moreover, NN denotes the total number of the particles. Mini-batch training is employed with a batch size of 16. Then, we further impose an exponentially increased weight γt\gamma_{t} to the estimated flow sequence:

γt=μTt,\gamma_{t}=\mu^{T-t}, (20)

where μ\mu is set to 0.8. The network is trained on six TITIAN XP GPUs for 40 epoches using Adam optimizer [37] with a initial learning rate of 0.001.

2.3 Datasets

2.3.1 Synthetic training datasets

Training data with ground-truth information plays an indispensable role in the network optimization with supervised learning strategies. However, the difficulty of recovering highly accurate ground-truth flow from real data enforces us turn to the simulated data from computational fluid dynamics (CFD), which is capable of generating various flow cases representing the true fluid mechanical characteristics by solving the Navier–Stokes equations [16].

To perform the flow supervision, we create the first synthetic 3D fluid flow dataset named FluidFlow3D, which is characterized by two consecutive particle sets 𝒫\mathcal{P}, 𝒬\mathcal{Q} and the corresponding ground-truth flow set 𝒱gt\mathcal{V}^{gt}. In particular, we first generate a particle set 𝒫\mathcal{P}, which comprises more than two thousand randomly distributed particles in the 3D observation volume. The observation volume is also defined in the random location of the CFD computational domain. Next, we use the public CFD data from Johns Hopkins Turbulence Database (JHTDB) [38], which is proven to provide physically correct simulated flow structures, to generate the ground-truth flow 𝒱gt\mathcal{V}^{gt}. This is achieved by applying the built-in functions to query the JHTDB database with the second order accurate Runge-Kutta integration scheme in time and the fourth order Lagrange interpolation in space, which are executed on the database nodes. The flow structures queried from JHTDB includes the incompressible isotropic turbulent flow, the incompressible magneto-hydrodynamic (MHD) turbulence, the fully developed turbulent channel flow and the transitional boundary layer flow. Therefore, we can get the next frame 𝒬0\mathcal{Q}_{0} of the given particle set 𝒫\mathcal{P} using the integrated trajectory with a designated time interval, subsequently get the the corresponding ground-truth flow 𝒱gt=𝒬0𝒫\mathcal{V}^{gt}=\mathcal{Q}_{0}-\mathcal{P}. The sketch of this generation procedure is illustrated in Figure 1(b). In addition to the data from JHTDB, several flow benchmarks, including the uniform flow and the Beltrami flow [39], are also generated to increase the diversity of the training dataset. Eventually, the synthetic training dataset consists of six categories of typical flow cases.

Furthermore, to enhance the generalization capabilities of the trained network for realistic experimental configurations, we augment the above flow cases with various flow conditions. The measurement performance in PTV experiments is highly limited by large dynamic velocity ranges and high particle densities. These two factors can be characterized by the particle displacement ratio ρ\rho, which has been widely used in the PTV community [40] and can comprehensively depict the flow conditions:

ρ=daumaxΔt,\rho=\frac{d_{a}}{u_{\max}\Delta t}, (21)

here dad_{a} and umaxΔtu_{\max}\Delta t denote the average distance between particles in the whole set (e.g., 𝒫\mathcal{P}) and the maximum displacement between two input sets, respectively. It is reported that the classical flow estimators can easily achieve satisfactory performance for flow cases with ρ1\rho\gg 1, while facing more challenges when ρ\rho gets smaller [40]. In our training dataset, we change the characteristic observation volume to vary dad_{a} (the number of particles remains constant) and alter the time interval Δt\Delta t to obtain different maximum displacements. As a result, this important parameter ρ\rho in the synthetic dataset ranges from 0.350.35 to 6.586.58, much beyond the normal range of real PTV experiments. Moreover, to mimic the real-world applications, we add some noises to create more realistic cases, which mainly arise from the physical motion of particles in and out of the field of view, the measurement noise and the occlusion. To this end, we shuffle the particles in 𝒬0\mathcal{Q}_{0} to obtain the second set 𝒬\mathcal{Q}, which directly disorders the correspondence. Then, we randomly replace n%n\% of the particles in 𝒬\mathcal{Q} with new random particles. Such noises eliminates n%n\% of the correspondence of 𝒫\mathcal{P}, while introducing new isolated particles in 𝒬\mathcal{Q} with no correspondence. Here we set nn with 1%1\% level of noise for this dataset. Finally, we generate a synthetic dataset with 16K training samples and 1.6K validation samples. A table summarizing the synthetic dataset is given in Table III.

2.3.2 Testing Dataset

A family of synthetic FluidFlow3D-derived datasets, a public turbulent cylinder wake flow dataset (denoted by CylinderFlow) [41] and a real-world dataset (denoted by SphericalIndentationFlow) across different domains are adopted to conduct the experimental evaluations.

FluidFlow3D-family. Apart from the aforementioned FluidFlow3D, we further transform FluidFlow3D to a new dataset with normalized spatial scale (denoted by FluidFlow3D-norm) to compute the unified evaluation metric among flow cases with different observation volumes. The FluidFlow3D-norm is used for the comparison with the state-of-the-art methods and ablation studies. To analyze the robustness, we create a synthetic subset FluidFlow3D-noise with the different noise level. Furthermore, another subset FluidFlow3D-ratio regarding various ρ\rho is also generated.

CylinderFlow. This public dataset contains Eulerian velocities and Lagrangian particle trajectories of the cylinder wake flow with a Reynolds number of 3900 [41]. Such flow data is also calculated using CFD. In addition, the fourth-order Runge-Kutta scheme in time and trilinear interpolations in space are utilized to transport about 200,000 particles to obtain the Lagrangian trajectories. Trajectories of two 3D observation volumes are available for the assessment of tracking algorithms.

DeformationFlow. This experiment is conducted by measuring the volumetric deformation of a soft polyacrylamide (PA) hydrogel exerted by the spherical indentation [30]. A stainless steel sphere is placed on the surface of the PA hydrogel, which leads to the indentation deformation due to the force of gravity. Then the 3D volumetric image of the hydrogel material as well as the embedded fluorescent particles are scanned both before and after the deformation with multiphoton microscopy.

2.4 Evaluation Metrics

To quantitatively assess and compare the performance of all the methods, we adopt the widely used four evaluation metrics in scene flow learning tasks [25, 27, 26, 28], including EPE, Acc Strict, Acc Relax and Outliers. The end-point-error EPE averaged over each particle is defined as:

EPE=1Ni=1N𝒱e𝒱gt2,EPE=\frac{1}{N}\sum_{i=1}^{N}\|\mathcal{V}^{e}-\mathcal{V}^{gt}\|_{2}, (22)

where 2\|\cdot\|_{2} denotes the l2l_{2} distance. The Acc Strict is calculated as the percentage of particles with the relative EPE error << 5% , while the Acc Relax is defined as the percentage of particles with the relative EPE error << 10% . In addition, the Outliers are denoted as the percentage of particles with EPE >> 0.3 unit or the relative error >> 10%. Moreover, in the comparison with tracking algorithms, two metrics are employed as the assessment principles for the tracking accuracy and tracking reliability, including the yield rate (EyE_{y}), the reliability rate (ErE_{r}):

Ey=NcNg,Er=NcNe,E_{y}=\frac{N_{c}}{N_{g}},\qquad E_{r}=\frac{N_{c}}{N_{e}}, (23)

where EyE_{y} denotes the ratio of correct matches (denoted by NcN_{c}) to all ground-truth matches (NgN_{g}), while ErE_{r} is the proportion of correct matches to all extracted matches (NeN_{e}).

3 Results

In this section, a great deal of evaluations on experiments are presented to comprehensively assess the performance of GotFlow3D in the complex flow learning task. We compare GotFlow3D to the state-of-the-art methods and verify the robustness to the noise as well as various displacement ratios. There followed the validation on a public fluid flow dataset, where the superiority of GotFlow3D to traditional PTV methods and the promotion of GotFlow3D to the tracking accuracy are investigated. Finally, a PTV experiment is conducted to further investigate the generalization of GotFlow3D to real-world experimental data.

3.1 Comparison with SOTA scene flow learners

TABLE I: Performance comparison between the state-of-the-art scene flow learning methods and GotFlow3D. The best results are marked in bold.
 
Methods EPE Acc Strict Acc Relax Outliers
 
FlowNet3D [25] 0.06229 19.34% 38.23% 61.77%
 
FLOT [26] 0.05865 24.99% 45.59% 54.41%
 
PointPWC-Net [27] 0.01723 46.75% 71.61% 28.39%
 
PV-RAFT [28] 0.01651 72.98% 83.69% 16.31%
 
GotFlow3D 0.00487 93.15% 96.38% 3.62%
 

We first present quantitative results on the test set of FluidFlow3D-norm by comparing our model with other four representative state-of-the-art scene flow learning methods, including the FlowNet3D [25], FLOT [26], PointPWC-Net [27] and PV-RAFT [28]. All of these baseline methods are trained on the training set of FluidFlow3D-norm with the training hyper-parameters reported in their articles.

The comparisons of the EPE, Acc Strict, Acc Relax and Outliers metrics of these methods are listed in Table 1. As revealed in the table, our proposed model outperforms other baseline methods and achieves the state-of-the-art performance on all of the four metrics. Specifically, GotFlow3D surpasses FlowNet3D and FLOT by more than an order of magnitude on the EPE, significantly boosts the Acc Strict and Acc Relax and brings the Outliers to a extremely low level.

Refer to caption
Figure 2: Flow estimation performance on testing examples in the FluidFlow3D dataset. Three flow cases are demonstrated here, including the Forced isotropic turbulence (left), Turbulent channel flow (middel) and Beltrami flow (right), while other cases can be found in Table IV. There are three kinds of visualization for each case: (top) the flow field estimated by GotFlow3D; (middle) the source particles 𝒫\mathcal{P} (red), transported particles 𝒫\mathcal{P^{\prime}} (green) using the learnt flow and the corresponding target particles 𝒬\mathcal{Q} (blue); (bottom) the ground-truth flow vectors (red) and the estimated flow vectors (cyan). In addition, the table shows the comparison of overall performances between GotFlow3D and other SOTA scene flow learners on these testing cases.

To visualize more detials, we conduct comparative experiments over different flow cases in FluidFlow3D-norm and verify the generalization and superiority of the proposed model. Three typical examples are demonstrated in Figure 2, while the errors for all the flow cases are given in Table IV. As shown in Figure 2 and Table IV, GotFlow3D achieves the state-of-the-art performance with the smallest EPE for all flow cases. In addition, GotFlow3D presents the highest Acc Strict (Relax) and the smallest Outlier except for the Uniform flow case. In the Uniform flow case, we observe that PV-RAFT shows comparable results with GotFlow3D. PV-RAFT is slightly better, and they both achieve over 99% accuracy and contain less than 0.02% outliers. We suspect that such a good performance in this relatively simple case is attributed to the GRU module employed both in GotFlow3D and PV-RAFT. However, in some of the complex flow cases (e.g., Forced isotropic turbulence, Forced MHD turbulence and Beltrami flow) which contain abundant fine-grained small-scale flow structures, the other flow estimators including PV-RAFT provide worse results, whereas GotFlow3D consistently performs the best on the EPE and Outliers even coping with the difficult datasets containing complicated flow scenes and diverse motion patterns.

To further depict the learning performance of GotFlow3D, we visualize the estimated flow from GotFlow3D with diverse fashions in Figure 2. Specifically, we plot the results of three test data samples from FluidFlow3D-norm, including the Forced isotropic turbulence (left), Turbulent channel flow (middle) and Beltrami flow (right). Moreover, we present the predicted flow fields (top) to show the capability of GotFlow3D to learn complex flow distributions, which involve numerous small-scale flow structures. In the middle row, the source particle 𝒫\mathcal{P} (red), transported particles 𝒫\mathcal{P^{\prime}} (green) using the learnt flow and the corresponding target particle 𝒬\mathcal{Q} (blue) are visualized to qualitatively demonstrate the tiny end point error, which also illustrates how GotFlow3D copes with the flow cases containing both small and large displacements. Furthermore, we also manifest both the ground-truth flow (red) and the flow estimations (cyan) in the bottom row. It can be observed that the estimations maintain highly consistency with the ground truth.

Refer to caption
Figure 3: Robustness assessment for GotFlow3D. (a) Performance of the state-of-the-art scene flow learners with regard to the displacement ratio ρ\rho. The right figure shows the mean (solid lines) and standard deviation (shadows) values, which are computed over 100 samples. (b) Performance regarding to various noise levels. (c) Recurrent flow refinement estimated by GotFlow3D. Particles and motions in different iterations are visualized: (red) source particles; (blue) flow motion estimated in iteration 1; (green) updated residual flow from iteration 1 to 2; (purple) updated residual flow from iteration 2 to 32. Different viewing angles are demonstrated (i-iv). Moreover, the performance (EPE) of two recurrent learning models with respect to iteration numbers are shown in (v).

3.2 Robustness against experimental conditions

To further quantify the generalization and robustness of GotFlow3D facing extreme flow conditions, we conduct comparisons across different methods on dedicated datasets with wide range of velocity and varying noise levels. The metrics reported in this section are obtained by taking the average over 100 samples for each investigated condition.

To investigate the influence of the dynamic velocity range, we first present the assessment with respect to different displacement ratios ρ\rho on the subset FluidFlow3D-ratio. This data subset is generated by simulating the Beltrami flow with varied velocity magnitudes, which leads to a series of ρ\rho ranging from 0.33 to 1.15, covering most situations in real experiments. We demonstrate the Acc Strict, Acc Relax and Outliers of all flow learners in Figure 3(a)(i), also plot the corresponding EPEs as a function of ρ\rho in Figure 3(a)(ii). As shown, GotFlow3D leads the way in EPE, Acc Strict (Relax) and Outliers compared to other methods while surpasses the FlowNet3D and FLOT to a great extent. When facing the flow condition with a extreme low ρ\rho of 0.33, GotFlow3D consistently shows impressive performances. GotFlow3D presents a relatively higher variance on EPE than PV-RAFT and PointPWC-Net in the case when ρ=0.33\rho=0.33 in Figure 3(a)(ii), which may be due to that the ρ\rho of 0.33 is lower than the ρ\rho of the training set (from 0.35 to 6.58). Other than that, the EPE curve of GotFlow3D nearly remains flat with low variances when ρ\rho increases from 0.43 to 1.15. We can observe that PV-RAFT also shows comparable results and approaches GotFlow3D when facing conditions with large displacement ratio (i.e., the case where flow estimators are easier to track particles). Nevertheless, the outperformance of GotFlow3D convincingly reveals the robustness toward diverse flow cases with wide dynamic velocity ranges.

Likewise, the performance of all these flow estimators are demonstrated with respect to various noise levels defined in Section 2.3.1. This comparison is assessed on the subset FluidFlow3D-noise, which is also generated using Beltrami flow with mean ρ\rho of 0.71. We investigate the noise level ranging from 0 % to 10%, which is reasonable to mimic the noises in real experiments. The evaluation metrics among different learning methods are shown in Figure 3(b)(i), while the EPEs are ploted against the noise in Figure 3(b)(ii). It is obviously seen that the metrics of all methods inevitably get worse with the increase of noise level, which is expected. Furthermore, it can be observed that GotFlow3D achieves the best performance on EPE, Acc Strict (Relax) and Outliers among all investigated noise levels and surpasses FlowNet3D as well as FLOT by a large margin (more than 50%) on Acc Strict (Relax) and Outliers. When it comes to a condition with high level noise, GotFlow3D does not possess much superiority over PV-RAFT while the gap between them gradually narrows. The impressive performance indicates the robustness of GotFlow3D facing some degree of noise levels.

To better illustrate the effects of the recurrent flow estimation, we investigate the performance of the network inference with different iteration numbers in the recurrent flow updating. The iteration number is a user-defined parameter in inference stage, which can be set higher to learn flow with large displacements and can be defined lower to reduce the computational cost. This investigation is assessed on two flow subsets from FluidFlow3D-ratio: a more difficult case of ρ=0.43\rho=0.43 and an easier case of ρ=0.90\rho=0.90, and the comparison is only performed between the learning methods with recurrent unit (i.e., PV-RAFT and GotFlow3D). As shown in Figure 3(c)(v), both methods yield a remarkable improvement on EPE as the iteration number increases to 6. However, the error of PV-RAFT becomes slightly higher when the iteration number exceeds 6 (marked with a blue dashed line) for both values of ρ\rho. On the other hand, the optimal iteration number of GotFlow3D is 10 (marked with a purple dashed line) for both the flow cases. These curves indicate that the iteration number has a profound impact on the performance of recurrent approaches. However, it can be found that the GotFlow3D is more accurate and less sensitive in different configurations. It also indicates that it is better to chose an iteration number close to the pre-defined value of the training process while applied to real flow scenarios. To understand the recurrent mechanism more intuitively, we visualize the recurrently updated flow of GotFlow3D by plotting trajectories of the transported particles in several iterations. As shown in Figure 3(c)(i-iv), the source particles (red) are transported using the estimated flow in iteration 1 (blue), the updated residual flow from iteration 1 to 2 (green) and the updated residual flow from iteration 2 to 32 (purple). It can be observed that the major flow motion is estimated in the first two iterations while the residual flow continues to refined slowly in the remaining iterations.

In addition to the aforementioned assessments, an ablation study regarding different modules in GotFlow3D is also performed to demonstrate the effectiveness of each module. The results are summarized in Table V. We also compute the inference time of GotFlow3D to present the computational efficiency. GotFlow3D takes about 0.1 s to estimate the flow of one sample containing about 2000 particles on a NVIDIA RTX 3090 GPU, which is promising for online flow visualization and measurement.

3.3 GotFlow3D for shedding flow with a high Reynolds number

PTV is one of the most primary techniques to analyse 3D turbulent flows by tracking particles. In this section, we perform further comparisons between GotFlow3D and other advanced open-source tracking algorithms on a public dataset CylinderFlow. We should note that GotFlow3D is a motion estimator, which does not perform particle matching and linking as the open-source PTV algorithms do. Therefore, we present the assessment by generalizing GotFlow3D as a customized motion predictor to provide existing PTV algorithms with additional motion prior information (i.e., initialization), which is supposed to promote the particle matching between adjacent frames. Two open-source particle tracking algorithms, i.e., T-PT [42] and SerialTrack [30], are implemented as evaluation benchmarks, which have been proven two of the state-of-the-art double-frame 3D tracking algorithms. T-PT is a topology-based particle tracking algorithm that encodes relative spatial neighbors as feature descriptors to deal with large displacements, while SerialTrack resolves large deformation and rotational motion with scale and rotation invariant augmented particle tracking. Both the two algorithms have been successfully applied to reveal complex motion behaviors in biological and physical applications [15, 43].

TABLE II: Performance comparison with other recently-developed particle tracking algorithms for CylinderFlow dataset. The best results are marked in bold. The numbers (30, 60 and 90) in the first raw represent the time interval between two particle sets, and the corresponding displacement ratios are shown in the brackets. The terms “GotFlow3D + X” means that GotFlow3D is applied to provide initialization for the PTV algorithms.
 
30 (ρ=0.1.06\rho=0.1.06) 60 (ρ=0.52\rho=0.52) 90 (ρ=0.35\rho=0.35)
 
Methods NcN_{c} NgN_{g} NeN_{e} EyE_{y} ErE_{r} NcN_{c} NgN_{g} NeN_{e} EyE_{y} ErE_{r} NcN_{c} NgN_{g} NeN_{e} EyE_{y} ErE_{r}
 
T-PT [42] 161727 195504 161875 82.72 99.91 118865 193903 119046 61.30 99.85 98202 192375 98324 51.05 99.88
 
GotFlow3D + T-PT 192196 195504 192450 98.31 99.86 166232 193903 166908 85.73 99.60 131451 192375 132087 68.33 99.52
 
SerialTrack [30] 181334 195504 189493 92.75 95.69 123643 193903 187647 63.76 65.89 95334 192375 187501 49.55 50.84
 
GotFlow3D + SerialTrack 192350 195504 194303 98.39 98.99 178556 193903 190297 92.08 93.83 154322 192375 187211 80.21 82.43
 

To better show the promotion brought to these PTV algorithms from GotFlow3D, we conduct three comparisons with different frame intervals, including the first frame and the 30th frame (denoted as experiment 30), the first frame and the 60th frame (denoted as experiment 60), the first frame and the 90th frame (denoted as experiment 90). Larger time interval generally leads to relatively larger displacement motion, and the three investigated experiments correspond to ρ\rho of 1.06 (30), 0.52 (60) and 0.35 (90), respectively. For evaluation of PTV algorithms, the yield rate (EyE_{y}) and reliability rate (ErE_{r}) are used as metrics for comparison, which strongly indicate the tracking accuracy and tracking reliability in PTV applications. As shown in Table 2, the EyE_{y} and the correct matches NcN_{c} of all methods decrease as the frame interval gets longer. T-PT consistently presents the highest ErE_{r} among all the three experiments, showing the tracking reliability of this method. However, the EyE_{y} of T-PT is relatively low, which means it misses a large number of particle tracks. SerialTrack outperforms T-PT on EyE_{y} and NcN_{c} in conditions of high ρ\rho (experiment 30 and 60) but is slightly surpassed by T-PT in the challenging experiment 90. Impressively, the integrated implementations named “GotFlow3D+T-PT” and “GotFlow3D+SerialTrack” improve the EyE_{y} performance of T-PT and SerialTrack to an extremely high level, indicating that much more correct matches are extracted. In particular, the proposed GotFlow3D boosts the EyE_{y} of T-PT by 15-24% and that of SerialTrack by 6-30%. The highest improvement of EyE_{y} regarding T-PT comes from the experiment 60 (about 24%), while that regarding SerialTrack is obtaiend from the experiment 90 (about 30%). It can be seen that GotFlow3D yields larger promotion coping with the conditions with relatively larger displacements. When facing the condition with short frame interval of 30 (i.e., high ρ\rho value), “GotFlow3D+T-PT” and “GotFlow3D+SerialTrack” achieve a nearly perfect performance with both EyE_{y} and ErE_{r} approaching 100%. The ErE_{r} of “GotFlow3D+T-PT” and “GotFlow3D+SerialTrack” surpasses SerialTrack but is not superior to T-PT, since GotFlow3D also boosts the extraction of more matches, which may lead to the decrease of ErE_{r}.

Refer to caption
Figure 4: GotFlow3D applied for CylinderFlow dataset, estimating the shedding flow motion over a cylinder. (a) Overview of the 3D vorticity iso-surfaces colored by velocity magnitude in the entire domain, with a specific subdomain outlined by a red box. (b) Flow field in the subdomain estimated by GotFlow3D, where the color represents the displacement magnitude. (c) Ground-truth flow field, where the color represents the displacement magnitude. (d) Residual error between GotFlow3D and ground-truth. The EPE in this case is approximately 0.0058. (e-h) The results of different particle tracking algorithms, including (e) T-PT, (f) GotFlow3D+T-PT, (g) SerialTrack, (h) GotFlow3D+SerialTrack. Correct vectors (blue), missed vectors (green) and wrong vectors (orange) are shown in the plots.

To visually validate the contribution of GotFlow3D, we plot the estimated cylinder flow as well as the ground-truth flow of the experiment 60. An overview of the flow field in a large domain is shown in Figure 4(a), where a subdomain, that we mainly focus on, is outlined. As shown in Figure 4(b), GotFlow3D itself has fulfilled an excellent recovery of the cylinder wake flow structure, which is nearly aligned with the ground truth (Figure 4c). We also present the residual flow (i.e., the error between estimated flow and the ground truth) in Figure 4(d), where the error mainly exists in a small region behind the cylinder. Furthermore, the tracking results of the two open-source PTV algorithms as well as their combinations with GotFlow3D are visually illustrated in Figure 4(e-h). We present the correct matches in blue vectors, missed matches in green vectors (i.e., the complementary set of correct matches in possible matches) and wrong matches in red vectors (i.e., the complementary set of correct matches in extracted matches), respectively. It can be seen that T-PT strives to maintain a high tracking reliability, and consequently misses plenty of potential matches, which however are successfully brought back by GotFlow3D. On the contrary, SerialTrack focuses on the extraction of more potential matches, and therefore results in abundant wrong matches, which are also well corrected by GotFlow3D. It is worth noting that although our synthetic training dataset cannot cover all flow scenarios (e.g., the cylinder flow), the impressive performance GotFlow3D makes it a promising method to be generalized to more potential and realistic applications.

3.4 GotFlow3D for deformation of sphere indentation

Refer to caption
Figure 5: GotFlow3D applied for deformative flow of sphere indentation. (a) Reference and deformed configuration of the sphere indentation, showing the deformation of the flow motion. (b-e) Tracking vector fields of (b) T-PT, (c) SerialTrack, (d) GotFlow3D+T-PT and (e) GotFlow3D+SerialTrack. With GotFlow3D combined, the tracking algorithms can extract more potential matches and demonstrate more flow structures.

To further investigate the generalization to real experiments, we apply GotFlow3D to recover the 3D spherical indentation deformation data DeformationFlow. An illustration of the reference and deformed experimental configurations is shown in Figure 5(a). About 35000 particles (plotted as red dots) are simultaneously traced in the investigated domain of 1024 voxels ×\times 1024 voxels ×\times 445 voxels. The investigated flow motion of the hydrogel deformation exhibits an approximate displacement ratio of ρ=0.49\rho=0.49 according to our estimation. The velocity vectors computed by T-PT, SerialTrack, the integrated “GotFlow3D+T-PT” and “GotFlow3D+SerialTrack” are presented in Figure 5 for comparison. Parameters of SerialTrack are set as the recommended ones from the original paper [30].

The recovered displacements of these methods are visualized in Figure 5(b-e). It can be observed that “GotFlow3D+T-PT” (Figure 5d) and “GotFlow3D+SerialTrack” (Figure 5e) yield denser 3D flow fields than T-PT and SerialTrack (Figure 5a-b). The improvement of spatial resolution comes from the flow prediction of GotFlow3D, which contributes to covering more fine details of the revealed complex flow structures. With GotFlow3D combined, the tracking algorithms can extract more potential matches and demonstrate more flow structures. Moreover, the estimations of “GotFlow3D+T-PT” and “GotFlow3D+SerialTrack” are nearly consistent with the results of T-PT and SerialTrack in the hydrogel region of relatively small displacements, which is far away from the sphere and mainly distributed with blue vectors. However, both original T-PT and SerialTrack fail to deal with the flow in the region near the contact surface between the sphere and the hydrogel. They provide sparse flow fields with small magnitudes, but relatively large particle displacements are expected. On the contrary, “GotFlow3D+T-PT” and “GotFlow3D+SerialTrack” show a great capability in extracting large deformation fields, especially in the contact surface region. Through the comparison, we can conclude that GotFlow3D, which is trained using synthetic flow, can be effectively generalized to more complex flow scenarios in real-world experiments and show excellent performance in resolving large displacements in particle tracking.

4 Discussion

In this paper, we introduce a complete pipeline to employ the deep neural network called GotFlow3D to learn the 3D fluid motion from two consecutive unorganized particle sets. The contributions of this work are manifold. A static-dynamic fusion graph neural network (SDGNN) is first proposed to derive the distinctive pointwise feature of particle sets by fusing the extracted geometric information from the static spatial graph and the dynamic feature graph. We present a novel approach to retrieve the adaptive point-voxel optimal transport (AdaPT) plan for the iterative flow learning. The retrieval of the AdaPT plan effectively provides the long-range and small-scale correspondence of particle pairs. We generate the first 3D synthetic fluid flow dataset named FluidFlow3D for the network training based on the physical simulations from CFD. This dataset covers a great deal of complex non-rigid motion under different flow conditions beyond the common rigid motion in scene flow tasks. The generated dataset is available for broader research in both the fluid mechanics community and the computer vision community. Comprehensive evaluations on synthetic and real-world experimental data have shown that our approach rivals state-of-the-art scene flow networks and other advanced PTV methods. In particular, our approach demonstrates competitive accuracy to the prior art, excellent robustness to various flow conditions and impressive generalization to real-world applications.

We devise a new framework to learn 3D complex fluid flow motion from consecutive particle sets in particle tracking experiments, which draws inspiration from some pioneering learning-based networks. However, there are obvious differences between GotFlow3D and the existing scene flow learners. The design of the proposed network only follows the basic similar concept of RAFT to estimate the flow in an iterative manner. In fact, RAFT performs the flow estimation on regular images, while our GotFlow3D targets at irregular particle sets. PV-RAFT adopts a similar iterative flow update to deal with point clouds, which however pays more attention to the basic point-voxel correlation fields. In comparison, we propose to adaptively retrieve a more efficient AdaPT plan to boost the recurrent flow learning guided by optimal transport. Both FLOT and GotFlow3D seek guidance from optimal transport theory, but implement that in different ways. The former directly employs optimal transport to compute a coarse flow, whereas the latter performs the AdaPT plan retrieval to extract more reliable correspondence information for the subsequent iterative flow refinement. Apart from the above-demonstrated differences, the innovation of the proposed GotFlow3D is also embodied in the design of the SDGNN, which makes the first attempt to combine features from the dynamic feature graph and that of the static spatial graph in flow motion learning tasks. Such combination of new modules presents a significant improvement on the accuracy of flow estimation according to the ablation study shown in Table V.

In summary, a novel deep learning based model - GotFlow3D - is proposed for learning complex flow motion from two particle sets, which provides a new perspective for introducing deep learning approaches in the further analysis of particle tracking applications encountered in a wide range of biological and physical systems. We present the first synthetic dataset for further deep learning developments in PTV community. GotFlow3D shows state-of-the-art accuracy on the PTV database and surpasses the existing scene flow neural networks on most flow cases. Robustness analysis further validates the stable performance of GotFlow3D under challenging flow cases and extreme experimental conditions, which are normally encountered in reality. GotFlow3D also achieves a significant performance on the experiment of cylinder flow, which does not even exist in the training database, and brings a great improvement to the existing PTV approaches. Moreover, GotFlow3D has been generalized to a real-world particle tracking application of sphere indentation, where GotFlow3D shows the superiority of improving the spatial resolution and resolving large displacements.

Future extensions to deep learning approaches in PTV analysis include Bayesian framework of GotFlow3D, which could provide not only flow estimation but also the uncertainty quantification of the measurement. Moreover, GotFlow3D is currently designed as an universal tool, which simply consumes spatial coordinates of particles. More properties of particles (e.g., intensity) could be utilized to further improve the correspondence searching and make GotFlow3D customized tools in specific applications. It is also a promising research topic to incorporate physics-informed constraints (e.g., governing equations of the flow motion) in specific applications while training the neural networks.

Appendix Table III: Configuration of various flow cases in the synthetic dataset. We generate a training dataset with more than 16000 samples, each of which contains two particle sets and the corresponding ground truth flow. The synthetic dataset consists of six kinds of flow scenarios generated by using computational fluid dynamics (CFD), where some of them are extracted from the Johns Hopkins Turbulence Databases. Here, we list the name of flow motion patterns, the range of observation volume, the range of duration time for particle trajectories in the simulation and the range of displacement ratio ρ\rho. By randomly selecting the parameters in the ranges for each data item, we generate a dataset mimicking diverse conditions in PTV experiments.
Flow case Observation Volume Δt\Delta t ρ\rho
Forced isotropic turbulence {(0.25π\pi, 0.25π\pi, 0.25π\pi), (0.5π\pi, 0.5π\pi, 0.5π\pi)} {0.03, 0.04, 0.05} [0.36, 2.50]
Forced MHD turbulence {(0.25π\pi, 0.25π\pi, 0.25π\pi), (0.5π\pi, 0.5π\pi, 0.5π\pi)} {0.06, 0.07, 0.08} [0.58, 2.33]
Transitional boundary flow {(0.5π\pi, 1, 0.5π\pi) , (π\pi, 2, π\pi)} {0.12, 0.16, 0.20} [0.48, 2.24]
Turbulent channel flow (0.25π\pi, 0.5, 0.25π\pi) {0.03, 0.04, 0.05, 0.06, 0.07, 0.08} [0.51, 1.63]
Beltrami flow (2, 2, 2) { [0.9, 1.0], [1.1, 1.2], [1.3, 1.4] } [0.35, 6.58]
Uniform flow (0.25π\pi, 0.25π\pi, 0.25π\pi) {0.03, 0.04, 0.05} [0.51, 3.11]
Appendix Table IV: Performance comparison between the state-of-the-art scene flow learning methods and GotFlow3D on different flow cases in the synthetic dataset. The best results are marked in bold. We present the performance of these methods with metrics of EPE, ACC Strict, ACC Relax and Outliers. GotFlow3D achieves the best performance on most of the flow cases except the Uniform flow, where PV-RAFT is slightly superior to GotFlow3D on ACC Strict (Relax) as well as Outliers. In some of the complex flow cases (e.g., Forced isotropic turbulence, Forced MHD turbulence and Beltrami flow) which contain small-scale flow structures, the existing flow estimators including PV-RAFT provide worse results, whereas GotFlow3D consistently performs the best.
Method Forced isotropic turbulence Forced MHD turbulence Transitional boundary flow
EPE Acc Strict Acc Relax Outliers EPE Acc Strict Acc Relax Outliers EPE Acc Strict Acc Relax Outliers
FlowNet3D 0.12465 0.20% 1.52% 98.48% 0.09400 0.06% 0.45% 99.55% 0.02003 67.23% 91.01% 8.99%
FLOT 0.13090 0.14% 1.15% 98.85% 0.09842 0.04% 0.28% 99.72% 0.01715 70.37% 94.63% 5.37%
PointPWC-Net 0.02719 18.47% 57.06% 42.94% 0.01714 8.46% 33.53% 66.47% 0.01163 89.13% 98.29% 1.71%
PV-RAFT 0.04190 51.37% 62.82% 37.18% 0.02594 41.44% 61.37% 38.63% 0.00324 99.65% 99.94% 0.063%
GotFlow3D 0.01153 86.62% 90.74% 9.26% 0.00596 83.005% 91.81% 8.19% 0.00222 99.87% 99.97% 0.028%
Method Turbulent channel flow Beltrami flow Uniform flow
EPE Acc Strict Acc Relax Outliers EPE Acc Strict Acc Relax Outliers EPE Acc Strict Acc Relax Outliers
FlowNet3D 0.03670 24.89% 70.78% 29.22% 0.04435 2.30% 13.73% 86.27% 0.03748 25.41% 79.23% 20.77%
FLOT 0.03541 26.89% 72.72% 27.28% 0.02417 13.11% 42.22% 57.78% 0.02059 68.25% 96.45% 3.55%
PointPWC-Net 0.01266 87.51% 98.41% 1.59% 0.01616 24.36% 62.19% 37.81% 0.02140 64.31% 97.32% 2.68%
PV-RAFT 0.00653 93.71% 97.83% 2.17% 0.00899 69.69% 91.00% 9.00% 0.00316 99.87% 99.99% 0.015%
GotFlow3D 0.00241 99.13% 99.74% 0.26% 0.00291 95.03% 98.44% 1.56% 0.00244 99.85% 99.98% 0.019%
Appendix Table V: Ablation study on different modules in GotFlow3D. We conduct a set of ablation experiments to demonstrate the effectiveness of each module in GotFlow3D, including the static and/or dynamic graph neural network for feature extraction, optimal transport and/or AdaPT module for correspondence retrieval. A green tick means the corresponding module is used, while a red cross means not used. It can be seen from Networks 5-7 that the proposed dynamic GNN greatly improves the performance and the fusion of two kinds of graph (i.e., SDGNN ) further boosts the accuracy to a higher level. One can also observe from Networks 1-2 that the replacement of correlation fields with the optimal transport reduces the error to some extent by modeling the correspondence searching with the transport plan optimization. As shown by the results of Networks 2-5, both proposed adaptive transport plan 𝒯p\mathcal{T}^{*}_{p} and 𝒯v\mathcal{T}^{*}_{v} are better than the point-voxel based lookup, and the integration of the 𝒯p\mathcal{T}^{*}_{p} and 𝒯v\mathcal{T}^{*}_{v} (Netwrok 5) performs the best.
Method Module options Metrics
Feature Extraction Correspondence Retrieval EPE Outliers
Static Graph Dynamic Graph SDGNN Correlation Fields Optimal Transport Point-Voxel Lookup AdaPT
𝒯p\mathcal{T}^{*}_{p}+𝒯v\mathcal{T}_{v} 𝒯p\mathcal{T}_{p}+𝒯v\mathcal{T}^{*}_{v} 𝒯p\mathcal{T}^{*}_{p}+𝒯v\mathcal{T}^{*}_{v}
Network1 0.01651 16.39%
Network2 0.01208 10.24%
Network3 0.01177 9.85%
Network4 0.01036 9.43%
Network5 0.00916 7.96%
Network6 0.00603 4.46%
Network7 0.00487 3.63%

Acknowledgments

This work was supported in part by the National Key R&D Program of China (No. 2019YFB1705800), in part by the National Natural Science Foundation of China under grant No. 61973270, in part by the Foundation for Innovative Research Groups of the National Natural Science Foundation of China under grant No. 61621002.

References

  • [1] M. Kemp, “Leonardo da Vinci’s laboratory: studies in flow,” Nature, vol. 571, no. 7765, pp. 322–324, 2019.
  • [2] D. Dabiri and C. Pecora, Particle Tracking Velocimetry.   IOP Publishing Bristol, 2020.
  • [3] A. Kopitca, K. Latifi, and Q. Zhou, “Programmable assembly of particles on a Chladni plate,” Science advances, vol. 7, no. 39, p. eabi7716, 2021.
  • [4] B. Ferdowsi, C. P. Ortiz, M. Houssais, and D. J. Jerolmack, “River-bed armouring as a granular segregation phenomenon,” Nature communications, vol. 8, no. 1, pp. 1–10, 2017.
  • [5] D. L. Hu, B. Chan, and J. W. Bush, “The hydrodynamics of water strider locomotion,” Nature, vol. 424, no. 6949, pp. 663–666, 2003.
  • [6] B. He, K. Doubrovinski, O. Polyakov, and E. Wieschaus, “Apical constriction drives tissue-scale hydrodynamic flow to mediate cell elongation,” Nature, vol. 508, no. 7496, pp. 392–396, 2014.
  • [7] H. Mestre, T. Du, A. M. Sweeney, G. Liu, A. J. Samson, W. Peng, K. N. Mortensen, F. F. Stæger, P. A. Bork, L. Bashford et al., “Cerebrospinal fluid influx drives acute ischemic tissue swelling,” Science, vol. 367, no. 6483, p. eaax7171, 2020.
  • [8] Z. Zhang, M. Hwang, T. J. Kilbaugh, A. Sridharan, and J. Katz, “Cerebral microcirculation mapped by echo particle tracking velocimetry quantifies the intracranial pressure and detects ischemia,” Nature communications, vol. 13, no. 1, pp. 1–15, 2022.
  • [9] M. Guo, A. J. Ehrlicher, M. H. Jensen, M. Renz, J. R. Moore, R. D. Goldman, J. Lippincott-Schwartz, F. C. Mackintosh, and D. A. Weitz, “Probing the stochastic, motor-driven properties of the cytoplasm using force spectrum microscopy,” Cell, vol. 158, no. 4, pp. 822–832, 2014.
  • [10] Y. Peng, Z. Liu, and X. Cheng, “Imaging the emergence of bacterial turbulence: Phase diagram and transition kinetics,” Science advances, vol. 7, no. 17, p. eabd1240, 2021.
  • [11] S. Schuerle, A. P. Soleimany, T. Yeh, G. Anand, M. Häberli, H. Fleming, N. Mirkhani, F. Qiu, S. Hauert, X. Wang et al., “Synthetic and living micropropellers for convection-enhanced nanoparticle transport,” Science advances, vol. 5, no. 4, p. eaav4803, 2019.
  • [12] H. Punzmann, N. Francois, H. Xia, G. Falkovich, and M. Shats, “Generation and reversal of surface flows by propagating waves,” Nature physics, vol. 10, no. 9, pp. 658–663, 2014.
  • [13] P. Y. Huang, S. Kurasch, J. S. Alden, A. Shekhawat, A. A. Alemi, P. L. McEuen, J. P. Sethna, U. Kaiser, and D. A. Muller, “Imaging atomic rearrangements in two-dimensional silica glass: watching silica’s dance,” Science, vol. 342, no. 6155, pp. 224–227, 2013.
  • [14] F. Pereira, H. Stüer, E. C. Graff, and M. Gharib, “Two-frame 3D particle tracking,” Measurement science and technology, vol. 17, no. 7, p. 1680, 2006.
  • [15] S. E. Leggett, M. Patel, T. M. Valentin, L. Gamboa, A. S. Khoo, E. K. Williams, C. Franck, and I. Y. Wong, “Mechanophenotyping of 3D multicellular clusters using displacement arrays of rendered tractions,” Proceedings of the National Academy of Sciences, vol. 117, no. 11, pp. 5655–5663, 2020.
  • [16] M. Raffel, C. Willert, F. Scarano, C. Kähler, S. Wereley, and J. Kompenhans, Particle Image Velocimetry: A Practical Guide.   Cham: Springer International Publishing, 2018.
  • [17] C. Cierpka, B. Lütke, and C. J. Kähler, “Higher order multi-frame particle tracking velocimetry,” Experiments in Fluids, vol. 54, no. 5, pp. 1–12, 2013.
  • [18] S. L. Brunton, B. R. Noack, and P. Koumoutsakos, “Machine learning for fluid mechanics,” Annual Review of Fluid Mechanics, vol. 52, pp. 477–508, 2020.
  • [19] S. Cai, S. Zhou, C. Xu, and Q. Gao, “Dense motion estimation of particle images via a convolutional neural network,” Experiments in Fluids, vol. 60, no. 4, pp. 1–16, 2019.
  • [20] S. Cai, J. Liang, Q. Gao, C. Xu, and R. Wei, “Particle image velocimetry based on a deep learning motion estimator,” IEEE Transactions on Instrumentation and Measurement, vol. 69, no. 6, pp. 3538–3554, 2019.
  • [21] J. Liang, S. Cai, C. Xu, and J. Chu, “Filtering enhanced tomographic PIV reconstruction based on deep neural networks,” IET Cyber-Systems and Robotics, vol. 2, no. 1, pp. 43–52, 2020.
  • [22] C. Lagemann, K. Lagemann, S. Mukherjee, and W. Schröder, “Deep recurrent optical flow learning for particle image velocimetry data,” Nature Machine Intelligence, vol. 3, no. 7, pp. 641–651, 2021.
  • [23] K. Mallery, S. Shao, and J. Hong, “Dense particle tracking using a learned predictive model,” Experiments in Fluids, vol. 61, no. 10, pp. 1–14, 2020.
  • [24] C. R. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 652–660.
  • [25] X. Liu, C. R. Qi, and L. J. Guibas, “Flownet3D: Learning scene flow in 3D point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 529–537.
  • [26] G. Puy, A. Boulch, and R. Marlet, “Flot: Scene flow on point clouds guided by optimal transport,” in European conference on computer vision.   Springer, 2020, pp. 527–544.
  • [27] W. Wu, Z. Y. Wang, Z. Li, W. Liu, and L. Fuxin, “PointPWC-Net: Cost volume on point clouds for (self-) supervised scene flow estimation,” in European conference on computer vision.   Springer, 2020, pp. 88–107.
  • [28] Y. Wei, Z. Wang, Y. Rao, J. Lu, and J. Zhou, “PV-RAFT: point-voxel correlation fields for scene flow estimation of point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 6954–6963.
  • [29] J. Liang, S. Cai, C. Xu, T. Chen, and J. Chu, “DeepPTV: Particle Tracking Velocimetry for complex flow mmtion via deep neural networks,” IEEE Transactions on Instrumentation and Measurement, 2021.
  • [30] J. Yang, Y. Yin, A. K. Landauer, S. Buyuktozturk, J. Zhang, L. Summey, A. McGhee, M. K. Fu, J. O. Dabiri, and C. Franck, “SerialTrack: scale and rotation invariant augmented lagrangian particle tracking,” arXiv preprint arXiv:2203.12573, 2022.
  • [31] C. R. Qi, L. Yi, H. Su, and L. J. Guibas, “Pointnet++: Deep hierarchical feature learning on point sets in a metric space,” Advances in neural information processing systems, vol. 30, 2017.
  • [32] Y. Wang, Y. Sun, Z. Liu, S. E. Sarma, M. M. Bronstein, and J. M. Solomon, “Dynamic graph cnn for learning on point clouds,” Acm Transactions On Graphics (tog), vol. 38, no. 5, pp. 1–12, 2019.
  • [33] C. Villani, Optimal transport: old and new.   Springer, 2009, vol. 338.
  • [34] G. Peyré, M. Cuturi et al., “Computational optimal transport: With applications to data science,” Foundations and Trends® in Machine Learning, vol. 11, no. 5-6, pp. 355–607, 2019.
  • [35] M. Cuturi, “Sinkhorn distances: lightspeed computation of optimal transport,” Advances in neural information processing systems, vol. 26, 2013.
  • [36] Z. Teed and J. Deng, “Raft: Recurrent all-pairs field transforms for optical flow,” in European conference on computer vision.   Springer, 2020, pp. 402–419.
  • [37] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  • [38] Y. Li, E. Perlman, M. Wan, Y. Yang, C. Meneveau, R. Burns, S. Chen, A. Szalay, and G. Eyink, “A public turbulence database cluster and applications to study lagrangian evolution of velocity increments in turbulence,” Journal of Turbulence, no. 9, p. N31, 2008.
  • [39] C. R. Ethier and D. Steinman, “Exact fully 3D Navier–Stokes solutions for benchmarking,” International Journal for Numerical Methods in Fluids, vol. 19, no. 5, pp. 369–375, 1994.
  • [40] H. Maas, A. Gruen, and D. Papantoniou, “Particle tracking velocimetry in three-dimensional flows,” Experiments in fluids, vol. 15, no. 2, pp. 133–146, 1993.
  • [41] A. R. Khojasteh, S. Laizet, D. Heitz, and Y. Yang, “Lagrangian and Eulerian dataset of the wake downstream of a smooth cylinder at a Reynolds number equal to 3900,” Data in brief, vol. 40, p. 107725, 2022.
  • [42] M. Patel, S. E. Leggett, A. K. Landauer, I. Y. Wong, and C. Franck, “Rapid, topology-based particle tracking for high-resolution measurements of large complex 3D motion fields,” Scientific reports, vol. 8, no. 1, pp. 1–14, 2018.
  • [43] M. T. Scimone, H. C. Cramer III, E. Bar-Kochba, R. Amezcua, J. B. Estrada, and C. Franck, “Modular approach for resolving and mapping complex neural and other cellular structures and their associated deformation fields in three dimensions,” Nature protocols, vol. 13, no. 12, pp. 3042–3064, 2018.