1. Introduction
In a wide variety of practical situations, we wish to infer the state of a fluid flow from a limited number of flow sensors with generally noisy output signals. In particular, such knowledge of the flow state may assist within the larger scope of a flow control task, either in the training or application of a control strategy. For example, in reinforcement learning (RL) applications for guiding a vehicle to some target through a highly disturbed fluid environment (Verma, Novati & Koumoutsakos Reference Verma, Novati and Koumoutsakos2018), the system is partially observable if the RL framework only has knowledge of the state of the vehicle itself. It is unable to distinguish between quiescent and disturbed areas of the environment and to take actions that are distinctly advantageous in either, thus limiting the effectiveness of the control strategy. Augmenting this state with some knowledge of the flow may be helpful to improve this effectiveness.
The problem of flow estimation is very broad and can be pursued with different types of sensors in the presence of various flow physics. We focus in this paper on the inference of incompressible flows of moderate and large Reynolds numbers from pressure sensors, a problem that has been of interest in the fluid dynamics community for many years (Naguib, Wark & Juckenhöfel Reference Naguib, Wark and Juckenhöfel2001; Murray & Ukeiley Reference Murray and Ukeiley2003; Gomez et al. Reference Gomez, Lagor, Kirk, Lind, Jones and Paley2019; Sashittal & Bodony Reference Sashittal and Bodony2021; Iacobello, Kaiser & Rival Reference Iacobello, Kaiser and Rival2022; Zhong et al. Reference Zhong, Fukami, An and Taira2023). Flow estimation from other types of noisy measurements has also been pursued in closely related contexts in recent years, with tools very similar to those used in the present work (Juniper & Yoko Reference Juniper and Yoko2022; Kontogiannis et al. Reference Kontogiannis, Elgersma, Sederman and Juniper2022). Estimation generally seeks to infer a finite-dimensional state vector ${\textit {x}}$ from a finite-dimensional measurement vector ${\textit {y}}$. Since the state of a fluid flow is inherently infinite-dimensional, a central task of flow estimation is to approximately represent the flow so that it can be parameterized by a finite-dimensional state vector. For example, this could be done by a linear decomposition into data-driven modes (e.g. with proper orthogonal decomposition (POD) or dynamic mode decomposition) – in which the flow state comprises the coefficients of these modes – or by a generalized (nonlinear) form of this decomposition via a neural network (Morimoto et al. Reference Morimoto, Fukami, Zhang and Fukagata2022; Fukami & Taira Reference Fukami and Taira2023). Although these are very effective in representing flows close to their training set, they are generally less effective in representing newly encountered flows. For example, a vehicle's interaction with a gust (an incident disturbance) may take very different forms, but the basis modes or neural network can only be trained on a subset of these interactions. Furthermore, even when these approaches are effective, it is difficult to probe them for intuition about some basic questions that underlie the estimation task. How does the effectiveness of the estimation depend on the physical distance between the sensors and vortical flow structures, or on the size of the structures?
We take a different approach to flow representation in this paper, writing the vorticity field as a sum of $N$ (nearly) singular vortex elements. (The adjective ‘nearly’ conveys that we will regularize each element with a smoothing kernel of small radius.) A distinct feature of a flow singularity (in contrast to, say, a POD mode) is that both its strength and its position are degrees of freedom in the state vector, so that it can efficiently and adaptively approximate a compact evolving vortex structure even with a small number of vortex elements. The compromise for this additional flexibility is that it introduces a nonlinear relationship between the velocity field and the state vector. However, since pressure is already inherently quadratically dependent on the velocity, one must contend with nonlinearity in the inference problem regardless of the choice of flow representation. Another advantage of using singularities as a representation of the flow is that their velocity and pressure fields are known exactly, providing useful insight into the estimation problem.
To restrict the dimensionality of the problem and make the estimation tractable, we truncate the set of vortex elements to a small number $N$. In this manner, the point vortices can be thought of as an adaptive low-rank representation of the flow, each capturing a vortex structure but omitting most details of the structure's internal dynamics. To keep the scope of this paper somewhat simpler, we will narrow our focus to unbounded two-dimensional vortical flows, so that the estimated vorticity field is given by
where $\tilde {\delta }_{\epsilon }$ is a regularized form of the two-dimensional Dirac delta function with small radius $\epsilon$, and any vortex ${\textit {J}}$ has strength $\varGamma _{{\textit {J}}}$ and position described by two Cartesian coordinates $\boldsymbol {r}_{{\textit {J}}} = (x_{{\textit {J}}},y_{{\textit {J}}})$. As we show in Appendix A, the pressure field due to a set of $N$ vortex elements is given by
where $P_{\epsilon }$ is a regularized direct vortex kernel, representing the individual effect of a vortex on the pressure field, and $\varPi _{\epsilon }$ is a regularized vortex interaction kernel, encapsulating the coupled effect of a pair of vortices on pressure; their details are provided in Appendix A. This focus on unbounded two-dimensional flows preserves the essential purpose of the study, to reveal the important aspects of vortex estimation from pressure, and postpones to a future paper the effect of a body's presence or other flow contributors. Thus, the state dimensionality of the problems in this paper will be $n = 3N$, composed of the positions and strengths of the $N$ vortex elements.
In this limited context, we address the following question: Given a set of noisy pressure measurements at $d$ observation points (sensors) in or adjacent to an incompressible two-dimensional flow, to what extent can we infer a distribution of vortices? It is important to make a few points before we embark on our answer. First, because of the noise in measurements, we will address the inference problem in a probabilistic (i.e. Bayesian) manner: find the distribution of probable states based on the likelihood of the true observations. As we have noted, we have no hope of approximating a smoothly distributed vorticity field with a small number of singular vortex elements. However, as a result of our probabilistic approach, the expectation of the vorticity field over the whole distribution of estimated states will be smooth, even though the vorticity of each realization of the estimated flow is singular in space. This fact is shown in Appendix B.
Second, vortex flows are generally unsteady, so we ultimately wish to address this inference question over some time interval. Indeed, that has been the subject of several previous works, e.g. Darakananda et al. (Reference Darakananda, da Silva, Colonius and Eldredge2018), Le Provost & Eldredge (Reference Le Provost and Eldredge2021) and Le Provost et al. (Reference Le Provost, Baptista, Marzouk and Eldredge2022), in which pressure measurements were assimilated into the estimate of the state via an ensemble Kalman filter (Evensen Reference Evensen1994). Each step of such sequential data assimilation consists of the same Bayesian inference (or analysis) procedure: we start with an initial guess for the probability distribution (the prior) and seek an improved guess (the posterior). At steps beyond the first one, we generate this prior by simply advancing an ensemble of vortex element systems forward by one time step (the forecast), from states drawn from the posterior at the end of the previous step. The quality of the estimation generally improves over time as the sequential estimator proceeds. However, when we start the problem we have no such posterior from earlier steps. Furthermore, even at later times, as new vortices are created or enter the region of interest, the prior will lack a description of these new features.
This challenge forms a central task of the present paper: we seek to infer the flow state at some instant from a prior that expresses little to no knowledge of the flow. Aside from some loose bounds, we do not have any guess for where the vortex elements lie, how strong they are or even how many there should be. All we know of the true system's behaviour comes from the sensor measurements, and we therefore estimate the vortex state by maximizing the likelihood that these sensor measurements will arise. It should also be noted that viscosity has no effect on the instantaneous relationships between vorticity, velocity and pressure in unbounded flow, so it is irrelevant if the true system is viscous or not. To assess the success of our inference approach, we will compute the expectation of the vorticity field under our estimated probability distribution and compare it with the true field, as we will presume to know this latter field for testing purposes.
The last point to make is that the inference of a flow from pressure is often an ill-posed problem with multiple possible solutions, a common issue with inverse problems of partial differential equations. For example, we will find that there may be many candidate vortex systems that reasonably match the pressure sensor data, forming ridges or local maxima of likelihood, even if they are not the global maximum solution. As we will show, these situations arise most frequently when the number of sensors is less than or equal to the number of states, i.e. when the inverse problem is underdetermined to some degree. In these cases, we will find that adding even one additional sensor can address the underlying ill posedness. There will also be various symmetries that arise due to the vortex-sensor arrangement. In this paper, we will use techniques to mitigate the effect of these symmetries on the estimation task. However, multiple solutions may still arise even with such techniques, and we seek to explore this multiplicity thoroughly. Therefore, we adopt a solution strategy that explores all features of the likelihood function, including multiple maxima. We describe the probability-based formulation of the vortex estimation problem and our methodologies for revealing it in § 2. Then, we present results of various estimation exercises with this methodology in § 3, and discuss the results and their generality in § 4.
2. Problem statement and methodology
2.1. The inference problem
Our goal is to estimate the state of an unbounded two-dimensional vortical flow with vortex system (1.1), which we will call a vortex estimator, specified completely by the $n$-dimensional vector ($n=3N$)
where
is the 3-component state of vortex ${\textit {J}}$. The associated state covariance matrix is written as
where each $3 \times 3$ block represents the covariance between vortex elements ${\textit {J}}$ and ${\textit {K}}$
Note that $\varSigma _{{\textit {K}}{\textit {J}}} = \varSigma ^{{\rm T}}_{{\textit {J}}{\textit {K}}}$.
Equation (1.2) expresses the pressure (relative to ambient), ${\rm \Delta} p(\boldsymbol {r}) \equiv p(\boldsymbol {r}) - p_{\infty }$, as a continuous function of space, $\boldsymbol {r}$. Here, we also explicitly acknowledge its dependence on the state, ${\rm \Delta} p(\boldsymbol {r},{\textit {x}})$. Furthermore, we will limit our observations to a finite number of sensor locations, $\boldsymbol {r} = \boldsymbol {s}_{\alpha }$, for $\alpha = 1,\ldots,d$, and define from this an observation operator, $h: \mathbb {R}^{n} \rightarrow \mathbb {R}^{d}$, mapping a given state vector ${\textit {x}}$ to the pressure at these sensor locations
The objective of this paper is essentially to explore the extent to which we can invert function (2.5): from a given set of pressure observations ${\textit {y}}^{\ast } \in \mathbb {R}^{d}$ of the true system at $\boldsymbol {s}_{\alpha }$, $\alpha = 1,\ldots,d$, determine the state ${\textit {x}}$. In this work, the true sensor measurements ${\textit {y}}^{\ast }$ (the truth data) will be synthetically generated from the pressure field of a set of vortex elements in unbounded quiescent fluid, obtained from the same expression for pressure (1.2) that we use for $h({\textit {x}})$ in the estimator. Throughout, we will refer to the set of vortices that generated the measurements as the true vortices. However, there is inherent uncertainty in ${\textit {y}}^{\ast }$ due to random measurement noise $\varepsilon \in \mathbb {R}^{d}$, so we model the predicted observations ${\textit {y}}$ as
where $\varepsilon$ is normally distributed about zero mean, $\mathcal{N}(\varepsilon | 0,\varSigma _{\mathcal{E}})$, and the sensor noise is assumed independent and identically distributed, so that its covariance is $\varSigma _{\mathcal{E}} = \sigma _{\mathcal {E}}^{2} I$ with $I \in \mathbb {R}^{d \times d}$ the identity. We seek a probabilistic form of the inversion of (2.6) when set equal to ${\textit {y}}^{\ast }$. That is, we seek the conditional probability distribution of states based on the observed data, ${\rm \pi} ({\textit {x}}\,|\,{\textit {y}}^{\ast })$: the peaks of this distribution would represent the most probable state(s) based on the measurements, and the breadth of the peaks would represent our uncertainty about the answer.
From Bayes’ theorem, the conditional probability of the state given an observation, ${\rm \pi} ({\textit {x}}\,|\,{\textit {y}})$, can be regarded as a posterior distribution over ${\textit {x}}$,
where ${\rm \pi} _{0}({\textit {x}})$ is the prior distribution, describing our original beliefs about the state ${\textit {x}}$, and $L({\textit {y}}\,|\,{\textit {x}})$ is called the likelihood function, representing the probability of observing certain data ${\textit {y}}$ at a given state, ${\textit {x}}$. The likelihood function encapsulates our physics-based prediction of the sensor measurements, based on the observation operator $h({\textit {x}})$. Collectively, $L({\textit {y}}\,|\,{\textit {x}}) {\rm \pi}_{0}({\textit {x}})$ represents the joint distribution of states and their associated observations. The distribution of observations, ${\rm \pi} ({\textit {y}})$, is a uniform normalizing factor. Its value is unnecessary for characterizing the posterior distribution over ${\textit {x}}$, since only comparisons (ratios) of the posterior are needed during sampling, as we discuss below in § 2.4. Thus, we can omit the denominator in (2.7). We evaluate this unnormalized posterior at the true observations, ${\textit {y}}^{\ast }$, and denote it by $\tilde {{\rm \pi} }({\textit {x}}\,|\,{\textit {y}}^{\ast }) = L({\textit {y}}^{\ast }\,|\,{\textit {x}}) {\rm \pi}_{0}({\textit {x}})$.
Our goal is to explore and characterize this unnormalized posterior for the vortex system. Expressing our lack of prior knowledge, we write ${\rm \pi} _{0}({\textit {x}})$ as a uniform distribution within a certain acceptable bounding region $B$ on the state components (discussed in more specific detail below):
Following from our observation model (2.6) with Gaussian noise, the likelihood is a Gaussian distribution about the predicted observations, $h({\textit {x}})$,
where we have defined the covariance-weighted norm
Thus, our unnormalized posterior for the vortex estimator is given by
For practical purposes it is helpful to take the log of this probability, so that ratios of probabilities – some of them near machine zero – are assessed instead via differences in their logs. Because only differences are relevant, we can dispense with constants that arise from taking the log, such as the inverse square root factor. We note that the uniform probability distribution $\mathcal{U}_{n}({\textit {x}}\,|\,B)$ is uniform and positive inside $B$ and zero outside. Thus, to within an additive constant, this log posterior is
where $c_{B}({\textit {x}})$ is a barrier function arising from the log of the uniform distribution, equal to zero for any ${\textit {x}}$ inside of the restricted region $B$ of our uniform distribution, and $c_{B}({\textit {x}})= -\infty$ outside of $B$.
In the examples that we present in this paper, the pressure sensors will be uniformly distributed along a straight line on the $x$ axis, unless otherwise specified. and the set of vortices used for estimation purposes as the vortex estimator. To ensure finite pressures throughout the domain, both the true vortices and the vortex estimator are regularized as discussed in Appendix A.4, with a small blob radius $\epsilon = 0.01$ unless otherwise stated. To improve the scaling and conditioning of the problem, the pressure (relative to ambient) is implicitly normalized by $\rho \varGamma _{0}^{2}/L^{2}$, where $\varGamma _{0}$ is the strength of the largest-magnitude vortex in the true set and $L$ represents a characteristic distance of the vortex set from the sensors; all positions are implicitly normalized by $L$ and vortex strengths by $\varGamma _{0}$. Unless otherwise specified, the measurement noise is $\sigma _{\mathcal{E}} = 5 \times 10^{-4}$.
2.2. Symmetries and nonlinearity in the vortex-pressure system
As mentioned in the introduction, there are many situations in which multiple solutions arise due to symmetries. This is easy to see from a simple thought experiment, depicted in figure 1(a). Suppose that we wish to estimate a single vortex from pressure sensors arranged in a straight line. A vortex on either side of this line of sensors will induce the same pressure on the sensors, and a vortex of either sign of strength will as well. Thus, in this simple problem, there are four possible states that are indistinguishable from each other, and we would need more information about the circumstances of the problem to rule out three of them. Such symmetries arise commonly in the problems we will study in this paper.
The symmetry with respect to the sign of vortex strength is due to the nonlinear relationship between pressure and vorticity. However, it is important to note that this symmetry issue is partly alleviated by the nonlinear relationship, as well, because of the coupling that it introduces between vortices. Figure 2 depicts the pressure fields for two examples of a pair of vortices: one in which the vortices in the pair have equal strengths and another in which the vortices have equal but opposite strengths. Although the pressure in the vortex cores of both pairs is similar and sharply negative, the pressures outside the cores are distinctly different because the interaction kernel enters the sum with different sign. At the positions of the sensors, the pressure has a small region of positive pressure in the case of vortices of opposite sign. These differences are essential for inferring the relative signs of vortex strengths in sets of vortices. However, it is important to stress that the pressure is invariant to a change of sign of all vortices in the set, so we would still need more prior information to discriminate one overall sign from the other.
Another symmetry arises when there is more than one vortex to estimate, as in figure 1(b), because in such a case, there is no unique ordering of the vortices in the state vector. With each of the vortices assigned a fixed set of parameters, any of the $N!$ permutations of the ordering leads to the same pressure measurements. This vortex relabelling symmetry is a discrete analogue of the particle relabelling symmetry in continuum mechanics (Marsden & Ratiu Reference Marsden and Ratiu2013); it is also closely analogous to the non-identifiability issue of the mixture models that will be used for the probabilistic modelling in this paper. All of the $N!$ solutions are obviously equivalent from a flow field perspective, so this symmetry is not a problem if we assess estimator performance based on flow field metrics. However, the $N!$ solutions form distinct points in the state space and we must anticipate this multiplicity when we search for high-probability regions.
The barrier function $c_{B}({\textit {x}})$ in (2.12) allows us to anticipate and eliminate some modes that arise from problem symmetries, because we can use the bounding region $B$ to reject samples that fail to meet certain criteria. To eliminate some of the aforementioned symmetries, we will, without loss of generality, restrict our vortex estimator to search for vortices that lie above the line of sensors on the $x$ axis. For cases of multiple vortices in the estimator, we re-order the vortex entries in the state vector by their $x$ position at each Markov-chain Monte Carlo (MCMC) step to eliminate the relabelling symmetry. We also assume that the leftmost vortex has positive strength, which reduces the number of probable states by half; the signs of all estimated vortices can easily be switched a posteriori if new knowledge shows that this assumption is wrong.
2.3. The true covariance and rank deficiency
The main challenge of the vortex inference problem is that the observation operator is nonlinear, so the posterior (2.11) is not Gaussian. Thus, we will instead sample this posterior and develop an approximate model for the samples, composed of a mixture of Gaussians. However, we can obtain some important insight by supposing that we already know the true state and then linearizing the observation operator about it,
where $H \equiv \boldsymbol {\nabla }h({\textit {x}}^{\ast }) \in \mathbb {R}^{d \times n}$, the Jacobian of the observation operator at the true state. Then we can derive an approximating $n$-dimensional Gaussian model about mean ${\textit {x}}^{\ast }$ (plus a bias due to noise in the realization of the true measurements) with covariance
A brief derivation of this result is included in Appendix C.
We will refer to $\varSigma ^{\ast }_{{\textit {X}}}$ as the ‘true’ state covariance. It is useful to note that the matrix $H^{{\rm T}}\varSigma _{\mathcal{E}}^{-1}H$ is the so-called Fisher information matrix for our Gaussian likelihood (Cui & Zahm Reference Cui and Zahm2021), evaluated at the true state. In other words, it quantifies the information about the state of the system that is available in the measurements. Because all sensors have the same noise variance, $\sigma _{\mathcal{E}}^{2}$, then $\varSigma ^{\ast }_{{\textit {X}}} = \sigma _{\mathcal{E}}^{2}(H^{{\rm T}}H)^{-1}$. We can then use the singular value decomposition of the Jacobian, $H = U S V^{{\rm T}}$, to write a diagonalized form of the covariance,
Here, the eigenvalue matrix is $\varLambda = \sigma _{\mathcal{E}}^{2} D^{-1}$, where $D = S^{{\rm T}}S \in \mathbb {R}^{n \times n}$ is a diagonal matrix containing squares $s^{2}_{j}$ of the singular values $S$ of $H$ in decreasing magnitude up to the rank $r \leq \min (d,n)$ of $H$ and padded with $n - r$ zeros. The uncertainty ellipsoid thus has semi-axis lengths
along the directions $v_{j}$ given by the corresponding columns of $V$. Thus, the greatest uncertainty $\lambda ^{1/2}_{n}$ is associated with the smallest singular value $s_{n}$ of $H$. The corresponding eigenvector, $v_{n}$, indicates the mixture of states for which we have the most confusion.
In fact, the smallest singular values of $H$ are necessarily zero if $n > d$, i.e. when there are fewer sensors than states and the problem is therefore underdetermined. In such a case, $H$ has a null space spanned by the last $n-r$ columns in $V$. However, ${\textit {x}}^{\ast }$ is not a unique solution in these problems; rather, it is simply one element of a manifold of vortex states that produce equivalent sensor readings (to within the noise). The covariance $\varSigma ^{\ast }_{{\textit {X}}}$ evaluated at any ${\textit {x}}^{\ast }$ on the manifold reveals the local tangent to this manifold – directions near ${\textit {x}}^{\ast }$ along which we get identical sensor values. The true covariance $\varSigma ^{\ast }_{{\textit {X}}}$ will also be very useful for illuminating cases in which the problem is ostensibly fully determined ($n \leq d$), but for which the arrangement of sensors and true vortices nonetheless creates significant uncertainty in the estimate. In some of these cases, the smallest singular values may still be zero or quite small, indicating that the effective rank of $H$ is smaller than $n$.
2.4. Sampling and modelling of the posterior
The true covariance matrix (2.14) and its eigendecomposition will be an important tool in the study that follows, but we generally will only use it when we presume to know the true state ${\textit {x}}^{\ast }$ and seek illumination on the estimation in the vicinity of the solution. To characterize the problem more fully and explore the potential multiplicity of solutions, we will generate samples of the posterior and then fit the samples with an approximate distribution $\hat {{\rm \pi} }_{{\textit {y}}^{\ast }}({\textit {x}}) \approx {\rm \pi}({\textit {x}}\,|\,{\textit {y}}^{\ast })$ over ${\textit {x}}$. The overall algorithm is shown in the flowchart in figure 3 in the context of an example of estimating one vortex with three sensors. For the sampling task, we use the Metropolis–Hastings method (see, e.g. Chib & Greenberg Reference Chib and Greenberg1995), a simple but powerful form of MCMC. This method relies only on differences of the log probabilities between a proposed chain entry and the current chain entry to determine whether the proposal is more probable than the previous entry and should be added to the chain of samples. To ensure that the MCMC sampling does not get stuck in one of possibly multiple high-probability regions, we use the method of parallel tempering (Sambridge Reference Sambridge2014).
In practice, we generally have found good results in parallel tempering by using five parallel Markov chains exploring the target distribution raised to respective powers $3.5^{p}$, where $p$ takes integer values between $-4$ and $0$. We initially carry out $10^{4}$ steps of the algorithm with the MCMC proposal variance set to a diagonal matrix of $4\times 10^{-4}$ for every state component. Then, we perform $10^{6}$ steps with proposal variances set more tightly, uniformly equal to $2.5\times 10^{-5}$. The sample data set is then obtained from the $p = 0$ chain after discarding the first half of the chain (for burn in) and retaining only every $100$th chain entry of the remainder to minimize autocorrelations in the samples. On a MacBook Pro with a M1 processor, the overall process takes around 20 seconds in the most challenging cases (i.e. the highest-dimensional state spaces). There are other methods that move more efficiently in the direction of local maxima (e.g. such as hybrid MCMC, which uses the Jacobian of the log posterior to guide the ascent). However, the approach we have taken here is quite versatile for more general cases, particularly those in which the Jacobian is impractical to compute repetitively in a long sequential process.
To approximate the posterior distribution over ${\textit {x}}$ from the samples, we employ a Gaussian mixture model (GMM), which assumes the form of a weighted sum of $K$ normal distributions (the mixture components)
where $\bar {{\textit {x}}}^{(k)}$ and $\varSigma ^{(k)}$ are the mean and covariance of component $k$ of the mixture. Each weight $\alpha _{k}$ lies between 0 and 1 and represents the probability of a sample point ‘belonging to’ component $k$, so it quantifies the importance of that component in the overall mixture. For a given set of samples and a pre-selected number of components $K$, the parameters of the GMM ($\alpha _{k}$, $\bar {{\textit {x}}}^{(k)}$, and $\varSigma ^{(k)}$) are found via the expectation maximization algorithm (Bishop Reference Bishop2006). It is generally advantageous to choose $K$ to be large (${\sim }5$–$9$) because extraneous components are assigned little weight, resulting in a smaller number of effective components. It should also be noted that, for the special case of a single mode of small variance, a GMM with $K=1$ and a sufficient number of samples approaches the Gaussian approximation about the true state ${\textit {x}}^{\ast }$ described in § 2.3, with covariance $\varSigma ^{\ast }_{{\textit {X}}}$.
As we show in Appendix B, the GMM has a very attractive feature when used in the context of singular vortex elements because we can exactly evaluate the expectation of the vorticity field under the GMM probability,
where $\bar {\boldsymbol {r}}^{(k)}_{{\textit {J}}}$ and $\bar {\varGamma }^{(k)}_{{\textit {J}}}$ comprise the position and strength of the ${\textit {J}}$-vortex of mean state $\bar {{\textit {x}}}^{(k)}$ in (2.1), and $\varSigma ^{(k)}_{\boldsymbol {r}_{{\textit {J}}}\boldsymbol {r}_{{\textit {J}}}}$ and $\varSigma ^{(k)}_{\varGamma _{{\textit {J}}} \boldsymbol {r}_{{\textit {J}}}}$ are elements in the ${\textit {J}}{\textit {J}}$-block of covariance $\varSigma ^{(k)}$, defined in (2.4). Thus, under a GMM of the state, the expected vorticity field is itself composed of a sum of Gaussian-distributed vortices in Euclidean space (due to the first term in the square brackets), plus a sum of Gaussian-regularized dipole fields arising from covariance between the strengths and positions of the inferred vortex elements (the second term in square brackets).
3. Inference examples
Although we have reduced the problem by non-dimensionalizing it and restricting the possible states, there remain several parameters to explore in the vortex estimation problem: the number and relative configuration of the true vortices; the number of vortices used by the estimator; the number of sensors $d$ and their configuration; and the measurement noise level $\sigma _{\mathcal{E}}$. We will also explore the inference of vortex radius $\epsilon$ when this parameter is included as part of the state. In the next section, we will explore the inference of a single vortex, using this example to investigate many of the parameters listed above. Then, in the following sections, we will examine the inference of multiple true vortices, to determine the unique aspects that emerge in this context.
3.1. Inference of a single vortex
In this section, we will explore cases in which both the true vortex system and our estimator consist of a single vortex. We will use this case to draw insight on many of the parameters of the inference problem. For most of the cases, a single line of sensors along the $x$ axis will be used. The true vortex will remain on the $y = 1$ line and have unit strength. Many of the examples that follow will focus on the true configuration $(x_{1},y_{1},\varGamma _{1}) = (0.5,1,1)$. (The subscript 1 is unnecessary with only one vortex, but allows us to align the notation with that of § 2.1). Note, however, that we will not presume knowledge of any of these states: the bounding region of our prior will be $x \in (-2,2)$, $y \in (0.01,4)$, $\varGamma \in (0,2)$.
All of the basic tools used in the present investigation are depicted in figure 3, in which the true configuration is estimated with three sensors arranged uniformly along the $x$ axis between $[-1,1]$ with $\sigma _{\mathcal{E}} = 5\times 10^{-4}$. Figure 3(a) shows the ellipsoid for covariance $\varSigma ^{\ast }_{{\textit {X}}}$, computed at the true vortex state. This figure particularly indicates that much of the uncertainty lies along a direction that mixes $y_{1}$ and $\varGamma _{1}$; indeed, the eigenvector corresponding to the direction of greatest uncertainty is $(0.08,0.79,0.61)$. This uncertainty is intuitive: as a vortex moves further away from the sensors, it generates very similar sensor measurements if its strength simultaneously increases in the proportion indicated by this direction. In figure 3(b), the samples obtained from the MCMC method are shown. Here, and in later figures in the paper, we show only the vortex positions of these samples in Euclidean space and colour their symbols to denote the sign of their strength (blue for positive, red for negative). The set of samples clearly encloses the true state, shown as a block dot; the expected value from the samples is $(0.51,1.07,1.05)$, which agrees well.
The samples also demonstrate the uncertainty of the estimated state. The filled ellipse in this figure corresponds to the exact covariance of figure 3(a) and is shown for reference. As expected, the samples are spread predominantly along the direction of the maximum uncertainty. This figure also depicts an elliptical region for each Gaussian component of the mixture computed from the samples. These ellipses correspond only to the marginal covariances of the vortex positions and do not depict the uncertainty of the vortex strength. The weight of the component in the mixture is denoted by the thickness of the line. One can see from this plot that the GMM covers the samples with components, concentrating most of the weight near the centre of the cluster with two dominant components. The composite of these components is best seen in figure 3(d), in which the expected vorticity field is shown. In the remainder of this paper, this expected vorticity field will be used extensively to illuminate the uncertainty of the vortex estimation. Finally, figure 3(c) compares the true sensor pressures with those corresponding to the expected state from the MCMC samples. These agree to within the measurement noise.
3.1.1. Effect of the number of sensors
In the last section, we found that three sensors were sufficient to estimate a single vortex's position and strength. In this section we investigate how this estimate of a single vortex depends on the number of sensors. In most cases, these sensors will again lie uniformly along the $x$ axis in the range $[-1,1]$. Intuitively, we expect that if we have fewer sensors than there are states to estimate, we will have insufficient information to uniquely identify the vortex. Figure 4 shows that this is indeed the case. In this example, only two sensors are used to estimate the same vortex as in the previous example. The MCMC samples are distributed along a circular arc, but are truncated outside of the aforementioned bounding region. In fact, this arc is a projection of a helical curve of equally probable states in the three-dimensional state space. The samples broaden from the arc the further they are from the sensors due to the increase in uncertainty with distance. The true covariance, $\varSigma ^{\ast }_{{\textit {X}}}$, cannot reveal the full shape of this helical manifold, which is inherently dependent on the nonlinear relationship between the sensors and the vortex. However, the rank of this covariance decreases to 2, so that the uncertainty along one of its principal axes must be infinite. This principal axis is tangent to the manifold of possible states, as shown by a line in the plot.
What if there are more sensors than states? Figure 5(a–c) depicts expected vorticity fields for several cases in which there are increasing numbers of sensors arranged along a line, and figure 5(d) shows the expected vorticity when 5 sensors are instead arranged in a circle of radius 2.1 about the true vortex. (The choice of radius is to ensure that the smallest distance between the true vortex and a sensor is approximately 1 in all cases.) It is apparent that the uncertainty shrinks when the number of sensors increases from 3 to 4, but does so less notably when the number increases from 4 to 5. In figure 5(e), the maximum uncertainty is seen to drop by nearly half when one sensor is added to the basic set of 3 along a line, but decreases much more gradually when more than 4 sensors are used. The drop in uncertainty is more dramatic between 3 and 5 sensors arranged in a circle, but becomes more gradual beyond 5 sensors.
3.1.2. Effect of the true vortex position
It is particularly important to explore how the uncertainty is affected by the position of the true vortex relative to the sensors. We address this question here by varying this position relative to a fixed set of sensors with a fixed level of noise, $\sigma _{\mathcal{E}} = 5\times 10^{-4}$. Figure 6(a) depicts contours (on a log scale) of the resulting maximum length of the covariance ellipsoid, $\lambda _{n}^{1/2}$, based on four sensors placed on the $x$ axis between $-1$ and $1$. The contours reveal that there is little uncertainty when the true vortex is in the vicinity of the sensors, but the uncertainty increases sharply with distance when the vortex lies outside the extent of sensors. Indeed, one finds empirically that the rates of increase scale approximately as $\lambda _{n}^{1/2} \sim |y_{1}|^{5}$ and $\lambda _{n}^{1/2} \sim |x_{1}|^{6}$. This behaviour does not change markedly if we vary the number of sensors, as illustrated in figure 6(b). As the true vortex's $x$ position varies (and $y_{1}$ is held constant at 1), there is a similarly sharp rate of increase outside of the region near the sensors for 3, 4 or 5 sensors. However, although there is a small range of positions near $x_{1} = 0$ in which 3 sensors have less uncertainty than 4, there is generally less uncertainty at all vortex positions with increasing numbers of sensors. Furthermore, the uncertainty is less variable in this near region when 4 or 5 sensors are used.
3.1.3. Effect of sensor noise
From the derivation in § 2.3, we already know that the true covariance should depend linearly on the noise variance (2.16). In this section, we explore the effect of sensor noise on the estimation of a single vortex using MCMC and the subsequent fitting with a Gaussian mixture model. We keep the number of sensors fixed at 3 arranged along a line between $x = -1$ and $1$, and the true vortex in the original configuration, $(x_{1},y_{1},\varGamma _{1}) = (0.5,1,1)$. Figure 7 depicts the expected vorticity field as the noise standard deviation $\sigma _{\mathcal{E}}$ increases. Unsurprisingly, the expected vorticity distribution exhibits increasing breadth as the noise level increases. However, it is notable that this breadth becomes increasingly directed away from the sensors as the noise increases. Furthermore, the centre of the distribution lies somewhat further from the sensors than the true state, indicating a bias error. This trend toward increased bias error with increasing sensor noise is also apparent in other sensor numbers and arrangements.
3.1.4. Effect of true vortex radius
Throughout most of this paper, the radius of the true vortices is fixed at $\epsilon = 0.01$, and the estimator vortices share this same radius. However, for practical application purposes, it is important to explore the extent to which the estimation is affected by a mismatch between these. If the true vortex is more widely distributed, can an estimator consisting of a small-radius vortex reliably determine its position and strength? Furthermore, can the radius itself be inferred? These two questions are closely related, as we will show. First, it is useful to illustrate the effect of the vortex radius on the pressure field associated with a vortex, as in figure 8(a), which shows the vortex-pressure kernel $P_{\epsilon }$ for two different vortex radii, $\epsilon = 0.01$ and $\epsilon = 0.2$. As this plot shows, the effect of vortex radius is fairly negligible beyond a distance of 5 times the larger of these two vortex radii.
As a result of this diminishing effect of vortex radius, one expects that it is very challenging to estimate this radius from pressure sensor measurements outside of the vortex core. Indeed, that is the case, as figure 8(b) shows. This figure depicts the maximum length of the covariance ellipsoid as a function of true vortex radius, when four sensors along the $x$ axis are used to estimate this radius (in addition to vortex position and strength), for a true vortex at $(0.5,1)$ with strength 1. The uncertainty is far too large for the radius to be observable until this radius approaches $\epsilon = 1$. Even when the sensors are within the core of the vortex, they are confused between the blob radius and other vortex states. In fact, one can show that the dependence of the maximum uncertainty on $\epsilon ^{-3}$ arises because of the nearly identical sensitivity that the pressure sensors have to changes of blob radius and changes in other states, e.g. vertical position in this case. The leading-order term of the difference is proportional to $\epsilon ^{3}$. Of course, if we presume precise knowledge of the other states, then vortex radius becomes more observable.
The insensitivity of pressure to vortex radius has a very important benefit, because it ensures that, even when the true vortex is relatively broad in size, a vortex estimator with small radius can still reliably infer the vortex's position and strength. This is illustrated in figure 8(c), which depicts the contours of a true vortex of radius $\epsilon = 0.2$ (with the same position and strength as in (b)), and the expected vorticity contours from an estimate carried out with a vortex element of radius $\epsilon = 0.01$. It is apparent from this figure that the centre of the vortex is estimated very well. In fact, the mean of the MCMC samples is $(\bar {x}_{1},\bar {y}_{1},\bar {\varGamma }_{1}) = (0.51, 1.08, 1.04)$, quite close to the true values.
3.2. Inference of multiple vortices
The previous sections have illustrated the crucial aspects of estimating a single vortex. In this section, we focus on inferring multiple vortices. As in the case with a single vortex, the eigenvalues and eigenvectors of the true covariance ellipsoid $\varSigma ^{\ast }_{{\textit {X}}}$ will serve an important role in revealing many of the challenges in this context. However, some of the multiple vortex cases will have multiple possible solutions, and we must rely on the MCMC samples to reveal these. In the examples carried out in this section, we use the same uniform prior as in the previous section, except that now the prior vortex strengths can be negative (lying in the range $(-2,2)$).
3.2.1. Two true vortices with a two-vortex estimator
In this section, we will study the inference of a pair of vortices using a two-vortex estimator. The basic configuration of true vortices consists of $(x_{1},y_{1},\varGamma _{1}) = (-0.75,0.75,1.2)$ and $(x_{2},y_{2},\varGamma _{2}) = (0.5,0.5,0.4)$, both of radius $\epsilon = 0.01$. As in the previous section, sensors have noise $\sigma _{\mathcal{E}} = 5\times 10^{-4}$. In figure 9 we demonstrate the estimation of this pair of vortices with eight sensors. The estimator is very effective at inferring the locations and strengths of the individual vortices: the mean state of the samples is $(\bar {x}_{1},\bar {y}_{1},\bar {\varGamma }_{1}) = (-0.75,0.77,1.23)$ and $(\bar {x}_{2},\bar {y}_{2},\bar {\varGamma }_{2}) = (0.50,0.50,0.39)$, and the pressure field predicted by the expected state matches well with the true pressure field.
In this basic configuration, the vortices are widely separated so that the estimator's challenge is similar to that of two isolated single vortices, each estimated with four sensors. However, unique challenges arise as the true vortices become closer, as figure 10 shows. Here, we keep the strength and vertical position of each true vortex the same as in the basic case, but vary both vortices’ horizontal position – the left one is moved rightward and the right one is moved leftward – in such a manner that their average is invariant, $(x_{1}+x_{2})/2 = -0.125$. Three different numbers of sensors are used, $d = 6, 7, 8$, all uniformly distributed in the range $[-1,1]$. In figure 10(a), it is clear that using six sensors, although ostensibly sufficient to estimate the six states, is actually insufficient in a few isolated cases in which the maximum uncertainty becomes infinite. These cases are examples of rank deficiency in the vortex estimator. Importantly, this rank deficiency disappears when more than six sensors are used. An example of the estimator's behaviour in one of these rank-deficient configurations is depicted in figure 10(b,c). When six sensors are used (b), the MCMC samples are distributed more widely, along a manifold in the vicinity of the true state, with the eigenvector of the most-uncertain eigenvalue tangent to this manifold. However, when seven sensors are used (c), the MCMC samples are more tightly distributed around the true state.
Thus, we can avoid rank deficiency by using more sensors than states. As a demonstration, we show in the left panels of figure 11 the expected vorticity field that results from estimating four different true vortex configurations with eight sensors. In each case, the locations of the vortices are accurately estimated with relatively little uncertainty, even as the vortices become closer to each other than they are to the array of sensors. However, with closer vortices there is considerable uncertainty in estimating the strengths of the individual vortices, as exhibited in the right panels of figure 11, each corresponding to the vortex configuration on the left.
As the true vortices become even closer than in the examples in figure 11, multiple solutions emerge. This is illustrated in figure 12, depicting the extreme case of one vortex just above the other. The MCMC identifies three modes of the posterior, each representing a different candidate solution for the estimator. One mode consists of vortices of opposite sign to either side of the true set, shown in the top row of figure 12. The second mode, in the middle row, comprises vortices very near the true set, although the strengths of the vortices are quite uncertain, as evidenced by the long ridge of samples in the strength plot in figure 12(d). Finally, the bottom row shows a mode that has positive vortices further apart than in the other two modes.
It is natural to ask whether we can prefer one of these two candidate solutions over the other. One way to do so is to assess them based on their corresponding weights $\alpha _{k}$ in the mixture model, since each of these represents the probability of a given sample point belonging to that component. However, interpreting the mixture model's weights in this fashion requires that the MCMC has reached equilibrium, which can be challenging to determine with multimodal sampling. Instead, we follow the intuition that, if a mode is to be considered a more likely solution of the inference problem than another mode, then the samples belonging to that mode should be closer to the true observation. For this assessment we can compare the maxima of the log posterior (2.12) among the samples belonging to two modes. The mode with a significantly larger maximum (i.e. significantly closer to zero, since (2.12) is non-positive) is a superior candidate solution. For the two modes shown in figure 12, the maximum log posteriors are $-0.20$, $-0.11$ and $-14.83$, respectively, suggesting that the mode in the middle row is mildly superior to that of the top row and clearly superior to that of the bottom row. Indeed, the fact that this clearly inferior mode appears among the samples at all is likely due to incomplete MCMC sampling. Thus, in this example with two very closely spaced positive-strength vortices, the true solution is discernible from the two spurious solutions.
In figure 13 we carry out the same procedure of bringing two vortices closer together as in figure 11, but now we do so for one vortex of positive strength ($1.2$) and another of negative strength ($-1.0$). We get similar results as before, successfully estimating the vortex locations and strengths. The most challenging case among these is the first, in which the two vortices are furthest apart and near the extreme range of the sensors. Interestingly, no spurious solutions arise as the vortices become very close together, as they did in the previous example. In fact, when the two opposite-sign vortices are vertically aligned, as in figure 14, the estimator has no difficulty in identifying the individual vortices and their strengths. The figure ostensibly depicts two modes identified by the estimator, but in fact these modes are identical aside from the sign of their strengths. They remain distinct to the estimator only because they have eluded our simple mitigations for the relabelling and strength symmetries (of ordering the vortices by their $x$ position and assuming that the leftmost has positive strength). Clearly we could have chosen a different mitigation strategy to avoid this, but we include the separate modes here for illustration purposes.
3.2.2. Three true vortices
In this section we demonstrate the performance of the estimator on cases with three true vortices. In the first example, we will use three vortices in the estimator. Let the true state comprise ${\textit {x}}_{1} = (x_{1},y_{1},\varGamma _{1}) = (-0.5,0.5,1)$, ${\textit {x}}_{2} = (0.25,0.5,-1.2)$ and ${\textit {x}}_{3} = (0.75,0.75,1.4)$. Here, we apply the techniques shown to enhance performance in the previous sections: we use eleven sensors, two more than the marginal number, to avoid rank deficiency; and we choose a top candidate among the identified modes based on the maximum value of the log posterior. The resulting estimated solution has only a single candidate mode, whose mean is $\bar {{\textit {x}}}_{1} = (-0.47,0.53,1.10)$, $\bar {{\textit {x}}}_{2} = (0.25,0.51,-1.25)$, $\bar {{\textit {x}}}_{3} = (0.68,0.75,1.36)$. This is shown in figure 15. Both the expected vorticity field and the pressure field are captured very well by the estimator.
It is important to note that, in the previous example, the estimator identifies another mode (not shown) in which the signs of the rightmost two vortices are switched. However, this candidate solution was discarded on the basis of the maximum log-posterior criterion: the selected mode's value is $-0.72$, while the discarded mode's is smaller, $-0.85$. This slight difference is entirely due to the nonlinear coupling between the vortices via the interaction kernel. By restricting the leftmost vortex to be positive, the signs of the middle and rightmost vortices are established through this coupling. Because the algebraic decay of both types of kernels in (1.2) is the same, the ability to discriminate the signs of the vortices requires a degree of balance between the vortex positions and the sensors. As a counterexample, when two of the three vortices form a compact pair that is well separated from the third, the estimator tends to be less able to prefer one choice of sign for the vortex pair over the other. An example is shown in figure 16, in which the rightmost pair of vortices has opposite sign in each mode. The corresponding pressure fields shown on the right are nearly identical because the coupling of the pair with the leftmost vortex is much weaker than in the pair itself. Thus, these modes are indistinguishable by our maximum log-posterior criterion: in the absence of additional prior knowledge, we cannot discern one from the other.
The three-vortex estimator must explore a nine-dimensional space for the solution, a challenging task even with the various MCMC and symmetry mitigation techniques we have used in this paper. Thus, it is useful to restrict the estimator to search a lower-dimensional space, and the easiest way to achieve this is by using fewer vortices in the estimator. In figure 17, we illustrate the behaviour of a two-vortex estimator on the three-vortex configuration in figure 15, in variations in which the signs of the right two true vortices are changed. The range of strengths in the prior is expanded in this problem to $(-4,4)$. In the first case, the true vortex strengths are all positive. The two-vortex estimator identifies a single mode, with mean vortex states $\bar {{\textit {x}}}_{1} = (-0.54, 0.37, 0.64)$ and $\bar {{\textit {x}}}_{2} = (0.48, 0.74, 3.25)$, and a maximum log posterior of $-30.7$. In other words, the estimator places one vortex near (and slightly weaker than) the leftmost true vortex, and another vortex near the centre of the rightmost pair, with a strength roughly equal to the sum of the pair. In the second case, the two rightmost vortices are both negative, and the estimator produces an analogous result, aggregating the two negative vortices into a single vortex. The estimated state is $\bar {{\textit {x}}}_{1} = (-0.34, 0.69, 1.27)$ and $\bar {{\textit {x}}}_{2} = (0.40, 0.62, -3.07)$, with maximum log posterior $-20.4$, so that the rightmost pair is once again approximated by a single vortex with roughly the sum of the pair's strength.
The third case is the most interesting. Here, the true vortex configuration consists of positive, negative and positive vortices from left to right, so there is no pairing of like-sign vortices as in the previous two cases. The estimator identifies a solution consisting of $\bar {{\textit {x}}}_{1} = (0.40, 1.03, 3.50)$ and $\bar {{\textit {x}}}_{2} = (0.90, 0.97, -1.2)$. Neither of these vortices bears an obvious connection with one of the true vortices, so no aggregation is possible. The estimator has done the best in can in the lower-dimensional space available to it, aliasing the true flow onto a dissimilar flow state. The maximum log posterior is $-88.6$, significantly lower than in the other two cases.
4. Conclusions
In this paper, we have explored the inference of regularized point vortices from a finite number of pressure sensors with noisy measurements. By expressing the problem in a Bayesian (probabilistic) manner, we have been able to quantify the uncertainty of the estimated vortex state and to explore the multiplicity of possible solutions, which are expressed as multiple modes in the posterior distribution. We sampled the posterior with MCMC and applied Gaussian mixture modelling to develop a tractable approximation for the posterior from the samples. Mixture modelling allowed us to soft classify the samples into each mode. We reduced the multiplicity by anticipating many of the symmetries that arise in this inference problem – strength, relative position and vortex re-labelling – and then mitigated their influence through simple techniques: e.g. restricting the prior region, strictly ordering the vortices in the state vector by $x$ coordinate. The remaining multiplicity of solutions were identified by thoroughly exploring the prior region with help from the method of parallel tempering in MCMC. Where possible, the best candidate solution was discerned by monitoring the maximum log posterior in each mode. We have also made use of the largest eigenvalue and associated eigenvector of the true covariance matrix in order to illuminate many of the challenges of the inference.
On a variety of configurations of one, two or three true vortices, we have made several observations of this vortex inference problem. One must use at least as many sensors as there are estimator states in order to infer a unique vortex system rather than a manifold of equally possible states. Using one additional sensor guards against the risk of cases of rank deficiency, which arise occasionally when multiple vortices are used in the estimator. However, additional sensors do not significantly improve the uncertainty of the estimate. Uncertainty scales linearly with sensor noise. It also rises very rapidly, with the fifth or sixth power of distance, when the true vortex lies outside of a region of the sensors. The size of the vortex is exceptionally challenging to estimate because its effect on pressure is almost indistinguishable from other vortex states. However, this fact is also advantageous, for it allows us to use a small radius (nearly singular) vortex to accurately estimate the position and strength of a larger one. For systems of multiple vortices, the estimator relies on the nonlinear coupling between them to ascertain the sign of the strength of each. Even when multiple modes emerge, one can often discern the best candidate among the modes based on the criterion of maximum probability (i.e. the shortest distance to the true measurements). This approach fails in some cases when the vortices are imbalanced, such as when a pair of vortices is well separated from a third. When the estimator uses fewer vortices than in the true configuration, it identifies the most likely solution in the reduced state space. Often, this reduced-order estimate appears to be a natural aggregation of the true vortex state, but in some cases the estimator aliases the sensors onto a dissimilar vortex configuration when no aggregated one is possible.
It is important to reiterate that the static inference we have studied in this paper is useful both in its own right as a one-time estimate of the vorticity field and as part of a sequential estimation of a time-varying flow. In the latter context – for example, in an ensemble Kalman filter (da Silva & Colonius Reference da Silva and Colonius2018; Darakananda & Eldredge Reference Darakananda and Eldredge2019; Le Provost & Eldredge Reference Le Provost and Eldredge2021; Le Provost et al. Reference Le Provost, Baptista, Marzouk and Eldredge2022) – the inference comprises the analysis part of every step, when data are assimilated into the prediction. Some of the challenges and uncertainty of the static inference identified in this paper are overcome with advancing time as the sensors observe the evolving configuration of the flow and the forecast model predicts the state's evolution. Indeed, we speculate that the rank deficiency is partially mitigated in this manner. However, as new vortices are generated or enter the region of interest, the prior distribution obtained from the previous step will not be descriptive of these new features. Thus, our assumption of a non-informative prior remains relevant ever after the initial step of a flow, and the conclusions we have drawn in this work can guide an estimator that remains receptive to new vortex structures. Overall, most of the conclusions of this paper, including rank deficiency, the decay of signal strength with distance and the effects of vortex couplings, are impactful on a sequential filter's overall performance. It is also important to stress the fact that, when the vortex state can be unambiguously inferred from a set of sensors, it implies that the sensor data contain sufficient information to describe the flow. This has implications for RL-based flow control approaches (Verma et al. Reference Verma, Novati and Koumoutsakos2018), which treat the fluid flow as a partially observable Markov decision process. Pressure sensor data potentially make the process more Markovian, and therefore more amenable to learning a control policy (Renn & Gharib Reference Renn and Gharib2022). In short, there is less risk that control decisions made from measured pressures data are working on a mistaken belief about the flow state.
Finally, although we have addressed many substantive questions in this work, there are still several to address: In a realistic high Reynolds number flow, i.e. comprising a few dominant coherent structures amidst shear layers and small-scale vortices, can the estimator infer the dominant vortices? Our work here has exhibited some key findings that suggest that it can; in particular, weaker vortices with comparatively little effect on the velocity and small measured pressures are neglected by the estimator in favour of stronger vortices. However, this question deserves more thorough treatment. Also, how does the presence of a body affect the vortex estimation? Furthermore, when such a body is in motion, or subject to a free stream, can the vortices and the body motion be individually inferred? In the presence of a body, stationary or in motion, it is straightforward to expand the pressure–vortex relation presented here to develop an inviscid observation model that incorporates geometry, other flow contributors and their couplings (e.g. between a vortex and a moving body). The estimation framework presented here can readily accommodate such an enriched model. Indeed, in our prior work with an ensemble Kalman filter (Darakananda & Eldredge Reference Darakananda and Eldredge2019; Le Provost & Eldredge Reference Le Provost and Eldredge2021), we have found that the signal of a vortex can be enhanced in pressure sensors on a nearby wall due to the effect of the vortex's image, and that this inviscid observation model remains effective in a viscous setting, such as flow past a flat plate. However, there are challenges with estimating vortices near regions of large wall curvature, such as the edges of an airfoil, where large pressure gradients, viscous effects and the subsequent generation of new vortices complicate the flow process. In our previous work, we have approximately accounted for these processes by augmenting the state with a leading-edge suction parameter (Ramesh et al. Reference Ramesh, Gopalarathnam, Granlund, Ol and Edwards2014). However, it would also be worthwhile to investigate approaches in which the physics-based observation model is replaced or augmented with a data-driven approach, such as a trained neural network (Zhong et al. Reference Zhong, Fukami, An and Taira2023).
Funding
Support for this work by the National Science Foundation under Award number 2247005 is gratefully acknowledged.
Declaration of interests
The authors report no conflict of interest.
Appendix A. Pressure and vorticity in unbounded flow
A.1. Pressure from vorticity in unbounded flow
We start by writing the Poisson equation for pressure, obtained by taking the divergence of the incompressible Navier–Stokes equations and using the fact that velocity is divergence free. To emphasize the role of vorticity, we first re-write the convective term in the equations with the help of the vector identity $\boldsymbol {u}\boldsymbol {\cdot } \boldsymbol {\nabla }\boldsymbol {u} = \boldsymbol {\nabla } |\boldsymbol {u}|^{2}/2 - \boldsymbol {u} \times \boldsymbol {\omega }$, where $\boldsymbol {\omega } = \boldsymbol {\nabla }\times \boldsymbol {u}$, obtaining
The quantity $\boldsymbol {u}\times \boldsymbol {\omega }$ is the so-called Lamb vector (Lamb Reference Lamb1932). To formally solve this problem for the quantity in parentheses, we can make use of the free-space Green's function for the (negative) Laplace operator, i.e. the solution of
so that, by Green's theorem (and integration by parts), we get the pressure to within a uniform constant, $C$ (actually, any homogeneous solution of Laplace's equation could be added to this expression, but in this unbounded domain only a uniform constant will permit a finite pressure at infinity),
where $\mathrm {d}V(\boldsymbol {r}')$ denotes the infinitesimal volume element at point $\boldsymbol {r}'$, and the integral is taken over the entire space. Thus, the pressure at any observation point $\boldsymbol {r}$ is partly attributable to the velocity magnitude at $\boldsymbol {r}$ (the so-called dynamic pressure) and partly induced by the Lamb vector field. The dynamic pressure's role is simple and familiar: faster local flow speed is associated with lower pressure. The role of the Lamb vector is less familiar, so it is helpful to manipulate this integral further.
The Lamb vector field is clearly only distributed over regions that are vortical. To further elucidate the role of vorticity, we will make use of the fact that the velocity can itself be recovered from the vorticity field, via the Biot–Savart integral, plus any additional irrotational contributions represented by a scalar potential field. In an unbounded context this irrotational contribution is a possible uniform flow, $\boldsymbol {U}_{\infty }$. Thus,
where
For clarity in what follows, it is useful to denote the contribution from each infinitesimal volume element by
so the Biot–Savart integral can be written more compactly as
where the notation indicates that we are taking this integral over all points $\boldsymbol {r}'$ in space. Interestingly, this infinitesimal velocity contribution also appears in the Lamb vector integral in (A3), made apparent by permuting the triple product in the integral so that we can write it alternatively as
To eliminate the constant $C$, we note that, provided the uniform flow is steady, the limits as $|\boldsymbol {r}| \rightarrow \infty$ are $p(\boldsymbol {r}) \rightarrow p_{\infty }$ (the ambient pressure), $\boldsymbol {u}(\boldsymbol {r}) \rightarrow \boldsymbol {U}_{\infty }$ and $\mathrm {d}\boldsymbol {u}_{\boldsymbol {\omega }}(\boldsymbol {r},\boldsymbol {r}') \rightarrow 0$. Thus, $C = p_{\infty } + \frac {1}{2} \rho |\boldsymbol {U}_{\infty }|^{2}$ and
This form of the pressure is reminiscent of the familiar Bernoulli equation. In fact, in the special case of an inviscid flow comprising singular distributions of vorticity (e.g. point vortices or vortex filaments), then it can be shown that the integral is equivalent to $-\rho \partial \varphi /\partial t$, where $\varphi$ is the equivalent scalar potential field induced by the singular vortices. The unsteadiness of this potential field at the observation point $\boldsymbol {r}$ is brought about by the convection and tilting of the vortices by their local velocity field, associated with the presence of $\boldsymbol {u}$ inside the integral. Thus, under those special circumstances, (A9) is identically the Bernoulli equation, as one would expect. For the general case of distributed vorticity, the equation no longer reverts to Bernoulli, but the pressure still receives an essential ‘velocity-modulated’ contribution from the vorticity.
We could finish our derivation with expression (A9). However, in order to distinguish the contributions from the uniform flow and the vorticity, we introduce (A4) for the velocity. Simplifying the resulting expression, we arrive at
It is important to observe that steady uniform flow makes no contribution whatsoever to the pressure in incompressible flow; the pressure in an unbounded flow is entirely due to vorticity (if the uniform flow is unsteady, then its rate of change contributes to pressure. Also, in a slightly compressible flow, the uniform flow would affect the pressure at large distance by modifying the travel time and directivity of acoustic waves between the vorticity and the observer Powell Reference Powell1964). The coupling between $\boldsymbol {U}_{\infty }$ and $\boldsymbol {u}_{\boldsymbol {\omega }}$ that arises in the dynamic pressure is cancelled by the modulation by $\boldsymbol {U}_{\infty }$ inside the integral. Thus, only the modulation of vorticity by velocity induced by other vortex elements matters to pressure.
Equation (A10) shows that vorticity has a quadratic effect on pressure. To reveal this effect more clearly, we replace $\boldsymbol {u}_{\boldsymbol {\omega }}$ in (A10) by the Biot–Savart integral (A5), obtaining
This form reveals an essentially triadic relationship between vorticity and pressure, illustrated in figure 18: the pressure at $\boldsymbol {r}$ comprises a double sum of elementary interactions between vorticity at $\boldsymbol {r}'$ and $\boldsymbol {r}''$. Interestingly, a consequence of this relationship is that the pressure is invariant to a change of sign of the entire vorticity field.
A.2. Pressure from point vortices in the plane
To illustrate this triadic interaction in a simple setting, let us consider a two-dimensional vorticity field consisting of two point vortices
The integrals can be evaluated exactly by virtue of the properties of the Dirac delta function, and in this two-dimensional setting, $\boldsymbol {\nabla } G(\boldsymbol {r}) = -\boldsymbol {r}/(2{\rm \pi} |\boldsymbol {r}|^{2})$. The velocity induced by vortex ${\textit {J}}$ is $\boldsymbol {u}_{{\textit {J}}}(\boldsymbol {r}) = \boldsymbol {\nabla }G(\boldsymbol {r}-\boldsymbol {r}_{{\textit {J}}}) \times \varGamma _{{\textit {J}}} \boldsymbol {e}_{z}$, and the resulting pressure field can be written as
where we have defined a direct vortex kernel for vortex ${\textit {J}}$
and a vortex interaction kernel
which we have split into additive parts arising from the dynamic pressure term and the Lamb vector term, respectively. In the expression for $\varPi ^{(2)}$, we have used the fact that the Green's function's gradient is skew–symmetric: $\boldsymbol {\nabla }G(\boldsymbol {r}-\boldsymbol {r}') = -\boldsymbol {\nabla }G(\boldsymbol {r}'-\boldsymbol {r})$.
There are a few notable features of expression (A13). First of all, each vortex makes an independent contribution to pressure via its direct vortex kernel. This direct contribution to the pressure field is always negative, regardless of the sign of the vortex, and is radially symmetric about the centre of the vortex. These direct contributions are modified by the vortex interaction kernel, in a term that introduces the signs of the individual vortices into the pressure field. This kernel, $\varPi (\boldsymbol {r}-\boldsymbol {r}_{{\textit {J}}},\boldsymbol {r}-\boldsymbol {r}_{{\textit {K}}})$, is dependent only on the relative positions of the observation point $\boldsymbol {r}$ from each of the two vortex positions, $\boldsymbol {r}_{{\textit {J}}}$ and $\boldsymbol {r}_{{\textit {K}}}$. It is symmetric with respect to the members of the pair, ${\textit {J}}$ and ${\textit {K}}$, as is apparent from figure 19, which shows the kernel and its two additive parts.
The vortex interaction kernel is centred midway between the pair at $\bar {\boldsymbol {r}}_{{\textit {J}}{\textit {K}}} = (\boldsymbol {r}_{{\textit {K}}}+\boldsymbol {r}_{{\textit {J}}})/2$ and has directivity as indicated in figure 19(a). It is apparent that the interaction kernel has much less influence along the pair's axis (the $\tau$ direction); its primary influence is perpendicular to this line, in the $n$ direction. We can write the kernel exactly as
where we denote the unit vector from vortex ${\textit {J}}$ to vortex ${\textit {K}}$ by $\boldsymbol {\tau }_{{\textit {J}}{\textit {K}}} = (\boldsymbol {r}_{{\textit {K}}} - \boldsymbol {r}_{{\textit {J}}})/d_{{\textit {J}}{\textit {K}}}$, and $d_{{\textit {J}}{\textit {K}}} = |\boldsymbol {r}_{{\textit {K}}} - \boldsymbol {r}_{{\textit {J}}}|$ the distance between the vortices. This form for the interaction kernel clearly shows that it, like the direct kernel, is purely positive. It is for this reason that we have explicitly pulled the negative sign out of (A13), to more clearly reveal the dependence of the sign of pressure on the vortex strengths.
Alternatively, we can write the kernel in a manner that emphasizes its scaling and directivity, as
where by $\tilde {\boldsymbol {r}} = (\boldsymbol {r} - \bar {\boldsymbol {r}}_{{\textit {J}}{\textit {K}}})/d_{{\textit {J}}{\textit {K}}}$ we denote the vector from the vortex centre to the observation point, re-scaled by $d_{{\textit {J}}{\textit {K}}}$. This form distinguishes the scaling due to the pair's separation distance (the $1/d_{{\textit {J}}{\textit {K}}}^{2}$ factor) from the scale-invariant directivity pattern. It also makes it easy to identify the far-field behaviour, $|\tilde {\boldsymbol {r}}| \gg 1$, or $|\boldsymbol {r} - \bar {\boldsymbol {r}}_{{\textit {J}}{\textit {K}}}| \gg d_{{\textit {J}}{\textit {K}}}$
Thus, the vortex interaction kernel has the same geometric decay, $1/r^{2}$, as that of the direct contributions and cannot generally be ignored. However, as (A17) shows, the vortex interaction kernel's contribution becomes weaker in the near field, or equivalently, as the pair becomes further separated.
A.3. Systems of vortices
It is a simple matter to extend the example of two point vortices to a larger set of $N$ point vortices. Using the same notation, the pressure field for this set is
The first term provides the direct contributions from each vortex; the second term comprises a sum over all of the $N(N+1)/2$ unique pairs in the set, using the vortex interaction kernel defined earlier. We can write this as a quadratic form in the vector of vortex strengths, $\boldsymbol {\varGamma } = [\varGamma _{1}\ \varGamma _{2}\cdots \varGamma _{N}]^{{\rm T}}$:
where the elements of the symmetric positive semi-definite matrix $\boldsymbol{\mathsf{P}}(\boldsymbol {r})$ are
A.4. Regularized point vortices
In practice, we often rely on regularized point vortex elements rather than singular elements. There are two interpretations of the regularization process when viewed from the infinitesimal contribution (A6). In one, we view the vorticity field as comprising a set of smooth blobs of vorticity in place of singular distribution. In the other, we still interpret the vorticity as a set of singular elements, but the Green's function and its gradient are replaced by versions that are convolved with a smooth regularization kernel. This convolution process is ultimately still present in the blob interpretation, after the infinitesimal contributions to velocity are integrated over space, so the interpretations both result in the same velocity field.
A common choice of regularization kernel – and the one we use in this work – is the smooth algebraic form
where $\epsilon$ is the blob radius. From this kernel, it can be shown that the regularized Green's function gradient is
It is important to note that regularized vortices are approximate solutions to the incompressible Euler's equations (Hald Reference Hald1979), and as such, the pressure is determined in the same way as for singular point vortices. To determine this pressure, we utilize the second interpretation above, in which the vortices are still regarded as singular but the velocity they induce is regularized. As a result, we simply replace all instances of $\boldsymbol {\nabla }G$ in the pressure kernels (A14) and (A15) with the regularized version (A23), leading to regularized versions of these kernels that we denote by $P_{\epsilon }$ and $\varPi _{\epsilon }$, respectively.
A.5. Grid-based vorticity
Finally, we should emphasize that, although we have derived the equations in this section for a finite system of point vortices, the forms of these relations also hold for a grid-based representation of a continuous (and unbounded) two-dimensional vorticity field, as in a computational fluid dynamics simulation. In that context, the sums are carried out over all grid points with non-negligible vorticity: each vortex position represents the location of a grid point and its corresponding strength is replaced by the vorticity at that point, multiplied by the area of the surrounding grid cell. An example of a three-vortex system, whose pressure is computed on a Cartesian grid, is shown in figure 20. To emphasize that the form extends to distributions of vorticity and not just isolated point vortices, this example depicts Gaussian-distributed vortices. The overall pressure field in the centre panel is the sum of the fields in all of the satellite panels. The satellite panels on the right exhibit the same interaction kernel – shifted, rotated and re-scaled for each pair.
Appendix B. The expectation of vorticity under a Gaussian mixture model for vortex states
In this section, we derive the expectation of the vorticity field (2.18) when the vortex states are described by a GMM $\hat {{\rm \pi} }({\textit {x}})$ given by (2.17). (We omit the subscript ${\textit {y}}^{\ast }$ from this probability here for brevity.) We start with the singular vorticity representation (1.1), and seek to evaluate the integral
where the notation $\boldsymbol {\omega }(\boldsymbol {r}, {\textit {x}})$ explicitly represents the dependence of the singular vortex field on the state vector ${\textit {x}}$. It is sufficient to consider the integration of just a single Gaussian component, since the overall expectation will simply be a linear combination of the $K$ components. Thus, we seek the integral
Let us recall that the state and covariance are organized as shown in (2.1) and (2.3), respectively. The integration over ${\textit {x}}$ is multiplicatively decomposable into integrals over the states of the individual vortex elements, $\mathrm {d}{\textit {x}} = \mathrm {d}{\textit {x}}_{1}\,\mathrm {d}{\textit {x}}_{2}\cdots \mathrm {d}{\textit {x}}_{N}$. When the integral in (B2) for vortex ${\textit {J}}$ in the sum is carried out, the integrals over all states of the other vortices ${\textit {I}} \neq {\textit {J}}$ represent marginalizations of the probability distribution over these vortices. Using properties of Gaussians (Bishop Reference Bishop2006), it is easy to show that this marginalized distribution is simply a Gaussian distribution over the states of vortex ${\textit {J}}$
Now, we can decompose the integral into the individual states of vortex ${\textit {J}}$, $\mathrm {d}{\textit {x}}_{{\textit {J}}} = \mathrm {d}\boldsymbol {r}_{{\textit {J}}}\,\mathrm {d}\varGamma _{{\textit {J}}}$, and the state ${\textit {x}}_{{\textit {J}}}$ and covariance $\varSigma _{{\textit {J}}{\textit {J}}}$ partitioned accordingly, as in their definitions (2.2) and (2.4). To assist the calculations that follow, it is useful to write the joint probability distribution for the strength and position $\hat {{\rm \pi} }(\varGamma _{{\textit {J}}},\boldsymbol {r}_{{\textit {J}}}) = \mathcal{N}({\textit {x}}_{{\textit {J}}}\,|\,\bar {{\textit {x}}}_{{\textit {J}}},\varSigma _{{\textit {J}}{\textit {J}}})$ in the conditional form ${\rm \pi} (\varGamma _{{\textit {J}}},\boldsymbol {r}_{{\textit {J}}}) = {\rm \pi}(\varGamma _{{\textit {J}}}\,|\,\boldsymbol {r}_{{\textit {J}}}) {\rm \pi}(\boldsymbol {r}_{{\textit {J}}})$, where
Again, using properties of Gaussians, the conditional probability ${\rm \pi} (\varGamma _{{\textit {J}}}\,|\,\boldsymbol {r}_{{\textit {J}}})$ can be shown (by completing the square) to be
where the mean and covariance are, respectively,
In this partitioned form, we can evaluate the integrals over $\boldsymbol {r}_{{\textit {J}}}$ and $\varGamma _{{\textit {J}}}$ in (B3). The integral over $\boldsymbol {r}_{{\textit {J}}}$ is particularly easy to evaluate because of the properties of the Dirac delta function. As a result, $\boldsymbol {r}_{{\textit {J}}}$ is replaced everywhere by the observation point $\boldsymbol {r}$. We are thus left with the integral
where $\boldsymbol {r}_{{\textit {J}}}$ is replaced by $\boldsymbol {r}$ in the mean, $\mu _{\varGamma _{{\textit {J}}}\,|\,\boldsymbol {r}_{{\textit {J}}}}$. This final integral is simply the expectation of $\varGamma _{{\textit {J}}}$ over the conditional distribution, and its value for a Gaussian is the mean, $\mu _{\varGamma _{{\textit {J}}}\,|\,\boldsymbol {r}_{{\textit {J}}}}$. Thus, we arrive at
The final result (2.18) follows easily by introducing (B8) into the mixture model.
Appendix C. Linearization of the observation operator in Bayes theorem
Let us suppose that the prior is Gaussian rather than uniform, with mean ${\textit {x}}_{0}$ and covariance $\varSigma _{0}$. (Ultimately, we will allow this covariance to become infinitely large.) Thus,
The likelihood is also assumed Gaussian about the observation prediction $h({\textit {x}})$ with covariance $\varSigma _{\mathcal{E}}$, as in (2.9), but now we will linearize the observation operator about the true state ${\textit {x}}^{\ast }$, as in (2.13). This can be written as
where $H \equiv \boldsymbol {\nabla }h({\textit {x}}^{\ast })$ is the Jacobian of the observation at the true state, and $b = h({\textit {x}}^{\ast }) - H{\textit {x}}^{\ast }$.
With Gaussian prior and likelihood and a linear relationship between ${\textit {y}}$ and ${\textit {x}}$, the joint distribution over these variables is also Gaussian. We can make use of the properties of multivariate Gaussians to obtain all of well-known results that follow; the reader is referred to Bishop (Reference Bishop2006) for more details. The mean and covariance of the joint variable ${\textit {z}} = ({\textit {x}},{\textit {y}})$ are, respectively,
and
To obtain the Gaussian form of the posterior distribution, we seek the mean and covariance of the conditional ${\rm \pi} ({\textit {x}}\,|\,{\textit {y}})$, which is obtained by starting from the log of the joint distribution
and rewriting it as a quadratic form in ${\textit {x}}$ only, with ${\textit {y}}$ set equal to the true observation, ${\textit {y}}^{\ast }$, and assumed known. Again, using well-known identities involving the inverse of a partitioned matrix, we arrive at the conditional mean and covariance
and
These results balance the prior mean and covariance with the information gained from the observation, ${\textit {y}}^{\ast }$. However, if the prior covariance grows to infinity, $\varSigma _{0}^{-1} \rightarrow 0$, reflecting our lack of prior knowledge, then all dependence on the prior vanishes, and we end up with
and
in which we have also substituted the specific form of $b$ in our linearized model and simplified. The second term in the mean represents a bias error that arises when the true observation differs from the model evaluated at the true state, as from the error $\varepsilon ^*$ in a single realization of the measurements.