Hostname: page-component-cd9895bd7-jn8rn Total loading time: 0 Render date: 2024-12-21T04:39:45.998Z Has data issue: false hasContentIssue false

Single-snapshot machine learning for super-resolution of turbulence

Published online by Cambridge University Press:  12 December 2024

Kai Fukami*
Affiliation:
Department of Mechanical and Aerospace Engineering, University of California, Los Angeles, CA 90095 USA
Kunihiko Taira
Affiliation:
Department of Mechanical and Aerospace Engineering, University of California, Los Angeles, CA 90095 USA
*
Email address for correspondence: [email protected]

Abstract

Modern machine-learning techniques are generally considered data-hungry. However, this may not be the case for turbulence as each of its snapshots can hold more information than a single data file in general machine-learning settings. This study asks the question of whether nonlinear machine-learning techniques can effectively extract physical insights even from as little as a single snapshot of turbulent flow. As an example, we consider machine-learning-based super-resolution analysis that reconstructs a high-resolution field from low-resolution data for two examples of two-dimensional isotropic turbulence and three-dimensional turbulent channel flow. First, we reveal that a carefully designed machine-learning model trained with flow tiles sampled from only a single snapshot can reconstruct vortical structures across a range of Reynolds numbers for two-dimensional decaying turbulence. Successful flow reconstruction indicates that nonlinear machine-learning techniques can leverage scale-invariance properties to learn turbulent flows. We also show that training data of turbulent flows can be cleverly collected from a single snapshot by considering characteristics of rotation and shear tensors. Second, we perform the single-snapshot super-resolution analysis for turbulent channel flow, showing that it is possible to extract physical insights from a single flow snapshot even with inhomogeneity. The present findings suggest that embedding prior knowledge in designing a model and collecting data is important for a range of data-driven analyses for turbulent flows. More broadly, this work hopes to stop machine-learning practitioners from being wasteful with turbulent flow data.

Type
JFM Papers
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press.

1. Introduction

By gazing at a turbulent flow acquired from numerical simulation or experiment, we can admire the rich physics that involves swirling, stretching and diffusion. Turbulence also presents multi-scale characteristics over broad length scales (Davidson Reference Davidson2015). In high-Reynolds-number turbulent flows, the rich phenomena and characteristics are exhibited at any instance in time. We argue that even a single snapshot of turbulent flow can hold sufficient information to train machine-learning models. This paper poses a question of whether a commonly used big data set is required for training machine-learning models in studying turbulence.

There have been increased usages of modern machine-learning techniques to analyse, model, estimate and control turbulent flows (Brunton, Noack & Koumoutsakos Reference Brunton, Noack and Koumoutsakos2020). These applications include subgrid-scale modelling (Duraisamy, Iaccarino & Xiao Reference Duraisamy, Iaccarino and Xiao2019), reduced-order modelling (Racca, Doan & Magri Reference Racca, Doan and Magri2023), super resolution/flow reconstruction (Fukami, Fukagata & Taira Reference Fukami, Fukagata and Taira2021; Guastoni et al. Reference Guastoni, Güemes, Ianiro, Discetti, Schlatter, Azizpour and Vinuesa2021; Cuéllar et al. Reference Cuéllar, Güemes, Ianiro, Flores, Vinuesa and Discetti2024) and flow control (Duriez, Brunton & Noack Reference Duriez, Brunton and Noack2017; Park & Choi Reference Park and Choi2020). These machine-learning models require enormous amounts of training data, which are generally significantly larger than those necessitated by traditional analysis techniques.

However, it may be possible to extract important flow features without such large data sets since even a single turbulent flow snapshot contains multi-scale, scale-invariant structures. To achieve meaningful learning from a single snapshot, we consider training machine-learning models through subsampling and leveraging turbulent statistics. We further note that it is important that machine-learning models have appropriate architectures and learning formulation that fold in physics (Brunton & Kutz Reference Brunton and Kutz2019; Lee & You Reference Lee and You2019; Raissi, Perdikaris & Karniadakis Reference Raissi, Perdikaris and Karniadakis2019; Fukami, Fukagata & Taira Reference Fukami, Fukagata and Taira2023).

This study considers data-driven analysis using only a single training snapshot of turbulent flow. As examples, we perform machine-learning-based super-resolution analysis for two-dimensional decaying turbulence and three-dimensional turbulent channel flow. We show that flow reconstruction over a range of Reynolds numbers is possible with nonlinear machine learning by cleverly sampling data from a single snapshot. The present results show that a large data set is not necessarily needed for machine learning of turbulent flows.

This paper is organized as follows. The approach is described in § 2. Results from the single-snapshot super-resolution analysis are presented in § 3. Conclusions are offered in § 4.

2. Approach

The objective of this study is to show that it is possible to perform data-driven analysis of turbulent flows with a very limited amount of training data – even from a single snapshot. For the present analysis, we consider machine-learning-based super-resolution reconstruction of fluid flows (Fukami, Fukagata & Taira Reference Fukami, Fukagata and Taira2019). A machine-learning model $\mathcal {F}$ is trained to reconstruct a high-resolution flow field $\boldsymbol{q}_{HR}$ from low-resolution data $\boldsymbol{q}_{LR}$:

(2.1)\begin{equation} \boldsymbol{q}_{HR} = {\mathcal{F}}(\boldsymbol{q}_{LR};\boldsymbol{w}), \end{equation}

where $\boldsymbol{w}$ denotes the weights inside the model. In this study, the model $\mathcal {F}$ is trained with a collection of subdomains sampled from only a single snapshot of two-dimensional isotropic turbulence and three-dimensional turbulent channel flow. The model is then tested with independent snapshots. If the model $\mathcal {F}$ successfully learns the relationship between low- and high-resolution flow fields from a single training snapshot, we expect that the reconstruction would be possible even for independent testing conditions.

For machine-learning-based super resolution of turbulent flows, the model ${\mathcal {F}}$ needs to be carefully designed to accommodate a range of length scales while accounting for rotational and translational invariance of vortical structures (Fukami et al. Reference Fukami, Fukagata and Taira2021). This study uses the interconnected hybrid downsampled skip-connection/multi-scale (DSC/MS) model (Fukami et al. Reference Fukami, Fukagata and Taira2023) based on convolutional neural networks (CNN; LeCun et al. Reference LeCun, Bottou, Bengio and Haffner1998), as illustrated in figure 1. Between the layers $(l-1)$ and $(l)$, the CNN learns the nonlinear relationship between input and output data by extracting spatial features of given data through filtering operations,

(2.2)\begin{equation} c^{(l)}_{ijn}=\varphi\left(\sum_{m=1}^M\sum_{p=0}^{H-1}\sum_{q=0}^{H-1}h^{(l)}_{pqmn}c^{(l-1)}_{i+p-G,j+q-G,m}+b_n^{(l)}\right), \end{equation}

where $G=\lfloor H/2\rfloor$, $H$ is the width and height of the filter $h$, $M$ is the number of input channels, $n$ is the number of output channels, $b$ is the bias, and $\varphi$ is the activation function. By using a nonlinear function for $\varphi$, the convolutional networks can account for nonlinearlities in learning features from training data.

Figure 1. Interconnected DSC/MS model (Fukami et al. Reference Fukami, Fukagata and Taira2023) for super-resolution reconstruction of turbulent flows.

The DSC model (boxed in red) includes up/downsampling operations and skip connections, capturing rotational and translational invariance (Fukami, Goto & Taira Reference Fukami, Goto and Taira2024). The MS model (boxed in blue) consists of three different sizes of filter operations, enabling the model to learn a range of length scales in turbulent flows. Furthermore, these two networks are internally connected via skip connections (He et al. Reference He, Zhang, Ren and Sun2016) to enhance the correlation of the intermediate input and output from both subnetworks in the training process. We refer to Fukami et al. (Reference Fukami, Fukagata and Taira2023) and a sample code (http://www.seas.ucla.edu/fluidflow/codes.html) for further details on the present machine-learning model. In this study, model ${\mathcal {F}}$ is trained such that weights $\boldsymbol{w}$ are optimized through

(2.3)\begin{equation} \boldsymbol{w}^* = \underset{\boldsymbol{w}}{{\rm argmin}}\, \|\boldsymbol{q}_{HR}-{\mathcal{F}}(\boldsymbol{q}_{LR};\boldsymbol{w})\|_2. \end{equation}

While this study uses an $L_2$ norm for optimization, one can consider incorporating the knowledge from the governing equations into the cost function to better constrain the solution space (Raissi et al. Reference Raissi, Perdikaris and Karniadakis2019; Fukami et al. Reference Fukami, Fukagata and Taira2023).

3. Results

3.1. Example 1: two-dimensional decaying homogeneous isotropic turbulence

Two-dimensional decaying isotropic turbulence is first considered in the present single-snapshot super-resolution analysis. The present machine-learning model is trained with subdomains collected from a single snapshot and then assessed with test snapshots obtained by independent simulations. The flow field data are generated with direct numerical simulation (Taira, Nair & Brunton Reference Taira, Nair and Brunton2016) that numerically solves the two-dimensional vorticity transport equation,

(3.1)\begin{equation} \dfrac{\partial\omega}{\partial t}+\boldsymbol{u}\boldsymbol{\cdot}\boldsymbol{\nabla}\omega=\dfrac{1}{Re_0}\nabla^2 \omega, \end{equation}

where $\boldsymbol{u}=(u,v)$ represents the velocity field and $Re_0=u^*l_0^*/\nu$ is the initial Reynolds number. Here, $u^*$ is the characteristic velocity defined as the square root of the spatially averaged initial kinetic energy, $l_0^*=[2{\overline {u^2}}(t_0)/{\overline {\omega ^2}}(t_0)]^{1/2}$ is the initial integral length and $\nu$ is the kinematic viscosity. The overline denotes the spatial average. The computational domain is a biperiodic square with length $L=1$. We use the vorticity field $\omega$ as a data attribute in the present super-resolution analysis.

The baseline super-resolution analysis is performed with the model trained with a single snapshot shown in figure 2 with $Re_0=1580$. Various vortical structures, including counter-rotating and co-rotating vortices and shear layers, of different length scales are contained in this single snapshot. The number of computational grid points $N^2$ is set to $1024^2$, satisfying $k_{max}\eta \geq 1$, where $k_{max}$ is the maximum wavenumber and $\eta$ is the Kolmogorov length scale, to ensure that the direct numerical simulation (DNS) resolves all flow scales. The simulation for training data preparation is initialized with a distribution composed of randomly placed Taylor vortices (Taylor Reference Taylor1918) with random strengths and sizes. The snapshot is collected after the flow reaches the decaying regime.

Figure 2. Two-dimensional isotropic vorticity field. Red boxes are example flow tiles used for training.

The present training data comprise square-sized subdomain samples randomly collected from the single snapshot with four different sizes of $L_{sub}=\{0.03125, 0.0625, 0.125, 0.25\}$, as illustrated in figure 2. The subdomain data are then resized to be $N^2_{ML} = 128^2$ for the present data-driven analysis. The dependence of super-resolution reconstruction on the choice of a single snapshot is examined later.

Test snapshots in this study are prepared from three different simulations. The initial Reynolds numbers and the number of grid points are respectively $Re_0=\{80.4, 177, 442\}$ and $N=\{128, 256, 512\}$, satisfying $k_{max}\eta \geq 1$. These settings are intended to generate test snapshots that include a similar size of vortical structures to that in the subdomains of the single snapshot. Once the test snapshots are collected from the simulations, they are resized to be $N_{ML} = 128$. The present machine-learning model ${\mathcal {F}}$ reconstructs the high-resolution vorticity flow field of size $128^2$ from the corresponding low-resolution data of size $8^2$ generated by average pooling (Fukami et al. Reference Fukami, Fukagata and Taira2019). The input and output data are normalized by the instantaneous maximum value of absolute vorticity, $\max (|\omega |)$ to account for the magnitude difference of vorticity fields across the Reynolds number.

We apply the super-resolution model trained with a single snapshot to decaying turbulence at three different test $Re$. The reconstruction by the DSC/MS model is compared with bicubic interpolation, as shown in figure 3. Let us first use 2000 local tiles in total for training the baseline model. The value listed underneath each figure reports the $L_2$ error norm $\varepsilon = \|{\omega }_{HR}-{\mathcal {F}}({\omega }_{LR})\|_2/\|{\omega }_{HR}\|_2$. As the bicubic interpolation simply smooths the given low-resolution data, the reconstructed fields do not provide any fine-scale information, resulting in a high $L_2$ error.

Figure 3. Single snapshot super-resolution of two-dimensional decaying turbulence. Its accuracy is assessed with test snapshots from three different simulations. The instantaneous Taylor length scale $\lambda (t)$ for a representative test snapshot is reported along with each $Re_0$. The value underneath each contour is the $L_2$ error norm. The probability density function (p.d.f.) for test snapshots at each Reynolds number is also shown.

To improve the reconstruction of fine-scale structures, let us consider the DSC/MS model-based super resolution. The reconstructed fields by the DSC/MS model show improved agreement with the reference data. In addition to large-scale structures, rotational and shear-layer structures are also well represented compared with bicubic interpolation, reporting only 10–20 % $L_2$ error across the range of Reynolds numbers. Note that this level of error suggests accurate reconstruction that captures turbulent coherent structures since the spatial $L_2$ norm is a strict comparative measure (Anantharaman et al. Reference Anantharaman, Feldkamp, Fukami and Taira2023).

The reconstruction performance is also examined with the probability density function (p.d.f.) of the vorticity field, as presented in figure 3. For the case of $Re_0 = 80.4$, the curves obtained from both the bicubic interpolation and the DSC/MS model are in agreement with the reference data. However, the curve for the bicubic interpolation (coloured in orange) deviates for the tail of the distribution, implying the failure in reconstructing strong rotation structures with low probability.

As the test $Re$ increases, the bicubic interpolation starts struggling to reconstruct the vorticity across its distribution. This is because the smallest and largest scales spread wider by increasing the test Reynolds number. In contrast, the distributions obtained by the present DSC/MS model are almost indistinguishable compared with those with the reference DNS, supporting statistically accurate reconstruction. These results imply that even just a single turbulent flow snapshot contains a variety of vortical structures across different length scales, which can be extracted by the present machine-learning approach.

To further examine the reconstruction performance across spatial length scales, let us present in figure 4 the kinetic energy spectrum $E(k)$, where $k$ is the wavenumber. While the bicubic interpolation significantly underestimates the energy across the wavenumbers, the machine-learning model provides reasonable agreement up to $k\approx 200$ for $Re_0 = 80.4$ and 177, and $k\approx 100$ for $Re_0 = 442$. The difference in the high-wavenumber regime is due to the low correlation between the low- and high-wavenumber components, which is often observed in supervised learning-based super-resolution of turbulent flows (Fukami et al. Reference Fukami, Fukagata and Taira2023). A remedy for improved matching over the high wavenumber range could be attained by using algorithms such as generative learning (Kim et al. Reference Kim, Kim, Won and Lee2021; Yousif et al. Reference Yousif, Zhang, Yu, Vinuesa and Lim2023). The results here indicate that the current model can learn the energy distribution over the spatial length scales and Reynolds numbers from only a single snapshot.

Figure 4. Kinetic energy spectrum $E(k)$ of the reconstructed vorticity fields.

The successful reconstruction above is supported by the richness of vortical information contained in the training snapshot depicted in figure 2. In other words, the single snapshot to be used for training must be rich with information. To examine this point, we further consider 150 different flow fields generated by 20 different initial conditions with $N \in [128, 2048]$ and $Re_0 \in [40, 2050]$. We perform the single-snapshot training with these snapshots covering a variety of flow realizations regarding the size and shape of vortices, as shown in figure 5.

Figure 5. Example snapshots used for single-snapshot training. The value underneath each snapshot is the instantaneous Taylor length scale $\lambda (t)$.

To quantify the effect of the single-snapshot choice in training on the reconstruction performance for test data, we use the ratio of the Taylor length scale between training and test snapshots, $\lambda _{test}/\lambda _{single}$, where $\lambda$ represents the Taylor length scale and subscripts ‘test’ and ‘single’ denote test and training (single) snapshots, respectively. The relationship between this ratio and the reconstruction error across the different numbers of local tiles $n_s$ generated from a vorticity snapshot is presented in figure 6(a). For each case, a threefold cross-validation is performed and the averaged error is reported. The reconstruction improves for large $\lambda _{test}/\lambda _{single}$. In other words, large $\lambda _{test}$ (low test $Re$ snapshots) or small $\lambda _{single}$ (high training $Re$ snapshots) provides low reconstruction error.

Figure 6. (a) Relationship between the reconstruction error $\varepsilon$ and the ratio of the Taylor length scale between training and test snapshots $\lambda _{test}/\lambda _{single}$ across the number of training samples $n_s$. (b) Dependence of the reconstruction performance on the number of training samples $n_s$ and (c) reconstructed vorticity fields for each test Reynolds number in using a single snapshot with $\lambda _{single} = 0.0425$ shown in figure 2. The value underneath each contour in panel (c) is the $L_2$ error norm.

The error decreases by increasing the number of local tiles $n_s$ across the length-scale ratio. While this error reduction for large $n_s$ is expected, it is worth pointing out that lower $n_s$ is needed as the ratio $\lambda _{test}/\lambda _{single}$ increases to achieve the same level of reconstruction. In other words, quantitative reconstruction can be achieved with a smaller number of local tiles in the single-snapshot training with a small $\lambda _{single}$ that generally corresponds to a high-$Re$ field including many vortical structures. These observations imply that in addition to the number of local samples or snapshots, the amount of information contained in the training data should also be considered when analysing turbulent flows.

Let us focus on the baseline case of $\lambda _{single} = 0.0425$, depicted in figure 2, to further discuss the effect of the number of local tiles $n_s$ across the test Reynolds number, as shown in figure 6(b). The averaged error over cross-validation is reported while the maximum and minimum errors at each $n_s$ are shown with shading. Across $n_s$, the reconstruction error at a higher $Re$ is larger compared with lower $Re$ flows, likely because of larger differences in the vortical length scales appearing in the flow. As $n_s$ increases, the reconstruction performance is improved across the Reynolds number. Notably, the present model achieves qualitative reconstruction for large-scale structures even with merely 250 training samples, as presented in figure 6(c). Even in such a modest number of local tiles, there exist physical insights (relations) that can be extracted by the present super-resolution model.

Once $n_s$ exceeds 2000, the error curves across the Reynolds number plateau, implying that extracted data of vortical flows become redundant from the perspective of learning. While the above model is trained with randomly sampled local tiles from a single snapshot, the present nonlinear machine-learning model can achieve quantitative reconstruction even with a much smaller number of local subdomains by sampling them in a smart manner based on some knowledge of the vortical flows.

The idea here is to avoid sampling local tiles that are not informative. To preferentially sample informative local subdomains that include insightful rotational motions and shear layers, we consider the moments of rotation and strain tensors, $W$ and $D$. The two-dimensional p.d.f.s based on the mean (first moment), standard deviation (second moment, $\sigma$), skewness (third moment, $S$), and flatness (fourth moment, $F$) of $W$ and $D$ with $L_{sub} = 0.0625$ are presented in figure 7(a). The 97 % confidence interval is also depicted on each p.d.f. map. Compared with the first and second moment-based p.d.f.s, the third and fourth moment-based p.d.f.s provide a sharper distribution of snapshots, as observed from the difference in the size of 97 % confidence interval area. Furthermore, we observe that local tiles containing various structures such as flow fields (i) and (ii) appear in the region with high probability while less informative tiles such as flow fields (iii) and (iv) are seen in the area with lower probability when using the skewness.

Figure 7. P.d.f.-based sampling for single-snapshot training. (a) Two-dimensional p.d.f. of first to fourth moments of rotation and strain tensors with $L_{sub} = 0.0625$. For each p.d.f. map, 97 % confidence interval is shown. (b) Example local tiles corresponding to flow fields (i–iv) on each p.d.f. map. (c) Reconstruction with different data sampling with $n_s=250$. The value underneath each contour reports the $L_2$ error norm.

Based on the findings above, data sampling informed by the moment probability for single-snapshot training with $n_s=250$ is performed, as shown in figure 7(b). For comparison, the first and second moment-based samplings are also considered. While the lower-order moment-based training presents similar reconstruction performance to the case in which the location of subdomains is randomly determined, the third and fourth moment-based sampling models provide enhanced reconstruction with the same number of local tiles, revealing vortices and shear-layer structures with finer details. Note that the error level of the higher-order moment-based sampling with $n_s=250$ becomes the same as that of random sampling with $n_s=2000$, achieving significant reduction in the required number of training subsamples for accurate reconstruction. These observations suggest that machine-learning-based analyses traditionally recognized as expensive, data-hungry approaches can take advantage of the scale-invariant property in analysing turbulent vortical flows from much smaller data sets.

3.2. Example 2: turbulent channel flow

Next, we perform the single-snapshot-based super-resolution analysis for turbulent channel flow as a test case that holds spatial inhomogeneity. For the present analysis, we consider the DNS data set made available from the Johns Hopkins Turbulence Database (Perlman et al. Reference Perlman, Burns, Li and Meneveau2007). Similar to the case of two-dimensional homogeneous turbulence, the present model is trained with a collection of subdomains sampled from a single high-$Re$ snapshot and then evaluated with test snapshots obtained by an independent simulation. The current setting enables assessing whether the present model learns flow features of turbulent channel flow across the Reynolds number from a single snapshot.

The single snapshot used for training is produced at a very high friction Reynolds number $Re_\tau = u_\tau \delta /\nu$ of 5200, holding a range of length scales (Lee & Moser Reference Lee and Moser2015). The variables are normalized by the half-channel height $\delta$ and the friction velocity at the wall $y=0$, $u_\tau =(\nu \,{{\rm d}U}/{{\rm d}\kern 0.05em y}|_{y=0})^{1/2}$, where $U$ is the mean velocity. The size of the computational domain and the number of grid points are $(L_{x}, L_{y}, L_{z}) = (8{\rm \pi} \delta, 2\delta, 3{\rm \pi} \delta )$ and $(N_{x}, N_{y}, N_{z}) = (10\,240, 1536, 7680)$, respectively. Details of the numerical simulation set-up are provided by Lee & Moser (Reference Lee and Moser2015).

We consider an $x$$y$ sectional streamwise velocity field $u$ as the variable of interest. The subdomains used for training are randomly sampled from the $x$$y$ sectional fields at random spanwise locations, as illustrated in figure 8. Four different sizes of subdomains are considered in the streamwise direction, $L_{x,{sub}}^+ = \{814, 1628, 3257, 6514\}$, where the variables with superscript $+$ denote quantities in the wall unit. The subdomain size in the wall-normal direction varies as the data are collected from a non-uniform grid. The minimum and maximum heights of the subdomains are $(\min (L_{y,{sub}}^+), \max (L_{y,{sub}}^+)) = (66.3, 2646)$, respectively. These collected data are resized to be $N^2_{ML} = 128^2$ for the present data-driven analysis.

Figure 8. Streamwise velocity field of turbulent channel flow at $Re_\tau = 5200$. Blue boxes are example flow tiles used for training.

Test snapshots are prepared from a different DNS at $Re_\tau = 1000$ (Graham et al. Reference Graham2016), also available from the Johns Hopkins Turbulence Database. The size of the computational domain and the number of grid points for $Re_\tau = 1000$ are $(L_{x}, L_{y}, L_{z}) = (8{\rm \pi} \delta, 2\delta, 3{\rm \pi} \delta )$ and $(N_{x}, N_{y}, N_{z}) = (2048, 512, 1536)$, respectively. Details on the numerical simulation set-up for this test data are available from Graham et al. (Reference Graham2016). The present test data are randomly subsampled from the $x$$y$ sectional streamwise velocity field at random spanwise locations. The super-resolution model is trained to reconstruct the high-resolution velocity field of size $128^2$ from the corresponding low-resolution data of size $8^2$ generated by average pooling (Fukami et al. Reference Fukami, Fukagata and Taira2021). The input and output data of streamwise velocity fields are normalized by the friction velocity $u_\tau$ to learn a universal relation between the low- and high-resolution data of turbulent channel flow across the Reynolds number (Kim et al. Reference Kim, Kim, Won and Lee2021).

Let us apply the present machine-learning model trained with a single snapshot at $Re_\tau = 5200$ to test datasets. The baseline super-resolution model of the case of turbulent channel flow is trained with 2000 local tiles. The reconstructed turbulent flow fields for two representative flow tiles of test data with $(L_{x,{test}}^+, L_{y,{test}}^+) = (1570, 300)$ are presented in figure 9. Here, the first grid point in the wall-normal direction of the test snapshots $y_{0, {test}}$ is set to zero to examine the reconstruction performance near the wall. To assess whether the fluctuation component of the velocity field is captured, the $L_2$ error norm in the case of turbulent channel flow reported hereafter is normalized by the velocity fluctuation such that $\varepsilon ^\prime = \|{u}_{HR}-{\mathcal {F}}({u}_{LR})\|_2/\|{u}^\prime _{HR}\|_2$.

Figure 9. Single-snapshot super-resolution analysis of turbulent channel flow. Its accuracy is assessed with test snapshots collected from a different simulation at $Re_\tau = 1000$. The value underneath each contour plot is the $L_2$ error norm normalized by streamwise velocity fluctuation.

The reconstructed turbulent flow fields by the present machine-learning model are in agreement with the reference DNS data with as little as $5$ to $8\,\%$ error. For comparison, we also show reconstruction from bicubic interpolation, which can only smooth the given low-resolution velocity fields. In contrast, the DSC/MS model accurately reproduces the fine-scale structures in the flow fields.

Accurate reconstruction by the present machine-learning model is also evident from statistics of the streamwise velocity field. The DSC/MS model is superior to the bicubic method especially in reconstructing the low-speed component, as seen in the p.d.f. of $u^+$ shown in figure 10(a). The difference in the reconstruction performance between the DSC/MS model and the bicubic interpolation is further reflected in the high-order moments depicted in figure 10(bd). Note that the skewness $S(u^+)$ of the bicubic method almost matches the reference value as the low-resolution input does not hold any negative values thereby producing a distribution skewed towards positive values. The flatness $F(u^+)$ particularly captures the difference in the produced distributions, supporting successful reconstruction by the present machine-learning model.

Figure 10. Statistics of the streamwise velocity. (a) Probability density function (p.d.f.), (b) second, (c) third and (d) fourth moments of the streamwise velocity $u^+$. The negative value of the third moment $S(u^+)$ is presented in panel (c).

Turbulence statistics of the reconstructed velocity fields are also evaluated. The root-mean-square of streamwise velocity fluctuation $u_{rms}$ and the mean velocity profile across the wall-normal direction are presented in figures 11(a) and 11(b), respectively. The statistics obtained by the DSC/MS model accurately match the reference DNS across the $y$ direction while the bicubic interpolation produces overestimation of the velocity field at the viscous sublayer and part of the buffer layer of $y^+\lesssim 10$.

Figure 11. Turbulence statistics of the streamwise velocity. (a) Root-mean-square of streamwise velocity fluctuation $u_{rms}$. (b) Mean velocity profile across the wall-normal direction. The coefficients $\kappa$ and $B$ for the logarithmic law of the wall are set to 0.41 and 5, respectively. (c) Streamwise kinetic energy spectrum $E_{uu}^+(k_x^+)$ and (d) spatial two-point correlation coefficient $R_{uu}^+(x^+)$ at $y^+=10.4$.

To further examine the reconstruction performance near the wall, we assess the streamwise kinetic energy spectrum $E_{uu}^+(k_x^+)$, where $k^+_x$ represents the streamwise wavenumber, and the spatial two-point correlation coefficient $R_{uu}^+(x^+)$ at $y^+=10.4$, depicted in figures 11(c) and 11(d), respectively. The energy distribution across the wavenumber is well represented with the DSC/MS model. In addition, the decaying profile of the spatial two-point correlation coefficient over $x^+$ is accurately reproduced by the present machine-learning model, suggesting that the streamwise flow pattern in a flow field is super-resolved well with the DSC/MS model. These results imply that features of turbulent channel flow across $Re_\tau$ (Reynolds & Tiederman Reference Reynolds and Tiederman1967; Yamamoto & Tsuji Reference Yamamoto and Tsuji2018) can be successfully extracted by nonlinear machine learning from only a given single snapshot.

Here, let us assess the effect of input noise on single-snapshot super-resolution reconstruction. For the present assessment, the Gaussian noise $\boldsymbol{n}$ is given to a low-resolution input such that $\boldsymbol{q}_{LR,noise} = \boldsymbol{q}_{LR}+\boldsymbol{n}$, where the magnitude of noisy input $\gamma$ is given as $\gamma = \|\boldsymbol{n}\|/\|\boldsymbol{q}\|$. The relationship between the noise magnitude and the $L_2$ error norm normalized by streamwise velocity fluctuation $\varepsilon ^\prime$ is depicted with representative reconstructed fields in figure 12. The error increases with the magnitude of noise $\gamma$. While the present model accurately reconstructs fine structures in a flow field up to $\gamma \approx 0.2$, large-scale structures can be reconstructed even with 50 % noise, exhibiting reasonable robustness for the given noise levels.

Figure 12. Robustness of the machine-learning model trained with 2000 subsamples against noisy low-resolution flow field input. The magnitude of noise $\gamma$ and the $L_2$ error norm normalized by streamwise velocity fluctuation are shown underneath each contour.

At last, the dependence of super-resolution reconstruction on the number of local tiles $n_s$ for the turbulent channel flow is examined, as shown in figure 13. The averaged fluctuation-based error over three-fold cross-validation is shown with the maximum and minimum errors at each $n_s$ indicated by the shading. The reconstruction performance improves monotonically with increasing $n_s$. While the region far away from the wall is reasonably reconstructed with $n_s = 500$, more subdomains with the order of $O(10^3)$ are required for accurate reconstruction near the wall, likely because of the difference in flow complexities across the wall-normal direction of turbulent channel flow. In turn, these results suggest that nonlinear machine learning can extract physical insights of turbulent flows even with spatial inhomogeneity by sufficiently collecting training subsamples from a single turbulent flow snapshot.

Figure 13. Dependence of the reconstruction performance on the number of training subsamples $n_s$ for turbulent channel flow. The $n_s$ and the $L_2$ error norm normalized by streamwise velocity fluctuation are shown underneath each contour.

4. Concluding remarks

This study discussed how we can efficiently extract physical insights from a very limited amount of turbulent flow data with machine learning. We considered machine-learning-based super-resolution reconstruction with training data of a single turbulent flow snapshot, enabling the evaluation of whether a physical relationship between high- and low-resolution flow fields can be learned from limited available flow data. A convolutional network-based super-resolution model, the DSC/MS model, is trained with local flow subdomains collected from only a single turbulent flow snapshot and then assessed for test data generated from different simulations. With an example of two-dimensional decaying isotropic turbulence, we showed that training data for super-resolution analysis can be efficiently prepared from a single flow snapshot based on their statistical characteristics. We also performed the single snapshot-based super-resolution for turbulent channel flow, showing that it is possible to learn physical relations between low- and high-resolution flow fields from a single snapshot even with spatial inhomogeneity.

Although machine-learning-based analysis is often described as data-intensive, our findings indicate that it is possible to extract physical insights without over-relying on massive amounts of training data for studying turbulent flows. Capturing universal flow features across the Reynolds number such as scale-invariant characteristics is the key to successful turbulent flow reconstruction with data-driven techniques. The use of an appropriate model architecture with physics embedding depending on flows of interest is important. The current results also imply that redundancy of turbulent flows in not only space but also time can also be considered in sampling training data. By incorporating prior knowledge for developing a machine-learning model and collecting training data, we should be able to use smaller data sets to learn physics in a much smarter manner. We should stop being wasteful of turbulent flow data.

Acknowledgements

We thank the support from the US Air Force Office of Scientific Research (FA9550-21-1-0178) and the US Department of Defense Vannevar Bush Faculty Fellowship (N00014-22-1-2798). The computations for machine-learning analysis were performed on Delta GPU at the National Center for Supercomputing Applications (NCSA) through the ACCESS program (Allocation PHY230125).

Declaration of interests

The authors report no conflict of interest.

References

Anantharaman, V., Feldkamp, J., Fukami, K. & Taira, K. 2023 Image and video compression of fluid flow data. Theor. Comput. Fluid Dyn. 37 (1), 6182.CrossRefGoogle Scholar
Brunton, S.L. & Kutz, J.N. 2019 Data-driven Science and Engineering: Machine Learning, Dynamical Systems, and Control. Cambridge University Press.CrossRefGoogle Scholar
Brunton, S.L., Noack, B.R. & Koumoutsakos, P. 2020 Machine learning for fluid mechanics. Annu. Rev. Fluid Mech. 52, 477508.CrossRefGoogle Scholar
Cuéllar, A., Güemes, A., Ianiro, A., Flores, Ó., Vinuesa, R. & Discetti, S. 2024 Three-dimensional generative adversarial networks for turbulent flow estimation from wall measurements. J. Fluid Mech. 991, A1.CrossRefGoogle Scholar
Davidson, P.A. 2015 Turbulence: An Introduction for Scientists and Engineers. Oxford University Press.CrossRefGoogle Scholar
Duraisamy, K., Iaccarino, G. & Xiao, H. 2019 Turbulence modeling in the age of data. Annu. Rev. Fluid Mech. 51, 357377.CrossRefGoogle Scholar
Duriez, T., Brunton, S.L. & Noack, B.R. 2017 Machine Learning Control – Taming Nonlinear Dynamics and Turbulence. Springer.CrossRefGoogle Scholar
Fukami, K., Fukagata, K. & Taira, K. 2019 Super-resolution reconstruction of turbulent flows with machine learning. J. Fluid Mech. 870, 106120.Google Scholar
Fukami, K., Fukagata, K. & Taira, K. 2021 Machine-learning-based spatio-temporal super resolution reconstruction of turbulent flows. J. Fluid Mech. 909, A9.CrossRefGoogle Scholar
Fukami, K., Fukagata, K. & Taira, K. 2023 Super-resolution analysis via machine learning: a survey for fluid flows. Theor. Comput. Fluid Dyn. 37, 421444.CrossRefGoogle Scholar
Fukami, K., Goto, S. & Taira, K. 2024 Data-driven nonlinear turbulent flow scaling with Buckingham Pi variables. J. Fluid Mech. 984, R4.CrossRefGoogle Scholar
Graham, J., et al. 2016 A web services accessible database of turbulent channel flow and its use for testing a new integral wall model for LES. J. Turbul. 17 (2), 181215.CrossRefGoogle Scholar
Guastoni, L., Güemes, A., Ianiro, A., Discetti, S., Schlatter, P., Azizpour, H. & Vinuesa, R. 2021 Convolutional-network models to predict wall-bounded turbulence from wall quantities. J. Fluid Mech. 928, A27.CrossRefGoogle Scholar
He, K., Zhang, X., Ren, S. & Sun, J. 2016 Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778.Google Scholar
Kim, H., Kim, J., Won, S. & Lee, C. 2021 Unsupervised deep learning for super-resolution reconstruction of turbulence. J. Fluid Mech. 910, A29.CrossRefGoogle Scholar
LeCun, Y., Bottou, L., Bengio, Y. & Haffner, P. 1998 Gradient-based learning applied to document recognition. Proc. IEEE 86 (11), 22782324.CrossRefGoogle Scholar
Lee, M. & Moser, R.D. 2015 Direct numerical simulation of turbulent channel flow up to $Re_\tau = 5200$. J. Fluid Mech. 774, 395415.CrossRefGoogle Scholar
Lee, S. & You, D. 2019 Data-driven prediction of unsteady flow over a circular cylinder using deep learning. J. Fluid Mech. 879, 217254.CrossRefGoogle Scholar
Park, J. & Choi, H. 2020 Machine-learning-based feedback control for drag reduction in a turbulent channel flow. J. Fluid Mech. 904, A24.CrossRefGoogle Scholar
Perlman, E., Burns, R., Li, Y. & Meneveau, C. 2007 Data exploration of turbulence simulations using a database cluster. In Proceedings of the 2007 ACM/IEEE Conference on Supercomputing. ACM.CrossRefGoogle Scholar
Racca, A., Doan, N.A.K. & Magri, L. 2023 Predicting turbulent dynamics with the convolutional autoencoder echo state network. J. Fluid Mech. 975, A2.CrossRefGoogle Scholar
Raissi, M., Perdikaris, P. & Karniadakis, G.E. 2019 Physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 378, 686707.CrossRefGoogle Scholar
Reynolds, W.C. & Tiederman, W.G. 1967 Stability of turbulent channel flow, with application to Malkus's theory. J. Fluid Mech. 27 (2), 253272.CrossRefGoogle Scholar
Taira, K., Nair, A.G. & Brunton, S.L. 2016 Network structure of two-dimensional decaying isotropic turbulence. J. Fluid Mech. 795, R2.CrossRefGoogle Scholar
Taylor, G.I. 1918 On the dissipation of eddies. Meteorology, Oceanography and Turbulent Flow, pp. 96101. Cambridge University Press.Google Scholar
Yamamoto, Y. & Tsuji, Y. 2018 Numerical evidence of logarithmic regions in channel flow at $Re_\tau = 8000$. Phys. Rev. Fluids 3 (1), 012602.CrossRefGoogle Scholar
Yousif, M.Z., Zhang, M., Yu, L., Vinuesa, R. & Lim, H.C. 2023 A transformer-based synthetic-inflow generator for spatially developing turbulent boundary layers. J. Fluid Mech. 957, A6.CrossRefGoogle Scholar
Figure 0

Figure 1. Interconnected DSC/MS model (Fukami et al.2023) for super-resolution reconstruction of turbulent flows.

Figure 1

Figure 2. Two-dimensional isotropic vorticity field. Red boxes are example flow tiles used for training.

Figure 2

Figure 3. Single snapshot super-resolution of two-dimensional decaying turbulence. Its accuracy is assessed with test snapshots from three different simulations. The instantaneous Taylor length scale $\lambda (t)$ for a representative test snapshot is reported along with each $Re_0$. The value underneath each contour is the $L_2$ error norm. The probability density function (p.d.f.) for test snapshots at each Reynolds number is also shown.

Figure 3

Figure 4. Kinetic energy spectrum $E(k)$ of the reconstructed vorticity fields.

Figure 4

Figure 5. Example snapshots used for single-snapshot training. The value underneath each snapshot is the instantaneous Taylor length scale $\lambda (t)$.

Figure 5

Figure 6. (a) Relationship between the reconstruction error $\varepsilon$ and the ratio of the Taylor length scale between training and test snapshots $\lambda _{test}/\lambda _{single}$ across the number of training samples $n_s$. (b) Dependence of the reconstruction performance on the number of training samples $n_s$ and (c) reconstructed vorticity fields for each test Reynolds number in using a single snapshot with $\lambda _{single} = 0.0425$ shown in figure 2. The value underneath each contour in panel (c) is the $L_2$ error norm.

Figure 6

Figure 7. P.d.f.-based sampling for single-snapshot training. (a) Two-dimensional p.d.f. of first to fourth moments of rotation and strain tensors with $L_{sub} = 0.0625$. For each p.d.f. map, 97 % confidence interval is shown. (b) Example local tiles corresponding to flow fields (i–iv) on each p.d.f. map. (c) Reconstruction with different data sampling with $n_s=250$. The value underneath each contour reports the $L_2$ error norm.

Figure 7

Figure 8. Streamwise velocity field of turbulent channel flow at $Re_\tau = 5200$. Blue boxes are example flow tiles used for training.

Figure 8

Figure 9. Single-snapshot super-resolution analysis of turbulent channel flow. Its accuracy is assessed with test snapshots collected from a different simulation at $Re_\tau = 1000$. The value underneath each contour plot is the $L_2$ error norm normalized by streamwise velocity fluctuation.

Figure 9

Figure 10. Statistics of the streamwise velocity. (a) Probability density function (p.d.f.), (b) second, (c) third and (d) fourth moments of the streamwise velocity $u^+$. The negative value of the third moment $S(u^+)$ is presented in panel (c).

Figure 10

Figure 11. Turbulence statistics of the streamwise velocity. (a) Root-mean-square of streamwise velocity fluctuation $u_{rms}$. (b) Mean velocity profile across the wall-normal direction. The coefficients $\kappa$ and $B$ for the logarithmic law of the wall are set to 0.41 and 5, respectively. (c) Streamwise kinetic energy spectrum $E_{uu}^+(k_x^+)$ and (d) spatial two-point correlation coefficient $R_{uu}^+(x^+)$ at $y^+=10.4$.

Figure 11

Figure 12. Robustness of the machine-learning model trained with 2000 subsamples against noisy low-resolution flow field input. The magnitude of noise $\gamma$ and the $L_2$ error norm normalized by streamwise velocity fluctuation are shown underneath each contour.

Figure 12

Figure 13. Dependence of the reconstruction performance on the number of training subsamples $n_s$ for turbulent channel flow. The $n_s$ and the $L_2$ error norm normalized by streamwise velocity fluctuation are shown underneath each contour.