We present an open-source package that helps the user to perform a basic statistical analysis of a given turbulence dataset. We believe that this package is useful to the entire turbulence community. It can be used to estimate statistical quantities of turbulence such as the spectral density, turbulence intensity, integral length scale, Taylor microscale, Kolmogorov scale, and dissipation rate as well as velocity increments, structure functions, and their scaling exponents. Different well-known methods available in the literature were selected so that they can be compared. This package also performs an advanced analysis that includes the scale-dependent statistical description of turbulent cascade using the Fokker–Planck equation, consequently leading to the assessment of integral fluctuation theorems. As a precondition for the stochastic process approach, Markovian properties of the turbulent cascade in scale are tested. This is utilized to estimate the drift and diffusion coefficients of the Fokker–Planck equation. The total entropy production for each independent cascade trajectory can be determined using a Fokker–Planck equation. Estimating total entropy production admits a rigorous law of non-equilibrium stochastic thermodynamics, namely, the integral fluctuation theorem, which must be valid if Markov properties hold and the Fokker–Planck equation is correct. This approach to the turbulent cascade process has the potential for a new way to link the statistical description of turbulence, non-equilibrium stochastic thermodynamics, and local turbulent flow structures. Finally, we emphasize that the presented package can be used for the analysis of other data with turbulent-like complexity as well.

The phenomenon of turbulent flows is ubiquitous and has been studied by various methods and concepts over a long period of time. One of the central models for describing turbulent flows is the phenomenologically inspired energy cascade model proposed by Richardson and Lynch.1 In this model, turbulence is understood as the evolution of turbulent structures on different spatial or temporal scales. The central assumption of this cascade model is that kinetic energy is transferred through all scales of the inertial range by the repeated random breakup of eddies into increasingly smaller eddies by a cascade-like process. Dissipation of kinetic energy into heat will not take place in this inertial range of the cascade but on smaller scales in the so-called dissipation range. Based on this assumption, in the famous dimensional analysis initiated by the work of Kolmogorov2–4 and Obhukov5,6 in 1941 (known as the K41 theory), the cascade model by Richardson is extended by a theory for fully developed, isotropic, and homogeneous turbulence (HIT). This kind of turbulence, which is assumed to be stationary, represents an idealized case that should have universal features. Nonetheless, the intermittency phenomenon, which manifests itself by deviations from such a self-similar cascade process at small scales, is considered as one of the key signatures of turbulence and is still not entirely understood.7 Kolmogorov himself8 and Obukhov9 refined the K41 theory by allowing fluctuations of the transferred energy leading to a lognormal distribution for the local energy dissipation rate and, consequently, to intermittent, i.e., non-Gaussian velocity fluctuations. This theory is known as K62 theory. An overview of various intermittency correction models along with fractal and multifractal models can be found in Refs. 7 and 10.

To characterize various types of turbulent flows such as atmospheric flows, jet flows, boundary layer flows, grid turbulent flows, wake flows, and von Kármán flows, in a statistical way, it is beneficial to have a common data post-processing tool in turbulence. Hence, this open-source package is created to perform a basic statistical analysis of turbulence data (here, we focus on one-dimensional velocity time series). We are not aware of such a comprehensive compilation of standard analysis tools commonly used in turbulence research in a single user-friendly package. Along with the basic statistical analysis of turbulence data, the application to Fokker–Planck equations (FPEs) and integral fluctuation theorems (IFTs), which provide a comprehensive statistical description in terms of the complexity of turbulent velocity time series, is implemented into this package. This analysis is very complex and time-consuming by means of developing a post-processing algorithm. With this package, for the first time, the application of this new method in practice is possible for everyone easily and quickly.

Altogether, this open source package enhances the practicability and availability of both the standard analyses already established in turbulence research, as well as the extended analysis via Fokker–Planck equations and integral fluctuation theorems. We believe that this package should be shared with the turbulence community as it is of great interest, especially to young scientists and those getting acquainted with this topic. It possesses a high reuse potential and considerable efforts are put into estimating turbulence quantities by different alternative methods. The latter package property, thus, allows for a more consistent and accurate assessment of these turbulence quantities. Furthermore, the open source package is organized in such a way that other researchers can add further methods to our collection as well.

The open source package can also be used by researchers outside of the field of turbulence. Using this package may contribute to new insights into the complex characteristics of scale-dependent processes (in space or time), which occur in many (multi-) fractal systems, ranging from physics to meteorology, biology systems, finance, economy, surface science and medicine. The package can be applied to these types of data (one-dimensional time series), but it is important to keep in mind that a variety of turbulence-specific assumptions like the theory of HIT or the Taylor hypothesis of frozen turbulence are assumed as the analysis proceeds. Hence, the implications of turbulent specific aspects for other systems has to be evaluated critically. A central feature of the open source package is the validation of a Markov property of velocity fluctuations in scale (see details in Sec. III A). It has turned out that the method of analysis based on Markov processes can also be successfully applied to the characterization of rough surfaces,11–13 rogue waves,14–16 urban systems,17 and financial time series18–20 to name but a few. Such spatially disordered structures can be analyzed as scale-dependent stochastic processes (see also Refs. 21 and 22) and are related to a non-equilibrium thermodynamical description.23 It has to be verified whether the investigation by non-equilibrium quantities such as scale-dependent entropies24 also contributes to the characterization of extreme events in the data the user intends to analyze.

The central part of this paper is a detailed discussion of the procedure of how to use this package. This paper only covers the discussion of turbulent flows satisfying the assumption of homogeneous, isotropic turbulence and the Taylor hypothesis of frozen turbulence. In principle, the tools of this package can be applied to turbulence data in other configurations and turbulent flows with non-ideal HIT conditions, as shown for non-fully developed turbulence, which may lead to strong deviations of the linearity of the drift coefficient D ( 1 ) (the deterministic force) with different slopes for positive and negative increments as found in Ref. 25. Furthermore, turbulence with strong shear is studied in Refs. 26 and 27 for the turbulent wake. In this investigation, a pronounced structural changes in the drift coefficient is observed as one passes from not fully developed turbulence (near field) into fully developed turbulence (far field). A careful interpretation of the results is advised here as well.

A general overview of all functions to carry out various analyses is given as part of this discussion in Secs. II–V. All subroutines that can be accessed in this package and the setting of the relevant parameters are called in a logical order using the main function. Overall, the analysis consists of four parts: Part I: Standard turbulence analysis, Part II: Markov analysis, Part III: Entropy analysis, and Part IV: Consistency check. A question dialog box is displayed where the user can select the analyses to be performed (see Fig. 1).

FIG. 1.

Question dialog box that allows the user to select the analyses to be performed.

FIG. 1.

Question dialog box that allows the user to select the analyses to be performed.

Close modal

To use this package, it is beneficial if the user has an adequate knowledge about turbulent flows. Nevertheless, for the fundamental understanding of turbulent flows, we recommend standard books.10,28 For a quick and simple overview of the theoretical background associated with the turbulent cascade process, Fokker–Planck equation (FPE), integral fluctuation theorem (IFT), and its interpretation, we recommend the following review article by Peinke, Tabar, and Wächter.22 

Note that all subroutines are also accessible from the command line and can be included by the user in other applications. Additionally, all abbreviations used in this document are listed at the end in the list of nomenclature/abbreviations. Also, it should be kept in mind that many of the equations discussed in this paper refer to the theory of fully developed, isotropic, and homogeneous turbulence. Thus, it is advised to use the package cautiously, as the assumption in the experimental data may not be fulfilled and can lead to incorrect/misleading results. In general, velocity is a vectorial quantity (with respect to its coordinate system), and it is dependent on the physical location in space. In this paper, the streamwise component of velocity is considered and features are discussed for this single velocity component. Other components may be characterized similarly. First promising results for more component datasets are shown in Refs. 29–31.

This package has been used continuously within our lab since 2018. It has also been successfully used by a large number of students (in the context of practical exercises which were part of the fluid dynamics lecture at the University of Oldenburg) to ensure stability across different machines and operating systems. Furthermore, this package has proved its value in a number of research publications.22,27,32–37 The package itself, or parts of it, as well as results obtained by using the package for analyzing turbulent data, have also been presented at several international conferences.

This software is implemented in MATLAB® (2022a), which is a high-level, matrix-based programming language designed specifically for engineers and scientists to analyze data, develop algorithms, and create models. Furthermore, a Python version of this package is currently under development and “Part I: Standard turbulence analysis,” will be made available on GitHub soon. The package is available as free software, under the GNU General Public License (GPL) version 3 [source code and standalone applications (64-bit) for Windows, macOS and Linux], and a typical dataset can be downloaded from the repository on GitHub or Matlab File Exchange Server to replicate all the results presented in this article. Support is available at github.com/andre-fuchs-uni-oldenburg/OPEN_FPE_IFT, where questions can be posted. The following toolboxes should be included in your MATLAB license.

  • Curve Fitting Toolbox

  • Optimization Toolbox

  • Parallel Computing Toolbox

  • Signal Processing Toolbox

  • Statistics and Machine Learning Toolbox

As these MATLAB toolboxes are essential for the application of the package, the compatibility to Octave cannot be provided. Nonetheless, in order to enhance the accessibility, standalone applications (64-bit) for Windows, macOS and Linux are also created to run the MATLAB code on target machines that do not have a MATLAB license.

To demonstrate the application of this program, an typical dataset obtained in a turbulent air jet experiment by Renner et al.38 is used within this document. The local velocity was sampled by a time-resolved single hot-wire measurement (Dantec 55P01) with a spatial resolution of 1.25 mm and a maximal time resolution of 30 kHz. This dataset is composed of ten independent measured datasets, and the data acquisition comprises 1.25 × 10 7 samples in total with a sampling frequency of 8 kHz. The latter exemplary dataset is also included in the package itself. We also provide the generated plots/results allowing the user to verify whether the software is operating correctly or not. Table I lists all parameters that the user must enter during the analysis to reconstruct the results shown in this paper.

TABLE I.

Information that the user must enter during the analysis to reconstruct the results shown in this paper: sampling frequency Fs, integral length scale L, Taylor length scale λ, number of bins to be used to divide the velocity increment series inc _ bin and Einstein–Markov length ΔEM.

Fs L λ inc _ bin ΔEM
8000 Hz  0.067 m  0.0066 m  93  22 samples 
Fs L λ inc _ bin ΔEM
8000 Hz  0.067 m  0.0066 m  93  22 samples 

The system requirements (memory, disk space, processor, total run time) demanded by the script depend very much on the size of the dataset to be examined, the resolution in relation to the number of bins, and the available number of central processing units (CPUs). The required memory will be allocated to each MATLAB computational engines during parallel computations using multicore CPUs and computer clusters. A typical processing time for this dataset using four physical cores and 16 GB RAM is about 60 min.

In this section, various subroutines to perform basic statistical analysis of the turbulence data are presented. This mainly includes the verification of stationarity of datasets, filtering of the signal, estimation of turbulence length scales, and dissipation rate along with the assessment of structure functions from velocity increments.

In the very first step, the user is asked to load the dataset and to choose the folder path to save the results. Furthermore, the user has to input the sampling frequency of the data and the kinematic viscosity of the fluid.

uiimport is the first command in the program, which asks the user to select interactively the data file that will be used for the analysis (Fig. 2). At this point, it is possible to specify the percentage of the total data that will be used to perform the analysis (for example, the first 20% of the data). This feature is sometimes helpful if one wants to make a first fast check of data and program. Note that this parameter has a significant effect on the overall performance of the script. Based on our experience, several thousand large eddy turnover times, respectively, acquired cascade trajectories (see Sec. IV A), are typically required to perform the analysis successfully.

FIG. 2.

The import tool lets the user interactively select the data file that will be used for the analysis.

FIG. 2.

The import tool lets the user interactively select the data file that will be used for the analysis.

Close modal

Save_path opens a dialog box to navigate through local directories to select a folder for saving figures and files.

Save_name generates a pop-up dialog box to enter the name for the analyzed data and figures to be saved.

Fs generates a pop-up dialog box to enter the sampling frequency of the data in Hz.

Kin_vis generates a pop-up dialog box to enter the value of kinematic viscosity ν in m 2 / s of the fluid for which the experimental data has been acquired.

Increment_bin generates a pop-up dialog box to specify the number of bins to be used to divide the velocity increment series. Note that this parameter has a significant effect on the overall performance of the script. A first estimation of the number of bins is made using
(1)
where σ is the standard deviation of the data.

For many statistical quantities to characterize a dataset, it is required that the datasets are stationary. Thus, before performing the basic turbulence analysis of the dataset, the stationarity of the data using different functions and parameters as mentioned below is examined.

plot_stationarity. With this function, the stationarity of the data is verified. For this purpose, the following analysis checks whether the statistical properties do not changeover time (stationary process). Therefore, the data are subdivided into 20 sections and the mean, standard deviation, skewness, and kurtosis are plotted, respectively, in Fig. 3(a). In the title of Fig. 3(a), the number of NaNs (not a number) is printed. The origin of these NaNs may be, for example, due to the measurement apparatus. Regardless of the origin of these, no preprocessing is performed in the package up to this point that can lead to a NaN. If the data to be analyzed contains NaNs a pop-up dialog box is displayed, which allows the user to choose five different methods to fill these entries using fillmissing function of MATLAB. The user must choose one of the following options to fill missing entries: “nearest non-missing value,” “linear interpolation of neighboring, non-missing values,” “piecewise cubic spline interpolation,” or “moving average” and “moving median” using a moving window of a length, which can also be set by the user. In a future version of the package, an additional stochastic interpolation of gaps in the dataset will also be provided.39 

FIG. 3.

(a) For fixed subdivision in 20 sections (each section corresponds to a length of 5% of the data), the mean, standard deviation, skewness, and kurtosis are plotted. (b) Plot of the complete data itself. In the title, the number of nans and the turbulence intensity is printed.

FIG. 3.

(a) For fixed subdivision in 20 sections (each section corresponds to a length of 5% of the data), the mean, standard deviation, skewness, and kurtosis are plotted. (b) Plot of the complete data itself. In the title, the number of nans and the turbulence intensity is printed.

Close modal
In Fig. 3(b), the complete data itself is plotted. This figure is used for the qualitative validation of stationarity, as it is very easy to detect by eye already, for example, outliers or drift in the data. In the title of this figure, the turbulence intensity,
(2)
is printed, with u being the mean value of a velocity time series.

plot_pdf. This function plots the probability density function (PDF) of the data with the number of bins specified in the function increment_bin in Fig. 4. It also plots the Gaussian distribution with the same standard deviation and mean value as of the data. In the title, the range of the data (the difference between maximum and minimum values of sample data), the skewness, and flatness of the data are printed.

FIG. 4.

Probability density function (PDF) of the data. The gray dashed line corresponds to a Gaussian distribution with the same standard deviation and mean value (vertical black dashed line) as the data.

FIG. 4.

Probability density function (PDF) of the data. The gray dashed line corresponds to a Gaussian distribution with the same standard deviation and mean value (vertical black dashed line) as the data.

Close modal
spectrum. This function calculates the energy spectral density (ESD) of the time series using the fft function implemented in MATLAB, which computes the discrete Fourier transform of the time series using a fast Fourier transform algorithm. In the literature, the term power spectral density (PSD) is often used for stationary turbulence signals. Based on the following Matlab code fragment [Eqs. (3)–(4) and (6)], we distinguish between
(3)
and
(4)
where abs ( ) returns the absolute value, Ldata is the number of samples of the data, and Fs is the sampling frequency. The energy spectral density is normalized so that
(5)
with σ 2 being the variance of the timeseries and the single-sided spectrum (including Nyquist and excluding 0 Hz frequency),
(6)
In terms of turbulence, σ 2 is often understood as being one velocity component of the turbulent kinetic energy. In Fig. 5, the ESD with and without averaging (moving average with equally spaced frequency interval in log-space) is plotted as a function of frequency. In addition, the user can choose the range (called inertial range) of the spectrum to be used to fit Kolmogorov's f 5 / 3 prediction2 (represented by a black dashed line).
FIG. 5.

Energy spectral density (ESD) in the frequency domain. The red solid corresponds to the averaged energy spectral density in the frequency domain, and the yellow solid line corresponds to the averaged and filtered ESD using the low-pass filter. The black dashed line corresponds to the Kolmogorov's f 5 / 3 prediction.2 

FIG. 5.

Energy spectral density (ESD) in the frequency domain. The red solid corresponds to the averaged energy spectral density in the frequency domain, and the yellow solid line corresponds to the averaged and filtered ESD using the low-pass filter. The black dashed line corresponds to the Kolmogorov's f 5 / 3 prediction.2 

Close modal

low_freq generates a pop-up dialog box to select whether the data should be filtered (low-pass filter) or not. If a filter is to be applied, then in the next step the frequency in Hz at which the data will be filtered by a low-pass filter has to be specified (for example 1800 Hz). If the pop-up dialog box is denied, it is set to the value low_freq=Fs/2.

frequency_filter. This function returns the filtered data and the filtered energy spectral density in the frequency domain (see yellow curve in Fig. 5). This function uses the butter function, which returns the transfer function coefficients of an nth-order lowpass digital Butterworth filter, and the filtfilt function of MATLAB to low-pass filter the data at the previously set low_freq frequency. For a fixed cutoff frequency, there is a threshold of nth-order of Butterworth filter at which the filter can be applied successfully to the data. We have implemented a routine that tries to design the highest order for a fixed cutoff frequency. The filtered data are named “data_filter” (a variable in MATLAB) and will be used for all further data post-processing. From this point, the variable u is the standard deviation and u 2 is the variance of the filtered data. If the filtering was not performed in the previous step, the variable “data_filter” and “data” (which is the unfiltered data) are equal. By means of Taylor's hypothesis of frozen turbulence (discussed in detail in Sec. II C), the spectral energy density as a function of frequency can be transformed into the energy spectrum,
(7)
in the wave number domain, with
(8)
In addition, different representations/normalization of the energy (or dissipation) spectrum density with respect to (a) frequency f and (b) wave number k and dissipation spectral density with respect to (c) scale r and (d) wave number k are plotted in Fig. 6.
FIG. 6.

Different representation/normalization of the energy spectral density with respect to (a) frequency f and (b) wave number k and dissipation spectral density with respect to (c) scale r and (d) wave number k. The vertical dashed lines indicate the low-pass filter frequency in a different representation.

FIG. 6.

Different representation/normalization of the energy spectral density with respect to (a) frequency f and (b) wave number k and dissipation spectral density with respect to (c) scale r and (d) wave number k. The vertical dashed lines indicate the low-pass filter frequency in a different representation.

Close modal

Length scales are important for any analysis of turbulent data. For scales larger than the “large length scale,” called integral length scale, it is expected that only uncorrelated noise is present in the flow. The “small length scale” determine the end of the cascade where dissipation smooths out all fluctuations. In this context, the Taylor length scale can be considered as the scale where dissipative become noticeable for the first time, whereas the dissipation or Kolmogorov length is where the flow structures become fully dissipative. Thus interesting features of turbulence are expected in the interval between large and small length scale, which is called inertial range. As these scales are important for the data analysis and interpretation, here, we present different well-known estimation methods for each of these scales. It should be noted that this list of estimation methods is far from complete and other package users are strongly encouraged to add further standard methods to it.

Before scales are discussed in detail, the equivalence of scales in time and space by the use of Taylor's hypothesis of frozen turbulence40 is mentioned. A spatial distance r is related to temporal separation τ by the mean velocity,
(9)
The validity of the Taylor hypothesis is an essential assumption of the analyses conducted in this package. Thus, the requirement for the data is that it is non-normalized local velocity data. For other data for which no mean value can be determined, the mean value can be set to 1 m/s. The consequences for the changed dimension have to be worked out by the user. The minus sign in Eq. (9) is a consequence of the typical time measurement with a hot wire at one location of the flow. In a time step τ, the flow from a location upstream (– r, direction x ̂ is defined downstream) is transported to the sensor. With the knowledge of the sampling frequency, all scales can also be expressed in units of samples, or sample steps, respectively. Thus all scales can be represented either as the number of samples, seconds or meters. For example, consider a hot wire signal with a sampling frequency of 50 kHz that is characterized by a mean velocity of 10 m/s and an integral length scale of 0.1 m. According to the basic equation of velocity = distance/time, the integral length scale of 0.1 m corresponds to a given duration 0.01 s. Considering the sampling frequency, this time is equal to 500 samples.
The accuracy of the Taylor hypothesis depends both upon the properties of the flow and on the statistic being measured. In grid turbulence, this method is quite accurate, and higher-order corrections can be made if the σ / u 1 (e.g., Refs. 28 and 41). As can be seen from the title of Fig. 3(b), the fluctuations are small compared with the mean velocity ( σ / u 0.17) for the dataset analyzed in this paper and, therefore, the method is applicable here. From the data series u(t) or u(x) (where u denotes the velocity component in direction of the mean flow), we construct the longitudinal increments,
(10)
labeled by the time-separation τ. According to Taylor's hypothesis, this temporal velocity increment is related to a spatial velocity increment by u r = u ( x + r ) u ( x ) = u τ. From the velocity increments, nth order structure functions,
(11)
are determined, where n is an integer.

length_scales. In this function, the integral length scale L, Taylor length scale λ, Kolmogorov length scale η, mean energy dissipation rate ε , normalized energy dissipation rate C ε and the local Taylor–Reynolds number R e λ are estimated using different methods of calculation. These parameters are used for the further processing of data (solving the Fokker–Planck equation and extracting cascade trajectories). If the fundamental length scales are already known from previous analysis, this calculation can be skipped and the values of integral, Taylor, Kolmogorov length scale in m, and energy dissipation rate in m 2 / s 3 can be entered in the pop-up dialog boxes. This also enables the package user to study the effect of different scale choices on further results of the analysis. The entered length scales will be rounded toward the nearest integer in sample units. Here, the Taylor hypothesis is used, so the link between mean velocity, sampling frequency, and a time interval is used to convert this interval into a length scale (see details in Sec. II C). The proposed value in the pop-up dialog box is the median length scale for all methods.

It should be noted that the following collection of methods for the estimation of fundamental length scales and the dissipation rate is far from being complete. When creating the package, only methods that are well-known to us have been taken into account and the list can be readily extended in future versions of the package. The different methods for estimating these parameters are numbered sequentially. While each method may share the same assumptions and hypotheses, the different methods can lead to inconsistent results. In our experience, this typically happens with experimental data, which may be characterized by some observational noise, as well as when the assumptions made are not perfectly met for these datasets, to which the methods have different sensitivities. Furthermore, for example, the number of methods for determining the Taylor length and dissipation rate differs. At the end of the estimation of the respective length scale, the results are presented in a single figure to compare them with each other. The comparison of the results of different methods allows the user to evaluate the consistency of the methods and adjust the proposed value in the pop-up dialog boxes (median length scale for all methods) accordingly. In an ideal case, all methods for one scale should give the same result, but in reality, different aspects of experimental data (e.g., measurement noise or insufficient number of samples) may change results significantly. In such cases, it is up to the user to find out the most reliable value.

The integral length scale L is estimated by using

  1. the energy spectrum density, which requires the range of frequency that will be used to linearly extrapolate the value of ESD at a frequency of 0 Hz (Refs. 42 and 43) (see Fig. 7). Therefore, the user will be asked to enter the f _ start and f _ end in Hz ( f _ start < f _ end). For example, f _ start = 0.03 Hz and f _ end = 2.7 Hz.
    (12)
  2. The autocorrelation coefficient R u ̃ u ̃ with respect to scales10,28,44 plotted in Fig. 8, namely,
    (13)

    where r is the scale in meter and u ̃ = u u is the fluctuating component of streamwise velocity. The cumulative integral gives the asymptotic value at a specific scale r, which is characterized by the integral length scale L.

    In the case of Lagrangian tracer particles where trajectories from direct numerical simulation (DNS) are oftentimes not long enough or in the case of non-ideal experimental data, Eq. (13) may not fully converge. As a result, the aforementioned method can lead to large errors in the estimation of the integral length. In the case of a non-monotonic decrease of R u ̃ u ̃ ( r ), the autocorrelation function is

  3. integrated up to the first zero-crossing of the autocorrelation function.45 

  4. integrated up to the first 1 / e crossing of the autocorrelation function.46 

    In particular, for experimental data, there may be additional measurement noise present in the data that may cause no zero crossing of the autocorrelation function. In this case, R u ̃ u ̃ ( r ) is

  5. fitted by an one-term exponential function,42,46
    (14)
    The fit region corresponds to the range of scales r [ 0 , r e ], with re is the scale of the first 1 / e crossing of the autocorrelation function. This fit range is indicated by the two black, vertical dashed lines in Fig. 8. The integral is calculated by using the coefficients a and b of the exponential fit,
    (15)

    Furthermore, the integral length scale L is estimated via

  6. the second order structure function,
    (16)
    which is linked to the autocorrelation coefficient R u ̃ u ̃ according to
    (17)

    where u can be calculated directly from the second order structure function. At a sufficiently large length scale compared to the energy injection scale of the experiment, the second order structure function tends to the asymptotic value of 2 u 2.47 

  7. The zero crossings proposed by Mora and Obligado.48 To verify that zero crossings are well resolved, the signal has to be filtered with a low-pass filter with a characteristic length ηc. This method consists of estimating the Voronoi tessellation of the 1D zero-crossings positions data set. It is compared with a random Poisson process, which has no correlations between the scales. The method proposes that the integral length scale is equal to the characteristic length ηc for which
    (18)

    Here, σvoro is the standard deviation of the Voronoi cells normalized by their mean value and σRPP is the equivalent value for an random Poisson process, that is equal to ( 1 / 2 ). In Fig. 9, the standard deviation (normalized) of the Voronoi cells as a function of the characteristic length ηc is plotted. Finally, the integral length scale is defined as the value of ηc that corresponds to σ voro / σ RPP = 1. If we observe that σ voro / σ RPP > 1 for all values of ηc, the method cannot provide the value of the integral length scale and longer signals are needed (nevertheless the extrapolation of the value remains possible). Note that the characteristic length ηc should not be confused with the Kolmogorov length scale η. To provide a direct link to Ref. 48, here, we used the exact same notation as in the original reference.

FIG. 7.

Representation of the linear extrapolation of ESD at a frequency of 0 Hz for estimating the integral length scale with the method according to Hinze42 and Roach.43 The two vertical dashed lines correspond to the range of frequency that will be used to linearly extrapolate (solid black line).

FIG. 7.

Representation of the linear extrapolation of ESD at a frequency of 0 Hz for estimating the integral length scale with the method according to Hinze42 and Roach.43 The two vertical dashed lines correspond to the range of frequency that will be used to linearly extrapolate (solid black line).

Close modal
FIG. 8.

Representation of the autocorrelation coefficient R u ̃ u ̃ as a function of scale in meter. The two vertical dashed lines correspond to the range of scale that will be used for the exponential fit (red solid line) for estimating the integral length scale with the method according to Hinze42 and Tritton.46 

FIG. 8.

Representation of the autocorrelation coefficient R u ̃ u ̃ as a function of scale in meter. The two vertical dashed lines correspond to the range of scale that will be used for the exponential fit (red solid line) for estimating the integral length scale with the method according to Hinze42 and Tritton.46 

Close modal
FIG. 9.

Standard deviation (normalized) of the Voronoi cells as a function of the characteristic length ηc (not to be confused with the Kolmogorov length scale η) of a low-pass filter for estimating the integral length scale with the method according to Mora and Obligado.48 

FIG. 9.

Standard deviation (normalized) of the Voronoi cells as a function of the characteristic length ηc (not to be confused with the Kolmogorov length scale η) of a low-pass filter for estimating the integral length scale with the method according to Mora and Obligado.48 

Close modal

The Taylor length scale λ is estimated by using

  1. a parabolic fit,
    (19)

    to the autocorrelation function, to estimate λ at the origin r = 0 (see Fig. 10). The range of the negative curvature is, therefore, used for the estimation (close to r = 0 the auto-correlation function has an inflection point). Since this method needs a well-resolved auto-correlation function, it strongly depends on the resolution of the sensor. It is also affected by low pass filtering.

    Assuming isotropy at small scales and Taylor's frozen turbulence hypothesis the Taylor length scale λ is estimated using the relation
    (20)

    The numerical differentiation in the denominator is approximated by

  2. the simple difference quotient. Due to the finite spatial resolution of the measuring sensor used in this typical experimental dataset and measurement noise, this method will yield an incorrect result. In order to correctly compute the derivatives, the spatial resolution must be of the order of the Kolmogorov microscale η (see Ref. 49), which is estimated via Eq. (32).

  3. The procedure proposed by Aronson and Löfdahl.50 Here the derivative of the velocity is approximated via the second order structure function S2 in the inertial range. In Fig. 11 the asymptotic behavior of
    (21)

    as a function of scale r is plotted. For the extrapolation we use a linear fit. The fit-region is determined for spatial lags that are larger than four times the scale that corresponds to the low-pass filter frequency low_freq. The larger limit of the fit-region must be set by the user.

  4. Using the dissipation spectrum,
    (22)

    This procedure has been proposed by Hinze.42 The upper limit of the integration is set to the low-pass filter frequency low_freq.

  5. From the dissipation spectrum, but here the upper limit of the integration of Eq. (22) is set to infinity (or the largest available wave number k) so that the full dissipation spectrum is used.

    Furthermore, the Taylor length scale λ is estimated via

  6. the zero crossings of the fluctuating velocity
    (23)

    where l is the average distance between zero-crossings. C is a constant of the order of unity that quantifies the level of non-Gaussianity of the derivative u ̃ / x. This method is discussed in Refs. 51–53. In Fig. 12, the density of zero-crossings ns times the average distance between zero crossings as a function of the characteristic length ηc is plotted.

    For values of ηc within the inertial range, a power law 2/3 is expected,51 and eventually for smaller filter sizes (or large 1 / η c) a plateau is reached. The presence of this plateau, related to the dissipative range, implies that the density of zero-crossings are well resolved, and therefore, l can be deduced using the trivial relation n s · l = 1. If the plateau is not reached, small scales are not resolved and the method cannot estimate the Taylor length scale. On the other hand, if after the plateau the value of n s · l increases again, it means that the cutoff frequency is too high and the analysis is affected by small scale noise. Initially, the constant C = 1 gives a good approximation of λ using Eq. (23).

  7. The zero crossings of the fluctuating velocity, but here C is defined as (see Ref. 54),
    (24)

    where | x u ̃ | is the mean and σ x u ̃ is the standard deviation of u ̃ / x, where u ̃ is filtered with the largest frequency within the plateau of n s · l. Latter method results in a better estimation of λ if u ̃ / x is resolved.

FIG. 10.

Autocorrelation function with parabolic fit as a function of the scale r for estimating the Taylor length scale. The vertical dashed line corresponds to the range of scale that will be used for the extrapolation (red solid line) by a parabolic fit.

FIG. 10.

Autocorrelation function with parabolic fit as a function of the scale r for estimating the Taylor length scale. The vertical dashed line corresponds to the range of scale that will be used for the extrapolation (red solid line) by a parabolic fit.

Close modal
FIG. 11.

Development of Eq. (21) with linear extrapolation (red solid line) for estimating the Taylor length scale with the method according to Aronson and Löfdahl.50 The two vertical dashed lines correspond to the range of scale that will be used for the extrapolation.

FIG. 11.

Development of Eq. (21) with linear extrapolation (red solid line) for estimating the Taylor length scale with the method according to Aronson and Löfdahl.50 The two vertical dashed lines correspond to the range of scale that will be used for the extrapolation.

Close modal
FIG. 12.

Density of zero-crossings times the average distance between zero crossings as a function of the characteristic length ηc of a low-pass filter for estimating the Taylor length scale using the method discussed in Refs. 51–54.

FIG. 12.

Density of zero-crossings times the average distance between zero crossings as a function of the characteristic length ηc of a low-pass filter for estimating the Taylor length scale using the method discussed in Refs. 51–54.

Close modal

The mean energy dissipation rate ε is estimated by its one-dimensional, isotropic surrogate using

  1. 1. & 2. method: either second or third order structure function. The estimation of the dissipation rate using this method relies on the transfer of energy within the inertial range. This method is particularly useful when the higher frequency content present in the flow is not fully resolved by the measurement device. This is generally the case when, e.g., the length of the hot wire is larger than the Kolmogorov length scale η [see Eq. (32)] of the flow under consideration. This function generates a pop-up dialog box to enter the value of the constant C 2 1.9 ± 0.2 (Refs. 55–57) used in the relation [see Eq. (25)] between second order structure function S 2 ( r ) and energy transfer rate ε ( r ) (transfer from one scale to another one) based on the assumption of homogeneous isotropic turbulence (HIT).28,40 The mean energy dissipation rate ε is calculated by determining the mean among 5 points closest to the peak (i.e., a small plateau) value of ε ( r ). We expect that this plateau becomes more pronounced with the increasing Reynolds number. In Fig. 13, the scale dependence of
    (25)
    (26)

    using either second or third order structure function, S 2 ( r ) = u r 2 or S 3 ( r ) = u r 3 , is plotted.

    Basically, the energy transfer mechanism in the turbulent cascade takes place in three different ways: First, the energy is injected at large scales in the turbulent flow. Second, the injected energy up to the integral length scale transfers across within the inertial range down to the scale at which viscous effects become significant. Third, the transferred energy across the inertial scales starts to dissipate into heat because of the dominant effects of fluid viscosity across the dissipation range. From the conservation of energy principle, it follows that in steady state the rate of injection of energy at large scales, the energy transfer rate across inertial scales, and the energy dissipation rate at small scales in the dissipation range must be equal. We note that Eq. (26) is an exact result in the inertial range and it requires stationarity, homogeneity, and isotropy of the turbulent flow. The lack of isotropy in turbulent flows leads to a modification of the prefactor 5 / 4. For further details, the corresponding literature should be studied, see, for example, Ref. 58. In the present context, however, we assume the underlying flow to be isotropic. Equations (25) and (26) lead to the same estimation of the dissipation rate by choosing the right value of the constant C2.

  2. By using the chosen Taylor length scale and the following relation:
    (27)

    which is valid for isotropic turbulence.59 

  3. As a consequence of the K41 phenomenology, the second order structure function implies an energy spectrum density of the form
    (28)

    Ck is the so-called Kolmogorov constant that remains undetermined in Kolmogorov's theory [typically C k 0.53 ± 0.01 (Refs. 9 and 60)]. Following Kolmogorov's k 5 / 3 prediction, a fit in the inertial range ( λ < r < L) according to Eq. (28) is used to estimate the mean energy dissipation rate ε (see Fig. 14).

  4. A fit in the inertial range ( λ < r < L) according to
    (29)

    with k0 being a shift for the argument of the function, i.e., a shift along the wave number axis, which is now used to estimate the mean energy dissipation rate ε (see Fig. 14).

  5. Through the dissipation spectrum,
    (30)
    Equation (30) is valid for isotropic turbulence. As proposed by Mora et al.,53 the dissipation spectrum is modeled for large wave number k using a power law (see Fig. 15),
    (31)

    with the coefficients a and b of the power law fit. The integration of Eq. (30) is divided in two parts. The first part is performed from k = 0 to the beginning of the range of wave numbers used for the fit. Then, the modeled dissipation spectrum is used for the second part of the integration from the beginning of the range of wave numbers used for the fit to the largest available wave number k.

FIG. 13.

Scale dependence of ε ( r ) using either second or third order structure function [see Eqs. (25) and (26)]. The black solid line marks the peak value of ε ( r ).

FIG. 13.

Scale dependence of ε ( r ) using either second or third order structure function [see Eqs. (25) and (26)]. The black solid line marks the peak value of ε ( r ).

Close modal
FIG. 14.

Energy spectral density (ESD) in the wave number domain. The red and yellow solid line corresponds to the fit according to Eqs. (28) and (29) for estimating the mean energy dissipation rate ε .

FIG. 14.

Energy spectral density (ESD) in the wave number domain. The red and yellow solid line corresponds to the fit according to Eqs. (28) and (29) for estimating the mean energy dissipation rate ε .

Close modal
FIG. 15.

Dissipation spectral density in wave number domain. The black dashed line corresponds to the power law fit to model the dissipation spectrum for large wave number k. The two black circles correspond to the range of wave number that will be used for fit.

FIG. 15.

Dissipation spectral density in wave number domain. The black dashed line corresponds to the power law fit to model the dissipation spectrum for large wave number k. The two black circles correspond to the range of wave number that will be used for fit.

Close modal
The Kolmogorov length scale η (the smallest size of the eddy in a given turbulent flow) is estimated by using the classical relation10 given by
(32)
In addition, the normalized energy dissipation rate C ε will be returned,61,62
(33)
(34)
with the local Taylor–Reynolds number,
(35)

R e λ allows for reasonable comparisons of experiments with different boundary conditions or energy injection mechanisms ( R e λ is independent of the integral length scale L).

At the end of this function, in Fig. 16, a vertical dashed line at the integral L, Taylor λ, and Kolmogorov length scale η will be added to energy spectral density and compensated energy spectral density.

FIG. 16.

ESD and compensated ESD in the frequency domain. Black dashed vertical lines correspond to the respective frequency of the fundamental length scales.

FIG. 16.

ESD and compensated ESD in the frequency domain. Black dashed vertical lines correspond to the respective frequency of the fundamental length scales.

Close modal

In the following, a test corresponding to the forward cascade using the third order structure function is performed. Furthermore, on the account of the comparability of various turbulent datasets, a common data normalization strategy is proposed, as mentioned below. Such normalization procedure is possible, as we have now introduced all the necessary scales.

Struc_flip_test. This function tests whether the data have to be flipped or not. Flipping the data here is referring to reversing the order of the elements. The decision of flipping of data depends on a simple relation of third order structure function S 3 ( r ) with the dissipation based on the assumption of homogeneous isotropic turbulence (HIT). In the present analysis, we assume the forward turbulent cascade. Thus, in the forward turbulent cascade, the third order structure function becomes negative as a consequence of irreversibility (i.e., no inverse cascade) within the turbulent cascade. Based on this we propose a rule of thumb that demands that the quantity S 3 ( r ) must be negative in the inertial range, taking into account that the increment time series is obtained according to Eq. (10). Vortex stretching is the physical phenomenon that causes the third order structure function to be negative. To verify this, S 3 ( r ) as a function of the scale r is plotted, from which it is possible to decide whether it is essential to flip the data or not.

Normalization. This function serves mainly to perform the normalization of the filtered data. Before doing so, it generates the pop-up dialog box, which asks the user whether to flip the data or not (based on the previous investigation). After that, the entire data can be normalized by the quantity
(36)
This normalization
(37)
is proposed by Renner et al.38 This function returns the variable in MATLAB named “m_data,” which is the mean of the filtered data before normalization and the variable “siginf,” which is equal to σ . In addition the user is asked whether the scale r should be given in units of Taylor length scale,
(38)
As ur and σ are of the same unit (m/s) and, respectively, r and λ have the same unit (m), the normalization of both the increment and the scale is used in the next part of this paper to obtain non-dimensionalized Kramers–Moyal coefficients (KMCs) as a function of velocity increment and scale in order to compare the results of different data sets. The normalization affects the KMCs as follows:
(39)
(40)
(41)
(42)
(43)
(44)
Note that the estimation of the Kramers–Moyal coefficients (KMCs) will be introduced in Sec. III C. The two functions D ( 1 , 2 ) defining the Fokker–Planck equation are called drift and diffusion coefficients, respectively, and can be estimated directly from measured data by an optimization procedure proposed in Refs. 32, 63, and 64. If the user utilizes normalized quantities, ur and r should not be interchanged with the previously unnormalized quantities, such as the ones used to determine the fundamental length scales.

In this section, the properties of the turbulent structures estimated from the velocity increments using the probability density function and scaling exponents of the higher-order structure functions are examined.

plot_increment_pdf. This function plots the probability density function of the velocity increments at the scale r = L, r = λ and r = η, as depicted in Fig. 17.

FIG. 17.

PDF of the velocity increments ur at the scale r = L , λ , η. The colored dashed line correspond to Castaing fits (shape factor Λr65) and gray dashed line to Gaussian fits. In the title, the skewness for each of the three scales is displayed.

FIG. 17.

PDF of the velocity increments ur at the scale r = L , λ , η. The colored dashed line correspond to Castaing fits (shape factor Λr65) and gray dashed line to Gaussian fits. In the title, the skewness for each of the three scales is displayed.

Close modal
plot_struc_function. As shown in Fig. 18, this function plots the kth order structure function of the velocity increments and of their absolute values,
(45)
(46)
with k = 2 7 for scales λ r L. By comparing S 3 ( r ) with the dashed red line in Fig. 18(b) possible deviations of the analyzed data from Kolmogorov's 4 / 5 are evident. This law represents an exact result in the inertial range and requires stationarity, homogeneity, and isotropy of the turbulent flow and a sufficiently high Reynolds number. In addition, this function plots the scaling exponent ζk (see Fig. 19), with
(47)
estimated using the extended-self-similarity (ESS) method introduced by Benzi et al.66 This method investigates the relative scaling laws between structure functions of different orders with k p. The most comfortable way is to apply this method for p =3 and thus ζ 3 = 1, leading to the investigation of S k ( S 3 ( r ) ) ( S 3 ( r ) ) ζ k. This method allows for determining the scaling exponent ζk for flows with small or even non-existent inertial ranges.
FIG. 18.

Course of the kth order structure function Sk and Tk for (a) k = 2 , 4 , 6 and (b) k = 3 , 5 , 7 as a function of scale. The dashed red line in (b) represents 4 / 5 law.

FIG. 18.

Course of the kth order structure function Sk and Tk for (a) k = 2 , 4 , 6 and (b) k = 3 , 5 , 7 as a function of scale. The dashed red line in (b) represents 4 / 5 law.

Close modal
FIG. 19.

Course of scaling exponent ζk as a function of order k of structure function. Dashed lines represent the scaling of a selected set of known intermittency models.

FIG. 19.

Course of scaling exponent ζk as a function of order k of structure function. Dashed lines represent the scaling of a selected set of known intermittency models.

Close modal

The scaling of a selected set of known intermittency models is also included in Fig. 19 for comparison. For this, the user is asked to specify the intermittency coefficient μ [experiments suggest a value of μ 0.227 (Ref. 10) and μ 0.26 (Ref. 67)] and the coefficient D =2.8 (Refs. 10 and 68–70) of the β-model proposed by Novikov and Stewart69 and introduced by Frisch et al.10,70

In the phenomenological model by Friedrich and Peinke,71 the turbulent energy cascade is interpreted as a Markov process of the velocity increments ur evolving in scale r. The chaotic property of a turbulent flow implies that the evolution of turbulent structures in the cascade exhibits a certain degree of randomness. For this reason, the turbulent cascade is taken as a stochastic process described by a Fokker–Planck equation and its Kramers–Moyal coefficients. The distinctive feature of this approach is that the stochastic processes are not considered in time but as a process evolving in scale r.71–75 A realization of this stochastic process is a sequence of velocity increments, u r 0 , u r 1 , , u r n with r 0 > r 1 > > r n at one location x or, respectively, t, see definition equation (10).

To use a Fokker–Planck equation for the cascade, the process has to exhibit the Markov property in scale. The Markov property reduces a general description of cascade trajectories by the joint multiscale pdf p ( u r 0 , u r 1 , , u r n ) to the knowledge of the one step or single conditioned pdfs p ( u r | u r 0 ) with r < r 0, which can be described by a general Kramers–Moyal forward expansion,
(48)
Furthermore, if Gaussian noise is dominant, the general Kramers–Moyal forward expansion of the Markov process reduces to the first two terms D ( 1 , 2 ). This can be tested by Pawula's theorem, which states that a vanishing fourth order Kramers–Moyal coefficient D ( 4 ) ( u r , r ) implies that higher-order coefficients vanish as well (see Ref. 76). The latter theorem has been used extensively in turbulence studies38,71,73,77 in order to reduce the full Kramers–Moyal expansion to a Fokker–Planck equation. In recent numerical studies, the asymptotic behavior of higher-order Kramers–Moyal coefficients has been invoked in order to better distinguish between various phenomenological models of turbulence.39,78
A stochastic description of the energy cascade process by a Fokker–Planck equation in scale can be expressed with r < r 0 according to
(49)
This equation is also referred to as forward or second Kolmogorov equation.38 The two functions D ( 1 , 2 ) defining the Fokker–Planck equation, which describes the diffusion processes, are called drift and diffusion coefficients, respectively. Both coefficients can be estimated directly from data, as shown below. The drift coefficient determines where the process will go to on average (purely deterministic). The diffusion coefficient determines the strength of the stochastic influence (uncertainty in the evolution of the process). In other words, the drift coefficient characterizes the deterministic evolution from large scale to small scales, whereas the diffusion coefficient expresses the interaction of additive and multiplicative noise within the turbulence cascade.

This approach, hereafter referred to as Markov analysis, achieves a comprehensive and effective characterization of the complexity of turbulence including/preserving the phenomenon of intermittency by stochastic equations. The correct estimation of the coefficients D ( 1 , 2 ) is crucial for an accurate description of the underlying scale dependent process by a Fokker–Planck equation. The validity of these coefficients is subsequently tested via the reconstruction of structure functions and probability density functions of velocity increments (see Sec. V) and by validating the integral fluctuation theorem (see Sec. IV).

A central assumption of the Markov analysis is that the turbulent cascade process (more precisely the statistics of the scale-dependent velocity increments) possesses a Markov property in scale. Experimental evidence show that the Markov property can be assumed to hold for the cascade coarse-grained by the Einstein–Markov length Δ E M 0.9 λ,38,79 which suggests that molecular friction causes the break-down of the Markov assumption. This finite step length can be seen in close analogy to the mean free path length of a Brownian diffusion process, which has to be so large, that two successive steps of the process can be considered as independent events.80 For Markovian processes, the relation
(50)
holds, which means that conditional PDF of velocity increments depends only on the increment at the next larger scale r n 1. From Eq. (50), one sees that the Markov property corresponds to a two increment closure for the joint multi-scale statistics. The two increments u r n = u ( x + r n ) u ( x ) and u r n 1 = u ( x + r n 1 ) u ( x ) are given for three velocity values at three locations x , x + r n , x + r n 1, and thus, we see that the Markov property is not only a two scale but also a three point closure. A direct relation to n-point statistics can be achieved if a further condition on u(x), i.e., p ( u r n | u r n 1 , , u r 0 , u ( x ) ) are included, for more details see Ref. 22. Motivated by the cascade picture from largest to smallest scale, this can be interpreted in terms of a large eddy, which entirely determines the properties of the next smaller eddy and which affects the following eddies only indirectly. In other words, the future state of the process does not depend on the entire past, but only on the present state of the process (the scale evolution of the increments is memoryless).

There are several methods to verify the Markov property using experimental data.81–83 In the following, two different methods are presented. The Wilcoxon test, which is a quantitative test and one of the most reliable procedures, and a qualitative/visual test to validate the Markov property. The advantage of these two methods is that they do not depend on the estimation of the Kramers–Moyal coefficients instead both use exclusively the underlying data for the verification.

wilcoxon_test. This function determines the Einstein–Markov length ΔEM.38 Above this length scale, the Markov properties hold and below this length scale, the Markov properties cease to hold. The Wilcoxon test is a parameter-free procedure to compare two empirically determined probability distributions (two multi-scale statistics of velocity increments, respectively, two datasets of conditioned velocity increments). As described above, Eq. (50) is valid for a Markovian processes. For finite datasets,
(51)
is commonly assumed to be a sufficient condition. Therefore, Eq. (50) is validated for the three different scales r 0 > r 1 > r 2, each separated by Δ r = Δ E M. For this chosen set of scales the normalized expectation value of the number of inversions of the conditional velocity increment as a function of Δ r is calculated (see Fig. 20). If the Markov properties holds the expectation value is equal to 1. A sufficient resolution of the measurement below Taylor's length scale is expected in order to perform this test. A detailed description of the test is given in Refs. 38 and 79.
FIG. 20.

Development of the normalized expectation value t ( r , Δ r ) / 2 / π (see Ref. 79 for details) as a function of Δ r in terms of samples (a) on the log –log scale and (b) lin–log scale. The vertical dashed line corresponds to the Taylor length scale λ in terms of samples.

FIG. 20.

Development of the normalized expectation value t ( r , Δ r ) / 2 / π (see Ref. 79 for details) as a function of Δ r in terms of samples (a) on the log –log scale and (b) lin–log scale. The vertical dashed line corresponds to the Taylor length scale λ in terms of samples.

Close modal

markov generates a pop-up dialog box to enter the number of samples that corresponds to Einstein–Markov length (proposed value in the pop-up dialog box is Δ E M 0.9 λ). It should be noted that if the resolution of the used sensor ceases at ΔEM, it is possible to enter the number of samples that correspond to a larger scale than ΔEM at which the data might be resolved in scale (for example samples corresponding to λ or 5 λ). A red vertical dashed line at the Einstein–Markov length will be added to the spectrum in the frequency domain.

min_events generates a pop-up dialog box to enter the minimum number of events/counts to occur in a single bin, which will make that specific bin valid for further processing. If the minimum number of events is equal to 400 all of the information in those specific bins in which the number of events/counts is less than 400 will be excluded for further post-processing of data. The provision of the minimum number of events/counts is to avoid the appearance of noise and hence to achieve better parameter fits. Based on the experience, we have fixed the minimum value of the min _ events to 400. Nonetheless, it is possible to increase/decrease this number on the basis of number of sample points or significant statistics.

conditional_PDF_markov. This function performs a qualitative/visual check for the validation of the Markov property based on the alignment or misalignment of the single conditioned p ( u r 2 | u r 1 ) and double conditioned p ( u r 2 | u r 1 , u r 0 ) probability density functions (PDFs) of datasets of velocity increments for a chosen set of three different scales r 0 > r 1 > r 2, each of which is separated by ΔEM. To do this, a pop-up dialog box is generated to enter the conditioned value for large scale increment u r 0, for example u r 0 = ± 1. Note that the condition u r 0 = 0 corresponds to the maximum number of statistics. This function also plots various representations of the single and double conditioned PDFs (shown in Fig. 21 are only two). If there is no good agreement between single conditioned and double conditioned PDF of velocity increments, it is possible to modify the Einstein–Markov length and/or the minimum number of events and repeat this check for the validation of the Markov property. In order to support this qualitative/visual check additionally in the title of Figs. 21(a) and 21(b), a weighted mean square error function in logarithmic space84 (analogous to Kullback–Leibler divergence of conditional probabilities),
(52)
is given. This error function is a logarithmic measure of the difference between the single conditioned A = p ( u r 2 | u r 1 ) and double conditioned B = p ( u r 2 | u r 1 , u r 0 ) conditional probability density function. The closer the match between the PDFs (black and red solid lines in Fig. 21), the smaller this distance measure. The distance measure defined in Eq. (52) is composed of the specific difference of each bin (specified in the function increment_bin) for all values of longitudinal velocity increments u r 1 and u r 2. The first sum is calculated for all bins with respect to u r 1 and the second sum is for all bins regarding u r 2. The specific values of the single, respectively, double conditioned probability density function in the respective bin are compared. Finally, the individual differences of each bin are summed up to a single value ξ Δ E M.
FIG. 21.

Visualization of Markov properties. [(a) and (b)] Contour plots showing single (black solid lines) and double conditioned PDFs (red solid lines, u r 0 = 0) of velocity increments for three different scales r 0 > r 1 > r 2, each of which is separated by ΔEM. (b) is a three-dimensional view of (a). In the title of [(a) and (b)], the error function ξ Δ E M [see Eq. (52)] is given. The dashed vertical black lines in (a) correspond to cut through the single and double conditioned PDFs at the marked points of u r 2. [(c)–(f)]: Cut through the single (black) and double (red) conditioned PDFs at the marked points of u r 2 [(c) and (d): double linear plot, (e) and (f): semi logarithmic plot].

FIG. 21.

Visualization of Markov properties. [(a) and (b)] Contour plots showing single (black solid lines) and double conditioned PDFs (red solid lines, u r 0 = 0) of velocity increments for three different scales r 0 > r 1 > r 2, each of which is separated by ΔEM. (b) is a three-dimensional view of (a). In the title of [(a) and (b)], the error function ξ Δ E M [see Eq. (52)] is given. The dashed vertical black lines in (a) correspond to cut through the single and double conditioned PDFs at the marked points of u r 2. [(c)–(f)]: Cut through the single (black) and double (red) conditioned PDFs at the marked points of u r 2 [(c) and (d): double linear plot, (e) and (f): semi logarithmic plot].

Close modal

Note that the results presented hereafter are related to a modified number of bins (changed from 93 to 201). The number of bins has been adjusted to get a more detailed view of the following figures (conditional PDFs and the Kramers–Moyal coefficients). These detailed illustrations will be described in the readme file in an exemplary manner. A smaller number of bins leads to slightly different results, but the general trend remains the same.

In this section, the conditional moments of the increments,
(53)
are estimated. Here, the most straightforward approach that performs a histogram-based estimation is chosen. To do so, the user is asked to input the number of scales (steps in the energy cascade) at which Markov analysis needs to be performed.

Scale_steps. This pop-up dialog box calculates the possible number of steps between the integral length scale and Taylor length scale, which are separated by the Markov length. For these steps, the conditional moments for different r and later on the Kramers–Moyal coefficients (KMCs) will be estimated.

multi_point. In this pop-up dialog box, it must be selected whether to perform the multipoint analysis or not (see annual review article by Peinke, Tabar and Wächter22). To do the Fokker–Planck analysis using the multiscaling approach, the question in this pop-up dialog box must be denied. If multipoint analysis should be performed, an additional condition on the increment must be specified in the next pop-up dialog box.

conditional_moment. This function estimates the kth order conditional moment M ( k ) ( u r , r , Δ r ), k = 1 4 for all scales 2 Δ E M < r L and for each bin (specified in the function scale_steps and increment_bin) for all values of longitudinal velocity increments ur. For a fixed scale r, the conditional moments are calculated for 5 different scales separations (colored circles in Fig. 22) Δ r = r r within the range of Δ E M Δ r 2 Δ E M. Here, the condition r < r is fulfilled.

FIG. 22.

First and second conditional moments M ( 1 , 2 ) ( u r , r , Δ r ) as a function of the scale separation Δ r. In addition, a linear extrapolation in Δ r (solid black line) of the first and second order conditional moments is plotted (see Chapter: Estimation of Kramers–Moyal coefficients). The vertical dashed lines and the colored circles limit the range used for the linear fit ( Δ E M Δ r 2 Δ E M).

FIG. 22.

First and second conditional moments M ( 1 , 2 ) ( u r , r , Δ r ) as a function of the scale separation Δ r. In addition, a linear extrapolation in Δ r (solid black line) of the first and second order conditional moments is plotted (see Chapter: Estimation of Kramers–Moyal coefficients). The vertical dashed lines and the colored circles limit the range used for the linear fit ( Δ E M Δ r 2 Δ E M).

Close modal

plot_conditional_moment. This function plots a higher resolved estimation of the first and second conditional moments M ( 1 , 2 ) ( u r , r , Δ r ) for 12 different scales separations Δ r within the range of 0 Δ r 3 Δ E M (see Fig. 22). For this purpose, a scale r and the number of a bin (value of the velocity increment ur) condition must be specified. The proposed value in the pop-up dialog box is r = L and u r 0. A possible deviation from a linear law for small values of Δ r is due to the Einstein–Markov length, as the Markov property ceases to hold for very small scale separations.

In this section, the Kramers–Moyal coefficients D ( k ) ( u r , r ) are estimated by means of the conditional moments that were calculated in Sec. III B.

KM_Calculation. This function calculates the Kramers–Moyal coefficients D ( k ) ( u r , r ) with k = 1 4 for all scales (specified in scale_steps) and for each bin (specified in the function increment_bin) for all values of velocity increments by a linear extrapolation in Δ r of the kth order conditional moments M k ( u r , r ) (see Fig. 22) and the function KM_plot_raw plots the coefficients [see Figs. 23(a) and 23(b)]. With r < r
(54)
Note that some publications use the prefactor −r to implicitly describe a logarithmic scaling of the scale evolution from large to small scales, which is advantageous for complex structures with self-similar properties. This prefactor is omitted here. From Eq. (53), it can be seen that the smallest scale at which the KMCs can be calculated is slightly larger than 2 Δ E M.
FIG. 23.

[(a) and (b)] Non-optimized Kramers–Moyal coefficients (a) D ( 1 ) ( u r , r ) and (b) D ( 2 ) ( u r , r ) with respect to scale r (with L 10 r / λ) and velocity increment obtained by the linear extrapolation method of Fig. 22. (c) Contour plots showing experimental pexp (black), non-optimized pstp (blue) and optimized conditional PDFs p stp , opti (red, first step of the pointwise optimization) using the short time propagator [see Eq. (55)] of velocity increments for a pair of two scales with r < r, each of which is separated by ΔEM. (d) is a three-dimensional view of (c) (only pexp and p stp , opti are shown). In the title of [(c) and (d)] the error function ξ Δ E M and ξ stp , opti is given. [(e) and (f)]: Non-optimized and optimized (first step of the pointwise optimization) Kramers–Moyal coefficients D ( 1 , 2 ) ( u r , r ) with respect to velocity increment ur for a fixed scale r = 2.7 λ = 3.2 Δ E M obtained by the optimization algorithm.

FIG. 23.

[(a) and (b)] Non-optimized Kramers–Moyal coefficients (a) D ( 1 ) ( u r , r ) and (b) D ( 2 ) ( u r , r ) with respect to scale r (with L 10 r / λ) and velocity increment obtained by the linear extrapolation method of Fig. 22. (c) Contour plots showing experimental pexp (black), non-optimized pstp (blue) and optimized conditional PDFs p stp , opti (red, first step of the pointwise optimization) using the short time propagator [see Eq. (55)] of velocity increments for a pair of two scales with r < r, each of which is separated by ΔEM. (d) is a three-dimensional view of (c) (only pexp and p stp , opti are shown). In the title of [(c) and (d)] the error function ξ Δ E M and ξ stp , opti is given. [(e) and (f)]: Non-optimized and optimized (first step of the pointwise optimization) Kramers–Moyal coefficients D ( 1 , 2 ) ( u r , r ) with respect to velocity increment ur for a fixed scale r = 2.7 λ = 3.2 Δ E M obtained by the optimization algorithm.

Close modal

As described above, a histogram-based estimation of the Kramers–Moyal coefficients is performed. We call this estimation pointwise due to the fact that the coefficients are determined for fixed scale r and fixed velocity increment of a bin. While this approach is one of the most straightforward methods to determine an initial estimate of D ( k ) ( u r , r ) from data, it is also sometimes characterized by a high degree of uncertainty. This is especially true for bins containing a relatively small number of events. Furthermore, the limit approximation in Eq. (54) (see also Fig. 22) leads to uncertainties in the absolute values of the Kramers–Moyal coefficients, whereas the functional forms of D ( k ) ( u r , r ) are commonly well estimated.

In this package, we address this problem by following a two-step optimization. First, a pointwise optimization of the experimentally estimated D ( 1 , 2 ) ( u r , r ) at each scale and value of velocity increment is performed using the short time propagator76 to find the best Fokker–Planck equation to reproduce the conditional PDFs as depicted in Figs. 23(c) and 23(d) (see Sec. III E). As shown in Ref. 64 for experimental data, there still remains an uncertainty in the coefficients D ( 1 , 2 ) using this first optimization. This uncertainty is used as a freedom to furthermore optimize the KMCs in the following second step. In this second step of the optimization (see details in Sec. IV B), the Kramers–Moyal coefficients are optimized relatively to the integral fluctuation theorem (see details in Sec. IV).

This procedure to obtain an optimal Fokker–Planck equation can be interpreted as follows: If the condition for a Markov process for the cascade is given, the whole process is uniquely defined by the transition probabilities p ( u r n | u r n 1 ) for increments from scale r n 1 to rn. This is the most central feature of the cascade process. In step 1, the best Fokker–Planck equation to model p ( u r n | u r n 1 ) is found. The remaining uncertainties in the coefficients of the Fokker–Planck equation are used in step 2 to get the best values for the integral fluctuation theorem (IFT). In this way, a Fokker–Planck equation, which is compatible with the Markovian cascade process is obtained. As the fulfillment of the IFT is an independent feature and a direct consequence of a Fokker–Planck equation, the here-obtained Fokker–Planck equation can be considered as the optimal solution. The methods presented here allow for a critical analysis of these points for any given data set.

KM_STP_optimization. This function performs the first step of the pointwise optimization of Kramers–Moyal coefficients D ( 1 , 2 ) ( u r , r ) at each scale and value of velocity increment to minimize possible uncertainties in the absolute values of the Kramers–Moyal coefficients. The purpose of this optimization is to find the best Fokker–Planck equation to reproduce the conditional PDFs as these are the essential part of the Markov process. This optimization procedure is proposed in Refs. 32, 63, and 64, and it includes the reconstruction of the conditional probability density functions p ( u r | u r ) via the short time propagator of Eq. (55) with r < r,
(55)
The scale step size Δ r = Δ E M leads to consistent results. Smaller steps than ΔEM do not significantly improve the results. The aim of this optimization is to minimize a weighted mean square error function in logarithmic space84 (analogous to the Kullback–Leibler divergence),
(56)
This error function is a logarithmic measure of the difference between the experimental pexp and reconstructed pstp conditional probability density function. The calculation of the distance measures ξ stp , opti is performed in an analogous way to Eq. (52) (see the detailed description of the function conditional_PDF_markov). The optimization procedure systematically changes D ( 1 , 2 ) ( u r , r ) until the error function is minimized. This optimization use the function fmincon implemented in MATLAB.

In addition, this function generates a pop-up dialog box whether an example optimization is required or not. If the pop-up dialog box is denied, then this function straightaway performs the optimization for all scales and all velocity increments without plotting anything. If the pop-up dialog box is confirmed, then the conditional PDFs will be plotted using different representations [see Figs. 23(c)–23(f)] to see the differences between optimized, non-optimized, and experimental conditional PDFs. Note that if the variable scale_steps is equal to 9, then it is possible to enter any scale number from 1 up to scale number 9 (smallest and largest scale, respectively).

FIT_KM. This function plots the optimized D ( 1 , 2 ) ( u r , r ) as shown in Figs. 24(a) and 24(b). In addition, this function performs the surface fits with a linear function for D ( 1 ) ( u r , r ) and a parabolic function for D ( 2 ) ( u r , r ) to the optimized and non-optimized KMCs interpreted in the Itô convention,85 
(57)
(58)
FIG. 24.

Optimized Kramers–Moyal coefficients (a) D ( 1 ) ( u r , r ) and (b) D ( 2 ) ( u r , r ) (black filled circles) and the surface fits with a linear function for D ( 1 ) ( u r , r ) and a parabolic function for D ( 2 ) ( u r , r ) [see Eqs. (57) and (58)] with respect to scale r and velocity increment ur. The colorbars indicate the mapping of the surface fits data values into the colormap. [(c)–(f)] Coefficients d i j ( r ) of the optimized Kramers–Moyal coefficients using the surface fits with a linear function for D ( 1 ) ( u r , r ) [see Eq. (57) and (c) for d 11 ( r )] and a parabolic function for D ( 2 ) ( u r , r ) [see Eq. (58) and (d) for d 20 ( r ), (e) for d 21 ( r ) and (f) for d 22 ( r )] with respect to scale.

FIG. 24.

Optimized Kramers–Moyal coefficients (a) D ( 1 ) ( u r , r ) and (b) D ( 2 ) ( u r , r ) (black filled circles) and the surface fits with a linear function for D ( 1 ) ( u r , r ) and a parabolic function for D ( 2 ) ( u r , r ) [see Eqs. (57) and (58)] with respect to scale r and velocity increment ur. The colorbars indicate the mapping of the surface fits data values into the colormap. [(c)–(f)] Coefficients d i j ( r ) of the optimized Kramers–Moyal coefficients using the surface fits with a linear function for D ( 1 ) ( u r , r ) [see Eq. (57) and (c) for d 11 ( r )] and a parabolic function for D ( 2 ) ( u r , r ) [see Eq. (58) and (d) for d 20 ( r ), (e) for d 21 ( r ) and (f) for d 22 ( r )] with respect to scale.

Close modal
Since the comparison of the D ( 1 , 2 ) ( u r , r ) on the basis of three-dimensional figures like in Figs. 23(a) and 23(b) is very cumbersome and difficult, these surface fits are used to quantify the dependencies on the increment and the scale. The coefficients d i j ( r ) in the fits are functions of scale r of the form
(59)
Using this type of fit, it is possible to quantify intermittency while using different formalism in the literature. The constraints of this surface fits were set in a physically and mathematically meaningful way: d 11 0 , d 20 0 and d 22 0. After fitting, this function plots in Figs. 24(c)–24(f) the parameters d11, d20, d21, and d22 as a function of scale for optimized and non-optimized D ( 1 , 2 ) ( u r , r ).
The calculation leading toward the integral fluctuation theorem will be evaluated in the remaining part of the script. As already discussed in Sec. III, the cascade process is interpreted as stochastic processes evolving in scale r.71–75 Based on velocity increments u r = u ( x + r ) u ( x ) and a hierarchical or cascade like ordering of scales L > L Δ r > > λ from the integral length scale L to the Taylor length scale λ at one location x, so-called cascade trajectories
(60)
can be extracted from the data series of velocities. Independent cascade trajectories are obtained by splitting the velocity time series into intervals of integral length scales and calculating the velocity increments with respect to the point x at the beginning of these intervals. Thus, the point x is shifted by multiples of L and [ u ( · ) ] corresponds to a sequence of velocity increments from the initial scale to the final scale (path through the state space) instead of a distinct value ur. An exemplification of some cascade trajectories or “cascade paths” is illustrated in Fig. 25. It should be kept in mind that a single trajectory [ u ( · ) ] is not one specific repeated breakup of turbulent structures due to the non-linear interactions in the flow. Instead, it is assumed that these individual trajectories are probes of the spatial structures of the flow field being composed of numerous different and simultaneously evolving cascade processes. Therefore, in this investigation, a single trajectory should not be taken as one isolated large eddy evolving down-scale, rather the main assumption is that a large number of these trajectories reflect the statistics caused by the cascade process.
FIG. 25.

Schematic representation of the phenomenologically inspired turbulent energy cascade process. Shown in gray dashed lines are different cascade trajectories taken by the cascade process between the integral length scale L (initial scale) and the Taylor length scale λ (final scale). The black solid trajectory represents the preferential path through the cascade, called instanton.

FIG. 25.

Schematic representation of the phenomenologically inspired turbulent energy cascade process. Shown in gray dashed lines are different cascade trajectories taken by the cascade process between the integral length scale L (initial scale) and the Taylor length scale λ (final scale). The black solid trajectory represents the preferential path through the cascade, called instanton.

Close modal

In the spirit of non-equilibrium stochastic thermodynamics,23 it is possible to associate a total entropy variation Δ S tot with every individual cascade trajectory [ u ( · ) ].23,24,32,86,87 As the effective dynamics of the cascade trajectories through scales are stochastic in scale due to the stochasticity of the variable ur itself, this entropy defined in the following is a stochastic entropy variation associated with the evolution of the cascade process given in terms of a Langevin corresponding to the above-mentioned Fokker–Planck equation. Therefore, entropy is defined here as a statistical or rather information-theoretic quantity, that is, not equivalent to otherwise defined thermodynamic entropy of the fluid.

In this section, agreement with the integral fluctuation theorem is addressed based on the estimation of the total entropy production for each of the turbulence cascade trajectories. The validity of the integral fluctuation theorem follows if the Fokker–Planck equation correctly describes a system.23 

trajec. A pop-up dialog box is generated to select if the start and end of the cascade trajectory should be adjusted. If the pop-up dialog box is denied, then [ u ( · ) ] starts at the integral length L and end at the Taylor length λ. If this is confirmed, at the beginning of function checkFT, a pop-up dialog box is generated to specify whether the start and/or the end of the cascade trajectory should be adjusted in multiples of integral, respectively, Taylor length scale.

In addition, a pop-up dialog box (see Fig. 26) is generated to select whether the total entropy should be calculated for overlapping or independent cascade trajectories. Independent cascade trajectories are obtained by splitting the velocity time series into intervals of integral length scales. For overlapping cascade trajectories, it is no longer guaranteed that the trajectories are independent. As a result, the detected structures are analyzed several times. Thus, higher accuracy is incorrectly obtained for the derived quantities. This should be taken into account when interpreting the results.

FIG. 26.

Question dialog box that allows the user to select whether the total entropy should be calculated for overlapping or independent cascade trajectories.

FIG. 26.

Question dialog box that allows the user to select whether the total entropy should be calculated for overlapping or independent cascade trajectories.

Close modal

dr_ind. A pop-up dialog box is generated to define the separation of scales Δ r referred to the sequence from large to small scales in the cascade trajectory. The proposed value in the pop-up dialog box is equal to the Einstein–Markov length ΔEM.

data_length. A pop-up dialog box is generated to select the percentage of the data length that should be used to perform the calculation of the total entropy variation (for example, the first 20% of the data).

checkFT. The set of measured cascade trajectories results in a set of total entropy variation values Δ S tot (the same number of entropy values as the number of trajectories). This function calculates the system entropy,
(61)
medium entropy,
(62)
and the total entropy variation,
(63)
for all independent cascade trajectories. The numerical differentiation is approximated by the central difference quotient,
(64)
This numerical differentiation is performed for every individual extracted cascade trajectory [ u ( · ) ] in a sequence from large to small scales. The integration in scale is approximated by using rectangles and a mid-point rule discretization of the scale intervals; therefore, the integral takes the average of beginning and end of the discretization interval. The probabilities of start and end of the cascade trajectories, uL and u λ, can be estimated from the given data. The results depend slightly on the discretization rules and convention. However, the overall statements do not depend on it.
Plot_entropy. This function plots the empirical average e Δ S tot N of Δ S tot as a function of the number, N (sample size), of cascade trajectories [ u ( · ) ] with error bars as depicted in Fig. 28(a). Furthermore, the probability density functions of the system, medium, and total entropy are plotted in Fig. 28(b) together with the value of Δ S tot , which should be larger than 0. The integral fluctuation theorem (IFT) expresses the integral balance between the entropy-consuming ( Δ S tot < 0) and the entropy-producing ( Δ S tot > 0) cascade trajectories and states,
(65)

This function performs a pointwise optimization of D ( 1 , 2 ) ( u r , r ) with respect to the integral fluctuation theorem, which utilizes the experimental uncertainty in the determination of the Kramers–Moyal coefficients. The separation of scales/step increment (in samples) associated with the sequence from large to small scales in the (independent) cascade trajectories is set to a minimum step increment of 1 sample. Note that we use here a separation, that is, less than, or equal to the Einstein–Markov length of ΔEM. This can be justified by the fact that the stochastic equation is a continuous model for the cascade process, which bridges the corresponding coarse graining of the turbulent data with the finite Einstein–Markov length. Thereby, the stochastic model processes are not changed to processes with correlated noise, which could be a topic for further research. Furthermore, the defined separation of scales referred to the sequence from large to small scales in the cascade trajectory can be modified by the function dr_ind.

iter. This pop-up dialog box is generated to enter the maximum number of iteration, which will be performed for the optimization.

Tol_D1, tol_D2. This pop-up dialog box is generated to specify the constraints/tolerance in percent of the coefficients d i j ( r ), which will be used to perform the optimization.

OPTI_IFT_dij. This function performs the optimization of D ( 1 , 2 ) ( u r , r ) at each scale and at each value of velocity increment in order to satisfy the integral fluctuation theorem with minimum possible error and plots the optimized dij as a function of scale (see Fig. 27). The optimization procedure systematically changes D ( 1 , 2 ) ( u r , r ) until the error function,
(66)
is minimized. Within the optimization process, the user is asked, which d i j ( r ) should be optimized. This optimization uses the function fmincon implemented in MATLAB.
FIG. 27.

Coefficients d i j ( r ) of the optimized Kramers–Moyal coefficients using the surface fits with a linear function for D ( 1 ) ( u r , r ) [see Eq. (57) and (a) for d 11 ( r )] and a parabolic function for D ( 2 ) ( u r , r ) [see Eq. (58) and (b) for d 20 ( r ), (c) for d 21 ( r ) and (d) for d 22 ( r )] with respect to scale.

FIG. 27.

Coefficients d i j ( r ) of the optimized Kramers–Moyal coefficients using the surface fits with a linear function for D ( 1 ) ( u r , r ) [see Eq. (57) and (a) for d 11 ( r )] and a parabolic function for D ( 2 ) ( u r , r ) [see Eq. (58) and (b) for d 20 ( r ), (c) for d 21 ( r ) and (d) for d 22 ( r )] with respect to scale.

Close modal

Combining the functions checkFT and plot_entropy with dr_ind=1 and imposing overlapping cascade trajectories and optimized Kramers–Moyal coefficients, the results presented in Fig. 28 are obtained for the calculation of Δ S tot.

FIG. 28.

[(a) and (c)] Empirical average e Δ S tot N of Δ S tot as a function of the number N (sample size) of cascade trajectories [ u ( · ) ] with error bars. According to the integral fluctuation theorem (IFT), the empirical average has to converge to the horizontal dashed line. [(b) and (d)] Probability density function of the system Ssys, medium Smed and total entropy variation Stot. For [(a) and (b)] the optimized coefficients d i j ( r ) derived from the first step of the pointwise optimization using the short-time propagator76 and [(c) and (d)] d i j ( r ) derived from the second step of the pointwise optimization toward the integral fluctuation theorem are applied.

FIG. 28.

[(a) and (c)] Empirical average e Δ S tot N of Δ S tot as a function of the number N (sample size) of cascade trajectories [ u ( · ) ] with error bars. According to the integral fluctuation theorem (IFT), the empirical average has to converge to the horizontal dashed line. [(b) and (d)] Probability density function of the system Ssys, medium Smed and total entropy variation Stot. For [(a) and (b)] the optimized coefficients d i j ( r ) derived from the first step of the pointwise optimization using the short-time propagator76 and [(c) and (d)] d i j ( r ) derived from the second step of the pointwise optimization toward the integral fluctuation theorem are applied.

Close modal

As it can be seen in Fig. 27, this optimization is a fine-tuning of the coefficients, but by comparing Figs. 28(a)–28(d), its impact on the IFT and the PDF of the system, medium and total entropy is clearly evident. In this comparison, it must be taken into account that the separation of scales from large to small scales in the cascade trajectory is different and that overlapping cascade trajectories are investigated here.

In this section, the reconstruction of structure functions, i.e., moments of PDFs of velocity increments at different scales using the estimated KMCs is performed. The comparison of reconstructed and experimental structure functions allows addressing the validity of the estimated KMCs.

The overall scheme is: if the Markov properties are fulfilled and the cascade process can be described by a Fokker–Planck equation (i.e., Pawula theorem is fulfilled), then the knowledge of the Kramers–Moyal coefficients allow to determine:

  1. the conditional PDF given by the short time propagator (consequence of Markov property),

  2. the unconditioned PDF (consequence of the definition of conditioned probabilities),

  3. the kth order structure function (consequence of the definition),

  4. the entropy production of each cascade path (consequence of the definition),

  5. and to check the agreement with the IFT (consequence of the validity of the Fokker–Planck description).

recon_struc_pdf. In addition to verifying the consistency with the integral fluctuation theorem, the validity of the estimated drift and diffusion coefficients is subsequently tested via the reconstruction of structure functions and probability density functions of velocity increments at different scales is performed. The comparison of reconstructed and experimental structure functions allows addressing the validity of the estimated KMCs.

Initially, a pop-up dialog box (see Fig. 29) is generated to select the Kramers–Moyal coefficients to be used for the reconstruction.

FIG. 29.

Question dialog box that allows the user to select which of the Kramers–Moyal coefficients to be used for the reconstruction.

FIG. 29.

Question dialog box that allows the user to select which of the Kramers–Moyal coefficients to be used for the reconstruction.

Close modal
Using the hierarchical ordering of scales L r 0 > r 1 > > r n λ (each of which is separated by ΔEM), the kth order structure function can be reconstructed at the respective scale, for example, r0
(67)
The reconstructed unconditional PDF of velocity increments at scale r0
(68)
is estimated by integrating the reconstructed conditional PDF p stp ( u r 0 | u L ), using the short time propagator in Eq. (55), that solely depends on the Kramers–Moyal coefficients D ( 1 , 2 ).

From Eq. (68), it can be seen that the unconditional PDF at the larger scale is needed to perform the reconstruction at the next smaller scale in the hierarchical ordering. In a second pop-up dialog box (see Fig. 30), the user must select from the following two scenarios:

  • In the first scenario, Eq. (68) is solved using the unconditional PDF at the larger scale estimated directly from the experimental data at each scale independently. In the case, the optimized D ( 1 , 2 ) (conditional PDF) or optimized D ( 1 , 2 ) (IFT) is chosen, and this method often results in very good agreement, because the optimization of the drift and diffusion coefficient is performed at each scale independently. The non-optimized D ( 1 , 2 ) leads to poorer agreement in the experimental and reconstructed structure functions and PDFs.

  • In the second scenario, Eq. (68) is solved using the unconditional PDF at the larger scale from the experimental data only at scale L in the first step and an iterative process is used toward smaller scales. In other words, the entire turbulence cascade is reconstructed using only the initial solution at the largest scale from the experimental data. Typically, this method leads to very good agreement at large scales, whereas the error between experiment and reconstruction increases toward smaller scales using the iterative approach. As mentioned earlier, this is due to the fact that the drift and diffusion coefficients are estimated locally at each scale by optimization. Accordingly, in this iterative approach, minor deviations at the largest scale influence the reconstruction at the next smaller scales. This error may be reduced by performing a scale dependent optimization from large to small scales. Overall, the convergence of the statistics at large increment values, which are rarely encountered, should also be taken into account when making this comparison.

FIG. 30.

Question dialog box that allows the user to select if the reconstruction should be iterative (2. scenario) or not (1. scenario).

FIG. 30.

Question dialog box that allows the user to select if the reconstruction should be iterative (2. scenario) or not (1. scenario).

Close modal

As just described, the structure functions and the probability density functions of velocity increments are obtained from pstp. In Fig. 31, the comparison of experimental and reconstructed kth order structure function with k = 2 7 for scales λ r L are shown. In Fig. 32, the reconstructed probability density functions of velocity increments are plotted.

FIG. 31.

Comparison of experimental (black symbols) and reconstructed (red solid lines) kth order structure function for [(a) and (c)] k = 2 , 4 , 6 and [(b) and (d)] k = 3 , 5 , 7 for scales λ r L. Reconstruction using the first (top) and second (bottom) scenario.

FIG. 31.

Comparison of experimental (black symbols) and reconstructed (red solid lines) kth order structure function for [(a) and (c)] k = 2 , 4 , 6 and [(b) and (d)] k = 3 , 5 , 7 for scales λ r L. Reconstruction using the first (top) and second (bottom) scenario.

Close modal
FIG. 32.

Comparison of experimental (black symbols) and reconstructed (red solid lines) PDF of velocity increments at various scales. The velocity increment ur is normalized using σ (left) or the root mean square of the velocity increment time series at the corresponding scale rms ( u r ) (right). For better visualization, the PDFs are shifted in the vertical direction. Reconstruction using the first [(a) and (b)] and second [(c) and (d)] scenario. The black arrow indicates the direction of decreasing scales.

FIG. 32.

Comparison of experimental (black symbols) and reconstructed (red solid lines) PDF of velocity increments at various scales. The velocity increment ur is normalized using σ (left) or the root mean square of the velocity increment time series at the corresponding scale rms ( u r ) (right). For better visualization, the PDFs are shifted in the vertical direction. Reconstruction using the first [(a) and (b)] and second [(c) and (d)] scenario. The black arrow indicates the direction of decreasing scales.

Close modal

We have presented a user-friendly, open-source package, which helps to perform standard analyses of given turbulent data. In addition, the stochastic equations describing the scale-dependent cascade process in turbulent flows through Fokker–Planck equations and their application to the integral fluctuation theorem can be estimated. Altogether, we believe that this package is of interest especially to young scientists and newcomers to a stochastic approach to turbulence, as it greatly enhances the practicability and availability of both the standard and the advanced analyses introduced in this paper.

This package is designed to analyze turbulent data, but it can also be used by researchers outside of the field of turbulence for the investigation of the statistical properties of complex time series such as financial, surface height, or seismic time series to name but a few. Using this package may contribute to new insights into the complex characteristics of length/timescale-dependent processes ranging from physics to meteorology, biological systems, finance, economy, surface science, and medicine; see also Ref. 21.

The overall structure of the package is shown in Fig. 33. As can be seen from this figure, the fundamental length scales are key variables for all subsequent analyses. Therefore, a collection of the methods for the estimation of the fundamental length scales and the dissipation rate is presented. This collection does not claim to be complete, and it would be of benefit if other researchers add further methods to our collection. The direct comparison of different methods allows the user to evaluate the consistency of the methods. We include a comprehensive collection of basic analyses already established in turbulence research. The advanced analysis will result in the stochastic equations describing the scale-dependent cascade process in turbulent flows by Fokker–Planck equations. The consistency check, including the IFT, between the results of the Fokker–Planck equation and the data can be performed.

FIG. 33.

Structure of the open-source package.

FIG. 33.

Structure of the open-source package.

Close modal

As discussed in the paper, the estimation of the stochastic equations [in particular via the limit approximation in Eq. (54) (see also Fig. 22)] leads to uncertainties of the Kramers–Moyal coefficients. In this package, we address this problem by a two-step optimization as an example. If the condition for a Markov process for the cascade is given, which is the central feature of this advanced analysis, the first step of the optimization finds the best Fokker–Planck equation to reproduce the transition probabilities. In the second step, the optimization of the Kramers–Moyal coefficients with respect to the integral fluctuation theorem, which must be fulfilled if the Fokker–Planck description is correct, is performed. Note that the optimization may be extended to other quantities like structure functions. It is always important to perform a final consistency check of the optimized Fokker–Planck equation to quantify the improvements, which is indicated by the double arrows in Fig. 33.

Potential usage of the Fokker–Planck analysis includes the derivation of equations for structure functions describing the moments of probability density functions of velocity increments at different scales. Furthermore, the estimated Fokker–Planck equation is the basis to the generation of synthetic or surrogate time series. Here, the estimated drift and diffusion coefficients can be used to setup a Langevin equation, which can then be deployed for time series enhancement or forecasting. Such a type of synthetic time series can be generated by either a forward in time88 or a central in space scheme.89 

This paper only covers the discussion of turbulent flows satisfying the assumption of homogeneous, isotropic turbulence and the Taylor hypothesis of frozen turbulence. In principle, the tools of this package can be applied to turbulence data in other configurations and turbulent flows with non-ideal HIT conditions as shown for non-fully developed turbulence25 or turbulence with shear; cf. Refs. 26, 27, and 90. Future work will also be devoted to benchmark the here-proposed package on the basis of exactly parametrized synthetic turbulence data.91 

The software resulted from funded research. We acknowledge financial support by Volkswagen Foundation (VolkswagenStiftung): 96528. This work was partly funded by the German Federal Ministry for Economic Affairs and Energy in the scope of the projects EMUwind (No. 03EE2031A/C). We acknowledge the following people for helpful discussions and testing pre-version of the package: A. Abdulrazek, J. Ehrich, A. Engel, A. Girard, G. Gülker, P. G. Lind, D. Nickelsen, N. Reinke, M. Obligado, and T. Wester.

The authors have no conflicts to disclose.

André Fuchs: Writing – original draft (lead); Writing – review & editing (equal). Swapnil Kharche: Writing – review & editing (equal). Aakash Patil: Writing – review & editing (equal). Jan Friedrich: Writing – review & editing (equal). Matthias Wachter: Writing – review & editing (equal). Joachim Peinke: Writing – review & editing (equal).

The package is available as free software, under the GNU General Public License (GPL) version 3. The package [source code and standalone applications (64-bit) for Windows, macOS, and Linux] and the data that support the findings of this study are openly available in the repository on GitHub92 or the Matlab File Exchange Server.93 Support is available at the GitHub repository,92 where questions can be posted and generally receive quick responses from the authors.

Name: OPEN_FPE_IFT

Persistent identifier: GitHub

https://github.com/andre-fuchs-uni-oldenburg/OPEN_FPE_IFT

Persistent identifier: Matlab File Exchange Server

https://www.mathworks.com/matlabcentral/fileexchange/80551-open_fpe_ift

Publisher: André Fuchs

Version published: 5.0

Date published: 15/06/22

Operating system: Windows, macOS, and Linux

Programming language: MATLAB

Latin symbols
C ε

normalized turbulent kinetic energy dissipation rate

D ( k ) ( u r , r )

kth order Kramers–Moyal coefficients

E(f)

energy spectral density in the frequency domain

E(k)

energy spectral density in the wave number domain

Ekin

total kinetic energy

Fs

sampling frequency

k0

shift of the argument of the fit function, i.e., along the wave number axis

K

kurtosis

L

integral length scale

M ( k ) ( u r , r , Δ r )

kth order conditional moment

p ( u r )

probability density function of ur

p ( u r | u r )

conditioned PDF of velocity increments for a pair of two scales with r < r

r = τ u

spatial length scale

Re

Reynolds number

R e λ

Taylor Reynolds number

S

skewness

Ti

turbulence intensity

u

streamwise velocity

u

mean streamwise velocity

u ̃

streamwise velocity fluctuations

u

standard deviation

u τ ( t )

temporal velocity increment at time scale τ

u r ( t ) = u τ ( t )

spatial velocity increment at length scale r

[ u ( · ) ]

cascade trajectory

Greek symbols
Δ S med [ u ( · ) ]

medium entropy

Δ S sys [ u ( · ) ]

system entropy

Δ S tot [ u ( · ) ]

total entropy variation

ζk

scaling exponent of structure functions

ε

mean energy dissipation rate

η

Kolmogorov length scale

λ

Taylor length scale

μ

intermittency coefficient

ν

kinematic viscosity of the fluid

ρ

fluid density

σ

standard deviation of u

Latin symbols
C ε

normalized turbulent kinetic energy dissipation rate

D ( k ) ( u r , r )

kth order Kramers–Moyal coefficients

E(f)

energy spectral density in the frequency domain

E(k)

energy spectral density in the wave number domain

Ekin

total kinetic energy

Fs

sampling frequency

k0

shift of the argument of the fit function, i.e., along the wave number axis

K

kurtosis

L

integral length scale

M ( k ) ( u r , r , Δ r )

kth order conditional moment

p ( u r )

probability density function of ur

p ( u r | u r )

conditioned PDF of velocity increments for a pair of two scales with r < r

r = τ u

spatial length scale

Re

Reynolds number

R e λ

Taylor Reynolds number

S

skewness

Ti

turbulence intensity

u

streamwise velocity

u

mean streamwise velocity

u ̃

streamwise velocity fluctuations

u

standard deviation

u τ ( t )

temporal velocity increment at time scale τ

u r ( t ) = u τ ( t )

spatial velocity increment at length scale r

[ u ( · ) ]

cascade trajectory

Greek symbols
Δ S med [ u ( · ) ]

medium entropy

Δ S sys [ u ( · ) ]

system entropy

Δ S tot [ u ( · ) ]

total entropy variation

ζk

scaling exponent of structure functions

ε

mean energy dissipation rate

η

Kolmogorov length scale

λ

Taylor length scale

μ

intermittency coefficient

ν

kinematic viscosity of the fluid

ρ

fluid density

σ

standard deviation of u

1.
L. F.
Richardson
and
P.
Lynch
,
Weather Prediction by Numerical Process
(
Cambridge University Press
,
2007
).
2.
A. N.
Kolmogorov
, “
Dissipation of energy in locally isotropic turbulence
,”
Dokl. Akad. Nauk SSSR
32
,
16
(
1941
).
3.
A. N.
Kolmogorov
, “
The local structure of turbulence in incompressible viscous fluid for very large Reynolds numbers
,”
Dokl. Akad. Nauk SSSR
30
,
301
(
1941
).
4.
A. N.
Kolmogorov
, “
On degeneration of isotropic turbulence in an incompressible viscous liquid
,”
Dokl. Akad. Nauk SSSR
31
,
538
540
(
1941
).
5.
A.
Obukhov
, “
On the distribution of energy in the spectrum of turbulent flow
,”
Dokl. Akad. Nauk SSSR
5
,
453
(
1941
).
6.
A.
Obukhov
, “
Spectral energy distribution in a turbulent flow
,”
Dokl. Akad. Nauk SSSR
5
,
453
(
1941
).
7.
K. R.
Sreenivasan
and
R. A.
Antonia
, “
The phenomenology of small-scale turbulence
,”
Annu. Rev. Fluid Mech.
29
,
435
(
1997
).
8.
A. N.
Kolmogorov
, “
A refinement of previous hypotheses concerning the local structure of turbulence in a viscous incompressible fluid at high Reynolds number
,”
J. Fluid Mech.
13
,
82
(
1962
).
9.
A. M.
Oboukhov
, “
Some specific features of atmospheric turbulence
,”
J. Fluid Mech.
13
,
77
(
1962
).
10.
U.
Frisch
,
Turbulence
(
Cambridge University Press
,
1995
).
11.
G. R.
Jafari
,
S. M.
Fazeli
,
F.
Ghasemi
,
S. M. V.
Allaei
,
M. R. R.
Tabar
,
A. I.
zad
, and
G.
Kavei
, “
Stochastic analysis and regeneration of rough surfaces
,”
Phys. Rev. Lett.
91
,
226101
(
2003
).
12.
S. M.
Fazeli
,
A. H.
Shirazi
, and
G. R.
Jafari
, “
Probing rough surfaces: Markovian versus non-Markovian processes
,”
New J. Phys.
10
,
083020
(
2008
).
13.
M.
Waechter
,
F.
Riess
,
H.
Kantz
, and
J.
Peinke
, “
Stochastic analysis of surface roughness
,”
Europhys. Lett.
64
,
579
(
2003
).
14.
A.
Hadjihosseini
,
J.
Peinke
, and
N. P.
Hoffmann
, “
Stochastic analysis of ocean wave states with and without rogue waves
,”
New J. Phys.
16
,
053037
(
2014
).
15.
A.
Hadjihosseini
,
M.
Wächter
,
N. P.
Hoffmann
, and
J.
Peinke
, “
Capturing rogue waves by multi-point statistics
,”
New J. Phys.
18
,
013017
(
2016
).
16.
A.
Hadjihoseini
,
P. G.
Lind
,
N.
Mori
,
N. P.
Hoffmann
, and
J.
Peinke
, “
Rogue waves and entropy consumption
,”
Europhys. Lett.
120
,
30008
(
2018
).
17.
J.
Lengyel
,
S.
Alvanides
, and
J.
Friedrich
, “
Modelling the interdependence of spatial scales in urban systems
,”
Environ. Plann. B: Urban Anal. City Sci.
(published online
2022
).
18.
C.
Renner
,
J.
Peinke
,
R.
Friedrich
et al, “
Markov properties of high frequency exchange rate data
,”
Int. J. Theor. Appl. Finance
03
,
415
(
2000
).
19.
C.
Renner
,
J.
Peinke
, and
R.
Friedrich
, “
Evidence of Markov properties of high frequency exchange rate data
,”
Physica A
298
,
499
(
2001
).
20.
A. P.
Nawroth
,
R.
Friedrich
, and
J.
Peinke
, “
Multi-scale description and prediction of financial time series
,”
New J. Phys.
12
,
083021
(
2010
).
21.
R.
Friedrich
,
J.
Peinke
,
M.
Sahimi
, and
M. R. R.
Tabar
, “
Approaching complexity by stochastic methods: From biological systems to turbulence
,”
Phys. Rep.
506
,
87
(
2011
).
22.
J.
Peinke
,
M.
Tabar
, and
M.
Wächter
, “
The Fokker–Planck approach to complex spatiotemporal disordered systems
,”
Annu. Rev. Condens. Matter Phys.
10
,
107
(
2019
).
23.
U.
Seifert
, “
Stochastic thermodynamics, fluctuation theorems and molecular machines
,”
Rep. Prog. Phys.
75
,
126001
(
2012
).
24.
D.
Nickelsen
and
A.
Engel
, “
Probing small-scale intermittency with a fluctuation theorem
,”
Phys. Rev. Lett.
110
,
214501
(
2013
).
25.
S.
Lück
,
J.
Peinke
, and
R.
Friedrich
, “
Uniform statistical description of the transition between near and far field turbulence in a wake flow
,”
Phys. Rev. Lett.
83
,
5495
(
1999
).
26.
B.
Reisner
,
C.
Renner
,
S.
Lück
,
J.
Peinke
,
F.
Chillá
, and
R.
Friedrich
, “
A new method to characterize inhomogeneous turbulence
,” in
Fundamental Problematic Issues in Turbulence
(
Birkhäuser
,
Basel
,
1999
), pp.
361
364
.
27.
N.
Ali
,
A.
Fuchs
,
I.
Neunaber
,
J.
Peinke
, and
R. B.
Cal
, “
Multi-scale/fractal processes in the wake of a wind turbine array boundary layer
,”
J. Turbul.
20
,
93
(
2019
).
28.
S. B.
Pope
,
Turbulent Flows
(
Cambridge University Press
,
2000
).
29.
M.
Siefert
and
J.
Peinke
, “
Different cascade speeds for longitudinal and transverse velocity increments of small-scale turbulence
,”
Phys. Rev. E
70
,
015302
(
2004
).
30.
M.
Siefert
and
J.
Peinke
, “
Joint multi-scale statistics of longitudinal and transversal increments in small-scale wake turbulence
,”
J. Turbul.
7
,
N50
(
2006
).
31.
M.
Siefert
,
J.
Peinke
, and
R.
Friedrich
, “
A simple relation between longitudinal and transverse increments
,” in
Springer Proceedings in Physics
(
Springer-Verlag
,
2005
), pp.
63
66
.
32.
N.
Reinke
,
A.
Fuchs
,
D.
Nickelsen
, and
J.
Peinke
, “
On universal features of the turbulent cascade in terms of non-equilibrium thermodynamics
,”
J. Fluid Mech.
848
,
117
(
2018
).
33.
A.
Fuchs
,
N.
Reinke
,
D.
Nickelsen
, and
J.
Peinke
, in
Turbulent Cascades II
, edited by
M.
Gorokhovski
and
F. S.
Godeferd
(
Springer International Publishing
,
Cham
,
2019
), pp.
17
25
.
34.
A.
Fuchs
,
S. M. D.
Queirós
,
P. G.
Lind
,
A.
Girard
,
F.
Bouchet
,
M.
Wächter
, and
J.
Peinke
, “
Small scale structures of turbulence in terms of entropy and fluctuation theorems
,”
Phys. Rev. Fluids
5
,
034602
(
2020
).
35.
A.
Fuchs
,
M.
Obligado
,
M.
Bourgoin
,
M.
Gibert
,
P.
Mininni
, and
J.
Peinke
, “
The entropy and fluctuation theorems of inertial particles in turbulence
,” arXiv:2104.03136 (
2021
).
36.
A.
Fuchs
,
C.
Herbert
,
J.
Rolland
,
M.
Wächter
,
F.
Bouchet
, and
J.
Peinke
, “
Instantons and the path to intermittency in turbulent flows
,”
Phys. Rev. Lett.
129
,
034502
(
2022
).
37.
R.
Örlü
,
A.
Talamelli
,
J.
Peinke
, and
M.
Oberlack
, “
Progress in turbulence IX
,” in
Proceedings of the ITI Conference in Turbulence
(
Springer Nature
,
2021
), Vol.
267
.