|
|
(21 intermediate revisions not shown) |
Line 1: |
Line 1: |
- | == Foreword ==
| + | {{Introduction to turbulence menu}} |
- | | + | |
| Much of the study of turbulence requires statistics and stochastic processes, simply because the instanteous motions are too complicated to understand. This should not be taken to mean that the govering equations (usually the Navier-Stokes equations) are stochastic. Even simple non-linear equations can have deterministic solutions that look random. In other words, even though the solutions for a given set of initial and boundary conditions can be perfectly repeatable and predictable at a given time and point in space, it may be impossible to guess from the information at one point or time how it will behave at another (at least without solving the equations). Moreover, a slight change in the intial or boundary conditions may cause large changes in the solution at a given time and location; in particular, changes that we could not have anticipated. | | Much of the study of turbulence requires statistics and stochastic processes, simply because the instanteous motions are too complicated to understand. This should not be taken to mean that the govering equations (usually the Navier-Stokes equations) are stochastic. Even simple non-linear equations can have deterministic solutions that look random. In other words, even though the solutions for a given set of initial and boundary conditions can be perfectly repeatable and predictable at a given time and point in space, it may be impossible to guess from the information at one point or time how it will behave at another (at least without solving the equations). Moreover, a slight change in the intial or boundary conditions may cause large changes in the solution at a given time and location; in particular, changes that we could not have anticipated. |
| | | |
| + | Most of the statistical analyses of turbulent flows are based on the idea of an ensemble average in one form or another. In some ways this is rather inconvenient, since it will be obvious from the definitions that it is impossible to ever really measure such a quantity. Therefore we will spendlast part of this chapter talking about how the kind of averages we can compute from data correspond to the hypotetical ensemble average we wish we could have measured. In later chapters we shall introduce more statistical concepts as we require them. But the concepts of this chapter will be all we need to begin a discussion of the averaged equations of motion in <font color="orange">page ???</font> |
| | | |
- | Here will be introduced the simple idea of the ''ensemble average''
| + | Further details about statistical analysis in turbulence can be found in: |
- | | + | |
- | | + | |
- | | + | |
- | Most of the statistical analyses of turbulent flows are based on the idea of an ensemble average in one form or another. In some ways this is rather inconvenient, since it will be obvious from the definitions that it is impossible to ever really measure such a quantity. Therefore we will spendlast part of this chapter talking about how the kind of averages we can compute from data correspond to the hypotetical ensemble average we wish we could have measured. In later chapters we shall introduce more statistical concepts as we require them. But the concepts of this chapter will be all we need to begin a discussion of the averaged equations of motion in Chapter 3
| + | |
- | | + | |
- | == The ensemble and Ensemble Average ==
| + | |
- | | + | |
- | === The mean or ensemble average ===
| + | |
- | | + | |
- | The concept of an ''ensebmble average'' is based upon the existence of independent statistical event. For example, consider a number of inviduals who are simultaneously flipping unbiased coins. If a value of one is assigned to a head and the value of zero to a tail, then the average of the numbers generated is defined as
| + | |
- |
| + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | X_{N}=\frac{1}{N} \sum{x_{n}}
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | where our <math> n </math> th flip is denoted as <math> x_{n} </math> and <math> N </math> is the total number of flips.
| + | |
- | | + | |
- | | + | |
- | | + | |
- | | + | |
- | Now if all the coins are the same, it doesn't really matter whether we flip one coin <math> N </math> times, or <math> N </math> coins a single time. The key is that they must all be ''independent events'' - meaning the probability of achieving a head or tail in a given flip must be completely independent of what happens in all the other flips. Obviously we can't just flip one coin and count it <math> N </math> times; these cleary would not be independent events
| + | |
- | | + | |
- | Unless you had a very unusual experimental result, you probably noticed that the value of the <math> X_{10} </math>'s was also a random variable and differed from ensemble to ensemble. Also the greater the number of flips in the ensemle, thecloser you got to <math>X_{N}=1/2 </math>. Obviously the bigger <math> N </math> , the less fluctuation there is in <math> X_{N} </math>
| + | |
- | | + | |
- | Now imagine that we are trying to establish the nature of a random variable <math> x </math>. The <math>n</math>th ''realization'' of <math> x </math> is denoted as <math> x_{n}</math>. The ''ensemble average'' of <math> x </math> is denoted as <math> X </math> (or <math> \left\langle x \right\rangle </math> ), and ''is defined as''
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | | + | |
- | X = \left\langle x \right\rangle \equiv \frac{1}{N} \lim_{N \rightarrow \infty} \sum^{N}_{n=1} x_{n}
| + | |
- | | + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | Obviously it is impossible to obtain the ensemble average experimentally, since we can never an infinite number of independent realizations. The most we can ever obtain is the ariphmetic mean for the number of realizations we have. For this reason the arithmetic mean can also referred to as the ''estimator'' for the true mean ensemble average.
| + | |
- | | + | |
- | Even though the true mean (or ensemble average) is unobtainable, nonetheless, the idea is still very useful. Most importantly,we can almost always be sure the ensemble average exists, even if we can only estimate what it really is. The fact of its existence, however, does not always mean that it is easy to obtain in practice. All the theoretical deductions in this course will use this ensemble average. Obviously this will mean we have to account for these "statistical differenced" between true means and estimates when comparing our theoretical results to actual measurements or computations.
| + | |
- | | + | |
- | In general, the <math> x_{n}</math> couldbe realizations of any random variable. The <math> X </math> defined by equation 2.2 representsthe ensemble average of it. The quantity <math> X </math> is sometimes referred to as the ''expacted value '' of the random variables <math> x </math> , or even simple its ''mean''.
| + | |
- | | + | |
- | For example, the velocity vector at a given point in space and time <math> x^{\rightarrow},t </math> , in a given turbulent flow can be considered to be a random variable, say <math> u_{i} \left( x^{\rightarrow},t \right) </math>. If there were a large number of identical experiments so that the <math> u^{\left( n \right)}_{i} \left( x^{\rightarrow},t \right) </math> in each of them were identically distributed, then the ensemble average of <math> u^{\left( n \right)}_{i} \left( x^{\rightarrow},t \right) </math> would be given by
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | \left\langle u_{i} \left( x^{\rightarrow} , t \right) \right\rangle = U_{i} \left( x^{\rightarrow} , t \right) \equiv \lim_{N \rightarrow \infty} \frac{1}{N} \sum^{N}_{n=1} u^{ \left( n \right) }_{i} \left( x^{\rightarrow} , t \right)
| + | |
- | | + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | Note that this ensemble average, <math> U_{i} \left( x^{\rightarrow},t \right) </math> , will , in general, vary with independent variables <math> x^{\rightarrow} </math> and <math> t </math>. It will be seen later, that under certain conditions the ensemble average is the same as the average wich would be generated by averaging in time. Even when a time average is not meaningful, however, the ensemble average can still be defined; e.g., as in non-stationary or periodic flow. Only ensemble averages will be used in the development of the turbulence equations here unless otherwise stated.
| + | |
- | | + | |
- | === Fluctuations about the mean ===
| + | |
- | | + | |
- | It is often important to know how a random variable is distributed about the mean. For example, figure 2.1 illustrates portions of two random functions of time which have identical means, but are obviuosly members of different ensembles since the amplitudes of their fluctuations are not diswtributed the same. it is possible to distinguish between them by examining the statistical properties of the fluctuations about the mean (or simply the fluctuations) defined by:
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | X^{'}= x- X
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | It is easy to see that the average of the fluctuation is zero, i.e.,
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | \left\langle x^{'} \right\rangle = 0
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | On the other hand, the ensemble average of the square of the fluctuation is ''not'' zero. In fact, it is such an important statistical measure we give it a special name, the '''variance''', and represent it symbolically by either <math> var \left[ x \right] </math> or <math> \left\langle \left( x^{'} \right) ^{2} \right\rangle </math>
| + | |
- | The ''variance'' is defined as
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | var \left[ x \right] \equiv \left\langle \left( x^{'} \right) ^{2} \right\rangle = \left\langle \left[ x - X \right]^{2} \right\rangle
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | = \lim_{N\rightarrow \infty} \frac{1}{N} \sum^{N}_{n=1} \left[ x_{n} - X \right]^{2}
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | Note that the variance, like the ensemble average itself, can never really be measured, since it would require an infinite number of members of the ensemble.
| + | |
- | | + | |
- | | + | |
- | It is straightforward to show from eqaution 2.2 that the variance in equation 2.6 can be written as
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math> | + | |
- | var \left[ x \right] = \left\langle x^{2} \right\rangle - X^{2}
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | Thus the variance is the ''second-moment'' minus the square of the ''first-moment'' (or mean). In this naming convention, the ensemble mean is the ''first moment''.
| + | |
- | | + | |
- | | + | |
- | The variance can also referred to as the ''second central moment of x''. The word central implies that the mean has been subtructed off before squaring and averaging. The reasons for this will be clear below. If two random variables are identically distributed, then they must have the same mean and variance.
| + | |
- | | + | |
- | The variance is closely related to another statistical quantity called the ''stardard deviation'' or root mean square (''rms'') value of the random variable <math> x </math> , which is denoted by the simbol, <math> \sigma_{x} </math>. Thus,
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | \sigma_{x} \equiv \left( var \left[ x \right] \right)^{1/2}
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | or
| + | |
- | | + | |
- | <table width="100%"><tr><td>
| + | |
- | :<math>
| + | |
- | \sigma^{2}_{x} = var \left[ x \right]
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
- | | + | |
- | ----------------------------------------
| + | |
- | | + | |
- | | + | |
- | === Higher moments ===
| + | |
- | | + | |
- | Figure 2.2 illustrates tow random variables of time which have the same mean and also the same variances, but clearly they are still quite different. It is useful, therefore, to define higher moments of the distribution to assist in distinguishing these differences.
| + | |
| | | |
- | The <math> m </math>-th moment of the random variable is defined as
| + | *[[Introduction to turbulence/Statistical analysis/Ensemble average|Ensemble average]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Ensemble average#Mean or ensemble average|Mean or ensemble average]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Ensemble average#Fluctuations about the mean|Fluctuations about the mean]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Ensemble average#Higher moments|Higher moments]] |
| + | *[[Introduction to turbulence/Statistical analysis/Probability|Probability]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Probability#Histogram and probability density function|Histogram and probability density function]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Probability#Probability distribution|Probability distribution]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Probability#Gaussian (or normal) distributions|Gaussian (or normal) distributions]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Probability#Skewness and kurtosis|Skewness and kurtosis]] |
| + | * [[Introduction to turbulence/Statistical analysis/Multivariate random variables#Multivariate random vaiables|Multivariate random vaiables]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Multivariate random variables#Joint pdfs and joint moments|Joint pdfs and joint moments]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Multivariate random variables#The bi-variate normal (or Gaussian) distribution|The bi-variate normal (or Gaussian) distribution]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Multivariate random variables#Statistical independence and lack of correlation| Statistical independence and lack of correlation]] |
| + | * [[Introduction to turbulence/Statistical analysis/Estimation from a finite number of realizations|Estimation from a finite number of realizations]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Estimation from a finite number of realizations#Estimators for averaged quantities|Estimators for averaged quantities]] |
| + | ** [[Introduction to turbulence/Statistical analysis/Estimation from a finite number of realizations#Bias and convergence of estimators| Bias and convergence of estimators]] |
| + | * [[Introduction to turbulence/Statistical analysis/Generalization to the estimator of any quantity|Generalization to the estimator of any quantity]] |
| | | |
- | <table width="100%"><tr><td>
| + | {{Turbulence credit wkgeorge}} |
- | :<math>
| + | |
- | \left\langle x^{m} \right\rangle = \lim_{N \rightarrow \infty} \frac{1}{N} \sum^{N}_{n=1} x^{m}_{n}
| + | |
- | </math>
| + | |
- | </td><td width="5%">(2)</td></tr></table>
| + | |
| | | |
- | It is usually more
| + | {{Chapter navigation|Nature of turbulence|Reynolds averaged equations}} |
Much of the study of turbulence requires statistics and stochastic processes, simply because the instanteous motions are too complicated to understand. This should not be taken to mean that the govering equations (usually the Navier-Stokes equations) are stochastic. Even simple non-linear equations can have deterministic solutions that look random. In other words, even though the solutions for a given set of initial and boundary conditions can be perfectly repeatable and predictable at a given time and point in space, it may be impossible to guess from the information at one point or time how it will behave at another (at least without solving the equations). Moreover, a slight change in the intial or boundary conditions may cause large changes in the solution at a given time and location; in particular, changes that we could not have anticipated.
Most of the statistical analyses of turbulent flows are based on the idea of an ensemble average in one form or another. In some ways this is rather inconvenient, since it will be obvious from the definitions that it is impossible to ever really measure such a quantity. Therefore we will spendlast part of this chapter talking about how the kind of averages we can compute from data correspond to the hypotetical ensemble average we wish we could have measured. In later chapters we shall introduce more statistical concepts as we require them. But the concepts of this chapter will be all we need to begin a discussion of the averaged equations of motion in page ???
This text was based on "Lectures in Turbulence for the 21st Century" by Professor William K. George, Professor of Turbulence, Chalmers University of Technology, Gothenburg, Sweden.