Laserfiche WebLink
<br />SOME STATISTICAL TOOLS IN HYDROLOGY <br /> <br />. <br /> <br />The interval in brackets is called a confidence <br />interval and the extremes are called confidence <br />limits. Note that the above relation holds only <br />if we use cr instead of 8. If we use 8, then l-e <br />is a function of the S8ID.ple size, and the <br />appropriate probability statement is <br /> <br />P[(X -t8/{N)<I'< (X + t8/.JN)J=I-e=O,68, <br /> <br />where Student's t is 1.09 for 10 degrees of <br />freedom, for example. The width of the con- <br />fidence interval increases as the level of signifi- <br />cance decreases, For eX81llple, the 95-percent <br />confidence limits (e=O,05) are <br /> <br />(X -2,Z38/{N)<I'<(X +2.238NN), <br /> <br />. <br /> <br />where 2,23 is from the t table for 10 degrees of <br />freedom, <br />The confidence interval described in the <br />probability statement is a random interval, <br />not a specific one. The probability statement <br />(for e=0.05) means that 95 percent of a large <br />number of intervals similarly obtained wonld <br />include the true mean, This probability statement <br />cannot be extended to one specific interval <br />because a specific interval either contains the <br />true mean or it does not and the probability <br />is either one or zero, The true mean is not a <br />variable, it is unique. <br />But we are interested in making a probability <br />statement about one specific interval, We may <br />say that the probability of our obtaining a <br />ran dom interval which includes the true mean <br />is 0,95, or that we have 95-percent confidence <br />that the interval obtained includes the true <br />mean. Ordinarily in hydrologic reports it is <br />only necessary to state the computed confi- <br />dence interval and its level, not to interpret <br />the meaning. See Mood (1950, p. 221-222) for <br />a precise statement of the interpretation of a <br />confidence interval. <br />Using the above theory, from a random <br />sample, we can compute an estimate of the <br />population mean and a measure of its relia- <br />bility. This is an example of statistical inference, <br />Returning to the sampling theory, consider <br />the distribution of variances of samples of size <br />1\ from a normal distribution. This distribu- <br />tion is not centered around cr' but is to the left <br />of it, as shown by the upper graph of figure 6, <br />Therefore 8' is known as a biased estimator of <br /> <br />" <br /> <br />. <br /> <br />5 <br /> <br />.r, It can be made unbiased by multiplying it <br />by N/(N-l) as shown in the lower sketch of <br />figure 6. The standard deviation of the sampling <br />distribution of 8'(^/N-l) can also be <br />computed, <br /> <br /> <br />S2 u2 <br /> <br /> <br />S2(N/N-l ) <br />u2 <br /> <br />Figure 6.-Distribution of variances of samples. <br /> <br />A further use of inference is in testing hypoth- <br />eses. One example will be given. Suppose we set <br />up the null hypothesis, H" that the mean of a <br />population is zero; that is, we hypothesize that <br />there is no difference statistically between the <br />mean and zero. This null hypothesis is written <br /> <br />H,:I'=O. <br /> <br />We draw a sample from this population and <br />compute the statistics of the sampling distribu- <br />tion of the mean. We need some estimate of the <br />hypothetical sampling distribution of means, <br />and so we define it os normal with mean zero <br />and standard deviation equal to 81 v'N os com- <br />puted from the sample (fig. 7). <br />Now if X (as computed from the sample) lies <br />within one standard deviation of zero, we would <br />conclude that there is no basis for doubting <br />the hypothesis. If, on the other hand, X were <br />two or three standard deviations away from <br />zero, we would conclude that it is unlikely that <br />the mean of the population is zero. For this <br />