Pooled variances
1. Pooled variances
Experimenters in general are weIl advised to subdivide lengthy measurements into a number of shorter ones, whenever feasible. This gives them the possibility — should anything have gone wrong — at least to be aware of the problem and to know roughly when it occurred. If the results do not indicate an anomaly, the partial results must be combined in such a way that the outcome characterizes the entire measuring period.
Problems with recent measurements have led us to re-consider the question of how experimental values for the variance, obtained in sequences of runs, should be used to arrive at a “best” overall estimate.
This problem occurs rather frequently and, no doubt, has been treated many times. A few years ago we looked at the special case of two samples [1]. In what follows these earlier findings are generalized to include an arbitrary number m of samples. It is shown that the contributions arising from the differences between the individual mean values can be incorporated in those terms that involve only the measured variances.
Suppose that we have to deal with a situation which is “under statistical control”. What we actually require is that the first two moments of the number of events, counted in a given time interval, show no significant trend with time.
Now consider measurements that have been performed on samples, of size , taken randomly from sorne stable population. Let the results be available in the form of me an values and variances (for a single measurement), which have been obtained by forming from the measurements the quantities
(1)
and
(2)
This allows us to form the weighted mean value
where
is the total number of measurements performed. Note that all the variances Equation (2) refer to a single measurement , not to a mean value .
Our problem is to evaluate, on the basis of the data given in Equation (1) and Equation (2), a value for the variance of a single measurement, denoted by . According to the definition of a variance, we have, considering all the measurements performed, the estimation
(3)
This expression will now be rearranged. In a first step we easily find
(4)
Since according to Equation (1)
we arrive, by using Equation (2), at the form
(5)
This is a useful identity. For the special case of samples, it follows from Equation (1) that
and similarly
Hence, the relation Equation (5) takes the form
(6)
in agreement with a result given in [1].
This procedure can be taken a step furthero To see this, let us recall the general formula for the variance of a weighted mean value , based on m measured results with statistical weights . This expression is known to be given by
(7)
Since in our case, i.e. for the measurement the sample sizes correspond to the weights we have the relation
(8)
as and .
Note, however, that the last equation is not an identity: it is a statistical relation. Thus, it does not always hold numerically. It is true on the average and requires that experimental conditions do not change.
If we take advantage of Equation (8), Equation (5) can be brought into the form
or
Hence, we arrive at the general formula
(9)
which allows us to obtain the required variance from those measured in the samples. This form is quite remarkable in that the experimental mean values , present in Equation (5), have disappeared.
For samples of equal size () we are readily led to the expression
(10)
for which even a knowledge of the sample size is no longer required.
A comparison of Equation (10) with Equation (3) shows that, at least for equal groups of measurements obtained in stable conditions, pooled estimates of both the expectation value and the variance are obtained by simply forming arithmetical means. The reader, in hindsight, should feel free to find this result either trivial or quite remarkable. It is likely, although not proven, that analogous relations hold for the central moments of higher order.
References
[1] J.W. Müller: “Moments d’échantillons superposés”, BIPM WPN-215 (1980).