The numerical value of the likelihood function at (*) can, in principle, be used as a check on whether one is using the correct type of function for f (; x). If one is using the wrong f, the likelihood function will be lower in height and of greater width. In principle, one can calculate, using direct probability, the distribution of (*) assuming a particular true f (0, x). Then the probability of getting an (*) smaller than the value observed would be a useful indication of whether the wrong type of function for f had been used. If for a particular experiment one got the answer that there was one chance in 104 of getting such a low value of (*), one would seriously question either the experiment or the function f (;x) that was used.
In practice, the determination of the distribution of (*) is usually an impossibly difficult numerical integration in N-dimensional space. However, in the special case of the least-square problem, the integration limits turn out to be the radius vector in p-dimensional space. In this case we use the distribution of S(*) rather than of (*). We shall first consider the distribution of S(0). According to Eqs. (23) and (24) the probability element is
Note that S = 2, where is the magnitude of the radius vector in p-dimensional space. The volume of a p-dimensional sphere is U p. The volume element in this space is then
Thus
The normalization is obtained by integrating from S = 0 to S = .
(30a) |
where S S(0).
This distribution is the well-known 2 distribution with p degrees of freedom. 2 tables of
for several degrees of freedom are commonly available - see Appendix V for plots of the above integral.
From the definition of S (Eq. (24)) it is obvious that 0 = p. One can show, using Eq. (29) that = 2p. Hence, one should be suspicious if his experimental result gives an S-value much greater than
Usually is not known. In such a case one is interested in the distribution of
Fortunately, this distribution is also quite simple. It is merely the 2 distribution of (p - M) degrees of freedom, where p is the number of experimental points, and M is the number of parameters solved for. Thus we haved
(31) |
Since the derivation of Eq. (31) is somewhat lengthy, it is given in Appendix II.
Example 8
Determine the 2 probability of the solution to Example 6.
According to the 2 table for one degree of freedom the probability of getting S* > 0.674 is 0.41. Thus the experimental data are quite consistent with the assumed theoretical shape of
Example 9 Combining Experiments
Two different laboratories have measured the lifetime of the K10 to be (1.00 ± 0.01) × 10-10 sec and (1.04 ± 0.02) × 1010 sec respectively. Are these results really inconsistent?
According to Eq. (6) the weighted mean is * = 1.008 × 10-10 sec. (This is also the least squares solution for KO.
Thus
According to the 2 table for one degree of freedom, the probability of getting S* > 3.2 is 0.074. Therefore, according to statistics, two measurements of the same quantity should be at least this far apart 7.4% of the time.