The problem of finding confidence intervals based on data from several independent studies or experiments is considered. A general method of finding confidence intervals by inverting a combined test is proposed. The combined tests considered are the Fisher test, the weighted inverse normal test, the inverse chi-square test and the inverse Cauchy test. The method is illustrated for finding confidence intervals for a common mean of several normal populations, common correlation coefficient of several bivariate normal populations, common coefficient of variation, common mean of several lognormal populations, and for a common mean of several gamma populations. For each case, the confidence intervals based on the combined tests are compared with the other available approximate confidence intervals with respect to coverage probability and precision. R functions to compute all confidence intervals are provided in a supplementary file. The methods are illustrated using several practical examples.
There are many areas of the medical, behavioral, and social sciences where we have data or summary results available from several independent studies, experiments, or laboratories all addressing the same objective. Major medical hypotheses are typically studied more than once, often by different clinical research teams in different locations. Analysis of results from different sources or results of earlier research to derive conclusions about the body of research is known as meta-analysis. Usually, the study is based on randomized, controlled clinical trials. Such problems also arise, for example, when two or more independent agencies are involved in estimating the effect of a new drug, when several measuring instruments are used to measure the products produced by the same production process to assess the average quality, or when different laboratories are used to measure the amount of toxic waste in a river. There is a need to combine the results from independent sources or experiments to make inference on the common objective of interest. Egger and Smith1 have noted that “A single study often cannot detect or exclude with certainty a modest, albeit clinically relevant, difference in the effects of two treatments. A trial may thus show no significant treatment effect when in reality such an effect exists.” Regarding multivariate meta-analysis, Jackson et al.2 have noted that effects are often multivariate rather than univariate and medical studies often examine multiple, and correlated, outcomes of interest to the meta-analyst.
In this article, we address the problem of testing and interval estimating a common parameter of interest based on independent data sets collected from different sources. Common parameters could be the mean of several populations, coefficient of variation (CV) of several measurements obtained using different instruments, common coefficient of correlation of several bivariate normal populations, and so on. Examples with data or summary statistics for testing/estimating a common mean of several populations are given in many articles. Meier3 has provided an example where data collected from four experiments that were used to estimate the mean percentage of albumin in the plasma protein of normal human subjects. Eberhardt et al.4 provided three examples each of which is concerned with estimating a chemical substance in non-fat milk powder by combining the results of different analytical methods. Skinner5 presented examples related to clinical trials and determining the density of nitrogen using different laboratory methods. In these situations, the variances are typically unequal due to the variation in different methods or measuring instruments and the objective is to estimate the common mean of measurements from different sources. These authors have provided solutions assuming normal models with the same mean but possibly different variances. The common mean problem has also been addressed assuming lognormal distributions (e.g. Tian and Wu6 and Krishnamoorthy and Oral7), and gamma distributions.8
Another common parameter of several populations that is of interest is the CV, which is a commonly used measure of variability in experimental studies. Smaller values of CV indicate higher precision of measurements or of the experimental results. The CV is a preferred measure of variability because, unlike the variance, it does not depend on units of measurement. If it is reasonable to assume that a common CV in various independent experimental results, then it is of interest to test or interval estimate the common CV. Tian9 and Forkman10 have considered the problem of estimating the common CV of several normal populations. This seems to be an important problem with applications as Tian’s research was attracted by many researchers and practitioners. Both papers present some approximate confidence intervals for the common CV of several normal populations.
Combining the results from independent sources for estimating a common correlation coefficient of several bivariate normal populations has received considerable attention in the literature; see Donner and Rosner,11 Tian and Wilding12 and Liu and Xu.13 Donner and Rosner have proposed an approximate interval estimation method based on a linear function of independent Fisher statistics. Tian and Wilding12 have described a generalized variable (GV) approach to find a confidence interval (CI) for a common correlation coefficient and Liu and Xu13 have proposed a solution for the problem based on confidence distribution.
In general, if independent tests were carried out in different locations for a common purpose of interest, then a popular method of combining the test results is Fisher’s14 method of combining -values of the independent tests to arrive at a single test; see Mathew et al.15 and the references therein. Fisher’s combination of -values has a chi-square distribution, and so it is easy to determine the -value or critical region of the combined test. Stouffer et al.16 proposed an alternative combined test that is based on independent normal -scores. Liptak17 has suggested to use a weighted average of the -scores. Folks18 and other authors have found that Fisher’s test performed well over a wide range of parameter values and that the Liptak-Stouffer inverse-normal method performed poorly. However, for the normal case, Whitlock19 has noted that the Liptak-Stouffer test with weights proportional to the degrees of freedom (df) is better than the Fisher test. Liu and Xie20 proposed Cauchy combination test that combines independent -values using inverse of Cauchy distributions. This Cauchy combined test tackles some dependency structures of the proposed test statistic in their paper.20 Recently, Krishnamoorthy et al.21 have considered a new test based on chi-square scores and compared several combined tests for a common mean of normal populations, common correlation coefficient, common CV and for a common mean of several lognormal populations. On the basis of extensive simulation study, they found that no test is uniformly better than the other tests for all problems. For some specific problems, they noted that some tests perform better than others over a wide range of parameter space.
Even though there are several papers that have addressed the testing problems of combining independent tests for various purposes, only a few papers considered the problems of interval estimating a common parameter of interest in various setups. For estimating a common mean of several normal populations, Fairweather22 and Jordan and Krishnamoorthy23 have provided CIs based on pivotal-based approaches. Tian and her co-authors have proposed CIs for a common correlation coefficient, a common CV and a common mean of several lognormal populations; see Tian,9 Tian and Wilding12 and Tian and Wu.6 These authors have proposed approximate CIs based on the GV approach by Weerahandi.24 Donner and Rosner11 have developed an approximate CI for a common correlation coefficient of several bivariate normal populations on the basis of a combination of Fisher’s -statistics.
Most of the CIs proposed in the literature are approximate and not simple to compute. Our point here is that even though all available combined tests are exact in the frequentist sense, CIs that can be obtained by inverting the exact combined tests are not proposed in the literature. The purpose of this article is to describe interval estimation procedures based on several combined tests and make them readily available to researchers and practitioners of statistics. We propose a numerical approach to find a CI for a common parameter of interest by inverting a combined test and illustrate the procedure for several problems. In the following section, we describe a numerical approach to find a CI for a common parameter of interest by inverting a combined test. The method is applied to find confidence intervals for (a) a common mean of several normal populations (Section 3), (b) a common CV (Section 4), (c) a common mean of several lognormal populations (Section 5), (d) a common correlation coefficient of several bivariate normal populations (Section 6, (e) a common mean of several gamma populations (Section 7). For each problem, we considered CIs based on Fisher’s combined test, weighted inverse normal test (Liptak-Stouffer test), inverse Cauchy test20 and inverse test. Furthermore, CIs based on these combined tests are also compared with other available approximate CIs in terms of coverage probability and precision. An illustrative example with real data are given for each of the estimation problems. To help practitioners compute these CIs, we provide R functions in a supplementary file. Some concluding remarks are given in Section 8. R functions to compute CIs using the Fisher, weighted inverse normal, inverse Cauchy, and inverse chi-square are provided in a supplementary file.
Confidence intervals from combined tests
Suppose that there are independent populations with a common parametric function of interest. Let us denote the common parametric function by , where is a known function and , , , are unknown parameters. Assume that a sample of size is available from the th population, . Assume also that the samples are independent. Consider testing
where is a specified value. Let denote the -value of a test for the above hypotheses based on the th sample, . The -values are based on independent samples, so they are independent uniform random variables. To arrive at a combined test based on all samples, some transformed -values are combined to obtain a single test statistic so that its null distribution does not depend on any unknown parameter. In Table 1, we provide some well-known combined tests, namely, Fisher’s test, inverse normal test, weighted inverse normal test, inverse Cauchy test and the inverse test. The notations used in the table are as follows:
standard normal distribution function
standard normal random variable
and ,
is the -value based on the th sample, .
is an observed value of
the standard Cauchy random variable with the distribution function
the 100 percentile of the chi-squared distribution with df = .
At the level , a combined test rejects the null hypothesis in favor of the alternative hypothesis if the -value is less than . The value of for which the -value of a combined test is equal to is a lower confidence limit for . Similarly, by inverting a combined test for vs. , we can find a upper confidence limit for . The interval is a two-sided CI for .
In general, a CI from a combined test can be obtained only by numerically. The following algorithm may be used to find a CI for a common parameter by inverting the Fisher combined test. Confidence intervals based on other combined tests can be obtained similarly.
Algorithm 1
Compute the -values ’s for testing vs. based on the sample of size from the th population, .
Compute the combined statistic .
Set the function .
Solve for .
The root of the equation in the preceding step is a lower confidence limit for .
To compute an upper confidence limit for , we use the -values ’s for testing vs. in Step 1 and then follow Steps 2 – 5. The % one-sided confidence limits form a 100% two-sided CI for the parameter . For a graphical illustration of Algorithm 1 for finding CIs for a common mean of several normal distributions, see Figure 1.
-values of the Fisher test as a function of for the data in Table 3; the horizontal line is line.
Common mean of several normal distributions
Let denote the (mean, variance) based on a sample of size from a distribution, . The variance is defined with the divisor ,
Confidence intervals
We first describe some available exact and approximate CIs for a common mean of several normal populations.
Confidence interval based on a weighted F statistics
Jordan and Krishnamoorthy23 have proposed a CI by inverting a linear combination of -test statistics for the normal mean To describe this CI, let denote the random variable with 1 and degrees of freedoms, and let denote the 100 percentile of . Let and Then
is an exact CI for . The above CI is centered at the Graybill-Deal25 estimator
of if for all . For , the percentile can also be estimated using Monte Carlo simulation or approximated using the method in Jordan and Krishnamoorthy.23 Since the variance of which is defined when the df or equivalently, for , the above CI is defined only when all sample sizes are at least 6.
Fiducial confidence intervals
Let be an observed value of , . The fiducial variable for based on the th sample is given by where ’s are independent random variables with df ; see Section 2.1 of Krishnamoorthy and Mathew.26 Let so that A combined fiducial variable for is given by
For a given set of statistics , the lower and upper 100 percentiles of form a fiducial CI for . To find this fiducial CI, we need to find the percentiles of , which can be estimated using Monte Carlo simulation. The percentiles of can also be approximated using the normal-based approximation27 as where is the quantile of the distribution with df = . Using these approximate percentiles, a fiducial CI for can be expressed as
It can be readily verified that the above CI simplifies to the usual interval when .
Remark 1. Krishnamoorthy and Lu28 have proposed a generalized CI for a common mean. Their CI is essentially based the fiducial approach and not in closed-form. It requires Monte Carlo simulation to estimate whereas the CI in (4) is simple and it requires only percentiles to compute.
Confidence intervals based on a combined test
To deduce a CI from a combined test given in Table 1, consider testing vs. , where is a common mean of independent normal populations. The -value of the usual -test based on the th sample is given by , where denotes the random variable with df , . Using these independent -values in any of the combined tests in Table 1 and Algorithm 1, a lower confidence limit for the common mean can be computed. Similarly, by inverting a left-tailed combined test, an upper confidence limit for can be obtained.
Normal: Coverage and precision studies
The individual tests based on independent samples are exact and so do the combined tests. Therefore, the CIs based on any combined test in Table 1 are exact. To compare the CIs based on combined tests in terms of precision and to judge the accuracy of the approximate fiducial CIs and the CIs based on the weighted statistics, we estimated the coverage probabilities and expected widths of all CIs and reported them in Table 2. Monte Carlo estimates of the coverage probabilities and expected widths are based on 10,000 simulations runs. These estimates were obtained for the case of and 4. The CIs based on the inverse normal combined test and approximate fiducial method are not included in the simulation study because our preliminary study (not reported here) indicated that the inverse normal combined test is inferior to those based on the weighted inverse normal test. Furthermore, the closed-form approximate fiducial CIs are liberal when sample sizes are small. They are satisfactory in terms of coverage and precision if all sample sizes are 30 or more. For such sample sizes, the approximate fiducial CIs are satisfactory in terms of coverage probability and are expected to be slightly shorter than the other CIs.
Coverage probabilities and expected widths (in parentheses) of 95% CIs for a common mean of several normal populations.
Weighted F
Fisher
W Inv Norm
Inv
Inv Cauchy
(6,6)
0.1
0.948(0.237)
0.946(0.242)
0.947(0.313)
0.948(0.254)
0.953(0.172)
0.3
0.952(0.683)
0.953(0.630)
0.950(0.656)
0.953(0.630)
0.952(0.490)
0.5
0.953(1.04)
0.951(0.908)
0.949(0.882)
0.950(0.890)
0.951(0.764)
0.8
0.950(1.43)
0.950(1.21)
0.950(1.13)
0.951(1.17)
0.950(1.06)
1
0.949(1.61)
0.948(1.35)
0.948(1.27)
0.948(1.31)
0.951(1.19)
(16,6)
0.1
0.951(0.356)
0.950(0.231)
0.951(0.540)
0.949(0.296)
0.948(0.237)
0.3
0.951(0.757)
0.949(0.537)
0.950(0.710)
0.950(0.573)
0.951(0.674)
0.5
0.951(0.909)
0.950(0.719)
0.950(0.798)
0.950(0.720)
0.952(1.05)
0.8
0.951(0.992)
0.950(0.876)
0.950(0.875)
0.950(0.847)
0.949(1.46)
1
0.952(1.02)
0.950(0.943)
0.951(0.911)
0.950(0.902)
0.954(1.67)
(6,16)
0.1
0.950(0.106)
0.950(0.122)
0.950(0.109)
0.950(0.116)
0.949(0.232)
0.3
0.950(0.318)
0.950(0.348)
0.950(0.313)
0.950(0.328)
0.950(0.613)
0.5
0.950(0.526)
0.950(0.547)
0.950(0.498)
0.950(0.515)
0.950(0.849)
0.8
0.949(0.825)
0.950(0.800)
0.950(0.753)
0.950(0.759)
0.950(1.01)
1
0.950(1.02)
0.951(0.943)
0.951(0.911)
0.951(0.902)
0.946(1.06)
(12,14)
0.1
0.951(0.128)
0.950(0.131)
0.951(0.147)
0.950(0.137)
0.952(0.128)
0.3
0.950(0.371)
0.950(0.359)
0.950(0.365)
0.950(0.359)
0.954(0.375)
0.5
0.951(0.582)
0.950(0.541)
0.951(0.526)
0.951(0.530)
0.950(0.593)
0.8
0.951(0.828)
0.950(0.745)
0.950(0.711)
0.950(0.721)
0.950(0.854)
1
0.951(0.948)
0.950(0.845)
0.950(0.808)
0.950(0.818)
0.952(0.984)
(20,20)
0.1
0.954(0.105)
0.951(0.106)
0.951(0.126)
0.952(0.114)
0.947(0.104)
0.3
0.950(0.301)
0.950(0.288)
0.951(0.304)
0.951(0.293)
0.957(0.301)
0.5
0.952(0.468)
0.951(0.433)
0.949(0.431)
0.949(0.427)
0.952(0.474)
0.8
0.951(0.654)
0.950(0.591)
0.951(0.570)
0.952(0.574)
0.949(0.673)
1
0.949(0.736)
0.949(0.665)
0.949(0.640)
0.948(0.645)
0.949(0.760)
Weighted F
Fisher
W Inv Norm
Inv
Inv Cauchy
(20,15,10,6)
(0.1,0.2,0.1)
0.947(0.119)
0.951(0.099)
0.952(0.127)
0.951(0.104)
0.948(0.120)
(0.3,0.2,0.1)
0.951(0.226)
0.947(0.153)
0.948(0.213)
0.949(0.165)
0.952(0.202)
(0.3,0.4,0.6)
0.948(0.342)
0.947(0.299)
0.952(0.314)
0.950(0.292)
0.947(0.351)
(0.6,0.7,0.6)
0.950(0.560)
0.951(0.445)
0.952(0.444)
0.952(0.426)
0.953(0.603)
(0.6,0.7,0.1)
0.951(0.358)
0.949(0.226)
0.949(0.359)
0.949(0.263)
0.950(0.259)
(0.6,0.7,0.9)
0.948(0.566)
0.951(0.471)
0.949(0.458)
0.951(0.447)
0.947(0.620)
(0.1,0.9,0.9)
0.957(0.134)
0.951(0.143)
0.953(0.180)
0.953(0.152)
0.954(0.137)
(1,1,1)
0.950(0.750)
0.951(0.617)
0.950(0.581)
0.950(0.579)
0.950(0.818)
(6,10,15,20)
(0.1,0.2,0.1)
0.950(0.088)
0.949(0.081)
0.949(0.074)
0.949(0.076)
0.950(0.097)
(0.3,0.2,0.1)
0.951(0.099)
0.950(0.098)
0.949(0.086)
0.949(0.092)
0.953(0.108)
(0.3,0.4,0.6)
0.950(0.330)
0.950(0.280)
0.949(0.274)
0.949(0.266)
0.950(0.363)
(0.6,0.7,0.6)
0.951(0.477)
0.950(0.408)
0.949(0.375)
0.949(0.380)
0.954(0.529)
(0.6,0.7,0.1)
0.951(0.108)
0.951(0.118)
0.949(0.111)
0.949(0.115)
0.951(0.115)
(0.6,0.7,0.9)
0.950(0.566)
0.950(0.471)
0.950(0.457)
0.950(0.446)
0.952(0.628)
(0.1,0.9,0.9)
0.950(0.133)
0.951(0.143)
0.951(0.180)
0.951(0.152)
0.947(0.178)
(1,1,1)
0.950(0.748)
0.951(0.617)
0.951(0.580)
0.951(0.578)
0.952(0.824)
(20,22,21,20)
(0.1,0.2,0.1)
0.948(0.076)
0.949(0.066)
0.951(0.067)
0.949(0.066)
0.951(0.080)
(0.3,0.2,0.1)
0.949(0.100)
0.948(0.089)
0.950(0.094)
0.950(0.090)
0.951(0.103)
(0.3,0.4,0.6)
0.944(0.247)
0.947(0.209)
0.949(0.203)
0.948(0.203)
0.954(0.263)
(0.6,0.7,0.6)
0.947(0.392)
0.947(0.321)
0.948(0.304)
0.947(0.306)
0.947(0.422)
(0.6,0.7,0.1)
0.949(0.117)
0.950(0.113)
0.948(0.140)
0.948(0.124)
0.950(0.114)
(0.6,0.7,0.9)
0.950(0.429)
0.951(0.352)
0.948(0.334)
0.950(0.336)
0.956(0.465)
(0.1,0.9,0.9)
0.950(0.112)
0.952(0.112)
0.954(0.137)
0.954(0.123)
0.947(0.108)
(1,1,1)
0.949(0.579)
0.948(0.468)
0.949(0.439)
0.950(0.444)
0.948(0.624)
Selenium content in non-fat milk powder using four methods.
Methods
Sample size
Mean
Variance
1. Atomic absorption Spectrometry
8
105.00
85.711
2. Neutron activation: Instrumental
12
109.75
20.748
3. Neutron activation: Radiochemical
14
109.50
2.729
4. Isotope dilution mass spectrometry
8
113.25
33.640
We observe from the reported results in Table 2 that the CIs based on the weighted statistics control the coverage probabilities very close to the nominal level for all cases. However, the expected widths of these CIs are larger than those based on combined tests. Comparisons of other CIs based on the Fisher, weighted inverse normal, inverse Cauchy, and inverse combined tests indicate that there is no clear-cut winner among them. For example, the CIs based on the inverse test and the weighted inverse normal tests are comparable and they are better than those based on the Fisher test in most cases. However, there are cases where the CIs based on the Fisher test are better than those based on the inverse test. See the results for and and and . Similarly, we see that the CIs based on the weighted inverse normal are slightly better than those based on the inverse test. For example, see the results for and and (1,.6,.7,.9). The inverse Cauchy CIs are, in general, wider than the inverse CIs for most cases. There are many cases where the inverse CIs are 30%–50% narrower than the corresponding inverse Cauchy CIs.
On an overall basis, the inverse CIs followed by the CIs based on the weighted inverse normal test are preferable to all other CIs. We can also recommend the approximate fiducial CIs when all sample sizes are 30 or more because they are not only simple to compute but also satisfactory in terms of coverage probability and better than others in terms of precision for such sample sizes.
Normal: Example
Four different analytical methods were used to estimate the selenium in non-fat milk powder and the mean and variance are reported in Table 3. The data are taken from Eberhardt et al.4 Application of Bartlett’s test by Jordan and Krishnamoorthy23 has shown that the variances are significantly different, and so the -test for the mean based on the pooled data is not appropriate.
The 90, 95 and 99% CIs based on different methods are reported in Table 4. All CIs are in agreement. The inverse , Fisher and the weighted inverse normal methods produced practically the same confidence interval. As indicated by the simulation studies in the preceding section, the inverse Cauchy intervals are wider than all other CIs for this example. Finding the 95% CI using the Fisher method is described graphically in Figure 1. The R function CI.normal.Cmean(n, xb, sq, cl,method) provided in the supplementary file was used to compute CIs by the Fisher, weighted inverse normal, inverse Cauchy and inverse methods.
Confidence intervals for the mean selenium content in non-fat milk powder.
90%
95%
99%
Method
CI
Width
CI
Width
CI
Width
Weighted F
(108.65, 110.48)
1.83
(108.45, 110.67)
2.22
(108.06, 111.06)
3.00
Apprx. Fid
(108.88, 110.33)
1.45
(108.72, 110.49)
1.77
(108.36, 110.84)
2.48
Fisher
(108.78, 110.47)
1.69
(108.60, 110.65)
2.05
(108.24, 111.02)
2.78
W Inv Normal
(108.80, 110.44)
1.64
(108.63, 110.62)
1.99
(108.27, 111.00)
2.73
Inverse
(108.80, 110.45)
1.65
(108.63, 110.63)
2.00
(108.27, 111.01)
2.74
Inv Cauchy
(108.52, 110.59)
2.07
(108.30, 110.76)
2.46
(107.87, 111.14)
3.27
Common coefficient of variation
The CV is defined as the ratio of the standard deviation to the mean, , and is defined only for distributions of positive random variables. For a normal population, the ratio of the mean to the standard deviation has to be on the order of three or more, for the probability of a negative value is negligible. This means that , or the CV must be at most in practical situations where the CV is a suitable measure of variability.29
A Test for a CV
Consider a sample from a population. The population CV is defined as Let be the (mean, variance) based on the sample and let be an observed value of . The sample CV is given by . To derive a test for based on the sample, we first note that where denotes the noncentral random variable with df and the noncentrality parameter . Let and . Hypothesis test or CI for can be readily obtained from the one for . In particular, for testing vs. , the -value is given by
where is an observed value of . A CI for can be obtained by inverting one-sided tests.
Confidence intervals for based on a combined test
Let denote the (mean, variance) based on the th sample, and be an observed value of , . Consider testing using the th sample. For an observed value , the null hypothesis will be rejected whenever the -value
and it is an exact level test. A combined test for can be obtained using a combinations of the -values in Table 1, and it can be inverted to find a lower confidence limit for . Similarly, by inverting a left-tailed test for , a lower confidence limit for can be obtained. Reciprocal of these one-sided confidence limits form a CI for .
Coefficient of variation: Coverage and precision studies
Forkman10 proposed an approximate CI for a common CV and Tian9 also proposed a generalized fiducial CI for a common CV. We did not include these CIs for comparison study as our preliminary studies indicated that these CIs are less accurate than those based on the combined tests (see Murshed30 for details).
To compare all other CIs, we chose samples of sizes from small to moderate, number of sources and 5, and and 0.40. From the reported coverage probabilities and expected widths in Table 5, we observe the following. Among the CIs based on combined tests, inverse Cauchy method produced CIs that are much wider than the other three CIs for almost all cases. There are only little differences among the precisions of the CIs based on the inverse test and the weighted inverse normal test. There are a few cases where the CIs based on the weighted inverse normal test are narrower than the corresponding CIs based on the inverse test. In general, Fisher’s CIs are little wider than the CIs based on the weighted inverse normal test and those based on the inverse test. Thus, the CIs based on the weighted inverse normal test are preferable to others.
Remark 2. Liu and Xu13 developed CIs for the common CV using the approach called confidence distributions which is “Neymannian interpretation” of Fisher’s fiducial distribution, and so closely related to the fiducial approach. The proposed CIs in their paper are also exact in the frequentist sense, except that Monte Carlo simulation is required to find CIs. The CIs based on confidence distributions and the ones based on our numerical approach with the Fisher combined test are the same. This is also supported by the agreement of coverage probabilities and expected widths reported in our Table 5 and those in Table 1 of Liu and Xu13 for the cases of sample sizes and .
Coverage probabilities and expected widths of 95% CIs for common coefficient of variation.
Fisher
W Inv Norm
Inv
Inv Cauchy
Fisher
W Inv Norm
Inv
Inv Cauchy
0.05
0.951(0.047)
0.952(0.042)
0.947(0.045)
0.957(0.064)
0.951(0.028)
0.949(0.029)
0.947(0.029)
0.953(0.044)
0.10
0.954(0.094)
0.945(0.084)
0.951(0.090)
0.954(0.129)
0.952(0.062)
0.949(0.057)
0.951(0.059)
0.950(0.081)
0.15
0.952(0.143)
0.950(0.128)
0.950(0.135)
0.952(0.200)
0.949(0.095)
0.947(0.088)
0.945(0.091)
0.950(0.126)
0.20
0.951(0.193)
0.955(0.174)
0.952(0.183)
0.950(0.276)
0.951(0.130)
0.953(0.120)
0.952(0.123)
0.947(0.172)
0.30
0.955(0.307)
0.952(0.276)
0.953(0.291)
0.951(0.422)
0.952(0.204)
0.952(0.189)
0.953(0.194)
0.952(0.278)
0.40
0.945(0.447)
0.948(0.398)
0.946(0.420)
0.949(0.497)
0.950(0.292)
0.951(0.270)
0.949(0.277)
0.948(0.405)
0.05
0.949(0.020)
0.951(0.019)
0.951(0.019)
0.951(0.027)
0.950(0.021)
0.950(0.019)
0.952(0.019)
0.954(0.026)
0.10
0.958(0.040)
0.956(0.038)
0.956(0.038)
0.955(0.051)
0.952(0.041)
0.952(0.039)
0.952(0.039)
0.952(0.051)
0.15
0.949(0.061)
0.950(0.058)
0.951(0.058)
0.955(0.078)
0.949(0.062)
0.952(0.059)
0.953(0.058)
0.951(0.076)
0.20
0.949(0.084)
0.950(0.079)
0.950(0.080)
0.948(0.107)
0.949(0.084)
0.948(0.080)
0.949(0.079)
0.950(0.105)
0.30
0.953(0.132)
0.953(0.125)
0.953(0.126)
0.945(0.169)
0.954(0.134)
0.952(0.127)
0.952(0.126)
0.952(0.166)
0.40
0.949(0.187)
0.949(0.176)
0.950(0.178)
0.949(0.241)
0.950(0.190)
0.950(0.180)
0.949(0.179)
0.948(0.236)
0.05
0.951(0.041)
0.949(0.036)
0.948(0.039)
0.954(0.064)
0.952(0.014)
0.951(0.013)
0.951(0.013)
0.954(0.021
0.10
0.952(0.082)
0.945(0.071)
0.955(0.077)
0.952(0.133)
0.951(0.029)
0.949(0.027)
0.949(0.027)
0.952(0.042
0.15
0.950(0.122)
0.944(0.107)
0.948(0.115)
0.951(0.205)
0.950(0.044)
0.952(0.041)
0.951(0.041)
0.952(0.062
0.20
0.950(0.164)
0.952(0.144)
0.951(0.154)
0.951(0.281)
0.947(0.060)
0.948(0.056)
0.947(0.056)
0.951(0.085
0.30
0.948(0.255)
0.950(0.225)
0.949(0.240)
0.948(0.369)
0.951(0.094)
0.949(0.088)
0.949(0.088)
0.949(0.135
0.40
0.951(0.366)
0.953(0.318)
0.952(0.342)
0.947(0.376)
0.952(0.134)
0.953(0.126)
0.952(0.126)
0.953(0.192
Example for coefficient of variation
For the purpose of promoting the quality and setting standards of medical laboratory technology, Hong Kong Medical Technology Association conducted a quality assurance program for medical laboratories in Hong Kong in 1989. In the specialty of hematology and serology, one normal and one abnormal hematology and serology blood samples were sent to participants for measurements of Hb, RBC, MCV, Hct, WBC, and Platelet in each survey. The summary statistics of data on normal blood samples collected from the third surveys of 1995 and 1996 are given in Fung and Tsang,31 and they are reproduced here in Table 6. Here, is the maximum likelihood estimator (MLE) of . The data were also analyzed by Tian9 and Krishnamoorthy and Lee.32
Summary statistics of normal blood samples.
1995
1996
HB
RBC
MCV
Hct
WBC
Plt
HB
RBC
MCV
Hct
WBC
Plt
65
65
63
64
65
64
73
73
72
72
73
71
13.39
4.409
84.13
0.3717
8.466
247.1
12.97
4.275
85.68
0.3658
7.818
227.31
0.2192
0.0826
3.390
0.0181
0.4655
17.91
0.2929
0.0863
2.946
0.0137
0.4798
21.61
0.0165
0.0189
0.0406
0.0490
0.0554
0.0730
0.0227
0.0203
0.0346
0.0378
0.0618
0.0957
The tests and CIs by Krishnamoorthy and Lee32 indicated that coefficients of variation of measurements in 1995 and 1996 are the same for variables RBC, MCV and WBC. For these three variables, we estimated CIs for the common CV using different approaches and reported them in Table 7. Since the numbers of measurements are quite large, all methods, except inverse Cauchy, produced similar CIs for common coefficients of variation of different pairs. A cursory glance of Table 7 shows that the inverse Cauchy CIs are somewhat different from others. R functions that were used to compute CIs for a common are provided in a supplementary file. The R function CI.common.CV(n, zetah, cl, method) provided in the supplementary file was used to compute CIs by the Fisher, weighted inverse normal, inverse Cauchy and inverse methods.
95% confidence intervals for a common CV of measurements from 1995 and 1996 surveys.
Method
RBC
Width
MCV
Width
WBC
Width
Fisher
(0.0175, 0.0226)
0.0051
(0.0337, 0.0428)
0.0091
(0.0525, 0.0676)
0.0151
W Inv Normal
(0.0177, 0.0224)
0.0047
(0.0334, 0.0426)
0.0092
(0.0528, 0.0671)
0.0143
Inverse
(0.0176, 0.0224)
0.0048
(0.0336, 0.0427)
0.0091
(0.0527, 0.0671)
0.0144
Inv Cauchy
(0.0172, 0.0234)
0.0062
(0.0339, 0.0428)
0.0089
(0.0523, 0.0691)
0.0168
Common mean of several lognormal populations
Consider log-normal populations with parameters . Assume that the means of these populations are the same , where , . Tian and Wu6 proposed a GV approach to find a CI for the common mean of several log-normal distributions. We shall now describe an efficient test for the lognormal mean and combined tests based on this one-sample test. The combined tests can be inverted to find CIs for a common mean of several lognormal populations.
Let be a log-transformed sample from a lognormal distribution with parameters and , . Recall that , . Since the common mean of lognormal distributions is given by with , , it is enough to find a CI for . Let denote the (mean, variance) based on a sample of size from the th population.
The modified likelihood ratio test for a lognormal mean
To describe the modified likelihood ratio test (MLRT) by Wu et al.33 for a single sample, let . The MLEs are where and . For fixed , the constrained maximum likelihood estimate of is Define
and
For testing vs. , the MLRT statistic is given by
which follows a standard normal distribution asymptotically. This asymptotic result has third-order accuracy, and is valid even for small samples. The -value of the MLRT is given by
where is the standard normal distribution function.
Combined tests for a common lognormal mean
To develop a combined test for a common mean of several lognormal distributions, let us suppose that the means of all the populations are the same. That is, , or equivalently, Let denote the -value for testing
based on the th sample. These -values can be combined using any of the methods described in Section 2 to arrive at a single test for the common mean of several lognormal populations. The combined tests may be inverted using Algorithm 1 to find a CI for a common mean of lognormal distributions.
Lognormal: Simulation studies
There are other approximate CIs based on the MOVER approach and fiducial approach.6 Our preliminary investigations indicated that these CIs are not so accurate as those based on combined tests, and so we do not include these CIs in our comparison study. Murshed30 has included these CIs for comparison study. We estimated the coverage probabilities and expected widths of all CIs based on combined tests for a common mean of several lognormal distributions and presented them in Table 8. We chose as 1, 2, and 3, and different combinations for values of . We considered only cases of and 4. All simulation estimates are based on 10,000 runs. Since the MLRT for testing a lognormal mean based on a single sample is very accurate even for small samples, the coverage probabilities of CIs based on combined tests are close to the nominal level for all cases. The inverse Cauchy CIs are wider than all CIs for almost all cases. The CIs based on the inverse test and the weighted inverse normal test are very similar in terms of coverage probability. The CIs based on the Fisher test and the inverse test are quite comparable, and both are equally efficient. The CIs based on the weighted inverse normal test are somewhat less efficient than the other two CIs based on the Fisher test and the inverse test for some cases; for example, see the results for in Table 8. On an overall basis, the CIs based on the weighted inverse normal test and the inverse are recommended for applications.
Coverage probabilities and expected widths (in parentheses) of 95% CIs for .
Fisher
W Inv Norm
Inv
Inv Cauchy
Fisher
W Inv Norm
Inv
Inv Cauchy
1
(0.2, 0.6)
0.948(1.04)
0.949(0.98)
0.949(1.01)
0.948(1.31)
0.950(0.61)
0.949(0.60)
0.949(0.60)
0.949(0.69)
1
(0.4, 1)
0.948(1.60)
0.949(1.49)
0.948(1.55)
0.949(2.10)
0.950(0.91)
0.952(0.88)
0.951(0.89)
0.950(1.05)
2
(0.7, 1.1)
0.946(2.13)
0.946(1.93)
0.945(2.04)
0.948(2.91)
0.948(1.18)
0.950(1.12)
0.950(1.14)
0.948(1.40)
2
(1, 1.3)
0.948(2.63)
0.949(2.36)
0.949(2.51)
0.947(3.67)
0.950(1.43)
0.951(1.35)
0.951(1.38)
0.951(1.73)
2
(1.3,1.5)
0.952(3.13)
0.949(2.80)
0.952(2.99)
0.948(4.44)
0.948(1.67)
0.948(1.57)
0.948(1.61)
0.950(2.04)
3
(0.4, 0.7)
0.949(1.44)
0.947(1.32)
0.949(1.39)
0.949(1.94)
0.944(0.82)
0.947(0.79)
0.946(0.80)
0.950(0.98)
3
(1.4,1.7)
0.947(3.40)
0.948(3.04)
0.948(3.24)
0.948(4.83)
0.950(1.80)
0.950(1.69)
0.950(1.73)
0.946(2.20)
1
(0.2, 0.6)
0.949(0.39)
0.950(0.39)
0.950(0.38)
0.946(0.43)
0.952(0.31)
0.950(0.32)
0.951(0.31)
0.947(0.34)
1
(0.4, 1)
0.948(0.56)
0.948(0.57)
0.948(0.56)
0.947(0.64)
0.953(0.46)
0.953(0.46)
0.951(0.45)
0.949(0.51)
2
(0.7,1.1)
0.954(0.72)
0.955(0.70)
0.954(0.70)
0.951(0.83)
0.949(0.58)
0.948(0.57)
0.948(0.57)
0.949(0.66)
2
(1, 1.3)
0.950(0.86)
0.949(0.84)
0.949(0.84)
0.946(1.00)
0.946(0.70)
0.948(0.68)
0.947(0.68)
0.949(0.80)
2
(1.3,1.5)
0.955(1.00)
0.953(0.96)
0.953(0.96)
0.950(1.16)
0.952(0.81)
0.952(0.78)
0.953(0.78)
0.952(0.93)
3
(0.4, 0.7)
0.945(0.51)
0.948(0.51)
0.947(0.50)
0.952(0.59)
0.950(0.42)
0.950(0.41)
0.950(0.41)
0.950(0.47)
3
(1.4,1.7)
0.950(1.08)
0.951(1.04)
0.951(1.04)
0.948(1.25)
0.951(0.87)
0.950(0.84)
0.950(0.84)
0.952(1.00)
1
(0.2,0.6,0.5,0.5)
0.948(0.72)
0.951(0.67)
0.949(0.69)
0.945(1.14)
0.949(0.28)
0.948(0.26)
0.949(0.27)
0.952(0.27)
1
(0.6,0.8,0.9,0.5)
0.948(0.99)
0.950(0.88)
0.950(0.93)
0.944(1.57)
0.950(0.21)
0.948(0.20)
0.950(0.20)
0.948(0.37)
2
(0.1,1,0.9,1)
0.948(0.77)
0.952(0.81)
0.948(0.76)
0.945(1.06)
0.948(0.20)
0.948(0.23)
0.947(0.21)
0.946(0.23)
2
(0.5,0.5,1,1)
0.947(1.13)
0.946(1.05)
0.946(1.14)
0.949(1.82)
0.950(0.30)
0.949(0.29)
0.950(0.28)
0.950(0.39)
2
(0.1,0.1,2,2)
0.949(0.63)
0.950(0.84)
0.950(0.66)
0.946(0.90)
0.948(0.16)
0.947(0.19)
0.948(0.17)
0.948(0.19)
3
(0.1,0.1,1,0.5)
0.946(0.52)
0.948(0.56)
0.946(0.52)
0.951(0.80)
0.952(0.14)
0.949(0.15)
0.951(0.14)
0.948(0.18)
3
(1,2,3,4)
0.948(2.43)
0.950(2.42)
0.949(2.43)
0.948(4.38)
0.948(0.62)
0.950(0.65)
0.951(0.62)
0.947(0.79)
Lognormal: An example
We shall now illustrate the methods using the example given in Tian and Wu6 and also discussed in Krishnamoorthy and Oral.7 The pharmacokinetics data were obtained from alcohol interaction study in men. For illustrative purpose, we use the measurements on maximum concentration () and compare the active treatment groups considered in Tian and Wu.6 The statistics based on the log-transformed data are as follows:
The MLRT by Krishnamoorthy and Oral7 indicated that is tenable. Since the group means are not significantly different, it maybe desired to the find the common mean of these three groups.
We estimated 90, 95 and 99% CIs for the common mean of these three groups and reported them in the following Table 9.
Confidence intervals of the common lognormal mean.
90%
95%
99%
Method
CI
Width
CI
Width
CI
Width
Fisher
(13.28, 16.88)
3.60
(13.06, 17.25)
4.19
(12.63, 18.05)
5.42
W Inv Normal
(13.55, 16.48)
2.93
(13.32, 16.84)
3.52
(12.87, 17.60)
4.73
Inverse
(13.47, 16.60)
3.13
(13.23, 16.96)
3.73
(12.80, 17.74)
4.94
Inv Cauchy
(12.69, 17.83)
5.14
(12.31, 18.59)
6.28
(11.59, 20.38)
8.79
Note that the inverse normal method produced the shortest CIs while the inverse Cauchy method produced the widest CIs for all three nominal confidence levels. We used the R function CI.lnorm.cmean = function(n, xb, sq, cl, method) provided in the supplementary file to compute CIs by the Fisher, inverse chi-square and the weighted inverse normal methods.
Common correlation coefficient
The problem of estimating a common correlation coefficient of several bivariate normal populations has been addressed by Donner and Rosner,11 Paul34 and Tian and Wilding.12 These authors have proposed some approximate tests and confidence intervals for a common correlation coefficient. To describe the problem formally, let be a sample from a bivariate normal distribution with mean vector and variance-covariance matrix
Let be the sample variance-covariance matrix based on the th sample. Then the sample correlation coefficient is given by , In the following, we first describe the CI given in Donner and Rosner11 that was obtained by combining the Fisher statistics based on individual samples.
A CI for based on Fisher’s statistics
Using the Fisher -transformation of ’s, we have the asymptotic result that
where . Using this distributional result along with the fact that ’s are independent, under , where and Thus, this combined test rejects in favor of if , or equivalently the -value .
A CI for , follows from the distribution of , and is given by
where . As is an increasing function of , a CI for can be deduced from the above one for as Donner and Rosner11 have proposed this CI and we refer to this CI as the -interval for .
CIs based combined tests
Let be an observed value of . That is, , where is an observed value of , . Furthermore, the -value for testing on the basis of the th sample is given by
where is the standard normal distribution function. The above -values can be combined using one of the methods in Section 2 to arrive at a single test, and the test can be inverted to get a CI for the common . Remark 3. It should be noted that there are other tests for that are more accurate than the Fisher test are available in the literature; see Krishnamoorthy and Xia.35 However, the above Fisher’s test is simple and commonly used in applications, and so we proceed with the above test.
Common correlation: Coverage and precision studies
The CIs for a common correlation coefficient are based on combined tests where individual tests are approximate. Even though the individual Fisher tests are known to be accurate even for small samples, we need to compare the CIs based on the combined tests and Donner and Rosner’s11 approach in terms of precision. We estimated the coverage probabilities and expected widths of all CIs using Monte Carlo simulation with 10,000 runs and reported them in Table 10. The estimated coverage probabilities in Table 10 indicate that the CIs based on all combined tests are practically exact. The inverse Cauchy CIs are much wider than all other CIs for all cases. The estimated coverage probabilities and expected widths indicate that the Donner and Rosner CIs and the ones based on the weighted inverse normal test are practically the same in terms of coverage probability and precision. The CI based on the inverse test and the one based on the weighted inverse normal test are similar in most cases, and the latter is better than the former in a few cases. Overall, Donner and Rosner’s CI is preferable to others for simplicity, accuracy and efficiency.
Coverage probabilities and expected widths of 95% CIs for a common correlation coefficient.
Measurements on diastolic and systolic blood pressures of proposita girls for age groups 6-8, 9-11, and 12-14 yielded the following results: group 1: ; group 2: ; and group 3: . These data have also been analyzed by many researchers in the context of making inference on interclass and intraclass correlations. Application of the Fisher test implied that the homogeneity of correlation coefficients among these three groups is tenable.12 So we can use the above statistics to find a CI for the common correlation coefficient of these three groups.
We computed the 95% CIs for based on different methods as follows. The CIs Donner and Rosner11 method is (.300, .904); Fisher’e method produced (.198, .921); weighted inverse normal yielded (.304, .905); the one based on the inverse is (.243, .914). Inverse Cauchy method produced . The weighted inverse normal method and the Donner and Rosner method produced similar CIs. We already noticed in our simulation studies that the inverse Cauchy intervals are expected to be wider than the others, and for this example, it produced CIs that are wider than those based on the other methods. As the method by Donner and Rosner11 is not only simple but also better than all other methods, we provide R code based on the method by Donner and Rosner to compute the CI for a common . See CI.common.corrl(n, r, cl) in the supplementary file.
Common mean of several gamma populations
Let and denote respectively the arithmetic mean and geometric based on a sample of size from a gamma distribution with the shape parameter and the scale parameter , say, gamma. The mean of the gamma distribution is given by .
The MLRT of the mean based on a single sample
There are several approximate tests available for the mean of a gamma distribution. Among all the tests, the MLRT by Fraser et al.36 appears to be the best, and so we develop a method finding CI by combining the -values of independent MLRTs.
It can be readily verified that the log-likelihood function is given by
The MLE is the solution of the equation where is the digamma function. The MLE of is . The signed likelihood ratio test statistic is given by
and is the MLE of at . This constrained MLE is obtained as the root of the equation The MLRT is given by
where is defined in (15), and The MLRT for rejects the null hypothesis if the -value where is the standard normal random variable.
Combined tests for a common mean
Let be a sample from a gamma distribution with the shape parameter and the scale parameter , . Assume that the means of these gamma distributions are the same. That is, for all . Let be the -value of the MLRT test based on the th sample of size from the th population, . A combined test for the common mean can be obtained by combining the individual -values using a combination of -values given in Table 1. A CI based on a combined test can be obtained using Algorithm 1.
Gamma distribution: Simulation studies
We have carried out some simulation studies to judge the accuracy of the CIs based on Fisher’s combined test, inverse test, the weighted inverse normal test and the inverse Cauchy test. Other fiducial CIs by Krishnamoorthy and Wang37 and Yan8 are approximate and they are in general inferior to other CIs based on combined tests. So these CIs are not included in our simulation studies here, but available in Murshed’s30 Ph.D. dissertation. The simulation estimates of the coverage probabilities and expected widths of 95% CIs are reported in Table 11 for some sample sizes ranging from small to moderately large. As the MLRT based on a single sample is highly accurate even for small samples, the combined tests based on them are also highly accurate in terms of coverage probability. All the CIs based on combined tests control the coverage probabilities very close to the nominal level 0.95 for all cases. Comparison of expected widths clearly indicates that the inverse Cauchy method produced wider CIs for all cases considered in the table. Fisher’s CIs are slightly wider than the other two CIs based on inverse normal and inverse combined tests in most cases, and it is slightly better than the CIs based on weighted inverse normal test in some cases; see the results for in Table 11. There is no clear-cut winner between the CIs based on the inverse test and the weighted inverse normal test. These two comparable CIs with higher precisions are preferable among all CIs.
Coverage probabilities and expected widths of 95% CIs for common mean of several gamma populations.
Fisher
W Inv Norm
Inv
Inv Cauchy
Fisher
W Inv Norm
Inv
Inv Cauchy
3
0.949(1.95)
0.951(1.85)
0.952(1.89)
0.951(2.34)
0.946(1.67)
0.949(1.57)
0.948(1.60)
0.951(1.93)
5
0.952(3.25)
0.951(3.08)
0.951(3.13)
0.952(3.91)
0.950(2.79)
0.949(2.63)
0.951(2.68)
0.945(3.20)
7
0.947(4.57)
0.946(4.32)
0.946(4.41)
0.948(5.44)
0.954(3.89)
0.953(3.67)
0.952(3.74)
0.948(4.46)
10
0.952(6.56)
0.951(6.20)
0.953(6.31)
0.950(7.76)
0.949(5.56)
0.949(5.25)
0.947(5.35)
0.945(6.40)
12
0.952(7.85)
0.952(7.44)
0.953(7.51)
0.950(9.37)
0.948(6.70)
0.947(6.32)
0.946(6.44)
0.949(7.75)
15
0.951(9.81)
0.949(9.27)
0.949(9.22)
0.953(11.6)
0.952(8.38)
0.951(7.89)
0.950(8.03)
0.951(9.69)
3
0.949(1.37)
0.951(1.37)
0.950(1.32)
0.948(1.56)
0.948(1.29)
0.948(1.24)
0.948(1.25)
0.949(1.48)
5
0.950(2.29)
0.950(2.29)
0.950(2.20)
0.946(2.60)
0.950(2.15)
0.950(2.06)
0.951(2.08)
0.949(2.47)
7
0.951(3.20)
0.949(3.20)
0.950(3.08)
0.952(3.64)
0.947(3.00)
0.947(2.89)
0.948(2.90)
0.952(3.47)
10
0.950(4.58)
0.954(4.59)
0.952(4.41)
0.946(5.22)
0.951(4.32)
0.950(4.15)
0.950(4.18)
0.946(4.95)
12
0.950(5.49)
0.952(5.49)
0.951(5.28)
0.950(6.24)
0.952(5.19)
0.947(4.98)
0.948(5.02)
0.949(5.95)
15
0.951(6.88)
0.949(6.87)
0.948(6.61)
0.951(7.82)
0.952(6.47)
0.954(6.23)
0.953(6.27)
0.948(7.39)
3
0.948(1.15)
0.948(1.07)
0.950(1.09)
0.951(1.61)
0.948(0.71)
0.949(0.66)
0.950(0.67)
0.947(0.92)
5
0.951(1.90)
0.950(1.77)
0.949(1.81)
0.944(2.69)
0.951(1.19)
0.949(1.10)
0.949(1.12)
0.951(1.53)
7
0.949(2.67)
0.952(2.49)
0.950(2.54)
0.949(3.79)
0.945(1.66)
0.951(1.54)
0.948(1.56)
0.953(2.15)
10
0.951(3.82)
0.949(3.56)
0.950(3.64)
0.950(5.37)
0.950(2.38)
0.953(2.20)
0.951(2.24)
0.944(3.07)
12
0.950(4.60)
0.951(4.28)
0.949(4.38)
0.953(6.47)
0.949(2.85)
0.947(2.64)
0.947(2.68)
0.950(3.70)
15
0.951(5.74)
0.951(5.35)
0.950(5.47)
0.951(8.09)
0.951(3.57)
0.948(3.30)
0.949(3.35)
0.950(4.61)
Gamma distribution: An example
The data in the following Table 12 are measurements of chloride concentration in spring water samples from two types of rocks in Sierra Nevada, California and Nevada. The data are reported in Feth et al.38 Gamma distributions are routinely used to analyze such pollution measurement data. Yan8 has used the data to find a CI for the common mean concentrations in both sites. The MLEs are , , and .
Chloride concentration (mg/liter) in water.
Site
1 Granodiorite
6.0
0.5
0.4
0.7
0.8
6.0
5.0
0.6
1.2
1.0
0.2
1.2
1.0
0.3
0.1
0.1
0.4
3.2
2 Quartz monzonite
1.0
0.2
1.2
1.0
0.3
0.1
0.1
0.4
3.2
0.3
0.4
1.8
0.9
0.1
0.2
0.3
0.5
We computed confidence intervals for the common mean of chloride concentrations in both sites using all methods and reported them in Table 13. For this example, the inverse Cauchy CIs are the shortest for all three nominal confidence levels. The 95% generalized CI reported in Yan8 is (0.524, 1.366) with width 0.842, which is somewhat different from other 95% CIs in Table 13. We used the R function CI.common.mean.gamma(n, xb, xd, sq, cl,method) provided in the supplementary file to compute the CIs based on the Fisher, inverse normal, inverse chi-square and the inverse Cauchy methods.
90%, 95% and 99% CI for the mean chloride concentration.
90%
95%
99%
Method
CI
Width
CI
Width
CI
Width
Fisher
(0.905, 1.37)
0.465
(0.845, 1.50)
0.655
(0.741, 1.79)
1.049
W Inv Normal
(0.811, 1.55)
0.739
(0.766, 1.67)
0.904
(0.686, 1.93)
1.244
Inverse
(0.848, 1.47)
0.623
(0.799, 1.59)
0.791
(0.711, 1.85)
1.139
Inv Cauchy
(0.969, 1.25)
0.281
(0.914, 1.36)
0.446
(0.794, 1.69)
0.896
Concluding remarks
A typical method of finding a CI for a parameter of interest is inverting a test for the parameter. If the test is exact, in the sense that the null distribution of the test statistic does not depend on any unknown parameter, then the CI that is obtained by inverting the test is also exact. This well-known approach to find CIs based on data from different independent resources was seldom investigated in the present context. In this article, we investigated such method and shown that efficient exact CIs for a common parameter of interest can be readily obtained by inverting a combined test. Eventhough there are a few exact combined tests available in the literature, the CIs based on them were seldom utilized. The methods that we considered in this article are conceptually simple, but they involve numerical computations. To help practitioners and other researchers, we provided R code in a supplementary file that can be used in a straightforward manner to compute the CIs based on the Fisher, inverse chi-square, and weighted inverse normal methods.
We also investigated (not reported here) robustness of the interval estimation procedures. For example, as both lognormal and gamma distributions are used to model positive right-skewed data, we investigated numerically if the CIs for a common mean of several gamma populations can be used to estimate a common mean of several lognormal populations. Our investigation indicated that such gamma-based CIs are too liberal when they are used for lognormal populations. In general, the combined tests and the method of finding CIs are valid only when the model assumptions are satisfied. This is because the individual tests (such as the -tests for normal and the LRTs for the gamma and log-normal) that were used to combine are valid only for the assumed models.
It should be clear that the proposed method of finding a CI for a common parameter of interest can be readily applied to practical problems where an exact test or highly accurate test such as the MLRT based on a single sample is available. For example, our approach can be readily extended to find the confidence region of a common mean vector of several multivariate normal distributions. We are currently extending the proposed approach to the case of multivariate normal and plan to publish the work elsewhere.
Supplemental Material
sj-txt-1-smm-10.1177_09622802231217644 - Supplemental material for Confidence estimation based on data from independent studies
Supplemental material, sj-txt-1-smm-10.1177_09622802231217644 for Confidence estimation based on data from independent studies by Kalimuthu Krishnamoorthy and Md Monzur Murshed in Statistical Methods in Medical Research
Footnotes
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
ORCID iD
Kalimuthu Krishnamoorthy
Acknowledgements
The authors are thankful to two reviewers for providing useful comments and suggestions.
References
1.
EggerMSmithGD. Meta-analysis potentials and promise. BMJ1997; 315: 1371–1374.
2.
JacksonDRileyRWhiteIR. Multivariate meta-analysis: Potential and promise. Stat Med2011; 30: 2481–2498.
3.
MeierP. Variance of a weighted mean. Biometrics1953; 9: 59–73.
4.
EberhardtKRReeveCPSpiegelmanCH. A minimax approach to combining means, with practical examples. Chemometr Intell Lab Syst1989; 5: 129–148.
5.
SkinnerJB. On combining studies. Drug Inf J1991; 25: 395–403.
6.
TianLWuJ. Inferences on the common mean of several lognormal populations: The generalized variable approach. Biometrical J2007; 49: 944–951.
7.
KrishnamoorthyKOralE. Standardized LRT for comparing several lognormal means and confidence interval for the common mean. Stat Methods Med Res2017; 26: 2919–2937.
8.
YanL. Confidence interval estimation of the common mean of several gamma populations. PLoS ONE2022; 17: e0269971. DOI: 10.1371/journal.pone.0269971.
9.
TianL. Inferences on the common coefficient of variation. Stat Med2005; 24: 2213–2220.
10.
ForkmanJ. Estimator and tests for common coefficients of variation in normal distributions. Commun Stat - Theory Methods2009; 38: 233–251.
11.
DonnerARosnerB. On inferences concerning a common correlation coefficient. Appl Stat1980; 29: 69–76.
12.
TianLWildingGE. Confidence interval estimation of a common correlation coefficient. Comput Stat Data Anal2008; 52: 4872–4877.
13.
LiuXXuX. A note on combined inference on the common coefficient of variation using confidence distributions. Electron J Stat2015; 9: 219–233.
14.
FisherRA. Statistical Methods for Research Workers. Edinburgh: Oliver and Boyd, 1932.
15.
MathewTSinhaBKZhouL. Some statistical procedures for combining independent tests. J Am Stat Assoc1993; 88: 912–919.
16.
StoufferSASuchmanEADeVinneyLC, et al. The American Soldier 1 Adjustment during Army Life. Princeton: Princeton University Press, 1949.
17.
LiptakT. On the combination of independent tests. Magyar Tudom Ãanyos Akad Aemia Matematikai Kutat Ao, Intezetenek Kozlemenyei1958; 3: 171–197.
18.
FolksJL. Combination of independent tests. Handbook Stat, P. R. Krishnaiah and P. K. Sen, eds. 1984; 4: 113–121.
19.
WhitlockMC. Combining probability from independent tests: The weighted Z-method is superior to fisher’s approach. J Evol Biol2005; 18: 1368–1373.
20.
LiuYXieJ. Cauchy combination test: A powerful test with analytic -value calculation under arbitrary dependency structures. J Am Stat Assoc2020; 115: 393–402.
21.
KrishnamoorthyKLvSMurshedMM. Combining independent tests for a common parameter of several continuous distributions: a new test and power comparisons. Commun Stat-Simul Comput2022. DOI: 10.1080/03610918.2022.2058546.
22.
FairweatherWR. A method of obtaining an exact confidence interval for the common mean of several normal populations. Appl Stat1972; 21: 229–233.
23.
JordanSMKrishnamoorthyK. Exact cofidence intervals for the common mean of several normal populations. Biometrics1996; 52: 78–87.
24.
WeerahandiS. Exact Statistical Methods for Data Analysis. New York, NY: Springer, 1995.
KrishnamoorthyKMathewT. Inferences on the means of lognormal distributions using generalized -values and generalized confidence intervals. J Stat Plan Inference2003; 115: 103–121.
27.
KrishnamoorthyK. Modified normal-based approximation to the percentiles of linear combination of independent random variables with applications. Commun Stat – Simul Comput2016; 45: 2428–2444.
28.
KrishnamoorthyKLuY. Inferences on the common mean of several normal populations based on the generalized variable method. Biometrics2003; 59: 237–247.
29.
JohnsonNLWelchBL. Application of the noncentral t-distribution. Biometrika1940; 31: 362–389.
30.
MurshedMM. Tests and Confidence Intervals Based on Data from Independent Sources. Ph.D. disserattion, Spring 2024. Department of Mathematics, University of Louisiana at Lafayette, LA 70504, USA.
31.
FungWKTsangTS. A simulation study comparing tests for the equality of coefficients of variation. Stat Med1988; 17: 2003–2014.
32.
KrishnamoorthyKLeeM. Improved tests for the equality of normal coefficients of variation. Comput Stat2013; 29: 215–232.
33.
WuJWongACMJiangG. Likelihood-based confidence intervals for a log-normal mean. Stat Med2003; 22: 1849–1860.
34.
PaulSR. Estimation and testing significance for a common correlation coefficient. Commun Stat Theory Methods1988; 17: 39–53.
35.
KrishnamoorthyKXiaY. Inferences on correlation coefficients: One-sample, independent and correlated cases. J Stat Plan Inference2007; 137: 2362–2379.
36.
FraserDASReidNWongA. Simple and accurate inference for the mean of the gamma model. Can J Stat1997; 25: 91–99.
37.
KrishnamoorthyKWangX. Fiducial inference on gamma distribution: Uncensored and censored cases. Environmetrics2016; 27: 479–493.
38.
FethJHRobertsonCEPolzerWL. Sources of mineral constituents in water from granite rocks, Sierra Nevada, California and Nevada. United States Geological Survey Water Supply Paper 1964. 1535. https://pubs.usgs.gov/wsp/1535i/report.pdf.
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.