Skip to main content
  • Research Article
  • Open access
  • Published:

Hartung-Knapp-Sidik-Jonkman approach and its modification for random-effects meta-analysis with few studies

Abstract

Background

Random-effects meta-analysis is commonly performed by first deriving an estimate of the between-study variation, the heterogeneity, and subsequently using this as the basis for combining results, i.e., for estimating the effect, the figure of primary interest. The heterogeneity variance estimate however is commonly associated with substantial uncertainty, especially in contexts where there are only few studies available, such as in small populations and rare diseases.

Methods

Confidence intervals and tests for the effect may be constructed via a simple normal approximation, or via a Student-t distribution, using the Hartung-Knapp-Sidik-Jonkman (HKSJ) approach, which additionally uses a refined estimator of variance of the effect estimator. The modified Knapp-Hartung method (mKH) applies an ad hoc correction and has been proposed to prevent counterintuitive effects and to yield more conservative inference. We performed a simulation study to investigate the behaviour of the standard HKSJ and modified mKH procedures in a range of circumstances, with a focus on the common case of meta-analysis based on only a few studies.

Results

The standard HKSJ procedure works well when the treatment effect estimates to be combined are of comparable precision, but nominal error levels are exceeded when standard errors vary considerably between studies (e.g. due to variations in study size). Application of the modification on the other hand yields more conservative results with error rates closer to the nominal level. Differences are most pronounced in the common case of few studies of varying size or precision.

Conclusions

Use of the modified mKH procedure is recommended, especially when only a few studies contribute to the meta-analysis and the involved studies’ precisions (standard errors) vary.

Peer Review reports

Background

Random-effects meta-analysis is most commonly performed based on an underlying hierarchical model including two unknowns as parameters: the effect μ, which is the figure of primary interest, and the between-study variance (heterogeneity) τ 2, which is a nuisance parameter. Inference then is usually done sequentially, by first deriving an estimate of the heterogeneity variance, \(\hat {\tau }^{2}\), and then determining the effect estimate \(\hat {\mu }\) by conditioning on the estimate \(\hat {\tau }^{2}\) [1, 2]. A large number of different estimators for the heterogeneity variance is available (see e.g. [36]), and effect estimation may be done based on a simple normal approximation, or by utilizing a Student-t distribution [7] with an additionally refined estimator of the variance of \(\hat {\mu }\) [812]. While the normal model may be motivated by asymptotic arguments, in actual applications the number of estimates to be combined is commonly small [13, 14] and hence the estimation uncertainty in the between-study variance τ 2 is substantial, so that an adjustment is appropriate and in fact improves operating characteristics [712, 15].

The problem of deriving estimates from only a small number of data sources is a common problem especially in fields of application where empirical information is sparse due to the rarity of the condition in question. The rarity of a disease is often accompanied with a low (commercial) interest or incentive, which is why such diseases are also known as orphan diseases. According to the European Commission, a disease is designated orphan status when the prevalence is ≤5 in 10 000 [16]. While by definition any individual rare disease has a low prevalence, there is a large number of these, eventually affecting a substantial fraction of an estimated 6–8 % of the total population [17], and with that posing a challenge to health care systems worldwide.

The European Medicines Agency acknowledges the particular obstacles in rare diseases research but points out that there is no fundamental difference between rare and more common diseases and hence no “paradigm change” when it comes to regulatory issues. Because of the common small-sample settings, the importance of sophisticated methods is emphasized, and meta-analyses of good quality randomised controlled clinical trials are still considered the highest level of evidence [18]. The problems encountered in rare diseases research often call for special statistical methods, especially with respect to study designs [17, 19, 20]. Meta-analyses are particularly important in this field due to the lack of large trials, while these will commonly still be faced with the problem of small numbers of available studies. Between-study heterogeneity then is anticipated, since the gathered pieces of evidence are likely to differ with respect to study designs, types of control groups or treatment allocation [17, 19, 20], Unkel S, Röver C, Stallard N, Benda N, Posch M, Zohar S, et al., Systematic reviews in paediatric multiple sclerosis and Creutzfeldt-Jakob disease exemplify shortcomings in methods used to evaluate therapies in rare conditions, Submitted. Small studies have in fact epirically been found to exhibit more heterogeneity than large trials [21]. Consequently, the use of methods suitable for few studies and marginally significant findings is of crucial importance here.

With an estimated incidence of 2–20 cases per 100 000 population, juvenile idiopathic arthritis (JIA) is an example of a rare disease [22]. In the following, we will use a meta-analysis in JIA [23] as a case study to illustrate the different methods discussed below.

In the following sections, we will first describe the methods used, then show the results of a simulation study, and demonstrate the different types of analyses in an example data set, before closing with conclusions and recommendations.

Methods

Random-effects meta-analysis

Meta-analysis is very commonly performed via a random-effects approach, utilizing the normal-normal hierarchical model. Here the data are given in terms of a number k of estimates y i that are associated with some uncertainty given through standard errors s i >0 that are taken to be known without uncertainty. The estimates are assumed to measure trial-specific parameters θ i :

$$ y_{i} \sim \mathrm{N}\left(\theta_{i}, {s_{i}^{2}}\right) \qquad \text{for}\, i = 1,\ldots,k. $$
((1))

The parameters θ i vary from trial to trial around a global mean μ due to some heterogeneity variance between trials that constitutes an additive variance component to the model,

$$ \theta_{i} \sim \mathrm{N}\left(\mu, \tau^{2}\right) \qquad \text{for}\, i = 1,\ldots,k, $$
((2))

where τ 2≥0. The model may then be simplified by integrating out the parameters θ i , leading to the marginal expression

$$ y_{i} \sim \mathrm{N}\left(\mu,\, {s_{i}^{2}} + \tau^{2}\right) \qquad \text{for}\, i = 1,\ldots,k. $$
((3))

Among the two unknowns in the model, the overall mean μ, the effect, usually is the figure of primary interest, while the heterogeneity variance τ 2 constitutes a nuisance parameter. When τ 2=0, the model simplifies to the so-called fixed-effect model [1, 2].

The “relative amount of heterogeneity” in a meta-analysis may be expressed in terms of the measure I 2, which is defined as

$$ I^{2} = \frac{\tau^{2}}{\tau^{2}+\tilde{s}^{2}} $$
((4))

where \(\tilde {s}\) is some kind of “average” standard error among the study-specific s i [24]. In the following simulation studies, we will determine \(\tilde {s}^{2}\) as the arithmetic mean of squared standard errors.

Parameter estimation

If the value of the heterogeneity variance parameter τ 2 were known, the (conditional) maximum-likelihood effect estimate would result as the weighted average

$$ \hat{\mu} = \frac{\sum_{i} w_{i} \,y_{i}}{\sum_{i} w_{i}} $$
((5))

with “inverse variance weights” defined as

$$ w_{i} = \frac{1}{{s_{i}^{2}} + \tau^{2}} \qquad \text{for} \,i = 1,\ldots,k. $$
((6))

A common approach to inference within the random-effects model is to first estimate the heterogeneity variance τ 2, and subsequently estimate the effect μ conditional on the heterogeneity estimate. Note that the w i are effectively treated as “known” while in fact both the \({s_{i}^{2}}\) as well as τ 2 are only measured with some uncertainty (that depends on the size/precision of the ith individual study and the number of studies k). There is a wide range of different heterogeneity estimators available (see e.g. [36] for more details). In the following we will concentrate on some of the most common ones, the DerSimonian-Laird (DL) estimator, a moment estimator [25] with acknowledged shortcomings [26], the restricted maximum likelihood (REML) estimator [3, 27], and the Paule-Mandel (PM) estimator, an essentially heuristic approach [28, 29]. Software to compute the different estimates is provided e.g. in the metafor and meta R packages [30, 31].

Confidence intervals and tests

Normal approximation

Confidence intervals and, equivalently, tests for the effect μ are commonly constructed using a normal approximation for the estimate \(\hat {\mu }\). The standard error of \(\hat {\mu }\), conditional on a fixed heterogeneity variance value τ 2, is given by

$$ \hat{\sigma}_{\mu} = \sqrt{\frac{1}{\sum_{i} w_{i}}}. $$
((7))

A confidence interval for the effect μ then results via a normal approximation as

$$ \hat{\mu} \; \pm \; \hat{\sigma}_{\mu} \; z_{(1-\alpha/2)} $$
((8))

where z (1−α/2) is the (1 − α/2)-quantile of the standard normal distribution, and (1 − α) is the nominal coverage probability [1, 2]. The normal approximation based on a heterogeneity variance estimate \(\hat {\tau }^{2}\) usually works well for many studies (large k) and small standard errors (small s i ), or negligible heterogeneity (small variance τ 2), but tends to be anticonservative otherwise [8, 9], Friede T, Röver C, Wandel S, Neuenschwander B, Meta-analysis of few small studies in orphan diseases, Submitted.

The Hartung-Knapp-Sidik-Jonkman (HKSJ) method

Hartung and Knapp [8, 9] and Sidik and Jonkman [10] independently introduced an adjusted confidence interval. In order to determine the adjusted interval, first the quadratic form

$$ q = \frac{1}{k-1} \sum_{i} w_{i}(y_{i}-\hat{\mu})^{2} $$
((9))

is computed [810]. The adjusted confidence interval then results as

$$ \hat{\mu} \; \pm \; \sqrt{q} \; \hat{\sigma}_{\mu} \; t_{(k-1); (1-\alpha/2)} $$
((10))

where t (k−1);(1−α/2) is the (1 − α/2)-quantile of the Student-t distribution with (k − 1) degrees of freedom. Note that \(q \hat {\sigma }^{2}\mu \) is derived from a non-negative and unbiased estimator of \(\frac {1}{\sum _{i} w_{i}}\) [32]. Confidence intervals based on the normal approximation may easily be converted to HKSJ-adjusted ones [12]. The method has also been generalized to the cases of multivariate meta-analysis and meta-regression [33].

The modified Knapp-Hartung (mKH) method

The HKSJ confidence interval (10) tends to be wider than the one based on the normal approximation (8), since the Student-t quantile is larger than the corresponding normal quantile, while q will tend to be somewhere around unity. However, q may in fact also turn out arbitrarily small, and if \(\sqrt {q}<\frac {z_{(1-\alpha /2)}}{t_{(k-1);(1-\alpha /2)}}\), then the modified interval will be shorter than the normal one, which may be considered counter-intuitive. A simple ad hoc modification to the procedure results from defining

$$ q^{\star} = \max\{1, q \} $$
((11))

[11] and using q instead of q to construct confidence intervals and tests. This will ensure a more conservative procedure. The modification was originally proposed in the meta-regression context, but the simple meta-analysis here constitutes the special case of an “intercept-only” regression.

Note that the PM heterogeneity variance estimator is effectively defined by choosing \(\hat {\tau }^{2}\) such that q (Eq. (9)) is = 1 (or less, if no solution exists) [28, 29], so that for the PM estimator the corresponding q value always equals q = 1.

Simulations

Since a (1 − α) confidence interval is supposed to cover the true parameter value with probability (1 − α), the calibration of such intervals may be checked by repeatedly generating random data based on known parameter values and then determining the empirical frequency with which true values are actually covered [34]. We performed such a Monte Carlo simulation comparing the HKSJ and mKH approaches using the setup that was introduced by IntHout et al. [12]. Data were simulated on a continuous scale, according to the random-effects model described above, with study-specific standard errors s i set to reflect certain scenarios with respect to the relative size of studies and their variation due to estimation uncertainty. Many meta-analyses are based on (discrete) count data, where the random-effects model assumptions only hold to an approximation that works well unless event probabilities are very low. Alternative methods have been proposed to deal with low event probabilities [35, 36], but low-event-rate effects were not considered in the present investigation. These simulations considered four different scenarios, namely meta-analyses (A) with trials of equal size, (B) with equally sized trials but including one small trial, (C) with 50 % large and small trials, and (D) equally sized trials and one large trial. The sizes of “small” and “large” trials (and hence, squared standard errors \({s_{i}^{2}}\)) here differ by a factor of ten, so that the associated standard errors differ by roughly a factor of 3; for more details see ([12], Appendix 2). Numbers of studies k considered here are in the range of 2–11, and the true levels of heterogeneity were I 2{0.00,0.25,0.50,0.75,0.90}. At each combination of parameters 10 000 meta-analyses were simulated. All simulations were performed using R [37].

Results

Simulations

Figure 1 shows the estimated coverage probabilities of the different confidence intervals based on the DL heterogeneity variance estimate. The corresponding figure for REML and PM look essentially the same, which is in line with the findings in [38], Friede T, Röver C, Wandel S, Neuenschwander B, Meta-analysis of few small studies in orphan diseases, Submitted. The HKSJ method works very well when the analyzed studies are of equal size (i.e., have equal standard errors), as can also be shown analytically [39], but coverage decreases in more imbalanced settings, especially for small numbers of studies. For the case of no heterogeneity (I 2=0) the HKSJ method also works fine, but if τ 2 was known a priori, in this case the fixed-effect model should work as well. The mKH procedure on the other hand is rather conservative for small k, but does not tend to inflate the type-I error substantially regardless of the underlying study sizes or true heterogeneity. For both methods, the dependence on the amount of heterogeneity is mostly a matter of whether τ 2 is =0 or >0.

Fig. 1
figure 1

Coverage probabilities of HKSJ and mKH 95 % confidence intervals. Probabilities are shown in dependence of the number k of studies and the amount of heterogeneity I 2. The four different scenarios A–D correspond to different amounts of imbalance between the study-specific standard errors s i . The DerSimonian-Laird (DL) method was used for estimation of the heterogeneity variance τ 2

Application of the modification obviously tends to widen the resulting confidence intervals. The ratio of interval lengths (which is equal to \(\frac {\sqrt {q^{\star }}}{\sqrt {q}}\)) is shown in Fig. 2. Most notably, the effect is largest for small heterogeneity and for few studies.

Fig. 2
figure 2

Ratios of lengths of HKSJ and mKH confidence intervals. (Same as \(\frac {\sqrt {q^{\star }}}{\sqrt {q}}\)). Lengths are shown in dependence of the number k of studies and the amount of heterogeneity I 2. Numbers are averaged over all 4 scenarios. The DL method was used for τ 2 estimation

The modification eventually only makes a difference in those cases where q turns out smaller than one. The fraction of intervals affected by the modification ranges between 31 % and 82 % for the DL heterogeneity variance estimator and for the scenarios investigated here, with an overall average of 61 %. For the other two estimators the fractions are 29–82 % with a mean of 62 % (REML), and 31–91 % with a mean of 74 % (PM). Again, the differences between the different estimators are rather small. With respect to the underlying simulation scenario, the probability decreases with increasing heterogeneity, since larger heterogeneity also leads to larger values of q.

Application to JIA example data

Hinks et al. [23] studied the occurrence of a particular genetic variant, CCR5, in juvenile idiopathic arthritis (JIA) patients in comparison with the general population. Their investigation included a meta-analysis of a small number (k=3) of available controlled studies looking into the association of JIA with this particular biomarker. The analysis was based on logarithmic odds ratios; the three estimates along with their standard errors are shown in a forest plot in Fig. 3. Here the largest standard error is 50 % larger than the smallest one. For these data, all three (DL, REML and PM) heterogeneity variance estimates turn out as \(\hat {\tau }^{2}\,=\,0\). A zero heterogeneity variance estimate is not uncommon, even when the actual heterogeneity is in fact substantial [14]. The associated q value is also small with q=0.31 (\(\sqrt {q}=0.55\)). The resulting three confidence intervals based on normal approximation, HKSJ adjustment and mKH method all differ in their lengths. The mKH interval is longest, and includes the zero log odds ratio (indicating no association between genetic marker and disease), while the other two intervals do not include zero. So the choice of procedure directly affects conclusions in this example.

Fig. 3
figure 3

Hinks et al. (2010) example data. A forest plot illustrating the three estimated log odds ratios and 95 % confidence intervals for the example data due to Hinks et al. [23]. The estimated amount of heterogeneity variance (according to DL, REML and PM estimators) is zero here. At the bottom the three alternative combined estimates based on normal approximation, HKSJ and mKH approach are shown

In this example, the three standard errors are rather similar (the largest is 50 % larger than the smallest), but while the heterogeneity variance estimates \(\hat {\tau }^{2}\) are all at zero, the 95 % Q-profile confidence interval [40] ranges up to τ 2=0.332, corresponding to I 2=0.90. Within the context of our simulations (Fig. 1), we cannot tell from the data which of the heterogeneity (I 2) scenarios we are in fact in, and as the standard errors are not exactly the same, it remains a matter of balancing the potential consequences whether one would rather risk losing on the side of (type I) error probability or power.

Conclusions

The HKSJ procedure ensures the coverage probability only when the included studies’ standard errors s i are similar; for unbalanced settings, the actual error probability tends to exceed the targeted one. With the standard definition of the correction factor q the results may sometimes be counterintuitive, since the corresponding CIs may turn out shorter than using the simple normal approximation; in fact they may get arbitrarily short. In case of no heterogeneity (τ=0) the HKSJ method also works well, however practically this is of limited relevance, as one can rarely tell (or convincingly argue) whether this condition holds.

The ad hoc modification of the mKH method aims at fixing these shortcomings and results in type-I error probabilities that are not grossly in excess of the pre-specified ones. Especially when the standard errors s i are of dissimilar magnitude, the mKH method can therefore be recommended. For few studies (small k), the modified procedure however tends to be very conservative, with very small error probabilities especially in the extreme case of meta-analysis of only k = 2 studies. In this extreme case the choice of methods may therefore be considered a matter of a power vs. type-I error probability tradeoff.

While meta-analyses of few studies are a particular problem in indications where there is only little evidence available (such as rare diseases), such circumstances are not as uncommon as one might expect. Turner et al. [13] and Kontopantelis et al. [14] investigated the analyses archived in the Cochrane Database and actually found a majority of them to be based on as few as k = 2 or k = 3 studies; so these constitute highly relevant cases for which the proper control of error rates is crucial.

The properties of either unmodified or modified method for the extreme case of k = 2 may be considered unsatisfactory, as it seems one has the choice of either falling short of or exceeding the targeted error probability; the problem has in fact been regarded as effectively unsolved [41]. The poor behaviour may be explained by the fact that performing a random-effects meta-analysis effectively means the estimation of first- and second-order statistics, and it is not overly surprising to find that this is a hard task when the data consist of as few as two samples that are only measured with uncertainty. Bearing this in mind, the use of Bayesian methods [42] and the consideration of external evidence on the likely magnitude of the heterogeneity [43] may be the way forward.

Abbreviations

DL:

DerSimonian-Laird

HKSJ:

Hartung-Knapp-Sidik-Jonkman

JIA:

juvenile idiopathic arthritis

mKH:

modified Knapp-Hartung

PM:

Paule-Mandel

REML:

restricted maximum likelihood

References

  1. Hedges LV, Olkin I. Statistical Methods for Meta-analysis. San Diego, CA, USA: Academic Press; 1985.

    Google Scholar 

  2. Hartung J, Knapp G, Sinha BK. Statistical Meta-analysis with Applications. Hoboken, NJ, USA: Wiley; 2008.

    Book  Google Scholar 

  3. Viechtbauer W. Bias and efficiency of meta-analytic variance estimators in the random-effects model. J Educ Behav Stat. 2005; 30(3):261–93. doi:10.3102/10769986030003261.

    Article  Google Scholar 

  4. Sidik K, Jonkman JN. A comparison of heterogeneity variance estimators in combining results of studies. Stat Med. 2007; 26(9):1964–81. doi:10.1002/sim.2688.

    Article  PubMed  Google Scholar 

  5. Panityakul T, Bumrungsup C, Knapp G. On estimating heterogeneity in random-effects meta-regression: A comparative study. J Stat Theory and Appl. 2013; 12(3):253–65. doi:10.2991/jsta.2013.12.3.4.

    Google Scholar 

  6. Veroniki AA, Jackson D, Viechtbauer W, Bender R, Bowden J, Knapp, G et al.Methods to estimate the between-study variance and its uncertainty in meta-analysis. Res Synth Methods. 2015. doi:10.1002/jrsm.1164.

  7. Follmann DA, Proschan MA. Valid inference in random effects meta-analysis. Biometrics. 1999; 55(3):732–7. doi:10.1111/j.0006-341X.1999.00732.x.

    Article  CAS  PubMed  Google Scholar 

  8. Hartung J, Knapp G. On tests of the overall treatment effect in meta-analysis with normally distributed responses. Stat Med. 2001; 20(12):1771–82. doi:10.1002/sim.791.

    Article  CAS  PubMed  Google Scholar 

  9. Hartung J, Knapp G. A refined method for the meta-analysis of controlled clinical trials with binary outcome. Stat Med. 2001; 20(24):3875–89. doi:10.1002/sim.1009.

    Article  CAS  PubMed  Google Scholar 

  10. Sidik K, Jonkman JN. A simple confidence interval for meta-analysis. Stat Med. 2002; 21(21):3153–9. doi:10.1002/sim.1262.

    Article  PubMed  Google Scholar 

  11. Knapp G, Hartung J. Improved tests for a random effects meta-regression with a single covariate. Stat Med. 2003; 22(17):2693–710. doi:10.1002/sim.1482.

    Article  PubMed  Google Scholar 

  12. IntHout J, Ioannidis JPA, Borm GF. The Hartung-Knapp-Sidik-Jonkman method for random effects meta-analysis is straightforward and considerably outperforms the standard DerSimonian-Laird method. BMC Med Res Method. 2014; 14:25. doi:10.1186/1471-2288-14-25.

    Article  Google Scholar 

  13. Turner RM, Davey J, Clarke MJ, Thompson SG, Higgins JPT. Predicting the extent of heterogeneity in meta-analysis, using empirical data from the Cochrane Database of Systematic Reviews. Int J Epidemiol. 2012; 41(3):818–27. doi:10.1093/ije/dys041.

    Article  PubMed Central  PubMed  Google Scholar 

  14. Kontopantelis E, Springate DA, Reeves D. A re-analysis of the Cochrane Library data: The dangers of unobserved heterogeneity in meta-analyses. PLoS ONE. 2013; 8(7):69930. doi:10.1371/journal.pone.0069930.

    Article  Google Scholar 

  15. Higgins JPT, Thompson SG. Controlling the risk of spurious findings from meta-regression. Stat Med. 2004; 23(11):1663–82. doi:10.1002/sim.1752.

    Article  PubMed  Google Scholar 

  16. European Commission. Communication from the Commission on Regulation (EC) No 141/2000 of the European Parliament and of the Council of 16 December 1999 on orphan medicinal products. Off J Eur Union. 2003; 46(C178):2–8.

    Google Scholar 

  17. Gagne JJ, Thompson L, O’Keefe K, Kesselheim AS. Innovative research methods for studying treatments for rare diseases: methodological review. BMJ. 2014; 349:6802. doi:10.1136/bmj.g6802.

    Article  Google Scholar 

  18. European Medicines Agency (EMEA). Guideline on clinical trials in small populations. CHMP/EWP/83561/2005. 2006. http://www.ema.europa.eu/docs/en_GB/document_library/Scientific_guideline/2009/09/WC500003615.pdf.

  19. Korn EL, McShane LM, Freidlin B. Statistical challenges in the evaluation of treatments for small patient populations. Sci Transl Med. 2013; 178. doi:10.1126/scitranslmed.3004018.

  20. Kesselheim AS, Myers JA, Avorn J. Characteristics of clinical trials to support approval of orphan vs nonorphan drugs for cancer. J Am Med Assoc. 2011; 305(22):2320–6. doi:10.1001/jama.2011.769.

    Article  CAS  Google Scholar 

  21. IntHout J, Ioannidis JPA, Borm GF, Goeman JJ. Small studies are more heterogeneous than large ones: a meta-meta-analysis. J Clin Epidemiol. 2015; 68(8):860–9. doi:10.1016/j.jclinepi.2015.03.017.

    Article  PubMed  Google Scholar 

  22. Prakken B, Albani S, Martini A. Juvenile idiopathic arthritis. The Lancet. 2011; 377(9783):2138–149. doi:10.1016/S0140-6736(11)60244-4.

    Article  Google Scholar 

  23. Hinks A, Martin P, Flynn E, Eyre S, Packham J, Barton A, et al.Association of the CCR5 gene with juvenile idiopathic arthritis. Genes & Immunity. 2010; 11(7):584–9. doi:10.1038/gene.2010.25.

    Article  CAS  Google Scholar 

  24. Higgins JPT, Thompson SG. Quantifying heterogeneity in a meta-analysis. Stat Med. 2002; 21(11):1539–58. doi:10.1002/sim.1186.

    Article  PubMed  Google Scholar 

  25. DerSimonian R, Laird N. Meta-analysis in clinical trials. Control Clin Trials. 1986; 7(3):177–88. doi:10.1016/0197-2456(86)90046-2.

    Article  CAS  PubMed  Google Scholar 

  26. Böhning D, Malzahn U, Dietz E, Schlattmann P, Viwatwongkasem C, Biggeri A. Some general points in estimating heterogeneity variance with the DerSimonian-Laird estimator. Biostat. 2002; 3(4):445–57. doi:10.1093/biostatistics/3.4.445.

    Article  Google Scholar 

  27. Raudenbush SW. Analyzing effect sizes: random-effects models In: Cooper H, Hedges LV, Valentine JC, editors. The handbook of research synthesis and meta-analysis. New York: Russell Sage Foundation: 2009. p. 295–315.

    Google Scholar 

  28. Paule RC, Mandel J. Consensus values and weighting factors. J Res Natl Bur Stand. 1982; 87(5):377–85.

    Article  Google Scholar 

  29. Rukhin AL, Biggerstaff BJ, Vangel MG. Restricted maximum-likelihood estimation of a common mean and the Mandel-Paule algorithm. J Stat Plan Infer. 2000; 83(2):319–30. doi:10.1016/S0378-3758(99)00098-1.

    Article  Google Scholar 

  30. Viechtbauer W. Conducting meta-analyses in R with the metafor package. J Stat Soft.2010;36(3) doi:10.18637/jss.v036.i03.

  31. Schwarzer G. Meta: Meta-analysis with R. R package version 3.7-1. 2014. http://CRAN.R-project.org/package=meta.

  32. Hartung J. An alternative method for meta-analysis. Biom J. 1999; 41(8):901–16. doi:http://dx.doi.org/10.1002/(SICI)1521-4036(199912)41:8%3C901:AID-BIMJ901%3E3.0.CO;2-W.

    Article  Google Scholar 

  33. Jackson D, Riley RD. A refined method for multivariate meta-analysis and meta-regression. Stat Med. 2014; 33(4):541–54. doi:10.1002/sim.5957.

    Article  PubMed Central  PubMed  Google Scholar 

  34. Dawid AP. The well-calibrated Bayesian. J Am Stat Assoc. 1982; 77(379):605–10. doi:10.1080/01621459.1982.10477856.

    Article  Google Scholar 

  35. Bradburn MJ, Deeks JJ, Berlin JA, Localio AR. Much ado about nothing: a comparison of the performance of meta-analytical methods with rare events. Stat Med. 2007; 26(1):53–77. doi:10.1002/sim.2528.

    Article  PubMed  Google Scholar 

  36. Kuß O. Statistical methods for meta-analyses including information from studies without any events—add nothing to nothing and succeed nevertheless. Stat Med. 2015; 34(7):1097–116. doi:10.1002/sim.6383.

    Article  PubMed  Google Scholar 

  37. R Core Team. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing; 2014. http://www.r-project.org/.

    Google Scholar 

  38. Sánchez-Meca J, Marín-Martínez F. Confidence intervals for the overall effect size in random-effects meta-analysis. Psychol Methods. 2008; 13(1):31–48. doi:10.1037/1082-989X.13.1.31.

    Article  PubMed  Google Scholar 

  39. Sidik K, Jonkman JN. Authors’ reply. Stat Med. 2004; 23(1):159–62. doi:10.1002/sim.1729.

    Article  Google Scholar 

  40. Viechtbauer W. Confidence intervals for the amount of heterogeneity in meta-analysis. Stat Med. 2007; 26(1):37–52. doi:10.1002/sim.2514.

    Article  PubMed  Google Scholar 

  41. Gonnermann A, Framke T, Großhennig A, Koch A. No solution yet for combining two independent studies in the presence of heterogeneity. Stat Med. 2015; 34(16):2476–80. doi:10.1002/sim.6473.

    Article  PubMed Central  PubMed  Google Scholar 

  42. Sutton AJ, Abrams KR. Bayesian methods in meta-analysis and evidence synthesis. Stat Methods Med Res. 2001; 10(4):277–303. doi:10.1177/096228020101000404.

    Article  CAS  PubMed  Google Scholar 

  43. Turner RM, Jackson D, Wei Y, Thompson SG, Higgins PT. Predictive distributions for between-study heterogeneity and simple methods for their application in Bayesian meta-analysis. Stat Med. 2015; 34(6):984–98. doi:10.1002/sim.6381.

    Article  PubMed Central  PubMed  Google Scholar 

  44. Innovative methodology for small populations research (InSPiRe). http://www.warwick.ac.uk/inspire.

Download references

Acknowledgements

This work has received funding from the European Union’s Seventh Framework Programme for research, technological development and demonstration under grant agreement number FP HEALTH 2013-602144 through the InSPiRe project [44].

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Christian Röver.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

TF conceived the concept of this study. CR carried out the simulations and drafted the manuscricpt. GK critically reviewed and made substantial contributions to the manuscript. All authors commented on and approved the final manuscript.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License(http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Röver, C., Knapp, G. & Friede, T. Hartung-Knapp-Sidik-Jonkman approach and its modification for random-effects meta-analysis with few studies. BMC Med Res Methodol 15, 99 (2015). https://doi.org/10.1186/s12874-015-0091-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12874-015-0091-1

Keywords