Elsevier

Journal of Clinical Epidemiology

Volume 69, January 2016, Pages 179-184.e5
Journal of Clinical Epidemiology

Review Article
Use of quality assessment tools in systematic reviews was varied and inconsistent

https://doi.org/10.1016/j.jclinepi.2015.06.023Get rights and content

Abstract

Objectives

To assess the use of quality assessment tools among a cross-section of systematic reviews (SRs) and to further evaluate whether quality was used as a parameter in the decision to include primary studies within subsequent meta-analysis.

Study Design and Setting

We searched PubMed for SRs (interventional, observational, and diagnostic) published in Core Clinical Journals between January 1 and March 31, 2014.

Results

Three hundred nine SRs were identified. Quality assessment was undertaken in 222 (71.8%) with isolated use of the Cochrane risk of bias tool (26.1%, n = 58) and the Newcastle-Ottawa Scale (15.3%, n = 34) most common. A threshold level of primary study quality for subsequent meta-analysis was used in 12.9% (40 of 309) of reviews. Overall, fifty-four combinations of quality assessment tools were identified with a similar preponderance of tools used among observational and interventional reviews. Multiple tools were used in 11.7% (n = 36) of SRs overall.

Conclusion

We found that quality assessment tools were used in a majority of SRs; however, a threshold level of quality for meta-analysis was stipulated in just 12.9% (n = 40). This cross-sectional analysis provides further evidence of the need for more active or intuitive editorial processes to enhance the reporting of SRs.

Introduction

What is new?

Key findings

  1. There is significant variation in the quality assessment tools used in systematic reviews (SRs) published in the Core Clinical Journals.

  2. A threshold level of study quality is relatively infrequently described as a prerequisite before undertaking meta-analysis.

What this adds to what was known?
  1. This cross-sectional analysis highlights the variation in methodology used by authors of interventional, observational, and diagnostic test reviews to grade study quality.

  2. There is a lack of clarity in relation to the specific level of quality necessary for inclusion in SRs.

What is the implication and what should change now?
  1. This study provides further evidence of the need for more active or intuitive editorial processes to enhance the reporting of SRs.

  2. Greater awareness, consistency, and acceptance of specific quality assessment tools are needed across SR types.

Systematic reviews (SRs) are engrained within evidence-based medicine being trusted to combine and appraise best available evidence in a robust and consistent manner. SRs should involve “a systematic approach to minimizing biases and random errors which is documented in a materials and methods section” [1]. Bias has been defined as a tendency for results of a primary study to differ from the results expected from a randomized trial, conducted on the same participant group that had no design flaws in its conduct. A key element in the SR process is appraisal of the risk of bias or methodological quality of the constituent primary studies to decide whether their results can be trusted and should contribute to eventual meta-analyses. Failure to make objective decisions concerning the inclusion of various study designs and the threshold level of quality required for a study to be included can introduce bias in the review process; moreover, inclusion of less robust studies risks leading to unreliable, often inflated estimates of treatment effects [2], [3].

In recent years, the Cochrane risk of bias tool has become established in the assessment of risk of bias in randomized designs [4]. However, until now, significant variation has existed in relation to tools used in nonrandomized designs [5]. Sanderson et al. identified 86 tools comprising 53 checklists and 33 scales in an electronic search undertaken in March 2005, with, for example, 23 unique tools developed to assess cohort studies and 19 for case–control studies. Most tools included items related to study variables (86%), design-related bias (86%), and confounding (78%), although certain aspects such as conflict of interest were underrepresented (4%) [5]. Similar results have been identified within SRs concerning the epidemiology of chronic disease with only 55% of reviews referring to quality assessment of primary studies overall [6]. More recently, the Cochrane Collaboration has espoused a new tool designed to appraise the risk of bias inherent in nonrandomized studies [7], although it has yet to find wider application. Similarly, in an assessment of interventional SRs, Moher et al. [8] identified 25 unique tools used in the evaluation of methodological quality of clinical trials. Many of these were found to lack accepted theoretical basis and varied in respect of their size, complexity, and regarding the weight attributed to individual methodological aspects such as randomization procedures or blinding. This assessment predated the development of the Cochrane risk of bias tool [4], however, but was echoed in a more recent analysis of interventional SRs within the field of general health research and physical therapy [9].

Notwithstanding the disagreement in respect of the appropriate tool to assess methodological quality, there is also variation regarding the threshold level of primary study quality or risk of bias above which studies may be included in quantitative synthesis. The Cochrane Collaboration, for example, advocates a domain-based assessment of bias and has suggested that a high risk of bias for any one of the individual domains would render a primary study at high risk of bias overall [4]. Consequently, a binary decision on whether to include a study in a meta-analysis may be made. A method of accounting for variations in study quality is the use of sensitivity analysis, whereby assessment of the overall findings to subjective decisions including inclusion of studies of varying design and quality [10].

The objectives of our study were to assess the frequency with which various quality assessment tools were used within SRs, to investigate whether threshold quality levels were applied when using these tools governing possible inclusion in meta-analysis, and to assess the use of sensitivity analysis based on methodological quality or risk of bias of primary studies within meta-analysis.

Section snippets

Data sources and eligibility

We included all types of SRs within the Core Clinical Journals in MEDLINE via PubMed from January 1, 2014, to March 31, 2014, using the command ‘jsubsetaim[All Fields]’ with search filters to identify meta-analyses and SRs. The Abridged Index Medicus or Core Clinical Journals is an online journal index encompassing 118 journals involving all clinical medicine and public health specialties. An a priori sample size calculation was not undertaken. Article types including narrative reviews or other

Results

We included 309 reviews in the analysis overall (Fig. 1). Of these, 210 were interventional, 84 were observational, and 15 were diagnostic test reviews. Within these, some form of quality or risk of bias assessment was undertaken in 222 reviews (71.8%, Table 1). The tools used most commonly in isolation were the Cochrane risk of bias tool (26.1%) and the Newcastle-Ottawa Scale (NOS) (15.3%, Appendix Table 1 at www.jclinepi.com). A threshold level of bias or primary study quality was stipulated

Main findings

A quality assessment tool was used in 71.8% (222 of 309) of SRs with the use of tools being equally prevalent among interventional, diagnostic, and epidemiologic reviews. Significant variation in the assessment tools used was observed with 58 different combinations identified. Typically, a single assessment tool was relied on (60.2%) although multiple tools were also occasionally used (11.6%). A threshold level of quality or bias was, however, prespecified within just 12.9% of reviews (40 of

Conclusion

In a representative sample of biomedical SRs, quality assessment tools were used in 72% of reviews with a threshold level of quality stipulated in just 13%. This cross-sectional analysis, therefore, provides further evidence of the need for active or intuitive editorial or publication processes to enhance the reporting of SRs.

References (22)

  • S. Sanderson et al.

    Tools for assessing quality and susceptibility to bias in observational studies in epidemiology: a systematic review and annotated bibliography

    Int J Epidemiol

    (2007)
  • Cited by (42)

    • Morphologic Changes of Cervical Musculature in Relation to Chronic Nonspecific Neck Pain: A Systematic Review and Meta-Analysis

      2022, World Neurosurgery
      Citation Excerpt :

      The NOS scale was used to evaluate the quality of all included studies combined with the Agency for Healthcare Research and Quality criteria (good, fair, and poor).38 The NOS scale is chosen to evaluate included studies because it is a quality evaluation tool for case–control studies, cohort studies, and cross-sectional studies, and it is validated and adaptable.39,40 The divergences in the evaluation process are resolved through discussion.

    • Microscopic changes in the spinal extensor musculature in people with chronic spinal pain: a systematic review

      2022, Spine Journal
      Citation Excerpt :

      The methodological quality of the included studies was assessed by two reviewers (SP and AS) independently using the Newcastle-Ottawa Scale (NOS) [42]. As there is no consensus on the optimal study quality or risk of bias (RoB) tools for observational studies [43], the NOS was chosen because it is validated, adaptable, and quick to complete [44,45]. Since all of the included studies were case-control studies, the NOS scale was modified to match the design of the evidence (Appendix 4).

    • Authors' response

      2020, American Journal of Orthodontics and Dentofacial Orthopedics
    View all citing articles on Scopus
    View full text