Skip to content
Publicly Available Published online by De Gruyter April 16, 2024

Using analytical performance specifications in a medical laboratory

  • Graham Ross Dallas Jones EMAIL logo

Abstract

Analytical performance specifications (APS) are used for the quantitative assessment of assay analytical performance, with the aim of providing information appropriate for clinical care of patients. One of the major locations where APS are used is in the routine clinical laboratory. These may be used to assess and monitor assays in a range of settings including method selection, method verification or validation, external quality assurance, internal quality control and assessment of measurement uncertainty. The aspects of assays that may be assessed include imprecision, bias, selectivity, sample type, analyte stability and interferences. This paper reviews the practical use of APS in a routine clinical laboratory, using the laboratory I supervise as an example.

Introduction

Our primary goal in the routine clinical pathology laboratory is to provide results which support medical decision making for the benefit of the patient. The issuing of results which do not reflect the true value of a substance in the patient (e.g. concentration of sodium in the circulation) has the potential to harm patients by supporting incorrect decisions. There are many decisions made in the routine laboratory to provide results which are fit for their clinical need. Analytical Performance Specifications (APS) can provide guidance on the amount of error that is acceptable. There are a range of factors, mainly analytical and pre-analytical, where a quantitative assessment of the required quality is needed (Table 1), and a number procedures undertaken in a laboratory where these factors are assessed (Table 2). This paper aims to describe the use of APS in the routine laboratory. Please note that this is a personal viewpoint and represents actions in the routine clinical chemistry laboratory under my supervision. I am certain there will be many alternate approaches to the use of APS in other laboratories and also many areas for improvement in our laboratory.

Table 1:

Factors that may affect quantitative results.

Assay bias
Assay imprecision
Assay selectivity
Assay interferences
Analyte stability
Sample container type
Table 2:

Laboratory procedures where analytical performance specifications may be applied.

External quality assurance
Internal quality control
Method/instrument selection
Method/instrument verification (for kit methods)
Method/instrument validation (for in-house methods or changes to kit methods)
Measurement uncertainty assessment
Result change protocol

There is now a long history of the processes to establish APS, with the Milan consensus document providing the higher level concepts in current use [1]. The three “Milan models” are 1. Effect on clinical decision making, 2. Based on biological variation and 3. State of the art. These concepts will be used in the discussion below.

By way of background, pathology laboratories in Australia have been subject to compulsory accreditation for many years. This is carried out by the National Association of Testing Authorities (NATA) using ISO 15189 [2] as the overarching standard, as well as Australian standards from the Australian Commission on Safety and Quality in Healthcare [3]. Laboratories in other countries may have different regulatory systems and other influences on their routine practice.

It is my experience that there is a wide variation in how different laboratories carry out any activities. These differences may be the result of careful local thought with good intentions. It may also be that local decisions are made without awareness of wider issues or recommendations. I follow up this statement with another opinion, that in the absence of evidence to the contrary, assume that laboratories act in different ways. The evidence for similar actions and outcomes comes from External Quality Assurance (EQA) activities. Following on from this, if we want laboratories to act in similar ways, this will not happen without common decision-making processes, for example in the preparation of laboratory standards, and assessment against those standards. Given these statements of belief (spoken only half in jest), we can assume that laboratories used APS in different ways in their processes decision-making. As indicated above, this paper discusses the ways APS are used in decision making in our laboratory.

Before proceeding further, it is necessary to explore what are possible sources for APS for use in the routine laboratory. APS are commonly established externally by a professional body. In practice for our laboratory, and for most Australian laboratories, this usually means APS from the Royal College of Pathologists of Australasia Quality Assurance Program (RCPAQAP). These have been in development over many years [4] and are freely available at the RCPAQAP website [5]. APS from other EQA organisations may be used, for example if enrolled on other programs, or from other professional bodies such as the CDC lipid program [6] or CDC steroid program [7]. Additionally, we make decisions using “in-house” APS. These are usually based on either within-subject biological variation (CVI) for assay imprecision (CVA) (Milan level 2) or state of the art (Milan level 3). I describe these as “in-house” as the quality standard (e.g. optimal, desirable, minimal, other), the data set and clinical decision for use are not prescribed externally. Before proceeding further, when considering the use of biological variation as a standard, it is important to recognize the tremendous work of the EFLM working group on biological variation in ensuring the availability of high quality data in this area [8, 9]. I also note in this paper using the concept of “state of the art” in a looser manner than the specific published definitions or opinions proposed for establishment of formal APS [10], referring to any comparison of laboratory analytical performance with that obtained in other routine laboratories.

Laboratory activities utilising (APS)

External quality assurance

External Quality Assurance (EQA) is a required laboratory activity under ISO 15189 and for numerical pathology results a quantitative assessment is made comparing a laboratory’s result against a target. The difference between these values is assessed and an APS provided by the EQA program is the usual tool for this process. The most commonly issued EQA program in Australia is the RCPAQAP where APS have been determined based on a combination of biological variation and state of the art [4]. The APS are applied in multiple locations in the EQA reports, including graphical reporting of recent results, monitoring trends over time and across the concentration range, and in summary reports highlighting findings (Figure 1). The use of APS in this setting can be seen as somewhat analogous to the use population reference intervals to quickly scan a pathology report. In the same way that reference intervals do not define the presence or absence of disease, APS usually do not define the presence of absence of a clinically important deviation in assay performance, but they do allow a rapid assessment of the report for results that should be given more consideration. When using an APS it is important to understand the basis for their development and how they should be used. An EFLM Task and Finish group following after the Milan consensus conference identified six elements needed to describe APS for use in EQA and to allow appropriate use and comparison with APS between different programs [11]. The APS provided in most RCPAQAP programs are used for total error (i.e. are applied to individual results), are “aspirational” (i.e. are aiming to improve laboratory testing, and are set on a combination of biological variation and state of the art. The approximate definition of state of the art used is that about 80 % of laboratories can achieve the standard [4, 10]. The effect of using APS set in this way should be to improve the less well-performing laboratories.

Figure 1: 
Example EQA report from the RCPAQAP showing the role of APS in graphical displays (highlighted in image by added dark arrows). (A) Frequency histogram, (B) Youden plot, (C) time sequence, (D) concentration plot. Reproduced with permission from the RCPAQAP.
Figure 1:

Example EQA report from the RCPAQAP showing the role of APS in graphical displays (highlighted in image by added dark arrows). (A) Frequency histogram, (B) Youden plot, (C) time sequence, (D) concentration plot. Reproduced with permission from the RCPAQAP.

The choice of APS to use in any setting is a key component of their use. While it is a dated reference, the paper from Ricos et al. in 1996 showed extreme differences between the APS used by different EQA providers in Europe with the widest APS typically 3 to 6 times wider than the narrowest [12]. As a wise statistician once said: never use a statistical test you do not understand. A relevant take on this would be advice to laboratories to “never use an APS you do not understand”.

One recent use of specific APS based on single components of the Milan models is analysis of the RCPAQAP Liquid Serum Chemistry program. This program uses freshly prepared, unaltered serum which is rapidly distributed to over 200 laboratories in Australia, which can be assumed to be commutable [13]. Using imprecision criteria relative to biological variation, there are two measurands (triglycerides and iron) for which the total country CV (CVC, the CV of results from all participating laboratories) meet the optimal level (CVC<0.25 × CVI), 8 measurands which meet the desirable level (CVC<0.5 × CVI) and another 13 which meet the minimal level (CVC<0.75 × CVI). With this knowledge, laboratories can advise that patients can be monitored across the country with these tests using results from any laboratories. For tests not meeting these standards, variation in results in a patient is likely to reflect between-laboratory differences rather than variation in the patient. This information, the combination of data with appropriate APS, allows us to answer clinical queries regarding questions about interpreting results derived from different laboratories. Australia has recommended using common reference intervals for a number of measurands [14]. The RCPAQAP Liquid serum chemistry program allows assessment of acceptable between-method bias using criteria relative to the reference interval width, which is a combination of within and between person biological variation combined with analytical variation. This data has supported the recommended reference intervals for the majority of method in routine use and identified methods which should be reviewed [13]. This information is vital in supporting decisions by laboratories when they consider adoption of these recommended common reference intervals.

It is worth noting that EQA programs, by design, include a state-of-the-art component. Any graphical output of results or of calculated performance (e.g. CV) allows visual assessment of results to answer the question “am I performing similarly to other laboratories”. Similar to most other EQA programs, the RCPAQAP provides z-scores for individual results, providing a quantitative state of the art assessment for results, consistent with a Milan 3 approach.

Internal quality control

The role of internal quality control (QC) is to ensure adequate assay performance in real time, and therefore, by definition, requires an assessment of what constitutes such performance. There is a wide-ranging literature on the assessment of assay performance against APS for total analytical error (TAE), and then the implementation of QC protocols based on that performance. This can be by calculation of sigma values (assay CV/(TAE – bias)) or assay capability (Assay CV/TAE) [15, 16]. In this setting the TAE APS are based on the concept that results deviation from the target by more that this value should be identified and avoided. An assay with a higher sigma value (e.g. 5 or 6), requires fewer QC samples, fewer QC rules and less stringent rules. An assay with a lower sigma value requires the opposite, with the concept that more effort should be taken to identify smaller changes sooner after they may have occurred. Clearly the sigma values, and therefore the QC practices based on these, are completely dependent on the value of the APS with a wider APS, for example regulatory APS such as CLIA guidelines used by CAP in the USA leading to different practices from labs setting sigma values on tighter limits [17]. Despite many years of personal interest in QC theory and practice [15, 18], our current laboratory practice is largely guided by “in-house” APS. The first comparison of achieved assay CV (i.e. during run-in of a QC material, or review over time) is whether the CVA is small relative to CVI (e.g. minimal standard or better). The smaller the ratio of CVA/CVI, the less tight we set the limit in the QC software program (effectively changing the multiple of SD used for flagging). For assays not meeting the minimal standard, the question is are we performing as well as can be expected. Here the comparison is with imprecision data from the manufacturer’s product information, as supplied in the Instructions for Use (IFU) or from EQA programs. If we are performing as well as the majority of other users, this is as good as can be expected. The frequency of QC is at least partly based on convenience, with many analytes using the same QC material, but a major factor is the stability of assays over time. Assays with a higher rate of fluctuation (serum calcium and bicarbonate (total carbon dioxide) come to mind) are tested with QC more frequently, and stable assays, (for example NT-proBNP) require less frequent QC.

Method selection

Assays can be developed in-house (so-called lab developed tests or in-house in vitro diagnostic devices) or they can be purchased from manufacturers (kit methods), with the latter being much more common in the routine laboratory. Analytical performance, e.g. bias, imprecision, reportable range, acceptable sample types, are factors that are included in the purchasing decision. Consideration of these performance factors can also be seen as a way of demonstrating to manufacturers that we value good analytical performance. For better or worse, such analytical considerations are only one factor in a purchase of analytical equipment, with menu, reliability, turn-around time, footprint, software, costs, maintenance requirements and other factors generally pushing down the analytical quality as a decision making factor. Indeed at our last purchase of automated biochemistry equipment, the analytical factors were scored at only 10 % of the total decision-making process. Assigning a higher value to this aspect may provide a better reflection of the importance of this issue.

When assessing equipment with multiple assays, it is important to consider how the analytical performance may affect clinical decision making. For example, with multi-parameter auto analysers, if one manufacturer has all assays with imprecision at the optimal level with respect to CVI, and another is all at the desirable level, the clinical benefit of the better analyser may be minimal. But if the number of poorly-performing assays on one analyser (e.g. CVA significantly greater than the minimal standard) is more than another, more clinical decisions may be affected, making this factor potentially a better selection criteria. The use of APS is important in these considerations, helping identify poorly performance assays based on EQA data or manufacturer’s supplied information.

Method verification/validation

Following purchase of an analytical system, there is a need to verify its performance in your laboratory. The key principle that should be applied here is whether the system is performing in your laboratory in the way that the manufacturer intended. For this purpose, the manufacturer’s claims in the IFU become the de facto APS.

In practice in our laboratory we use a range of APS in addition to those indicated by the manufacturer to assess a range of properties during assay verification.

Imprecision data produced during verification is generally compared with multiple APS. These include the manufacturers expected performance (from the IFU), within-subject biological variation, state of the art (e.g. using imprecision data from the RCPAQAP), and comparison with the previous method in use. While performing as expected for the manufacturer’s method is the pass/fail criteria, understanding how the assay performs against the other criteria can be useful to understand how it will be expected to perform in routine use. Bias relative to the expected performance of the method is generally assessed using QC and EQA samples, using the supplied APS for the QC material, even though these are often very wide, the EQA APS for that material. Bias is also assessed relative to the previous method with a method comparison study along with assessing the effect on flagging rates with reference intervals and clinical decision points. The finding of a significant bias relative to the previous method will lead to a need to change reference intervals, alert clinicians of the difference, and report results on different lines in paper and electronic systems.

Other aspects of performance that can be assessed include interferences (especially the measurable visible interferences e.g. haemolysis, lipaemia, icterus), sample container type (e.g. serum, heparin, gel tubes, EDTA) and analyte stability. When doing experiments to confirm or extend manufacturers claims and make decisions about implementation, APS are needed. In general, in our laboratory we use the RCPAQAP APS as these provide a robust starting point to use for any measurand. An example of sample type assessment in this manner is shown in Figure 2. Two important variations need to be considered. These are the APS used by manufacturers, which are often unstated (“hidden APS”), the need to consider the “accuracy/utility balance”, and the error budget.

Figure 2: 
Sample type comparison using RCPAQAP APS. (A) Urea (mmol/L), (B) urate (umol/L). X – lithium heparin, ◊ – EDTA, dashed lines – APS. For urea, heparin and EDTA considered equivalent to serum, for urate, only heparin acceptable.
Figure 2:

Sample type comparison using RCPAQAP APS. (A) Urea (mmol/L), (B) urate (umol/L). X – lithium heparin, ◊ – EDTA, dashed lines – APS. For urea, heparin and EDTA considered equivalent to serum, for urate, only heparin acceptable.

When a manufacturer’s IFU makes a claim about a factor that may affect the value of results, there must be an APS behind this claim. Examples of such claims are that serum and heparin are both acceptable sample types, that there is no interference up to an haemolysis level of 300 mg/dL (3 g/L) released haemoglobin, or that a sample is stable for up to three days at room temperature. In each case the manufacturer will be relying on experimental data and a change, defined by an APS, is not exceeded by the claimed conditions. I refer to these as “hidden” APS as they are often not specifically referenced by the manufacturer, or that they are not recognized by the reader of the IFU.

Examples from two manufacturers from current IFUs for haemolysis interference in serum potassium are as follows: haemoglobin of 20 mg/dL (0.2 g/L) with a criteria of a change in potassium of ≤0.1 mmol/L; haemoglobin of 50 mg/dL (0.5 g/L) with a criteria of change in potassium of +/− 0.14 mmol/L to 3.5 mmol/L then +/− 4 %. The units of mg/dL are those used by these manufacturers. If a laboratory accepts the stated haemolysis limit, they are accepting the manufacturer’s APS. Consideration is required as to the appropriateness of the APS for use. On many occasions the APS for an IFU claim is not supplied and the laboratory is unaware of the APS used to provide the limitation. On other occasions claims are made with references to external sources. A current claim from one manufacturer for both serum and heparin plasma being acceptable sample types for serum AST references a paper from 1989 where the average difference between these tube types is 28 % at a serum value of 23 U/L [19]. Therefor the unstated APS from this manufacturer for sample type could be seen as this value. There are several risks to good laboratory practice here. A laboratorian might accept the recommendations based on the supplied APS from the manufacturer without considering its relevance, or the laboratorian may accept the limitations without even knowing what performance they are accepting. Additionally the evidence for the limitation is commonly not available.

The next issue is the “accuracy/utility balance”. As shown in Figure 3, accepting tighter APS for these types of decisions is not a pure benefit. For example, if a tight limit for the effect of haemolysis is taken, and potassium results with haemolysis greater than the limit are with-held and a recollection requested. While only more accurate potassium results are released, the need for recollections will subject patients to more venepunctures, take the time of the collector and slow down the delivery of care. A study from the Australian Institute of Health Innovation has shown that every haemolysed sample adds an average of 18 min to every Emergency Department stay [20]. As an example of this trade-off, based on local data for serum potassium, an allowable haemolysis of 20 mg/dL (0.2 g/L, see above) would withhold 25 % of samples from the Emergency Department, a limit of 0.25 mmol/L (based on the RCPAQAP APS) would withhold 15 % of samples, however the limit we have selected of 100 mg/dL (1 g/L, with an effect of approximately 0.4 mmol/L) flags about 5 % of samples. The use of a wider APS in certain settings is a balance between the benefits of accuracy and the costs of not releasing results [21].

Figure 3: 
Illustration of the accuracy-utility balance.
Figure 3:

Illustration of the accuracy-utility balance.

There are of course alternatives to with-holding results [21, 22] and in this case, it is our practice to correct the potassium results for the effect of haemolysis for the Emergency Department, taking into account the uncertainty of this correction process [23].

There is another aspect to take into account as these various source of error are considered. If each source of error, e.g. unidentified bias, interference, analyte stability, tube type etc., is allowed to have an effect up to the selected APS, then in if all of these should occur in the same sample, with an effect in the same direction, the error could be multiples of the APS. The concept here is known as the error budget, where each of the error types must be small relative to the APS to allow for co-exiting errors. The concept of error budget has typically been applied only to the analytical phase [24], however clearly all factors that can affect the numerical value of results should be taken into account. Again there is a cost to this approach and consideration must be given to the likelihood of this occurring, against the cost of adopting very tight APS.

Measurement uncertainty

It is a requirement of ISO 15189 that laboratories calculate the Uncertainty of Measurement (MU) for all tests performed in the laboratory, and in Australia there is an NPAAC standard for this activity [25] in addition to an ISO standard (ISO/TC 20914) [26]. As well as calculating a value, it is necessary to assess whether this is satisfactory for routine use. Again for this purpose we use laboratory-selected APS. If the long term uncertainty of a test is small relative to CVI, then the test is accepted. If this standard is not met, then the performance of the test relative to other routine assays is assessed using EQA data. The performance is accepted if it is similar to that seen in comparison laboratories. These represent the concepts in Milan levels 2 and 3 respectively. Assays which do not meet either of these criteria are the ones that need to be addressed to try and improve performance. On some occasions where biological variation data is not available, for example therapeutic drug testing, we have determined in-house values for CVI against which to assess imprecision requirements [27, 28, 29].

Result change protocol

While there are many steps taken to avoid errors in the results released from the laboratory [30], on some occasions errors are found after a result has been released. For example after a QC failure with correction of the error and re-analysis of the samples, there may need to be correction of previously issued results. There needs to be a consideration of the magnitude of the change in deciding to alert the clinician and change the report. In recognizing that contacting physicians for result changes of a likely low clinical impact has a cost in time and interruption, we have established change limits for most tests to ensure consistent actions at all times. In general these change limits are based on the APS from the RCPAQAP, however many have different changes within the reference interval, on the understanding that a change wholly within the interval is, for many tests, unlikely to change management. For example, we do not reissue changes entirely within the reference interval for serum amylase, lipase, AST, ALT and chloride amongst other tests. This is an example of the accuracy-utility balance where a set of general APS have been modified for a specific purpose.

Conclusions

APS can be used to guide many decisions affecting laboratory performance and are used in many different procedures including method selection, verification, EQA, QC, MU, interferences, sample type and sample stability. In using APS for these decisions, what they mean, how they should be applied, their strengths and limitations and appropriateness for the issue in question. Additionally, laboratories can use the basic components which are used to derive APS (e.g. BV, State of the art) to determine acceptable laboratory performance. Laboratories need to be aware of assumptions and “hidden” APS in manufacturers’ information and consider the accuracy/utility balance and error budget when implementing performance assessment. Additionally, given the high likelihood of variation in laboratory practice, a survey for how laboratories use APS in their activities may allow us to learn from other laboratories about best practice in this field, with such a recent survey indicating that awareness and use of APS in the routine laboratory remains a work in progress [31].

Note: This paper is based on a presentation by the author at the 5th Symposium Cutting Edge of Laboratory Medicine in Europe CELME 2023, Prague, Czech Republic, October 12–13, 2023 Analytical Performance Specifications.


Corresponding author: Graham Ross Dallas Jones, Department of Chemical pathology, SydPath, St Vincent’s Hospital, 390 Victoria St, Darlinghurst, NSW, 2010, Australia; and Facult of Medicine, University of NSW, Kensington, Australia, E-mail:

Acknowledgments

I acknowledge the contributions of Antonella Screnci and Reshma Prasad for their significant contributions to the processes in the laboratory described in this article.

  1. Research ethics: Not applicable.

  2. Informed consent: Not applicable.

  3. Author contributions: The autor has accepted responsibility for the entire content of this manuscript and approved its submission.

  4. Competing interests: The author states no conflict of interest.

  5. Research funding: None declared.

  6. Data availability: Not applicable.

References

1. Sandberg, S, Fraser, FG, Horvath, AR, Jansen, R, Jones, G, Oosterhuis, W, et al.. Defining analytical performance specifications: consensus statement from the 1st strategic conference of the European federation of clinical chemistry and laboratory medicine. Clin Chem Lab Med 2015;53:833–5. https://doi.org/10.1515/cclm-2015-0067.Search in Google Scholar PubMed

2. Medical laboratories – requirements for quality and competence. ISO. International Standards Organisation; 2022, 15189.Search in Google Scholar

3. Pathology accreditation standards. Australian Commission on Safety and Quality in Healthcare. https://www.safetyandquality.gov.au/our-work/accreditation/pathology-accreditation-standards [Accessed 12 Apr 2024].Search in Google Scholar

4. Jones, GRD, Sikaris, K, Gill, J. Allowable limits of performance’ for external quality assurance programs – an approach to application of the Stockholm criteria by the RCPA quality assurance programs. Clin Biochem Rev 2012;33:133–9.Search in Google Scholar

5. RCPAQAP Chemical Pathology Analytical Performance Specifications. https://rcpaqap.com.au/resources/chemical-pathology-analytical-performance-specifications/ [Accessed 21 Apr 2024].Search in Google Scholar

6. CDC Lipid Standardization Program. Centre for disease control and prevention. https://www.cdc.gov/labstandards/csp/lsp.html [Accessed 21 Apr 2024].Search in Google Scholar

7. CDC Hormone Standardization Program for Testosterone and Estradiol. Centre for disease control and prevention. https://www.cdc.gov/labstandards/csp/hs_host.html [Accessed 21 Apr 2024].Search in Google Scholar

8. Aarsand, AK, Røraas, T, Bartlett, WA, Coşkun, A, Carobene, A, Fernandez-Calle, P, et al.. On behalf of the European federation of clinical chemistry and laboratory medicine (EFLM) working group on biological variation. Harmonization initiatives in the generation, reporting and application of biological variation data. Clin Chem Lab Med 2018;56:1629–36. https://doi.org/10.1515/cclm-2018-0058.Search in Google Scholar PubMed

9. Carobene, A, Aarsand, AK, Bartlett, WA, Coskun, A, Diaz-Garzon, J, Fernandez-Calle, P, et al.. The European biological variation study (EuBIVAS): a summary report. Clin Chem Lab Med 2022;60:505–17. https://doi.org/10.1515/cclm-2021-0370.Search in Google Scholar PubMed

10. Borrillo, F, Panteghini, M. State-of-the-art model for derivation of analytical performance specifications: how to define the highest level of analytical performance technically achievable. Clin Chem Lab Med 2024. https://doi.org/10.1515/cclm-2023-1286 [Epub ahead of print].Search in Google Scholar PubMed

11. Jones, GRD, Albarede, S, Kesseler, D, MacKenzie, F, Mammen, J, Pedersen, M, EFLM Task Finish Group. Analytical Performance Specifications for EQAS (TFG-APSEQA), et al.. Analytical performance specifications for external quality assessment – definitions and descriptions. Clin Chem Lab Med 2017;55:949–55. https://doi.org/10.1515/cclm-2017-0151.Search in Google Scholar PubMed

12. Ricos, C, Baadenhuijsen, H, Libeer, J-C, Hyltoft Petersen, P, Stockl, D, Thienpont, L, et al.. External quality assessment: currently used criteria for evaluating performance in European countries, and criteria for future harmonization. Eur J Clin Chem Clin Biochem 1996;34:159–65.Search in Google Scholar

13. Farrell, CJL, Jones, GRD, Sikaris, KA, Badrick, T, Graham, P, Bush, J. Sharing reference intervals and monitoring patients across laboratories – findings from a commutable external quality assurance program. Clin Chem Lab Med 2024. https://doi.org/10.1515/cclm-2024-0041 [Epub ahead of print].Search in Google Scholar PubMed

14. Koerbin, G, Sikaris, K, Jones, GRD, Flatman, R, Tate, JR, On Behalf of the AACB Harmonization Committee for Common Reference Intervals. An update report on the harmonization of adult reference intervals in Australasia. CCLM 2019;57:38–41. https://doi.org/10.1515/cclm-2017-0920.Search in Google Scholar PubMed

15. Jones, GRD, Calleja, J, Chesher, D, Parvin, C, Yundt-Pacheco, J, Mackay, M, et al.. Collective opinion paper on a 2013 AACB workshop of experts seeking harmonisation of approaches to setting a laboratory quality control policy. Clin Biochem Rev 2015;36:2015 87–95.Search in Google Scholar

16. Westgard, JO, Westgard, SA. Quality control review: implementing a scientifically based quality control system. Ann Clin Biochem 2016;53:32–50. https://doi.org/10.1177/0004563215597248.Search in Google Scholar PubMed

17. Keles, M. Evaluation of the clinical chemistry tests analytical performance with sigma metric by using different quality specifications - comparison of analyser actual performance with manufacturer data. Biochem Med 2022;32:010703. https://doi.org/10.11613/bm.2022.010703.Search in Google Scholar PubMed PubMed Central

18. Jones, GRD. Reevaluation of the power of error detection of westgard multirules. Clin Chem 2004;50:762–4. https://doi.org/10.1373/clinchem.2003.025585.Search in Google Scholar PubMed

19. Doumas, BT, Hause, LL, Simuncak, DM, Breitenfeld, D. Differences between values for plasma and serum in tests performed in the Ektachem 700 XR Analyzer, and evaluation of “plasma separator tubes (PST)”. Clin Chem 1989;35:151–3. https://doi.org/10.1093/clinchem/35.1.151.Search in Google Scholar

20. Vecellio, E, Li, L, Mackay, M, Gay, S, Lake, R, Georgiou, A, et al.. A benchmark study of the frequency and variability of haemolysis reporting across pathology laboratories. Australian Institute of Health Innovation; 2015. Available from: https://www.mq.edu.au/__data/assets/pdf_file/0003/680682/Vecellio-et-al-2014-Frequency-and-variability-of-haemolysis-reporting-across-pathology-laboratories-v7c.pdf.Search in Google Scholar

21. Cadamuro, J, Mrazek, C, Haschke-Becher, E, Sandberg, S. To report or not to report: a proposal on how to deal with altered test results in hemolytic samples. Clin Chem Lab Med 2017;55:1109–11. https://doi.org/10.1515/cclm-2016-1064.Search in Google Scholar PubMed

22. Lippi, G, Cervellin, G, Plebani, M. Reporting altered test results in hemolyzed samples: is the cure worse than the disease? Clin Chem Lab Med 2017;55:1112–4. https://doi.org/10.1515/cclm-2017-0011.Search in Google Scholar PubMed

23. Jones, GRD, Hawkins, RC. Correcting laboratory results for the effects of interferences: an approach incorporating uncertainty of measurement. Ann Clin Biochem 2015;52:226–31. https://doi.org/10.1177/0004563214533516.Search in Google Scholar PubMed

24. Badrick, T. Biological variation: understanding why it is so important? Pract lab Med 2021;23:e00199. https://doi.org/10.1016/j.plabm.2020.e00199.Search in Google Scholar PubMed PubMed Central

25. Requirements for the estimation of measurement of uncertainty (2007 edition). Australian Commission on Safety and Quality in Healthcare. Available from: https://www.safetyandquality.gov.au/publications-and-resources/resource-library/requirements-estimation-measurement-uncertainty-2007-editionSearch in Google Scholar

26. Medical laboratories — practical guidance for the estimation of measurement uncertainty. ISO/TS 20914:2019.Search in Google Scholar

27. Pauwels, S, Vermeerch, P, Desmet, K, Vanstapel, F. Practice oriented quality specification for therapeutic drug monitoring. Clin Chem 2014;60:559–60. https://doi.org/10.1373/clinchem.2013.217125.Search in Google Scholar PubMed

28. Jones, GRD. Estimates of within-subject biological variation derived from pathology databases: an approach to allow assessment of the effects of age, sex, time between sample collections, and analyte concentration on reference change values. Clin Chem 2019;65:579–88. https://doi.org/10.1373/clinchem.2018.290841.Search in Google Scholar PubMed

29. Chai, JH, Flatman, R, Teis, B, Sethi, SK, Badrick, T, Loh, TP. Indirect derivation of biological variation data and analytical performance specifications for therapeutic drug monitoring activities. Pathology 2019;51:281–5. https://doi.org/10.1016/j.pathol.2018.12.418.Search in Google Scholar PubMed

30. Çubukçu, HC, Vanstapel, F, Thelen, M, Bernabeu-Andreu, FA, van Schrojenstein Lantman, M, Brugnoni, et al.. Improving the laboratory result release process in the light of ISO 15189:2012 standard. Clin Chim Acta 2021;522:167–3.10.1016/j.cca.2021.08.013Search in Google Scholar PubMed

31. Ceriotti, F, Buoro, S, Pasotti, F. How clinical laboratories select and use analytical performance specifications (APS) in Italy. Clin Chem Lab Med 2024. https://doi.org/10.1515/cclm-2023-1314 [Epub ahead of print].Search in Google Scholar PubMed

Received: 2024-01-21
Accepted: 2024-04-03
Published Online: 2024-04-16

© 2024 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 9.6.2024 from https://www.degruyter.com/document/doi/10.1515/cclm-2024-0102/html
Scroll to top button