Open Access
22 February 2024 Deep learning performance on MRI prostate gland segmentation: evaluation of two commercially available algorithms compared with an expert radiologist
Erik Thimansson, Erik Baubeta, Jonatan Engman, Anders Bjartell, Sophia Zackrisson
Author Affiliations +
Abstract

Purpose

Accurate whole-gland prostate segmentation is crucial for successful ultrasound-MRI fusion biopsy, focal cancer treatment, and radiation therapy techniques. Commercially available artificial intelligence (AI) models, using deep learning algorithms (DLAs) for prostate gland segmentation, are rapidly increasing in numbers. Typically, their performance in a true clinical context is scarcely examined or published. We used a heterogenous clinical MRI dataset in this study aiming to contribute to validation of AI-models.

Approach

We included 123 patients in this retrospective multicenter (7 hospitals), multiscanner (8 scanners, 2 vendors, 1.5T and 3T) study comparing prostate contour assessment by 2 commercially available Food and Drug Association (FDA)-cleared and CE-marked algorithms (DLA1 and DLA2) using an expert radiologist’s manual contours as a reference standard (RSexp) in this clinical heterogeneous MRI dataset. No in-house training of the DLAs was performed before testing. Several methods for comparing segmentation overlap were used, the Dice similarity coefficient (DSC) being the most important.

Results

The DSC mean and standard deviation for DLA1 versus the radiologist reference standard (RSexp) was 0.90±0.05 and for DLA2 versus RSexp it was 0.89±0.04. A paired t-test to compare the DSC for DLA1 and DLA2 showed no statistically significant difference (p=0.8).

Conclusions

Two commercially available DL algorithms (FDA-cleared and CE-marked) can perform accurate whole-gland prostate segmentation on a par with expert radiologist manual planimetry on a real-world clinical dataset. Implementing AI models in the clinical routine may free up time that can be better invested in complex work tasks, adding more patient value.

1.

Introduction

Prostate cancer is the most common male cancer. The diagnostic paradigm shift to “MRI first,”1,2 meaning magnetic resonance imaging (MRI) before biopsy, has increased the number of prostate MRI examinations and the number of targeted biopsies. Accurate prostate gland segmentation on MRI is crucial for ultrasound-MRI fusion biopsy planning,3 which is an increasingly used technique for targeted biopsies with a transrectal or transperineal approach.4 High-quality gland segmentation is also crucial for prostate volume calculation and prostate specific antigen (PSA) density, for focal prostate cancer therapy based on MRI findings, for brachytherapy, and for external radiation therapy planning. Several studies have evaluated deep learning (DL) algorithm performance on prostate gland segmentation and have shown good performance5 but, because of their single institution settings without diverse clinical MRI datasets, the results have typically not been generalizable.6

The number of Food and Drug Association (FDA)-cleared artificial intelligence (AI) algorithms for use in radiology is rapidly increasing and this has raised concern in the radiology community about potential risks in implementing those AI models in clinical practice outside the context in which they were developed.7 Several recent review articles concerning AI use in prostate MRI have noted a lack of studies validating DL algorithms’ true performance in the clinical setting5,8,9 and the need for larger and diverse datasets and testing of the AI models in real-life settings before routine use in clinical practice. Typically, in prior studies the authors have designed, trained, and tested a self-developed DL algorithm,10,11 and the majority of studies are from single-center, single-scanner settings.6 The MRI data require preprocessing before DL algorithm use,12 and most prior studies have used small test sets of under 100 patients.6

Even though DL prostate gland segmentation is the most commonly studied AI application, only a few commercially available DL algorithms are available (7–11; the number varies in recent review compilations).5,13 To our knowledge, no previous study has evaluated commercially available FDA-cleared and CE-approved DL algorithm contour quality on a diverse real-life MRI dataset.

Primary aim: to compare prostate contour assessment by two commercially available FDA-cleared and CE-marked DL algorithms to an expert radiologist’s manual contours as a reference standard.

Secondary aim: to compare prostate volume measures by two commercially available DL algorithms with expert radiologist volume measures from manual planimetry as a reference standard.

2.

Material and Methods

2.1.

Study Design and Population

This retrospective multicenter study was approved by the local ethics review committee at Lund University (entry no. 2014-886) and the Swedish Ethical Review Authority (entry no. 2019-03674).

Assessed for eligibility were all consecutive patients who in 2018 underwent robot-assisted radical prostatectomy at Skåne University Hospital in Malmö. Patients who had an MRI of the prostate performed less than 1 year before the date for surgery were included. Two patients were excluded due to MRI performed at a hospital outside our health care region—one patient due to patient withdrawal and one patient for DL algorithm technical reasons—resulting in the inclusion of 123 patients in the study (Fig. 1).

Fig. 1

Study cohort.

JMI_11_1_015002_f001.png

2.2.

MRI Dataset and Technique

The MRI examinations were performed at seven hospitals using eight scanners, seven different scanner models from two vendors, two different field strengths (1.5 and 3T), and two T2 transaxial slice angulations. Different imaging acquisition parameters were used at different sites. All protocols included transverse, coronal, and sagittal T2-weighted turbo spin-echo images, which were used for prostate segmentation by an expert radiologist. The T2 transaxial imaging was used for DL prostate segmentation. The different scanners and T2 transaxials used are illustrated in the Supplementary Material.

2.3.

Manual Segmentation as the Reference Standard

A consultant radiologist (JE) highly experienced in MRI/ultrasound fusion biopsy pre-processing (>800  cases) at Skåne University Hospital in Malmö Sweden, a tertiary reference center, performed prostate gland segmentation by manual planimetry, using external software MIM® (version 7.1.2 MIM Software Inc, Cleveland, Ohio, United States). This workflow is used in daily clinical praxis. Time consumption was measured on a subset of the exams, which were randomly selected using a manual stopwatch. Whole-gland segmentation requires manual planimetry, a tedious and time-consuming process in which the radiologist fits a line to the prostate outer contour on all T2 transaxial slices using the coronal and sagittal slices as reference. These whole-gland segmentation contours (called Reference Standard expert, abbreviated RSexp) were used as reference standards in the study and saved as RTSS objects (a coordinate-based file format used for fusion biopsy preprocessing) and adequate for comparison with other prostate contours. JE was blinded to the DL contours.

2.4.

Deep Learning Segmentation

Two commercially available DL algorithms were evaluated. DL algorithm 1 (called DLA1) AI-Rad Companion Prostate MR VA20A_HF02, Siemens Healthcare AG, Erlangen, Germany) is an FDA-cleared (510 [k]) and CE-marked (Class IIa–MDD) machine-learning deep-learning algorithm that uses a convolutional neural network deep image-to-image (DI2IN) network. Learning based whole gland segmentation methodology for DLA1 is described by Yang14 and AI model network architecture is described by Winkel.15 DL algorithm 2 (called DLA2) MIM contour protégé® (version 7.1.7.M209-02 MIM Software Inc, Cleveland (OH), USA) is an FDA-cleared 510 [k]) and CE-marked (Class IIa – MDD) machine-learning deep-learning algorithm that uses a convolutional neural network based in the U-net architecture. The model comprises multiple layers of weights and biases that transform the input image into a segmentation mask for each structure at the final output layer. The resulting output is then post-processed to retain the single, largest connected component.16 For the study, the T2 transaxials were manually exported from the Picture Archive and Communications System (PACS) (Sectra IDS7, Linköping, Sweden) to DLA1 hosted on a local in-house server (A) and to DLA2 hosted on another locally hosted server (B). Neither of the algorithms was previously exposed to or trained on the images in the current study cohort. Both algorithms used non-annotated T2 transaxial images for whole-gland segmentation.1416 The resulting RTSSs with DLA1 contours were exported to a server (B) where the RTSSs for DLA2 and the reference standard were localized. Prostate volumes for DLA1 and DLA2 were automatically calculated from whole-gland segmentations on server B. The above-described workflow with manual export to servers outside PACS is not used in daily practice, and for this reason its time consumption was not measured. Examples of contours from DLA1, DLA2, and RFexp are shown in Fig. 2.

Fig. 2

Consecutive T2 transaxials with prostate contours by planimetry. Left column: DLA1 (red contour); middle column: DLA2 (yellow contour); right column: the reference standard (turquoise contour). Both DLA1 and DLA2 succeeded in differentiating the prostate pseudocapsule from the endopelvine fascia and extraprostatic fat (white arrows in the middle column), one of the known pitfalls when performing planimetry.

JMI_11_1_015002_f002.png

2.5.

Statistical Analysis

Two methods for comparing segmentation overlap between DL algorithms (DLAs) and reference standards were used, the DICE similarity coefficient (DSC) and the Jaccard Index. The DSC17 is the most commonly used measurement for prostate segmentation,18 and its value ranges from 0 to 1, with 0 indicating no overlap and 1 indicating perfect overlap. We also calculated the Hausdorff distance (HD), which is another widely used method19 for evaluating medical image segmentation. HD output is a value in mm representing how far two subsets of metric space are from each other, i.e., the greatest of all distances from a point on one contour to the closest point on the second contour. Mean and standard deviation (SD) are presented for all three methods, but since there is data skewness, we also present median and interquartile range IQR/min-max. A box plot was used to present the distribution of DSC. A paired t-test was used to compare the DSC for DLA1 and DLA2.

Bland–Altman plots were used to present method agreement when comparing prostate volume assessment by DLA1 and DLA2 with the expert radiologist as the reference standard.

Descriptive statistics were used to describe the study cohort. All statistical analyses were performed in R version 4.0.2.20

3.

Results

The cohort consisted of 123 patients with a median age of 66 years (range 45 to 76 years) and median preoperative PSA of 6.90  μg/L (min 0.88; max 39).

The DSC mean and SD for DLA1 versus the radiologist reference standard (RSexp) was 0.90±0.05 and for DLA2 versus RSexp it was 0.89±0.04. The DSC median and IQR/min-max for DLA1 was 0.91 (IQR 0.04, min-max 0.60 to 0.95) and for DL2 it was 0.91 (IQR 0.04, min-max 0.73 to 0.94). The mean Jaccard index was similar for DLA1 and DLA2 (0.81). The HD was slightly lower for DLA1 (7.1 mm) compared with DLA2 (7.9 mm). The mean, SD, IQR, and min-max for the DSC, Jaccard index, and HD are listed in Table 1 and presented as box plots in Fig. 3. A paired t-test to compare the DSCs for DLA1 and DLA2 showed no statistically significant difference (p=0.8).

Table 1

DSC, Jaccard index, and HD (mm) mean and SD and median and IQR/min-max for the reference standard (RSexp) and deep learning algorithms (DLA1 and DLA2).

NameMeanSDMedianIQRMinQ1Q3Max
DICE RFexp vs. DLA10.8950.04730.9050.03550.5950.8840.9200.948
DICE RFexp vs. DLA20.8940.03860.9050.04150.7260.8790.9200.941
Jaccard RFexp vs. DLA10.8130.06940.8260.05950.4240.7920.8510.901
Jaccard RFexp vs. DLA20.8110.05990.8270.06800.5690.7840.8520.889
Hausdorff RFexp vs. DLA17.112.806.163.803.284.958.7516.6
Hausdorff RFexp vs. DLA27.933.447.213.043.595.979.0123.6

Fig. 3

Box plot showing DICE coefficient, Jaccard Index, and HD (mm) for DLA1 and DLA2.

JMI_11_1_015002_f003.png

As shown in the Bland–Altman plot (Fig. 4), the observed prostate volume mean difference between DLA2 and RSexp was lower than the observed mean difference between DLA1 and RSexp (mean difference [95% limits of agreement]): DLA1, 3.53  mL (11.55; 4.50 mL); DLA2, 0.67 mL (6.41; 7.75 mL). DLA2 showed better precision as seen in narrower limits of agreement. DLA2 tended to overestimate the volumes of small- and medium-sized prostates, in contrast to DLA1, which tended to underestimate the volumes. Observations for planimetry by the expert radiologist were timed from start to finish in 14/123 patients, and the mean time consumption per case was 8 min and 4 s.

Fig. 4

Bland–Altman plot showing method agreement for prostate volume calculation for DLA1 and DLA2 with the reference standard prostate volume calculation from manual planimetry (PVexp).

JMI_11_1_015002_f004.png

4.

Discussion

In this study, we show that two commercially available FDA-cleared and CE-marked DL algorithms performed whole-gland segmentation accurately as compared with an expert radiologist’s manual planimetry as a reference standard in a diverse clinical MRI prostate dataset.

4.1.

Whole-Gland Segmentation: Our Results in Context

In our study with 123 patients (i.e., test set n=123), the DSC mean and SD for DLA1 was 0.90±0.05, and for DLA2, it was 0.89±0.04. Our results are similar to the so-called grand challenges which serve to benchmark and validate AI models. The Promise12 challenge18 in 2012, the NCI-ISBI 201321 challenge and the MSD22 challenge all showed DSC close to 0.90 (0.87–0.92). Aside from challenges, previous studies by Turkbey in 201323 and Lee11 also demonstrated similar results (DSC 0.87–0.92). Three recent studies, however with smaller test set sizes, confirm similar agreements, Salvaggio in 202124 DSC 0.90, test set n=10; Sanford in 202012 DSC 0.93, test set n=2983; and Cuocolo in 202125 DSC 0.91 whole gland (Enet), test set n=105 from the ProstateX challenge. All DSCs and test set sizes are summarized in a concise table in the Supplementary Material. Thus, in our analysis of a multicenter, multiscanner clinical dataset by two commercially available DLAs, we observed the same performance that has previously been shown in segmentation challenges and recently published studies (Salvaggio single scanner, Sanford single center, and Cuocolo with 3T only).

Although both DL1 and DL2 are based on convolutional neural networks the differences in neural network architecture and training data can possibly explain at least part of those observed differences in performance between the two tested algorithms. The heterogeneity in the MRI dataset could possibly also influence the performance and explain the observed outliers in segmentation and prostate volume assessment. However, no evident motion artifacts on these specific MRI exams could be noted.

The limitations of test set size,23,26,27 lack of validation on external cohorts,11,28 and single-institution datasets11,12,23 were approached by Sorensen,6 who trained a DL model that was retrospectively tested (test set n=100 internal cases and n=56 external cases). The authors reported DSC 0.92 for ProGNet, DSC 0.85 for U-Net, and DSC 0.89 for a radiology technician, with expert segmentations used as the reference standard. In conclusion, the levels of agreement measured by DSC have remained consistently around 0.90 over the past decade. However, our study, along with Sorensen’s, demonstrates that this level of agreement holds true not only for single-center evaluations but also for multicenter real world data comparisons. In our current investigation, this level of agreement extends to two commercially available products. While our results may not be groundbreaking, they indicate that the method is robust for this relatively straightforward task and could represent a low-hanging fruit to assist and streamline the clinical workflow for radiologists.

Radiologist interreader variability was examined by Becker et al.;29 the author group showed an interreader DSC of 0.86 between six readers, with the highest variability found in the anatomical apex region. We have previously published interreader variability values for an experienced and an inexperienced radiologist performing manual planimetry for prostate volume calculation on an overlapping cohort. We found a small but statistically significant underestimation of prostate volumes by the inexperienced reader.30

In line with Bezinque et al.,31 who noted the lack of published performance data for DynaCad, published data on commercially available algorithms for prostate gland segmentation are still, with few exceptions, limited to conference abstracts and product white papers. The DL algorithm from the MIM contour protégé white paper16 reports a DSC of 0.94 from internal validation and testing. The Lucida Pi v 1.0 EuSoMII meeting presentation 202032 showed a DSC of 0.92 on a test set with n=10 from PROMISE 12, and the Quantib Prostate 1.3 white paper33 presents internal testing of prostate volume.

4.2.

Prostate Base and Apex: Areas with Segmentation Challenges

To get an impression of where the algorithms face challenges, we looked into some cases with poor algorithm performance and found that the majority of differences in segmentation between algorithms and manual planimetry were found in the base and the apex of the gland. This is in line with prior studies.24,34 It is also our experience from clinical work with preprocessing for fusion biopsies that delineating these regions of the gland with certainty is challenging. One possible explanation for these challenges is the partial volume effect in the through-plane direction, further accentuated by relatively thick slices (3 mm) and commonly non-perpendicular slice planes with respect to the longitudinal axis of the prostate (with interindividual variation depending on the patient’s anatomical conditions). One example is shown in Fig. 5.

Fig. 5

Challenges in segmentation. (a) Mid-gland, (b) base, and (c) apex. DLA1: red contour, DLA2: yellow contour, and reference standard: turquoise contour. Mid-gland, the contours are aligned; in the base and apex, larger variations in segmentation are noticeable.

JMI_11_1_015002_f005.png

4.3.

Demand for AI Model Validation and Data Overlap in Public Datasets

Several recent review articles highlight the need for adequate validation studies on AI models.5,8,13 The authors ask for well-curated diverse datasets,35 highlight the lack of large-scale multicenter prospective clinical studies,13 and propose that AI models be tested in real-life settings before routine use.8 Our current study was designed to address these concerns with a multicenter diverse dataset (including both 1.5 T and 3 T) evaluated in a real-life setting. In this context, an important concept is that training and test data must never be mixed when validating AI models.

For DL1 and probably for several other commercially available prostate segmentation algorithms, the Promise 12,18 ProstateX,36 and other public datasets have been included in the early training of the algorithm. According to the company behind DL2, no public datasets were included in its early training. Data overlap is another issue in several of the public datasets,13 exemplified by the inclusion of ProstateX data in the latest challenge, PICAI22.36 Together these emphasize the fact that validation studies ideally should be performed on clinical datasets where previous non-exposure of the algorithm can be guaranteed, as is the case in the current study.

4.4.

Limitations

Our study has several limitations. Only two commercially available DL algorithms for whole-gland segmentation were included. According to a comprehensive comparison,13 11 vendors offer segmentation, the majority bearing FDA clearance. However, it was outside the scope of the current study to include all available algorithms, since this validation on a multicenter dataset is a resource-intensive process and the dataset is limited. The size of our test set is rather small but on par with previous studies, and it was dominated by one vendor. The MRI examinations in our study are all from a radical prostatectomy cohort which does not reflect the true clinical patient mix in a radiology department. The study’s heterogeneity, with both 1.5 T and 3 T cameras, slice angulation differences, and a multicenter, multiscanner setup, improves its generalizability. Regarding the reference standard we used (manual planimetry by one expert radiologist), there is an ongoing discussion13 on how to obtain the best reference standard when validating AI models, since there is significant interreader variability.27,29 As for now, we know of no better reference standard than the one we used, and this is in line with most previous studies. We know that experts also disagree and this puts a limit to how accurately the algorithms perform, since they in principle cannot achieve better performance than the inter-expert variation. Although the performance of the two tested DL-algorithms have shown good potential to facilitate prostate volume measurements and fusion biopsy planning, the precision in segmentations do not live up to the standards required in radiation oncology where significantly lower error margins are required.

This retrospective study design should be seen as a first validation step. Future AI model validation should be designed as prospective, multivendor, multicenter trials, a planned next step for our institution in collaboration with other centers.

4.5.

Outlook

A previous study30 by our group has shown that a commercially available DLA performs similarly to the radiologist-dependent standard method for prostate volume calculation. This can allow radiologist resources to be reallocated from the time-consuming and tedious task of manually measuring and calculating prostate volumes on MRI. A next step could be to use DL in the preprocessing of fusion biopsies. Although we will still have a radiologist confirming AI-processed contours (“human-in-the-loop”), this workflow clearly has the potential to save radiologists’ time. In our clinical context, we have tried this concept by comparing the time consumed in fusion biopsy pre-processing (whole-gland segmentation and lesion segmentation) by a radiologist only versus a radiologist plus AI (with whole-gland contours already in place). The trend is that substantial time saving is possible if the radiologist is presented with whole-gland contours from AI. Sorensen et al.6 prospectively implemented AI in the fusion biopsy process in 11 patients but did not report time measurements.

Looking forward, an AI-driven workflow including prostate gland segmentation (prostate volume, PSAD, and accurate outer contours), lesion detection, and PI-RADS characterization (including accurate lesion contouring) can be expected. However, these steps must be validated in a true clinical setting.

5.

Conclusion

In this study, we show that two commercially available DL algorithms (FDA-cleared and CE-marked) can perform accurate whole-gland prostate segmentation on a par with expert radiologist manual planimetry on a real-world clinical dataset. Implementing AI models in the clinical routine may free up time that can be better invested by the radiologist in complex work tasks, adding more patient value.

Disclosures

ET has received speaker fees from Siemens Healthcare. SZ has received speaker fees from Siemens Healthcare and Pfizer. AB has received grants/research supports from Astellas, Bayer, and Ferring. AB has received honoraria or consultation fees from Astellas, AstraZeneca, Bayer, Janssen, Merck, Novartis, Sandoz, SAM Nordic and Telix. AB has participated in company sponsored speaker’s bureau by Astellas, Bayer, and Janssen. AB is stock shareholder in LIDDS Pharma, Glactone Pharma, and WntResearch.”

All other authors declare no conflicts of interest.

Code and Data Availability

The data utilized in this study were obtained from Skåne Region. Data are available from the authors upon request and with permission from Skåne Region.

Acknowledgments

We gratefully acknowledge statistician Sara Jesperson at Forum Söder, Jessica Sternisa clinical implementation specialist, and Steven Lelie business development manager, both at MIM software Inc., and Oskar Strand, operations manager, Siemens Healthineers. We also thank Björn Svensson and Henrik Granberg, both clinical engineers at Region Skåne. This study has received funding by grants from Governmental funding for clinical research (ALF) (Grant No. 4611) and scholarship from Stig and Ragna Gorthon Foundation (Grant No. 2020-2665). Open access funding was provided by Lund University.

References

1. 

V. Kasivisvanathan et al., “MRI-targeted or standard biopsy for prostate-cancer diagnosis,” N. Engl. J. Med., 378 1767 –1777 https://doi.org/10.1056/NEJMoa1801993 NEJMAG 0028-4793 (2018). Google Scholar

2. 

H. U. Ahmed et al., “Diagnostic accuracy of multi-parametric MRI and TRUS biopsy in prostate cancer (PROMIS): a paired validating confirmatory study,” Lancet, 389 815 –822 https://doi.org/10.1016/S0140-6736(16)32401-1 LANCAO 0140-6736 (2017). Google Scholar

3. 

G. A. Sonn, D. J. Margolis and L. S. Marks, “Target detection: magnetic resonance imaging-ultrasound fusion-guided prostate biopsy,” Urol. Oncol., 32 903 –911 https://doi.org/10.1016/j.urolonc.2013.08.006 (2014). Google Scholar

4. 

A. Sidana et al., “Fusion prostate biopsy outperforms 12-core systematic prostate biopsy in patients with prior negative systematic biopsy: a multi-institutional analysis,” Urol. Oncol., 36 341.e1 –341.e7 https://doi.org/10.1016/j.urolonc.2018.04.002 (2018). Google Scholar

5. 

B. Turkbey and M. A. Haider, “Deep learning-based artificial intelligence applications in prostate MRI: brief summary,” Br. J. Radiol., 95 20210563 https://doi.org/10.1259/bjr.20210563 BJRAAP 0007-1285 (2022). Google Scholar

6. 

S. J. C. Soerensen et al., “Deep learning improves speed and accuracy of prostate gland segmentations on magnetic resonance imaging for targeted biopsy,” J. Urol., 206 604 –612 https://doi.org/10.1097/JU.0000000000001783 (2021). Google Scholar

7. 

B. Allen et al., “Evaluation and real-world performance monitoring of artificial intelligence models in clinical practice purchase: try it, buy it, check it,” J. Am. Coll. Radiol., 18 1489 –1496 https://doi.org/10.1016/j.jacr.2021.08.022 (2021). Google Scholar

8. 

R. Suarez-Ibarrola et al., “Artificial intelligence in magnetic resonance imaging-based prostate cancer diagnosis: where do we stand in 2021?,” Eur. Urol. Focus, 8 409 –417 https://doi.org/10.1016/j.euf.2021.03.020 (2021). Google Scholar

9. 

T. Penzkofer et al., “ESUR/ESUI position paper: developing artificial intelligence for precision diagnosis of prostate cancer using magnetic resonance imaging,” Eur. Radiol., 31 9567 –9578 https://doi.org/10.1007/s00330-021-08021-6 (2021). Google Scholar

10. 

R. Bardis et al., “Segmentation of the prostate transition zone and peripheral zone on MR images with deep learning,” Radiol. Imaging Cancer, 3 e200024 https://doi.org/10.1148/rycan.2021200024 CANCAR 0008-543X (2021). Google Scholar

11. 

D. K. Lee et al., “Three-dimensional convolutional neural network for prostate MRI segmentation and comparison of prostate volume measurements by use of artificial neural network and ellipsoid formula,” Am. J. Roentgenol., 214 1229 –1238 https://doi.org/10.2214/AJR.19.22254 AJROAM 0092-5381 (2020). Google Scholar

12. 

T. H. Sanford et al., “Data augmentation and transfer learning to improve generalizability of an automated prostate segmentation model,” Am. J. Roentgenol., 215 1403 –1410 https://doi.org/10.2214/AJR.19.22347 AJROAM 0092-5381 (2020). Google Scholar

13. 

M. R. S. Sunoqrot et al., “Artificial intelligence for prostate MRI: open datasets, available applications, and grand challenges,” Eur. Radiol. Exp., 6 35 https://doi.org/10.1186/s41747-022-00288-8 (2022). Google Scholar

14. 

D. Yang et al., “Automatic liver segmentation using an adversarial image-to-image network,” Lect. Notes Comput. Sci., 10435 507 –515 https://doi.org/10.1007/978-3-319-66179-7_58 (2017). Google Scholar

15. 

D. J. Winkel et al., “AI-Rad companion prostate MR VA20A_HF02, Siemens Healthcare AG, Erlangen, Germany,” Diagnostics, 10 https://doi.org/10.3390/diagnostics10110951 (2020). Google Scholar

16. 

H. Wan, Automated Contouring Using Neural Networks Contour ProtegeAI, MIM Software Inc., Cleveland, Ohio (2020). Google Scholar

17. 

L. R. Dice, “Measures of the amount of ecologic association between species,” Ecology, 26 297 –302 https://doi.org/10.2307/1932409 ECGYAQ 0094-6621 (1945). Google Scholar

18. 

G. Litjens et al., “Evaluation of prostate segmentation algorithms for MRI: the PROMISE12 challenge,” Med. Image Anal., 18 359 –373 https://doi.org/10.1016/j.media.2013.12.002 (2014). Google Scholar

19. 

S. S. Chandra et al., “Patient specific prostate segmentation in 3-D magnetic resonance images,” IEEE Trans. Med. Imaging, 31 1955 –1964 https://doi.org/10.1109/TMI.2012.2211377 ITMID4 0278-0062 (2012). Google Scholar

20. 

R Core Team, R: A Language and Environment for Statistical Computing, Version 4.0.2, R Core Team, Vienna (2020). Google Scholar

21. 

Challenge N-I, NCI-ISBI 2013 Challenge - Automated Segmentation of Prostate Structures, The Cancer Imaging Archive( (2015). Google Scholar

22. 

M. Antonelli et al., “The medical segmentation decathlon,” Nat. Commun., 13 4128 https://doi.org/10.1038/s41467-022-30695-9 NCAOBW 2041-1723 (2022). Google Scholar

23. 

B. Turkbey et al., “Fully automated prostate segmentation on MRI: comparison with manual segmentation methods and specimen volumes,” Am. J. Roentgenol., 201 W720 –W729 https://doi.org/10.2214/AJR.12.9712 AJROAM 0092-5381 (2013). Google Scholar

24. 

G. Salvaggio et al., “Deep learning network for segmentation of the prostate gland with median lobe enlargement in T2-weighted MR images: comparison with manual segmentation method,” Curr. Probl. Diagn. Radiol., 51 328 –333 https://doi.org/10.1067/j.cpradiol.2021.06.006 CPDRDS 0363-0188 (2021). Google Scholar

25. 

R. Cuocolo et al., “Deep learning whole-gland and zonal prostate segmentation on a public MRI dataset,” J. Magn. Reson. Imaging, 54 452 –459 https://doi.org/10.1002/jmri.27585 (2021). Google Scholar

26. 

R. Cheng et al., “Fully automated prostate whole gland and central gland segmentation on MRI using holistically nested networks with short connections,” J. Med. Imaging, 6 024007 https://doi.org/10.1117/1.JMI.6.2.024007 JMEIET 0920-5497 (2019). Google Scholar

27. 

B. Wang et al., “Deeply supervised 3D fully convolutional networks with group dilated convolution for automatic MRI prostate segmentation,” Med. Phys., 46 1707 –1718 https://doi.org/10.1002/mp.13416 MPHYA6 0094-2405 (2019). Google Scholar

28. 

F. Zabihollahy et al., “Automated segmentation of prostate zonal anatomy on T2-weighted (T2W) and apparent diffusion coefficient (ADC) map MR images using U-Nets,” Med. Phys., 46 3078 –3090 https://doi.org/10.1002/mp.13550 MPHYA6 0094-2405 (2019). Google Scholar

29. 

A. S. Becker et al., “Variability of manual segmentation of the prostate in axial T2-weighted MRI: a multi-reader study,” Eur. J. Radiol., 121 108716 https://doi.org/10.1016/j.ejrad.2019.108716 EJRADR 0720-048X (2019). Google Scholar

30. 

E. Thimansson et al., “Deep learning algorithm performs similarly to radiologists in the assessment of prostate volume on MRI,” Eur. Radiol., 33 2519 –2528 https://doi.org/10.1007/s00330-022-09239-8 (2022). Google Scholar

31. 

A. Bezinque et al., “Determination of prostate volume: a comparison of contemporary methods,” Acad. Radiol., 25 1582 –1587 https://doi.org/10.1016/j.acra.2018.03.014 (2018). Google Scholar

32. 

J. Suchanek, “Multi-stage ai analysis system to support prostate cancer diagnostic imaging,” in EuSoMII Virtual Annu. Meeting, (2020). Google Scholar

33. 

A. van Engelen, “Improving the clinical pathway of prostate cancer with AI,” (2021). Google Scholar

34. 

R. Colvin et al., “Which measurement method should be used for prostate volume for PI-RADS? A comparison of ellipsoid and segmentation methods,” Clin. Imaging, 80 454 –458 https://doi.org/10.1016/j.clinimag.2021.09.003 CLIMEB 0899-7071 (2021). Google Scholar

35. 

M. Hosseinzadeh et al., “Deep learning-assisted prostate cancer detection on bi-parametric MRI: minimum training data size requirements and effect of prior knowledge,” Eur. Radiol., 32 2224 –2234 https://doi.org/10.1007/s00330-021-08320-y (2022). Google Scholar

36. 

O. D. Geert Litjens et al., ProstateX Challenge Data, Cancer Imaging Archive( (2017). Google Scholar

Biography

Erik Thimansson is working as a consultant radiologist at Helsingborg Hospital in the urogenital section and is a PhD candidate at the Department of Translational Medicine, Diagnostic Radiology, Lund University, Sweden. He received his MD degree in 2003, did his residency at Sahlgrenska University Hospital, and became a specialist in radiology in 2010. He has leading positions in the Organised Prostate Cancer Testing program regionally and nationally, is an associate member of the National Swedish Prostate Cancer Registry steering committee and works together with PACS companies to develop structured reporting tools for prostate MRI including AI-models and is engaged as a teacher at the national prostate MRI course. His current research interests include the role of prostate MRI in organized prostate cancer testing and the evaluation of artificial intelligence in MRI reading. He is a member of the Swedish Society of Radiology and the European Society of Radiology.

Biographies of the other authors are not available.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Erik Thimansson, Erik Baubeta, Jonatan Engman, Anders Bjartell, and Sophia Zackrisson "Deep learning performance on MRI prostate gland segmentation: evaluation of two commercially available algorithms compared with an expert radiologist," Journal of Medical Imaging 11(1), 015002 (22 February 2024). https://doi.org/10.1117/1.JMI.11.1.015002
Received: 16 May 2023; Accepted: 30 January 2024; Published: 22 February 2024
Advertisement
Advertisement
KEYWORDS
Prostate

Evolutionary algorithms

Magnetic resonance imaging

Image segmentation

Artificial intelligence

Deep learning

Biopsy

Back to Top