Next Article in Journal
Cost-Effective Full-Color 3D Dental Imaging Based on Close-Range Photogrammetry
Next Article in Special Issue
Interobserver Agreement in Automatic Segmentation Annotation of Prostate Magnetic Resonance Imaging
Previous Article in Journal
Multi-Dataset Comparison of Vision Transformers and Convolutional Neural Networks for Detecting Glaucomatous Optic Neuropathy from Fundus Photographs
Previous Article in Special Issue
Correction of Arterial-Phase Motion Artifacts in Gadoxetic Acid-Enhanced Liver MRI Using an Innovative Unsupervised Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Anatomical Prior-Based Automatic Segmentation for Cardiac Substructures from Computed Tomography Images

1
Shien-Ming Wu School of Intelligent Engineering, South China University of Technology, Guangzhou 511400, China
2
Department of Radiology, The Third Affiliated Hospital of Guangzhou Medical University, Guangzhou 510150, China
3
Guangzhou Janus Biotechnology Co., Ltd., Guangzhou 511400, China
4
School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou 510275, China
5
Guangdong Province Key Laboratory of Computational Science, Sun Yat-sen University, Guangzhou 510275, China
6
State Key Laboratory of Oncology in South China, Guangzhou 510060, China
*
Authors to whom correspondence should be addressed.
Xinyi Li is a co-author, Xuefang Wang and Xinyi Li contributed equally to this work.
Bioengineering 2023, 10(11), 1267; https://doi.org/10.3390/bioengineering10111267
Submission received: 18 July 2023 / Revised: 12 October 2023 / Accepted: 24 October 2023 / Published: 31 October 2023
(This article belongs to the Special Issue Artificial Intelligence-Based Diagnostics and Biomedical Analytics)

Abstract

:
Cardiac substructure segmentation is a prerequisite for cardiac diagnosis and treatment, providing a basis for accurate calculation, modeling, and analysis of the entire cardiac structure. CT (computed tomography) imaging can be used for a noninvasive qualitative and quantitative evaluation of the cardiac anatomy and function. Cardiac substructures have diverse grayscales, fuzzy boundaries, irregular shapes, and variable locations. We designed a deep learning-based framework to improve the accuracy of the automatic segmentation of cardiac substructures. This framework integrates cardiac anatomical knowledge; it uses prior knowledge of the location, shape, and scale of cardiac substructures and separately processes the structures of different scales. Through two successive segmentation steps with a coarse-to-fine cascaded network, the more easily segmented substructures were coarsely segmented first; then, the more difficult substructures were finely segmented. The coarse segmentation result was used as prior information and combined with the original image as the input for the model. Anatomical knowledge of the large-scale substructures was embedded into the fine segmentation network to guide and train the small-scale substructures, achieving efficient and accurate segmentation of ten cardiac substructures. Sixty cardiac CT images and ten substructures manually delineated by experienced radiologists were retrospectively collected; the model was evaluated using the DSC (Dice similarity coefficient), Recall, Precision, and the Hausdorff distance. Compared with current mainstream segmentation models, our approach demonstrated significantly higher segmentation accuracy, with accurate segmentation of ten substructures of different shapes and sizes, indicating that the segmentation framework fused with prior anatomical knowledge has superior segmentation performance and can better segment small targets in multi-target segmentation tasks.

Graphical Abstract

1. Introduction

According to a report from the World Health Organization (WHO), heart disease has been the leading cause of death worldwide for the past 20 years [1]. To effectively prevent and treat various heart diseases, accurate calculation, modeling, and analysis of the entire heart structure are essential for research and applications in the medical field. CT imaging, the current gold standard for the diagnosis of many diseases, is widely used in early diagnosis and screening of the brain and heart [2]. Using CT, it is possible to qualitatively and quantitatively assess the anatomical structure and function of the heart and provide support for diagnosis, disease monitoring, treatment planning, and recovery prediction. Cardiac image segmentation is the first step in these applications. It divides an image into multiple anatomically meaningful regions, from which quantitative metrics can be extracted [3]. Typically, the anatomical structures of interest in cardiac image segmentation include ten substructures: the left atrium (LA), right atrium (RA), left ventricle (LV), right ventricle (RV), superior vena cava (SVC), inferior vena cava (IVC), pulmonary artery (PA), pulmonary vein (PV), ascending aorta (AA), and descending aorta (DA). The circulatory system consists of the heart, arteries, and veins, which deliver blood throughout the body and are divided into pulmonary and systemic circulation. Cardiac substructures are the main forces of pulmonary and systemic circulation and are the basic elements for analyzing the entire heart structure and function.
Currently, segmentation of cardiac substructures in clinical practice relies heavily on manual labor. Radiologists manually draw contour lines of substructures, including the heart ventricles, atria, and vasculature, to obtain clinical parameters for analyzing heart function, such as ejection fraction and stroke volume. However, manual segmentation is often time-consuming and prone to subjective errors, resulting in a heavy workload for medical professionals and inconsistent segmentation results influenced by factors such as diagnostic experience. With a gradual increase in the incidence of cardiac-related diseases, manual segmentation cannot meet the great demand for medical treatment. A fast and accurate, fully automated cardiac segmentation algorithm is necessary to improve diagnostic efficiency, particularly in clinical applications of cardiovascular surgery planning and radiation therapy.
The early studies on automatic cardiac substructure segmentation were based on traditional machine learning methods and mainly relied on visible image features such as image texture, image edges, and pixel intensity to achieve segmentation of the heart target [4,5,6,7,8,9,10,11]. Hautvast et al. designed a new active contour model for segmentation of cardiac ventricles. This segmentation method obtains a constant contour by comparing it with the grayscale value of the vertical contour. The contour maintains a relatively constant position with the surrounding tissue to achieve better segmentation results [9]. Billet et al. used coupled and active models to restrict the motion of the endocardium and epicardium to achieve biventricular segmentation of the heart. This method uses special image features, image intensity, and gradient information to achieve better cardiac motion recovery and higher segmentation accuracy [10]. Senegas et al. proposed a complete Bayesian model that represented prior information regarding the shape and motion of the heart. The superiority of this method for segmentation was demonstrated by the probability distribution of the parameters and the results of Monte Carlo methods [11]. These segmentation methods are susceptible to the influence of similar feature tissues around the segmented heart target, due to factors such as noise. Traditional segmentation models rely on large amounts of prior knowledge. Modeling usually assumes that the heart is stationary and that the shape of the ventricle is cylindrical and requires manual feature design and segmentation. However, in actual conditions, different parts of the heart are constantly in motion; this leads to the weak performance generalization and poor robustness of the traditional models. The traditional methods also have a common insurmountable problem: feature extraction must be designed and performed manually; thus, it is highly dependent on the domain-specific knowledge of the researchers. This process produces a significant workload and greatly affects the quality of the features, which can lead to significant errors in the segmentation results. Most traditional methods are semi-automated methods that require manual interaction and cannot simultaneously segment three different cardiac structures; they can only customize them for a cardiac region. Thus, these methods do not satisfy the requirements for real-time processing.
In recent years, deep learning has been widely used in the field of medical image segmentation owing to its advantages in automatic feature extraction and nonlinear modeling. Its segmentation accuracy exceeds that of the traditional segmentation methods [12]. By integrating new technologies, deep learning achieves higher segmentation accuracy by learning from limited datasets, with fully automated batch processing of images, which greatly relieves the pressure on physicians in the interpretation of medical images. Many studies have achieved good results in the segmentation of the heart, especially in the segmentation of heart chambers (such as the atria and ventricles LV, RV, LA, and RV); many excellent algorithms have emerged [13,14,15,16,17,18,19,20]; however, they are mostly for one or a few substructures. Some studies have been based on the FCN [21], which uses convolutional layers to replace the last fully connected layer of the CNN to achieve pixel-level classification and end-to-end semantic segmentation of images. However, the pooling operation of the FCN causes loss of pixel position information, resulting in poor segmentation details. Olaf et al. proposed a U-Net based on an encoder–decoder structure, using skip connections between the corresponding layers of the encoder and decoder networks to help the decoder network to better recover the spatial information of the target pixels and achieve reasonable segmentation accuracy with fewer training iterations [22]. Its performance exceeds that of an FCN; a series of encoder–decoder networks derived from U-Net have subsequently been proposed and successfully applied to heart segmentation. In Reference [23], a dense version of U-Net with an initial module was developed to combine multiscale features for robust segmentation of images with large anatomical variations. Zotti et al. proposed an optimized U-Net model that embedded prior information on the shape of the heart to achieve a fully automated MRI heart image segmentation method [24]. To further improve the segmentation performance of the heart, some studies have guided 2D networks by introducing information on different time ranges of the cardiac cycle [25]. Researchers have also proposed multistage network segmentation methods that decompose the segmentation problem into subtasks to first locate the ROI (region of interest) and then connect the segmentation network to refine the segmentation [26]. These studies focused on segmentation of the atria and ventricles and can be used to evaluate their function and examine their volumetric changes, with great clinical significance for early diagnosis of heart disease. However, for computer-aided diagnosis, treatment, and the actual needs of physicians, it is often necessary to segment the entire heart.
With the emergence of the multimodality whole-heart segmentation (MM-WHS) challenge, an increasing number of methods have begun to focus on the automatic segmentation of cardiac substructures, benefiting from the availability of public datasets for whole-heart segmentation (including the four chambers, aorta, and pulmonary veins). Payer et al. [13] used a two-step segmentation process; one step located the heart position and the other narrowed the segmentation range for finer segmentation. Wang and Smedby [27] proposed a fully automatic algorithm for whole-heart segmentation by incorporating prior knowledge of the shapes of the heart substructures into a jointly trained convolutional neural network, achieving a good performance. Yang et al. [28] used a generic framework based on three-dimensional fully convolutional networks, focusing their work on model training and achieving good results. There are several problems with deep learning-based cardiac CT image segmentation. First, many previous studies used neural network models in cardiac substructure segmentation without considering cardiac anatomical structural information. Second, other segmentation methods introduce a small amount of prior information to enhance model performance by embedding additional modules into the neural network in the case of limited data. Through these novel designs, the performance of the model has been improved to some extent, but these methods lack effective anatomical significance from a mathematical perspective, such as using the perimeter of the target or the relationship between adjacent points as regularization to restrict the optimization process. However, some prior knowledge is difficult to represent using regularization [29]. Some methods represent a shape using parameters (such as the center of gravity and radius) or non-parametric methods (such as level set representation), which can perform better in segmenting organs with relatively fixed shapes. However, the shape of the same organ in different individuals may differ and cannot be described using the same parameters [30,31]. Some methods describe the shape using distributions (such as Gaussian and Gaussian mixtures), which represent the possibility of deformation on the main shape to some extent and partially solve the problem of differences in the shapes of the same organ in different individuals, but the statistical distribution may contain errors. In medical images, the basic structure of the body is the same, and the positions and shapes of the organs are generally fixed. These uniform image features are known as the anatomical priors of medical images and are helpful in analysis. Anatomical priors can handle segmentation errors caused by image noise and improve the algorithm reliability. Most methods are manually designed or use artificial prior information that lacks explanation and ignores the inherent anatomical prior structure of the image. However, this knowledge is helpful in improving network performance; suitable prior knowledge must be selected according to the algorithm. It is preferable to obtain prior knowledge through learning. Third, existing deep learning methods still have a common problem. They typically treat large and small substructures equally and rely on high-performance hardware platforms and large amounts of labeled data for model training. This leads to poor segmentation accuracy for small substructures with low contrast. Some methods assign higher weights to small organs in the loss function; however, these methods are not optimized for imbalanced organ segmentation and cannot solve the problem thoroughly.
In addition, inherent problems in cardiac CT images pose significant challenges to segmentation. There are many substructures in the heart; the shape and structure are complex, with large differences in size among the substructures. The substructures are close to each other, and the grayscale images of the structures are very similar, with weak contrast between them. Additionally, some substructures have blood inflow between them. Owing to differences in imaging principles and tissue characteristics, image formation is easily affected by factors such as noise and tissue movement. Moreover, the shape and position of the heart vary from person to person; even in the same person at different times or in different medical image slices, the shape and position of the heart may differ. Addressing these challenges by incorporating complex cardiac anatomical knowledge into the model to improve the segmentation accuracy without decreasing the model efficiency is a critical issue. We observed how physicians annotated the cardiac substructures. First, physicians delineate larger or more easily segmented substructures on a normal scale. For smaller substructures or those that are difficult to segment, physicians first determine the location of the small organ and then focus on the surrounding area to achieve a more accurate delineation, solving the problem of size differences and class imbalances. Physicians introduce prior anatomical knowledge of the heart structures during delineation to prevent annotation errors due to insufficient image information or imaging quality.
To address the limitations of the current deep learning segmentation models, our design, inspired by the approach of the doctors, incorporates prior anatomical knowledge of the position, shape, and scale of the cardiac substructure to separate different scales using a hierarchical segmentation approach to improve the segmentation performance of different scales, especially the sensitivity of small scales. We propose a novel framework for cardiac substructure segmentation that incorporates prior anatomical knowledge. The framework employs two consecutive segmentation steps using a cascade network from coarse to fine segmentation to separately address relatively easy and difficult-to-segment substructures. To improve the segmentation accuracy of small-scale substructures, we used the coarse segmentation results as prior information in combination with the original image to guide and train the fine segmentation network; we embedded anatomical knowledge of large-scale substructures into the fine segmentation network, which resulted in the efficient and accurate segmentation of ten cardiac substructures. The experimental results show that our proposed method has superior segmentation performance and can better segment small targets in multitarget segmentation tasks without requiring additional computation.
The remainder of this paper is organized as follows. Section 2 introduces the model structure and explains each component in detail. In Section 3, we describe the experiments, explore the effects of different grouping strategies on the segmentation results, and compare the results with those of the current mainstream models. Section 4 describes the typical problems encountered in this field and discusses the experimental results. Finally, we analyze the limitations of the experiment and provide prospects for future research.

2. Materials and Methods

2.1. Data Source and Data Analysis

With the approval of the Institutional Review Board, we collected cardiac CT data (DICOM) from 60 patients at the Third Affiliated Hospital of Guangzhou Medical University for this clinical study. Each case was annotated with ten structural contours (Figure 1), including the LV, RV, LA, RA, SVC, IVC, PA, PV, AA, and DA. The annotations were performed by two experienced radiologists, one mid-career doctor conducting the target contour annotation, and one senior doctor reviewing it. Statistical analysis of the annotations yielded the results shown in Figure 2. The proportion represents the average proportion of each substructure in each case; the average size refers to the average number of pixels occupied by each substructure in each slice of the case.
Several characteristics were identified from the experimental data analysis. First, the dataset is not large; the data are 3D, with a class imbalance, particularly for the smaller substructures. Second, there are targets of various sizes, with large differences between the largest (LV) and smallest (SVC) substructures. The left and right ventricles and atria were larger; the pulmonary veins and superior and inferior vena cava were smaller. Third, the cardiac substructures have diverse grayscale values, indistinct borders, irregular shapes, and non-fixed positions. Moreover, the background occupies the largest area in the CT images, which leads to significant imbalances between the substructures and between the substructures and the background, posing substantial challenges for neural network training for segmentation tasks.

2.2. Cardiac Substructures and Their Anatomical Information

The heart is located between the two lungs, near the middle of the chest, and slightly to the left. Its shape is similar to that of a peach and its size is roughly equivalent to that of a clenched right fist. A three-dimensional model and a sectional diagram of the entire cardiac anatomy are shown in Figure 3. The heart has four main chambers: the left ventricle, left atrium, right ventricle, and right atrium. In the human body, the left atrium and ventricle are on the left side; the right atrium and ventricle are on the right. The left and right sides of the heart are not connected and have similar structures, with upper empty spaces (atria) and lower spaces (ventricles). These four chambers comprise more than half the volume of the heart. The aorta is connected to the left ventricle; the pulmonary artery is connected to the right ventricle. The pulmonary veins connect to the left atrium; the superior and inferior vena cava connect to the right atrium. The right and left atria collect deoxygenated blood from the body and lungs, respectively. The right and left ventricles pump blood out of the heart for distribution to the body and lungs. Due to the interconnectedness of the cardiovascular system, CT can more clearly demonstrate the circulation of each component of the heart through continuous imaging. The anatomical information of the different substructures segmented in the cardiac images is presented in Table 1.
From Figure 3 and Table 1, we observe that the cardiac substructures have fixed relative positions and shapes. For example, in the four-chamber view of a transverse CT scan, the left ventricle is located in the upper left ventricle, the right ventricle in the upper right ventricle, the left atrium in the lower left ventricle, and the right atrium in the lower right ventricle. The left ventricle is located below the left atrium and to the left posterior of the right ventricle, whereas the right atrium is located to the right and front of the left atrium. The ascending aorta is connected to the left ventricle with a circular structure, and the pulmonary artery exhibits a tree structure. The inferior vena cava is typically located at the bottom of the heart. These insights provide the features of the various substructures of the heart; physicians can use these anatomical priors to annotate each structure with improved accuracy. We hope to design a segmentation model that integrates anatomical priors to improve the accuracy of the segmentation of cardiac substructures.

2.3. Anatomical Prior-Based Automatic Segmentation Framework

In this study, a cardiac substructure segmentation framework that integrates anatomical structure priors (Figure 4) was proposed to segment ten substructures in cardiac CT images. The framework consisted of two stages: coarse and fine segmentation. Larger or easier-to-segment substructures are segmented first to resolve the imbalance, before smaller or more challenging substructures are segmented. The coarse segmentation results are used as prior information; together with the original image, they form the input for the fine segmentation network. Anatomical knowledge, such as the location, shape, and scale of large-scale substructures, is embedded into a fine segmentation network through geometric constraints, helping to train small-scale substructures and improve the segmentation accuracy of small targets.
U-Net is particularly suitable for the semantic segmentation of datasets at a small scale due to its simple and symmetric structure. It can combine information from low- and high-level layers, fuse features of multiple scales, and achieve excellent segmentation results. Thus, our model uses U-Net as the backbone and divides the ten segmentation targets of the cardiac substructures into two groups, the coarse segmentation group (CG) and the fine segmentation group (FG), according to the characteristics of the cardiac substructures. These two segmentation subtasks are cascaded; the output of the coarse segmentation module serves as an input for the fine segmentation module. This cascading strategy facilitates training of each level of the network model. Using the results from the previous level of network segmentation to constrain the areas segmented by the subsequent level of the network can effectively reduce the number of false positives outside the defined area.

2.3.1. Large Substructure Segmentation Network (LS-Net)

The coarse segmentation network LS-Net (purple area in Figure 5) segments the substructures in the CG and consists of mainly symmetrical contraction and expansion pathways. The contraction pathway obtains image context information through convolution and pooling structures, whereas the expansion pathway uses upsampling and convolution structures to obtain a segmentation target mask. Skip connections combine both low- and high-dimensional features. The contraction pathway contains four encoding modules, each consisting of two 3 × 3 convolutional layers with a rectified linear unit activation function, followed by maximum pooling to downsample the operation. Each downsampling reduces the size of the feature map by half. The expansion pathway is structurally similar to the contraction pathway and contains four decoding modules, each consisting of two ReLU-activated 3 × 3 convolutional layers. The feature map is upsampled to twice its original size. Skip connections between the contraction and expansion pathways combine features at different levels. The output module of the network includes two 3 × 3 convolutional layers; a dropout layer is added to prevent overfitting. Finally, a SoftMax-activated 1 × 1 convolution maps all the feature maps to the target class to obtain the final segmentation result.
Compared with the original U-Net, there were two improvements: (1) To avoid gradient diffusion during network training, batch normalization was added to all convolution layers except the last layer. (2) A dropout layer was added before the output layer to prevent overfitting.

2.3.2. Small Substructure Segmentation Network (SS-Net)

The fine segmentation network SS-Net (blue area in Figure 5) segments the substructures in the FG group and consists of contraction and expansion pathways. The encoding and decoding modules are essentially the same as those of LS-Net; however, SS-Net cascades the segmentation results of LS-Net. We combined the mask output from LS-Net with the original slices and labels of the substructures in the FG group to form a multichannel input for network training. By encoding the segmentation results of large substructures, SS-Net embeds prior knowledge, such as the position, shape, and scale of large substructures. These high-resolution feature maps can help refine the segmentation results, particularly for small substructures.
We integrated the segmentation results of the small and large substructures to obtain the final prediction of all the substructures. This fully used the relatively stable anatomical information between substructures, reduced the difficulty of network training, and improved the segmentation accuracy.

2.3.3. Grouping Strategy

For the division of the coarse segmentation group and fine segmentation group for the ten substructures of the heart, we designed three grouping methods (Table 2) and selected one of them as the optimal segmentation grouping strategy based on the segmentation results. The first grouping method was established based on the anatomy of the heart and the proximity relationship between the substructures. The left ventricle, right atrium, ascending aorta, descending aorta, superior vena cava, and inferior vena cava were grouped as the CG group. The left atrium, right ventricle, pulmonary artery, and pulmonary vein were grouped as the FG group. The rationale for grouping was that there are two blood circulation systems in the heart during surgery: systemic circulation and pulmonary circulation. The blood vessels that make up the systemic circulation include the aorta, which originates from the left ventricle of the heart, and the veins that return blood to the heart, such as the superior and inferior vena cava. The blood from the left ventricle is ejected into the aorta, and then travels along the arteries to the capillaries throughout the body. Subsequently, it drains into the small veins, large veins, and ultimately returns to the right atrium via the superior and inferior vena cava. The second grouping method was based on the size of the substructures. Given the heart cavity structure, the heart cavity was grouped into the CG group, which included the left atrium, right atrium, left ventricle, and right ventricle. The other blood vessels were grouped into the FG group, which included the superior and inferior vena cava, pulmonary artery, and aorta. The basis for grouping was that the relative positions of the four larger targets (left and right atria and ventricles) were fixed, whereas the arteries, veins, and other blood vessels were relatively small, connected, and closely related. The third grouping method grouped the left atrium, left ventricle, and descending aorta in the CG group; the right atrium, right ventricle, superior vena cava, inferior vena cava, and pulmonary artery were included in the FG group. This method explored whether grouping substructures with no connection to each other could help to improve segmentation performance.

2.4. Loss Function

Owing to the large difference in the proportion of targets and background in cardiac CT images and the large differences in size between the substructures, a hybrid loss function that combined the Dice loss [32] and SD loss [33] was used. The hybrid loss function has the optimization characteristics of both loss functions, namely optimization processing for class imbalance and sensitivity to shape changes. The Dice loss is a region-related loss that focuses more on foreground areas during training. It can effectively improve the sensitivity of the network to target areas and is suitable for conditions with a pixel category imbalance. However, the drawback of the Dice loss is that it is not sensitive to boundary characterization and focuses mainly on the interior of the mask. As a measure of shape similarity, the SD loss can complement the deficiencies of the Dice loss in order to reduce training instability. The hybrid loss function is expressed as
L o s s = L o s s D i c e + α L o s s S D
L o s s D i c e = 1 1 / C c = 1 C i P i c G i c i ( P i c ) 2 + i ( G i c ) 2
L o s s S D = 1 / C c = 1 C 1 V i ( P i c G i c ) 2 ( D i c ) γ
where P i and G i represent the predicted SoftMax probabilities and the gold standard label of voxel i in channel C, respectively; D i is the corresponding normalized distance to the gold standard surface; and γ and α are parameters adjusting the surface error penalty; the weight of L o s s S D is set to 1.

3. Experiments and Results

3.1. Experimental Dataset

The experimental data for this study were provided by the Third Affiliated Hospital of Guangzhou Medical University and comprised 60 cardiac CT image samples. All scans were performed using a Siemens SOMATOM Force third-generation dual-source dual-energy spiral CT scanner with tube voltages ranging from 70 to 110 kV, current ranging from 12 to 85 mA, slice thicknesses of 5 mm, slice intervals of 5 mm, pitch of 1.2, and temporal resolution of 250 ms. The number of slices in each case ranged from 108 to 152, with an average of 130, and a slice size of 512 × 512 pixels. The dataset had a size of 60 × 130 × 512 × 512 (60 cases, 130 slices per case, and 512 × 512 pixels per slice), with a ratio of 6:2:2 for random partitioning into training, validation, and testing sets.

3.2. Image Preprocessing

The cardiac CT images were stored in Hounsfield units (HU) per pixel, with different tissue organs typically corresponding to different HU ranges. To improve the local contrast of the images, specific HU windows were first used to perform gray-level truncation of the original images. Normalization was performed by setting pixel values greater than 2048 to 2048 and those less than 0 to 0, followed by division by 2048. To reduce the risk of overfitting, the CT images were augmented. To avoid alteration of clinical features such as shape, the data augmentation methods included random horizontal flip, random horizontal and vertical translations (5%), and random scale (95–105%), with the aspect ratio preserved.

3.3. Functionally Gradient Porous Mandibular Prosthesis

The experimental hardware environment consisted of an Intel® Xeon® E5-2678 v3 CPU and an NVIDIA GTX-1080Ti GPU with 11 GB memory, running on the Ubuntu 18.04 operating system and programmed with Python 3.7. All the programs were implemented using the PyTorch 2.0.1 open-source framework. The network parameters in the experiment included an input image size of 512 × 512, a batch size of 12, and an initial learning rate set at 0.0001; the model optimizer was adaptive moment estimation (Adam: A Method for Stochastic Optimization) [34], and the maximum number of iterations was 300. During the training process, if the loss value did not decrease for ten consecutive rounds on the validation set, the learning rate was halved. To better verify the performance of the network model, five-fold cross-validation was used to obtain the average as the final experimental result.

3.4. Evaluation Metrics

The Dice similarity coefficient (DSC) [35] and Hausdorff distance (HD) [36], which are commonly used in segmentation tasks in medical imaging, were used as quantitative evaluation metrics for segmentation accuracy assessment. At the same time, Recall and Precision and a total of four evaluation metrics were used as quantitative evaluation indicators of the model. The formula for calculating Dice is expressed as
D i c e ( X , Y ) = 2 | G P | G | P |
where G and P represent the manually segmented mask and the prediction mask, respectively, using binary tags. Dice calculates the ratio of twice the intersection of two masks to their union, which reflects the similarity between the target region of segmentation and the annotated target region. The higher the similarity, the better the segmentation effect. Dice ranges from 0 to 1, where 1 represents the best segmentation, and 0 represents the worst segmentation.
The HD is the maximum distance between a point in the automatic segmentation group and the closest point in the corresponding manual segmentation group and is expressed in millimeters. It was used to calculate the distance between the boundary of the actual value and that of the predicted region. It can be expressed as
H D 95 = m a x ( h G , P , h P , G )
h G , P = max m i n | g p |
where G and P represent the volumes of the gold standard and predicted regions, respectively; their corresponding contour points are denoted as G = { g 0 , . . . . . . , g n } and P = { p 0 , . . . . . . , p n }, respectively. The H D 95 ranges from zero to infinity.
Recall is also known as sensitivity; it refers to the ability of a segmentation model to accurately identify and include all the relevant regions or structures of interest in an image. It quantifies the proportion of true positive pixels or voxels that are correctly classified as belonging to the target region, out of all the pixels or voxels that actually belong to the target region. Its calculation formula is as follows, where the true segmentation result is G and the actual segmentation result is P .
R e c a l l = G P G
Precision is expressed as
P r e c i s i o n = G P P
A higher DSC, Recall, Precision, and a lower H D 95 indicate more accurate segmentation. In this study, the DSC, Recall, Precision, and H D 95 were calculated for each case in three dimensions.

3.5. Experimental Results

3.5.1. Impact of Grouping Method for Cardiac Substructures on Segmentation Results

We divided the ten cardiac substructures into coarse and fine segmentation groups based on the grouping methods presented in Table 1. With no prior knowledge, we tested the segmentation accuracy of the cardiac substructures for the three grouping methods using two consecutive segmentation steps. The results in Table 3 indicate that the segmentation results of the second grouping method were the best. For the four relatively large substructures, the left atrium, right atrium, left ventricle, and right ventricle, the difference in the segmentation results was not significant, and the Dice coefficients were all greater than 0.8. For the three medium-sized substructures, including the ascending aorta, descending aorta, and pulmonary artery, most of the Dice coefficients were greater than 0.8. Regarding the three smaller substructures, the PV, IVC, and SVC, there were noticeable differences in the Dice coefficients between the three groups. The Dice coefficient for the IVC in the first group was 0.662; for the other two veins, SVC and PV, it was 0, indicating no predicted results. In the third group, the Dice coefficients for both the SVC and IVC were 0; for the PV, it was 0.275. In contrast, in the second group, the Dice coefficient for the SVC exceeded 0.6; although the Dice coefficients for the IVC and PV were not high, they indicated definite prediction results, surpassing the segmentation performance of the first and third groups. From the experimental results, the segmentation effect of the second group was the best; thus, we selected the second grouping method as the strategy for our segmentation model.

3.5.2. Segmentation Results

To verify the impact of our method on the segmentation performance, we compared the baseline U-Net, two-stage U-Net, and our method; the experimental results are shown in Table 4. It is observed that U-Net segmented most of the substructures in the CT images, including the four chambers of the heart, left atrium, right atrium, left ventricle, right ventricle, three arteries, ascending aorta, descending aorta, and pulmonary artery. These substructures had a high contrast, regular shape, and large size, resulting in better DSC, Recall, Precision, and HD scores. Three veins (PV, IVC, and SVC) were difficult to identify in the CT images due to their small size and low contrast. They were not segmented properly, leading to under-segmentation. The two-stage U-Net segmentation network divided the size of the structures into two groups: larger and smaller. This grouping made it easier to obtain clear segmentation results for smaller structures such as the PV, IVC, and SVC. Thus, we used the coarse segmentation results as the input to the fine segmentation network, embedded the anatomical structure of the four chambers into the next-level network, and guided and trained the small-scale substructures. The results showed that the proposed method significantly improved the segmentation accuracy of the heart substructures. The Dice scores of the three arteries (AA, DA, and PA) generally increased by more than 10%; the DSC of the IVC and SVC increased by over 80%; the IVC and SVC increased by more than 20%; and the PV increased by more than 40%, fully demonstrating the effectiveness of the proposed method for improving the accuracy of cardiac substructure segmentation.
Figure 6 shows the experimental results for some slices. Column (a) shows the expert manual segmentation results; column (b) shows the segmentation results of the baseline U-Net; column (c) shows the segmentation results of the two-stage U-Net, and column (d) shows the segmentation results obtained using the proposed method. The experimental results of our model are smooth without obvious under-segmentation or over-segmentation. The segmentation results for several substructures were closest to the ground truth, especially for the PV, IVC, and SVC, which were small but achieved a segmentation performance comparable to that of the larger substructures.

3.5.3. Comparison with Other Deep Learning Methods

To further validate the performance of the proposed method, it was compared with other deep learning methods, including 3D U-Net [37] and nnU-Net [38]. Three-dimensional U-Net replaces the two-dimensional operations in ordinary U-Net with three-dimensional operations, such as 3D convolution and 3D pooling, which can be conveniently applied to three-dimensional data. nnU-Net, an emerging fully automatic segmentation model in medical imaging in recent years, has achieved a fully automated process for data processing and model training. The cascaded 3D U-Net has achieved positive results in medical image segmentation and is an outstanding U-Net variant model.
Table 5 presents the segmentation results for the ten cardiac substructures in the CT dataset for 3D U-net, nnU-Net, and the proposed method. In traditional cardiac segmentation, such as in the atria and ventricles, our method showed a slight improvement in the RA, LV, LA, and RV. Our method demonstrated a significant advantage in the segmentation of PA, DA, and AA. In contrast, for the three classes with few samples and small targets, SVC, IVC, and PV, our segmentation performance had a more obvious advantage. The Dice scores of the 3D U-Net and nnU-Net segmentation were not larger than 0.6, indicating that they did not achieve adequate segmentation for these three targets. The analysis shows that our method has a higher segmentation accuracy than the other methods and achieves the best segmentation effect.

4. Discussion

In this study, we developed a cardiac substructure segmentation framework that combines anatomical structure priors to accurately segment ten cardiac substructures including the LA, RA, LV, RV, SVC, IVC, PA, PV, AA, and DA. The framework comprises a two-step cascade network from coarse to fine segmentation, incorporating coarse segmentation results as prior information and combining them with the original image to form a multichannel input. Thus, the anatomical knowledge of large-scale structures is embedded into a fine segmentation network to guide and train small-scale structures. Compared to existing segmentation models, our framework demonstrates superior segmentation performance, particularly in multi-target segmentation tasks involving small targets.
We found that the segmentation targets had significant differences in size (for example, the largest substructure (LV) was ten times larger than the smallest substructure (SVC)), unbalanced sample distribution, different grayscales, fuzzy boundaries, irregular shapes, and non-fixed substructure positions. It is difficult to accurately identify small substructures using current end-to-end segmentation neural network models. To address this issue, we proposed a framework that differentiates substructures with substantial size discrepancies and groups them for processing to mitigate the imbalance problem. If we treat large and small substructures equally and rely solely on high-performance hardware platforms and a large amount of labeled data to train the models, we may encounter lower segmentation accuracy for small substructures with low contrast. It is difficult to obtain large amounts of matched labeled data due to availability and the cost of medical imaging data and manual labeling. Thus, we divided the ten substructures into two groups: a coarse segmentation group for easy-to-segment substructures and a fine segmentation group for difficult-to-segment substructures. We explored the inherently stable relationships between heart substructures through experiments and selected the best grouping strategy for optimal model performance. As shown in Table 2, using the left atrium, right atrium, left ventricle, and right ventricle as the coarse segmentation group and the superior vena cava, inferior vena cava, pulmonary artery, pulmonary vein, ascending aorta, and descending aorta as the fine segmentation group was optimal. This is mainly because the relative positions of the four larger structures were fixed and their size differences were not significant, making them easier to segment. The fine segmentation group consisted of arteries and veins that were connected to each other, had a tight relationship, and did not differ significantly in size. For the three veins, the smallest substructures with the lowest contrast and the most challenging structures to segment, grouping large and small substructures separately can optimize the organ segmentation imbalance and simplify tasks, making it easier to train network models and significantly improve the segmentation results.
Hierarchical segmentation enhances performance; however, it remains insufficient for tiny substructures like the superior vena cava, inferior vena cava, and pulmonary veins. Their distribution is concentrated, and coarse segmentation results encompass these small substructures within the targeted area. This inspired us to use the large substructures surrounding small substructures as prior information in small substructure segmentation to further improve segmentation accuracy. We used the first-stage segmentation results and original images as inputs to form a multichannel input for the fine segmentation network, embedding prior knowledge of the position, shape, and scale of the large substructures into the fine segmentation network. These high-resolution features were used to refine the segmentation results. Our prior knowledge included the masks of the four larger substructures, atria, and ventricles, which contained the shape and positional information of the larger structures that served as important prior knowledge for training the smaller substructures through geometric constraints. However, our prior knowledge was obtained through deep learning without additional modules or manually designed or preconceived prior information, resulting in no significant computational burden or increased computation time. By introducing the necessary prior information and increasing the number of input channels, the performance of the model was strengthened with limited data without reducing the operating efficiency of the segmentation network. The results in Table 4 indicate that the segmentation framework incorporating anatomical priors can improve the sensitivity of the network to small target regions, handle erroneous segmentation caused by image noise, and increase the reliability of the algorithm.
We combined anatomical structure priors in our cardiac substructure segmentation framework, which is compatible with base networks such as ResNet, DenseNet, and U-Net, which are commonly used for multi-target segmentation. We chose U-Net as the backbone owing to its simple and symmetrical structure, suitability for small-sample data, and excellent segmentation performance combining multiple scales of features. To further verify the effectiveness of our approach, we compared it with other deep learning methods, including U-Net, 3D U-Net, and nnU-Net, which are currently the most widely used and best-performing deep learning networks in the field of medical image segmentation.
From Table 5, it is observed that although U-Net, 3D U-Net, and nnU-Net can roughly segment most of the substructures in the CT images, it is difficult to obtain accurate segmentation boundaries for the three small veins due to significant differences in grayscale, shape, and size features across different slices, making them prone to under-segmentation or poor prediction. Compared to the other three networks, our proposed method achieved results closer to those of expert manual segmentation. From the Dice evaluation metric, the segmentation accuracy of our method was significantly better than that of the other methods, particularly for the SVC, IVC, and PV substructures, which had limited samples and were small targets. There were two reasons: first, we divided the ten substructures into two groups based on size and difficulty, a coarse segmentation group (including larger substructures such as the four chambers) and a fine segmentation group (including smaller substructures such as arteries and veins). In the first stage, segmentation was performed on the coarse segmentation group; these four targets were relatively large with fixed relative positions, with better segmentation results. The first-stage segmentation results and the original image were used as inputs, forming a multichannel input for the fine segmentation network. First, prior knowledge of the location, shape, and scale of the larger substructures was embedded in the fine segmentation network, and high-resolution feature maps were used to refine the segmentation results. By cascading these two segmentation tasks, each level of the network model was easier to train. Using the segmentation results of the previous network to constrain the segmentation of the subsequent network can effectively reduce the number of false positives outside the boundary. Thus, the proposed coarse-to-fine cascaded segmentation framework is theoretically and practically reliable.
This study has several limitations. With the high cost of medical image data labeling, the size of the dataset was limited. The generalization ability of the model was not tested because the experiment lacked external test datasets; there may be differences in modeling accuracy owing to different CT scanning parameters. In future studies, we will continue to improve our method in order to address these issues. More data will be collected to enhance the generalization ability of the method. The segmentation network was adaptively adjusted according to specific characteristics of the cardiac substructures to further improve the overall segmentation accuracy and meet clinical application requirements.

5. Conclusions

This study proposes a cardiac substructure segmentation framework that integrates prior anatomical structure knowledge to address challenges such as varying grayscales, blurred boundaries, irregular shapes, and inconsistent positions of key substructures in cardiac CT images. The proposed method employs a cascading approach to automatically segment ten substructures in cardiac CT images, achieving accurate segmentation results for targets with different scales, from coarse to fine. The presented method leverages prior knowledge of the heart substructures’ location, shape, and size to separately process differently scaled structures. It employs a hierarchical segmentation approach to enhance segmentation performance, particularly for smaller structures. This segmentation model fuses the benefits of coarse-to-fine strategies and U-Net. By employing a cascading approach to progressively segregate larger to smaller substructures in cardiac CT images, the method effectively mitigates the issue of negligibly small substructures. Moreover, based on the coarse segmentation results, prior knowledge of the positions, shapes, and sizes of larger substructures was embedded into fine segmentation networks to refine the segmentation of smaller substructures. The proposed method eliminates the need for manual intervention and accurately segments substructures of varying sizes, shapes, and positions in cardiac CT images, even in cases of low contrast and blurred boundaries. The experiments demonstrated that this method can effectively identify small targets within substructures, outperforming existing methods in terms of segmentation accuracy.

Author Contributions

X.W., conceptualization, methodology software, writing—original draft; X.L., conducting the target contour annotation and anatomical guidance of cardiac CT images; R.D., review and editing; Y.Z., review and editing; T.S., reviewing the target contour annotation; Y.L., reviewing, supervision, project administration, funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the China Department of Science and Technology under Key Grant 2023YFE0204300, inpart by the R&D project of Pazhou Lab(HuangPu) under Grant 2023K0606, in part by the NSFC under Grant 82371476, Grant 12126610, Grant 82371917, Grant 81971691, Grant 81801809, Grant 81830052, Grant 81827802, and Grant U1811461, in part by the China Department of Science and Technology under Key Grant 210YBXM 2020109002, in part by the Guangzhou Science and Technology bureau under Grant 2023B03J1237, in part by the Science and Technology Innovative Project of Guangdong Province under Grant 2018B030312002, in part by Guangdong Province Key Laboratory of Computational Science at the Sun Yat-sen University under grant 2020B1212060032, in part by Key-Area Research and Development Program of Guangdong Province under Grant 2021B0101190003.

Institutional Review Board Statement

The prospective study has been approved by the institutional review board of The Third Affiliated Hospital of Guangzhou Medical University.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. Cardiovascular Diseases (CVDs). Available online: https://www.who.int/zh/news/item/09-12-2020-who-reveals-leading-causes-of-death-and-disability-worldwide-2000-2019 (accessed on 11 June 2021).
  2. Roberts, W.T.; Bax, J.J.; Davies, L.C. Cardiac CT and CT coronary angiography: Technology and application. Heart 2008, 94, 781–792. [Google Scholar] [CrossRef] [PubMed]
  3. Zhuang, X. Challenges and methodologies of fully automatic whole heart segmentation: A review. J. Healthc. Eng. 2013, 4, 371–408. [Google Scholar] [CrossRef] [PubMed]
  4. Peng, P.; Lekadir, K.; Gooya, A.; Shao, L.; Petersen, S.E.; Frangi, A.F. A review of heart chamber segmentation for structural and functional analysis using cardiac magnetic resonance imaging. Magn. Reson. Mater. Phy. 2016, 29, 155–195. [Google Scholar] [CrossRef] [PubMed]
  5. Auger, D.A.; Zhong, X.; Epstein, F.H.; Meintjes, E.M.; Spottiswoode, B.S. Semi-automated left ventricular segmentation based on a guide point model approach for 3D cine DENSE cardiovascular magnetic resonance. Cardiovasc. Magn. Reson. 2014, 16, 8. [Google Scholar] [CrossRef] [PubMed]
  6. Grosgeorge, D.; Petttjean, C.; Dacher, J.-N. Graph cut segmentation with a statistical shape model in cardiac MRI. Comput. Vis. Image Underst. 2013, 117, 1027–1035. [Google Scholar] [CrossRef]
  7. Ben Ayed, I.; Chen, H.M.; Punithakumar, K.; Ross, I.; Li, S. Max-flow segmentation of the left ventricle by recovering subject-specific distributions via a bound of the Bhattacharyya measure. Med. Image Anal. 2012, 16, 87–100. [Google Scholar] [CrossRef]
  8. Pednekar, A.; Kurkure, U.; Muthupillai, R.; Flamm, S.; Kakadiaris, I.A. Automated left ventricular segmentation in cardiac MRI. IEEE Trans. Biomed. Eng. 2006, 53, 1425–1428. [Google Scholar] [CrossRef]
  9. Hautvast, G.; Lobregt, S.; Breeuwer, M.; Gerritsen, F. Automatic contour propagation in cine cardiac magnetic resonance images. IEEE Trans. Med. Imaging 2006, 25, 1472–1482. [Google Scholar] [CrossRef]
  10. Billet, F.; Sermesant, M.; Delingette, H.; Ayache, N. Cardiac motion recovery and boundary conditions estimation by coupling an electromechanical model and cine-MRI data. In Proceedings of the International Conference on Functional Imaging and Modeling of Heart, Nice, France, 3–5 June 2009; pp. 376–385. [Google Scholar]
  11. Senegas, J.; Cocosco, C.A.; Netsch, T. Model-based segmentation of cardiac MRI cine sequences: A Bayesian formulation. In Medical Imaging 2004: Image Processing; International Society for Optics and Photonics: San Diego, CA, USA, 2004; pp. 432–443. [Google Scholar]
  12. Wang, J.; Xiaolei, D.; Zhou, P. Current Situation and Review of Image Segmentation. Recent Pat. Comput. Sci. 2017, 10, 70–79. [Google Scholar] [CrossRef]
  13. Payer, C.; Štern, D.; Bischof, H.; Urschler, M. Multilabel whole heart segmentation using cnns and anatomical label configurations. In International Workshop on Statistical Atlases and Computational Models of the Heart; Springer: Cham, Switzerland, 2017; pp. 190–198. [Google Scholar]
  14. Dangi, S.; Linte, C.A.; Yaniv, Z. A distance map regularized CNN for cardiac cine MR image segmentation. Med. Phys. 2019, 46, 5637–5651. [Google Scholar] [CrossRef]
  15. Bartoli, A.; Fournel, J.; Bentatou, Z.; Habib, G.; Lalande, A.; Bernard, M.; Boussel, L.; Pontana, F.; Dacher, J.-N.; Ghattas, B.; et al. Deep Learning-based Automated Segmentation of Left Ventricular Trabeculations and Myocardium on Cardiac MR Images: A Feasibility Study. Radiol. Artif. Intell. 2020, 3, e200021. [Google Scholar] [CrossRef]
  16. Wang, S.; Chauhan, D.; Patel, H.; Amir-Khalili, A.; da Silva, I.F.; Sojoudi, A.; Friedrich, S.; Singh, A.; Landeras, L.; Miller, T.; et al. Assessment of right ventricular size and function from cardiovascular magnetic resonance images using artificial intelligence. Cardiovasc. Magn. Reason. 2022, 24, 27–34. [Google Scholar] [CrossRef] [PubMed]
  17. Chen, W.; Huang, H.; Huang, J.; Wang, K.; Qin, H.; Wong, K.K.L. Deep learning-based medical segmentation of the aorta using XR-MSF-U-Net. Comput. Methods Programs Biomed. 2022, 225, 1070–1073. [Google Scholar] [CrossRef] [PubMed]
  18. Martín-Isla, C.; Campello, V.M.; Izquierdo, C.; Kushibar, K.; Sendra-Balcells, C.; Gkontra, P.; Sojoudi, A.; Fulton, M.J.; Weldebirhan Arega, T.; Punithakumar, K.; et al. Deep Learning Segmentation of the Right Ventricle in Cardiac MRI: The M&Ms Challenge. Biomed. Health Inform. 2023, 27, 3302–3313. [Google Scholar]
  19. Dang, H.; Li, M.; Tao, X.; Zhang, G.; Qi, X. LVSegNet: A novel deep learning-based framework for left ventricle automatic segmentation using magnetic resonance imaging. Comput. Commun. 2023, 208, 124–135. [Google Scholar] [CrossRef]
  20. Morris, E.D.; Ghanem, A.I.; Dong, M.; Pantelic, M.V.; Walker, E.M.; Glide-Hurst, C.K. Cardiac substructure segmentation with deep learning for improved cardiac sparing. Med. Phys. 2020, 47, 576–586. [Google Scholar] [CrossRef] [PubMed]
  21. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  23. Khened, M.; Kollerathu, V.A.; Krishnamurthi, G. Fully convolutional multi-scale residual DenseNets for cardiac segmentation and automated cardiac diagnosis using ensemble of classifiers. Med. Image Anal. 2019, 51, 21–45. [Google Scholar] [CrossRef] [PubMed]
  24. Zotti, C.; Luo, Z.; Lalande, A.; Jodoin, P.M. Convolutional Neural Network with Shape Prior Applied to Cardiac MRI Segmentation. IEEE J. Biomed. Health Inf. 2019, 23, 1119–1128. [Google Scholar] [CrossRef]
  25. Qin, C.; Bai, W.; Schlemper, J.; Petersen, S.E.; Piechnik, S.K.; Neubauer, S.; Rueckert, D. Joint learning of motion estimation and segmentation for cardiac MR image sequences. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018, Proceedings of the 21st International Conference, Granada, Spain, 16-20 September 2018; Springer: Cham, Switzerland, 2018; pp. 472–480. [Google Scholar]
  26. Vigneault, D.M.; Xie, W.; Ho, C.Y.; Bluemke, D.A.; Noble, J.A. Ω-Net (Omega-Net): Fully automatic, multi-view cardiac MR detection, orientation, and segmentation with deep neural networks. Med. Image Anal. 2018, 48, 95–106. [Google Scholar] [CrossRef]
  27. Wang, C.; Smedby, Ö. Automatic whole heart segmentation using deep learning and shape context. In International Workshop on Statistical Atlases and Computational Models of the Heart; Springer: Cham, Switzerland, 2017; pp. 242–249. [Google Scholar]
  28. Yang, X.; Bian, C.; Yu, L.; Ni, D.; Heng, P.A. 3d convolutional networks for fully automatic fine-grained whole heart partition. In International Workshop on Statistical Atlases and Computational Models of the Heart; Springer: Cham, Switzerland, 2017; pp. 181–189. [Google Scholar]
  29. Clough, J.R.; Oksuz, I.; Byrne, N.; Schnabel, J.A.; King, A.P. Explicit topological priors for deep-learning based image segmentation using persistent homology. In Proceedings of the 26th International Conference, Information Processing in Medical Imaging, IPMI, Hong Kong, China, 2–7 June 2019; pp. 16–28. [Google Scholar]
  30. Oktay, O.; Ferrante, E.; Kamnitsas, K.; Heinrich, M.; Bai, W.; Caballero, J.; Cook, S.A.; de Marvao, A.; Dawes, T.; O’Regan, D.P.; et al. Anatomically Constrained Neural Networks (ACNNs): Application to Cardiac Image Enhancement and Segmentation. IEEE Trans. Med. Imaging 2018, 37, 384–395. [Google Scholar] [CrossRef]
  31. Chen, X.; Williams, B.M.; Vallabhaneni, S.R.; Czanner, G.; Williams, R.; Zheng, Y. Learning active contour models for medical image segmentation. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 11632–11640. [Google Scholar]
  32. Milletari, F.; Navab, N.; Ahmadi, S.A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  33. Karimi, D.; Salcudean, S.E. Reducing the hausdorff distance in medical image segmentation with convolutional neural networks. IEEE Trans. Med. Imag. 2020, 39, 499–513. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Dice, L.R. Measures of the Amount of Ecologic Association Between Species. Ecology 2012, 26, 297–302. [Google Scholar] [CrossRef]
  36. Huttenlocher, D.P.; Klanderman, G.A.; Rucklidge, W.J. Comparing images using the hausdorff distance. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 850–863. [Google Scholar] [CrossRef]
  37. Ciçek, Ö.; Abdulkadir, A.; Lienkamp, S.S. 3D U-net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; pp. 424–432. [Google Scholar]
  38. Isensee, F.; Jaeger, P.F.; Kohl, S.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
Figure 1. Contour labeling of cardiac substructure. (AD) Plain chest CT (slice thickness: 5 mm, mediastinal window); (E,F) cardiac substructure segmentation based on chest CT scan; (A,E) level under tracheal carina; (B,F) ascending aortic root level; (C,G) four-cavity heart level; (D,H) coronary sinus level.
Figure 1. Contour labeling of cardiac substructure. (AD) Plain chest CT (slice thickness: 5 mm, mediastinal window); (E,F) cardiac substructure segmentation based on chest CT scan; (A,E) level under tracheal carina; (B,F) ascending aortic root level; (C,G) four-cavity heart level; (D,H) coronary sinus level.
Bioengineering 10 01267 g001
Figure 2. Proportion of each substructure and average size in dataset.
Figure 2. Proportion of each substructure and average size in dataset.
Bioengineering 10 01267 g002
Figure 3. Cardiac anatomy.
Figure 3. Cardiac anatomy.
Bioengineering 10 01267 g003
Figure 4. Cardiac substructure segmentation framework.
Figure 4. Cardiac substructure segmentation framework.
Bioengineering 10 01267 g004
Figure 5. Architecture of proposed LS-Net and SS-Net.
Figure 5. Architecture of proposed LS-Net and SS-Net.
Bioengineering 10 01267 g005
Figure 6. Segmentation comparisons for U-Net, two-staged U-Net, and the proposed method.
Figure 6. Segmentation comparisons for U-Net, two-staged U-Net, and the proposed method.
Bioengineering 10 01267 g006
Table 1. The performance parameters for porous scaffolds, covering porosity and peak stress.
Table 1. The performance parameters for porous scaffolds, covering porosity and peak stress.
SubstructuresAcronymAnatomical Position
Left VentricleLVThe left ventricle is located in the lower left part of the heart, below the left atrium, and in the left rear of the right ventricle, which is conical. In the four-chamber view of transverse CT, the left ventricle is located in the upper left.
Right VentricleRVThe right ventricle is located in the lower right part of the heart, in the anterior lower part of the right atrium. In the four-chamber view of transverse CT, the right ventricle is located in the upper right.
Left AtriumLAThe left atrium is located in the upper left part of the heart and is the most posterior heart cavity. In the four-chamber view of transverse CT, the left atrium is located in the lower left.
Right AtriumRAThe right atrium is located in the upper right part of the heart, on the right and anterior side of the left atrium. In the four-chamber view of transverse CT, the right atrium is located in the lower right.
Ascending AortaAAThe ascending aorta is connected to the left ventricle. In CT, the aorta is the largest cardiac blood vessel in the mediastinum and is presented in a circular structure in the transverse position.
Descending AortaDAIn CT, the descending aorta is located beside the spine and has a circular structure.
Pulmonary ArteryPAThe pulmonary artery starts from the bottom of the right ventricle. In the transverse CT, the left and right pulmonary arteries extend from the main pulmonary artery to both sides, presenting a tree structure.
Pulmonary VeinPVOn the axial CT, the pulmonary veins extend to both sides, showing a ‘reptile‘ shape due to the small diameter.
Inferior Vena CavaIVCIn the four-chamber view of transverse CT, the inferior vena cava is often located at the bottom of the heart and is closely related to the liver.
Superior Vena CavaSVCThe superior vena cava is located on the right side of the ascending aorta in the transverse CT; its diameter is smaller than that of the ascending aorta.
Table 2. Grouping methods of cardiac substructures.
Table 2. Grouping methods of cardiac substructures.
GroupCoarse Segment GroupFine Segment Group
1LV RA AA DA SVC IVCLA RV PV PA
2LA LV RA RVAA DA PA PV SVC IVC
3LA LV PV AA DARA RV IVC SVC PA
Table 3. Segmentation results for different grouping methods of cardiac substructures. Bold indicates the best result.
Table 3. Segmentation results for different grouping methods of cardiac substructures. Bold indicates the best result.
GroupingLARALVRVSVCIVCPAPVAADA
Group 10.8840.8460.8060.8510.00030.6620.8430.00030.8920.875
Group 20.8530.8510.8790.8680.6410.5200.8080.3620.8560.862
Group 30.8230.8370.8180.8320.00020.00050.7880.2750.8710.870
Table 4. DSC and HD95 comparisons for U-Net, two-staged U-Net, and the proposed method on all ten substructures in testing data. Bold indicates the best result or significant differences.
Table 4. DSC and HD95 comparisons for U-Net, two-staged U-Net, and the proposed method on all ten substructures in testing data. Bold indicates the best result or significant differences.
SubstructuresDSCRecallPrecisionHD/mm
U-NetTwo
-Staged
Proposed MethodU-NetTwo
-Staged
Proposed MethodU-NetTwo
-Staged
Proposed MethodU-NetTwo
-Staged
Proposed Method
LA0.8090.8530.8530.7790.8400.8400.8470.8500.8500.3550.3010.301
RA0.8370.8510.8510.8420.8750.8750.8340.8370.8370.4920.4220.422
LV0.8780.8790.8790.8700.8710.8710.8870.8890.8890.3520.3670.367
RV0.8560.8680.8680.8250.8730.8730.8560.8640.8640.4110.3830.383
SVC0.3550.6410.8160.5740.6500.8010.2500.6760.8323.8741.8870.345
IVC0.3020.5200.8010.2080.4590.8180.5510.7110.8763.9122.6030.366
PA0.8040.8080.9020.7640.7670.9090.8540.8640.8950.5480.5540.284
PV0.1900.3620.7970.2580.3790.7540.4220.5770.8355.2373.4310.351
AA0.8500.8560.9280.8230.8170.9030.8840.9300.9540.4790.3920.213
DA0.8560.8620.8730.8260.8290.9360.9090.9120.8180.4110.3460.312
Table 5. Comparisons of 3D U-net, nnU-Net, and the proposed method. Bold indicates the best result.
Table 5. Comparisons of 3D U-net, nnU-Net, and the proposed method. Bold indicates the best result.
SubstructuresDSCRecallPrecisionHD/mm
3D U-NetnnU-NetProposed Method3D U-NetnnU-NetProposed Method3D U-NetnnU-NetProposed Method3D U-NetnnU-NetProposed Method
LA0.8330.8420.8530.8150.8210.8400.8350.8440.8500.4730.3150.301
RA0.8200.8170.8510.8210.8320.8750.8170.8270.8370.5130.5210.422
LV0.8710.882 0.879 0.8630.878 0.871 0.8730.890 0.889 0.3570.3720.367
RV0.7920.8310.8680.7910.8290.8730.7840.8290.8640.5890.5510.383
SVC0.7120.7800.8160.7220.7500.9700.7230.7760.8320.6020.5740.345
IVC0.5680.7820.8010.5840.7740.8010.5910.7910.8762.8010.5740.366
PA0.7510.7920.9020.7420.7570.8180.7660.7840.8950.5860.5540.284
PV0.5270.5460.7970.5260.5350.9090.5210.5770.8353.1523.4310.351
AA0.8580.8760.9280.8460.8640.7540.8600.9100.9540.4470.3920.213
DA0.8020.8390.8730.8090.8210.9030.8130.8270.8180.3510.3460.312
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Li, X.; Du, R.; Zhong, Y.; Lu, Y.; Song, T. Anatomical Prior-Based Automatic Segmentation for Cardiac Substructures from Computed Tomography Images. Bioengineering 2023, 10, 1267. https://doi.org/10.3390/bioengineering10111267

AMA Style

Wang X, Li X, Du R, Zhong Y, Lu Y, Song T. Anatomical Prior-Based Automatic Segmentation for Cardiac Substructures from Computed Tomography Images. Bioengineering. 2023; 10(11):1267. https://doi.org/10.3390/bioengineering10111267

Chicago/Turabian Style

Wang, Xuefang, Xinyi Li, Ruxu Du, Yong Zhong, Yao Lu, and Ting Song. 2023. "Anatomical Prior-Based Automatic Segmentation for Cardiac Substructures from Computed Tomography Images" Bioengineering 10, no. 11: 1267. https://doi.org/10.3390/bioengineering10111267

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop