Next Article in Journal
Effects of a Person Centered Dementia Training Program in Greek Hospital Staff—Implementation and Evaluation
Previous Article in Journal
The Significance of Tau Aggregates in the Human Brain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Split-Attention U-Net: A Fully Convolutional Network for Robust Multi-Label Segmentation from Brain MRI

1
Research Institute, NEUROPHET Inc., Seoul 06247, Korea
2
Department of Radiology, Eunpyeong St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul 03312, Korea
3
Institute of Human Genomic Study, College of Medicine, Korea University, Ansan 15355, Korea
4
Department of Psychiatry, University of Iowa, Iowa City, IA 52242, USA
5
Department of Radiology, Yeouido St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul 07345, Korea
6
Department of Psychiatry, Yeouido St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul 07345, Korea
7
Department of Psychiatry, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul 06591, Korea
*
Authors to whom correspondence should be addressed.
Brain Sci. 2020, 10(12), 974; https://doi.org/10.3390/brainsci10120974
Submission received: 23 October 2020 / Revised: 30 November 2020 / Accepted: 7 December 2020 / Published: 11 December 2020
(This article belongs to the Section Neurotechnology and Neuroimaging)

Abstract

:
Multi-label brain segmentation from brain magnetic resonance imaging (MRI) provides valuable structural information for most neurological analyses. Due to the complexity of the brain segmentation algorithm, it could delay the delivery of neuroimaging findings. Therefore, we introduce Split-Attention U-Net (SAU-Net), a convolutional neural network with skip pathways and a split-attention module that segments brain MRI scans. The proposed architecture employs split-attention blocks, skip pathways with pyramid levels, and evolving normalization layers. For efficient training, we performed pre-training and fine-tuning with the original and manually modified FreeSurfer labels, respectively. This learning strategy enables involvement of heterogeneous neuroimaging data in the training without the need for many manual annotations. Using nine evaluation datasets, we demonstrated that SAU-Net achieved better segmentation accuracy with better reliability that surpasses those of state-of-the-art methods. We believe that SAU-Net has excellent potential due to its robustness to neuroanatomical variability that would enable almost instantaneous access to accurate neuroimaging biomarkers and its swift processing runtime compared to other methods investigated.

1. Introduction

Magnetic resonance imaging (MRI) provides detailed in vivo insights into the human brain’s morphology, which are essential for development, aging, and disease research [1,2,3,4]. To access measurements such as volume, thickness, or shape for structures, neuroanatomy must be segmented, which is a time-consuming issue when it is manually annotated. A manual delineation of neuroanatomy on MRI scans has long been regarded as the “ground truth”. Therefore, fully automation-based algorithms have been sought to reduce the manual effort. In the past, many whole-brain segmentation methods were proposed, including, but not restricted, to the region growing, clustering, and template-based deformation methods [5]. Template-based segmentation is an essential segmentation method that applies tissue labels to the unlabeled images using structural brain MRI as well as the corresponding manual segmentation. In the template-based segmentation models, the deformable registration methods typically operate by spatially deforming a template to fit a previously unlabeled target image [6,7]. A single template-based segmentation has been successfully applied in some commercial applications [6,7]. However, segmentation with a single template performs poorly when the target image has considerable inter-patient variation in neuroanatomy [8,9]. More studies have proposed multi-atlas-based schemes, which have become the actual standard template-based segmentation approach [10,11]. The multi-atlas-based segmentation methods can segment the brain from a clinically obtained T1-weighted MRI volume into more than 100 structural labels using a small number of manually annotated and representative scans. In practice, the multi-atlas segmentation methods are regarded as the standard for whole-brain segmentation methods because of their excellent accuracy and reproducibility. The patch-based methods have been proposed to handle the local variations from incomplete registrations [12,13,14,15,16]. Consequently, the multi-atlas label fusion-based methods have been studied to model the spatial relationships between atlases and target images in the three-dimension patches. However, one of the critical limitations of the conventional multi-atlas segmentation methods is their high complexity cost. Therefore, a multi-atlas label fusion is typically performed with a small number of templates. Many previous studies have been proposed to develop faster and robust multi-atlas segmentation approaches using unlabeled or automatically labeled data to utilize a larger number of templates or even previously unlabeled MRI images [17,18]. The machine learning-based methods have been incorporated into the multi-atlas segmentation method to replace fusion based on the traditional voting or statistical patch approaches. Several machine learning-based multi-atlas label fusion approaches have been developed to learn multi-atlas segmentation from a larger number of atlas templates [19,20,21,22,23]. The core idea of creating a deep-learning architecture from training image patches has contributed to and inspired various convolutional neural network (CNN) methods, including our proposed method.
Recently, deep-learning methods have been widely developed for whole-brain segmentation with multiple labels. The most straightforward strategy for segmenting the whole brain is to input the entire brain volume to a three-dimensional CNN-based segmentation framework, like U-Net [24] or V-Net [25]. However, it is inefficient to apply clinically used MRI images in State-Of-The-Art (SOTA) three-dimensional fully convolutional networks (FCNs) because the amount of memory in standard graphics processing units (GPUs) cannot handle the high resolution of the MRI images. Another challenge of utilizing CNN methods is that manually traced whole-brain MRI images with detailed structural annotations are rare commodities. Many studies have been developed to challenge the limits placed on training data by GPU memory restrictions. De Brebisson’s method proposed an integrated CNN method to learn two three-dimensional patches and their spatial coordinates for multi-label whole-brain segmentation [26]. This approach was subsequently extended to BrainSegNet [27], which employed 2.5-dimensional patches for training an FCN. Recently, DeepNAT [28] was proposed to perform multi-task learning on the three-dimensional patches. These methods design the whole brain segmentation as a per-voxel segmentation problem. More recently, more robust FCNs have been introduced to multi-label whole-brain segmentation from another perspective. QuickNAT is a two-dimensional approach developed to train an FCN framework using a number of auxiliary labels on initially unlabeled data [29]. Although the training was in a two-dimensional manner, this approach revealed a promising direction for leveraging the whole-brain segmentation architecture, using manually traced images and initially unlabeled data. FastSurfer developed a module containing competitive dense blocks and skip pathways for training multi-slice-based 2D CNNs [30].
However, the two-dimension-based segmentation approaches typically return low spatial consistency in the third dimension. Therefore, it would be fascinating to perform a three-dimensional FCN for whole-brain segmentation for higher spatial consistency [24,25]. Recently, multi-task learning [31] and semi-supervised learning [32] have been performed for whole-brain segmentation. Among such work, Rajchl et al. proposed a novel three-dimensional multi-task learning network with remarkably decent segmentation performance [31]. However, the network’s input image size is restricted to 128 × 128 × 128 because of the GPU memory restrictions, which misses about two-thirds of the spatial information available in a 1 mm isotropic brain volume in the Montreal Neurological Institute (MNI) template space [33]. Therefore, directly applying the three-dimensional FCN to whole-brain segmentation is still limited by current GPU capabilities. To overcome such challenges, Li et al. proposed a 3D CNN-based sliding window-based method, which utilized a single network to learn the patches at different spatial locations [34]. Additionally, the spatially localized atlas network tiles (SLANT) method was proposed to improve patch-based training using multiple independent 3D FCN methods [35].
Here, we propose Split-Attention U-Net (SAU-Net), a deep-learning architecture that can segment the whole brain into 33 structural labels in just 20–30 s on the GPU. The basic architecture is inspired by ResNeSt [36] and U-Net++ [37]. By applying the split-attention module of ResNeSt to the output of each pyramid level of the U-Net++ structure and evolving normalization instead of batch normalization (BN) [38] and rectified linear unit (ReLU) [39], the class differentiation between feature maps of the level is improved, and efficient three-dimensional training with a small number of batch sizes is possible. It is possible to train it more efficiently and it has strengths in the inference speed. Moreover, it can be known that the three-dimensional patch-based SAU-Net has a higher level of spatial information for accurate segmentation of neuroanatomical structures when compared to QuickNAT and FastSurfer.
Starting with an accurate three-dimensional whole-brain segmentation provided by a deep-learning framework, we perform cortical and sub-cortical reconstruction and fast mapping with an approach that quickly maps the cortex using label fusion functions. It also maps cortical labels and creates a complete FreeSurfer [40]-based brain segmentation alternative. Only 20–30 s of run time contributes to the whole brain segmentation. Hence, SAU-Net has the speed of a supervised deep-learning method with the convenience of a wide range of segmentation-based capabilities offered by conventional neural imaging pipelines.
We extensively validate our proposed segmentation method by assessing segmentation accuracy, qualitative analysis, speed, and test–retest reliability in several publicly available datasets. This is the work of a deep-learning approach that has been thoroughly tested. Even though it is similar to or faster than SOTA deep-learning approaches, SAU-Net increases sensitivity and ability, making it a reliable tool for future large-scale demographic work.

2. Materials and Methods

2.1. Subjects

We utilized eight brain MRI datasets in our experiments. Six datasets with manual annotations, IXI [41], ADNI [42], HCP [43], PPMI [44], AIBL [45], and the Catholic Aging Brain Imaging (CABI) database [46] of Yeouido St Mary’s Hospital, the Catholic University of Korea, were selected for training and to evaluate segmentation accuracy. The remaining two datasets, CoRR-BNU1 [47] and ChinaSet Multi-Center [48], were used to test the reliability of the segmentation scheme. Table 1 summarizes the number of subjects per dataset, scanner type, matrix size, pixel spacing, the annotated information, and purpose. IXI, ADNI, HCP, PPMI, AIBL, and CABI among our datasets are resampled to a size of 256 × 256 × 256 and pixel spacing of 1.0 × 1.0 × 1.0 (a dataset consisting of IXI, ADNI, HCP, PPMI, AIBL, and CABI is called a “mixed dataset”). Details about the acquisition protocol used for each of the public datasets can be found in their respective references.
Our training data were generated by using FreeSurfer (General Hospital Corporation, Boston, MA, USA, version 6.0) annotations with manual correction in the training datasets for pre-training (250 selected cases from the mixed dataset), and manual annotations (confirmed and corrected by four radiologists) from 48 selected cases of the pre-training dataset for fine-tuning. From the selected 48 cases, 30 cases were used for fine-tuning and 18 cases were only used for testing. For the robust training, the total mixed data are composed of a total of 300 scans which are collected from ten different MRI scanners with 30 scans each, also, the selected 18 scans are collected from nine MRI scanners with two scans each. Table 2 shows the summary of various MRI scanners; the SKYRA scanner was excluded due to failure of the inference process on QuickNAT and FastSurfer (refer to Section 2.5).

2.2. Pre-Processing and Augmentation

2.2.1. Pre-Processing

The pre-processing tasks employed in brain image processing include image registration, zero padding, and histogram matching-based intensity normalization. The first step is a rigid registration from the target image to the MNI305 template (256 × 256 × 256) using 3D rigid registration with an Euler transform (SimpleITK) [49,50,51,52]. Next, zero padding is performed: 16 × 16 × 16 padding is used for training and 24 × 24 × 24 is used for testing. The intensities of the obtained scans vary across different devices and can even vary across different scans from the same device. Therefore, to further normalize the intensities across different scans, a histogram matching-based intensity normalization method is employed [53]. In this process, an MRI brain image with a protocol corresponding to a patients’ population are given as input. In the pre-training step, histogram transformation landmark parameters are first learned from MNI-registered image data (250 cases). The images are then transformed using the parameters learned during pre-training. This transformation is image dependent and needs to be done for each given image. All the pre-processing steps were implemented using TorchIO [54].

2.2.2. Data Augmentation for Training

To increase the robustness of the segmentation model, input images for pre-training and fine-tuning were randomly processed using the 3D data augmentation module (“Transforms”) of TorchIO [54] with the following augmentation types and settings:
  • Gaussian noise: Zero mean noise for standard deviation randomly chosen from the range 0 to 0.2
  • Bias field: Uniformly distributed in the range −0.5 to 0.5
  • Affine transform: A scale factor range of 0.9 to 1.1, rotation degree range of −10 to 10, trilinear interpolation (shearing and translation were not applied)
  • Elastic deformation: A probabilistic spin was given to the basic affine transform. Since the affine transform is linear, the image is uniformly deformed as linearity. On the other hand, elastic deformation deforms the image in different directions for each pixel. It is also common in medical imaging data since these are data observed in living things. [55] The parameter is pre-defined as seven control points for each axis X, Y, and Z.

2.3. SAU-Net Architecture

The architecture of SAU-Net is largely based on an encoder/decoder which has a skip pathway with a pyramid level similar to that of the U-Net++ architecture with an encoder sub-layer or backbone followed by a decoder sub-layer, as shown in Figure 1a [24,37]. The last layer is a classifier block using softmax. The proposed SAU-Net architecture includes re-designed skip pathways between all encoder and decoder blocks with the same spatial resolution [37]. Each convolution layer is constructed using Evolving Normalization–Activation Layers (EvoNorm) [38] and 3D convolution (Figure 1b). The number of convolution layers depends on the pyramid level, and a 3D ResNeSt block [36] is used at the output of the skip pathways. Essentially, the EvoNorm-based convolution and 3D ResNeSt blocks bring the semantic level of the encoder feature maps closer to that of the corresponding feature maps in the decoder. Based on this architecture, EvoNorm with group normalization is used in the convolution layer of the skip pathways instead of Batch Normalization (BN) and Rectified Linear Unit (ReLU) activation fucntion, and a 3D ResNeSt block is used in the skip connections instead of the existing 3D convolution layer [56]. This process is convenient for improving training efficiency with small batch size and segmentation accuracy for small subcortical structures.

2.3.1. ResNeSt Block with Split-Attention

In the proposed architecture, the 3D ResNeSt block is used in the skip connections to efficiently overcome the disadvantages of small batch sizes and enhance the response of the feature map. Its architecture is shown in Figure 2a. The split-attention module inside the ResNeSt block is a computational unit consisting of a group of feature maps and split-attention operations [36]. Figure 2b presents an overview of a split-attention module.
  • Groups of feature maps: The feature can be divided into four groups, and the number of groups of feature maps is given by a cardinality K = 2 and a radix R = 2 [36,57]. The total number of feature groups G is four (G = KR).
  • ResNeSt block (Figure 2a): The representations of a cardinal group are concatenated along the channel dimension. As in existing residual blocks, the split-attention module’s final output is produced by a shortcut connection. For blocks with a stride, an appropriate transformation is applied to the shortcut connection to align the output shapes.
  • Split-attention (Figure 2b): An integrated representation for each cardinal group can be obtained by combining it via an element-wise summation across multiple split layers [58,59]. Global semantic information with channel-wise features can be gathered by global average pooling across spatial dimensions [58,59]. A weighted combination of the cardinal group representation is aggregated using channel-wise softmax, where each feature map channel is produced using a weighted combination over split features.

2.3.2. Skip Pathways with Pyramid Levels

Each skip pathway among the nodes consists of a convolution block concatenated with three convolution and ResNeSt block layers. A convolution layer is preceded by a concatenation layer that combines the output from the previous convolution layer of the same concatenation block with the corresponding up-sampled output layer of the lower dense block [37]. Formally, we formulate the split-attention module-based skip pathway as follows:
x p ,   q = { C ( x p i ,   q ) ,                                                                             q = 0 R ( [ [ x p , k ] k = 0 q = 1 ,   U ( x p + i , q 1 ) ] ) ,                   q > 0          
where function C ( · ) is a convolution operation followed by an EvoNorm activation function, U ( · ) denotes an up-sampling layer, R ( · ) denotes a ResNeSt block, and [∙] denotes the layer with the concatenation function. All prior feature responses accumulate and reach the current node because we make use of a concatenated convolution block along each skip pathway. Finally, the ResNeSt block is applied to the concatenated convolution block as an input. Figure 1a illustrates this structure, showing how the feature maps travel through the top skip pathway of U-Net++.

2.3.3. Classifier Block

The feature map output by the final ResNeSt block is passed to the classifier block, which is a convolutional layer with a 1 × 1 × 1 kernel size that maps the input to an N channel feature map, where N is the number of classes (34 structure labels with non-labels (zero value) in our experiments).

2.4. Training with Limited Annotated Data

Our model produces a segmentation result by taking 96 × 96 × 96 patches randomly from the image. Because the 3D MRI volume would not fit into GPU memory using standard deep-learning networks, we randomly employed 128 independent instances of our model as 3D patches to cover the entire MNI space. Our model was a sub-network, whose resolution is a compromise between memory limitations and spatial resolution. For each instance of our model, we modified its decoder part to be compatible with 33-label output. As shown in Figure 1, 33 3D output channels have been employed in each classifier layer.
Although the amount of unlabeled data is rapidly growing, access to labeled data is still limited because of the intense effort needed for manual annotation. Simultaneously, deep-learning success is mainly led by supervised learning, whereas unsupervised approaches still need much research. In our proposed approach, we use the two-step training procedure shown in Figure 3. The first step is pre-training on large unlabeled datasets with auxiliary labels. In this step, we use a large neuroimaging dataset (public dataset) and process it with an existing tool to create auxiliary labels. We employ the widely used FreeSurfer to obtain auxiliary segmentations, but other tools could be used, depending on the application. We pre-train on this large dataset with auxiliary labels, which results in a network that imitates the FreeSurfer segmentations. Pre-training enforces a strong prior on the network, in which robustness to data heterogeneity is encouraged by the diversity of the selected cases. The next step is fine-tuning with a limited amount of manually labeled data: in this step, we take the pre-trained model and fine-tune it with small amount of data that have manual annotations. Instead of learning all filters from scratch, fine-tuning only focuses on rectifying the discrepancies between the auxiliary and manual labels.

2.5. Aggregation

When separating the entire MNI space into randomly selected 128 subspaces, if the subspaces overlap (12 × 12 × 12), this provides more than one inference result for a single voxel. A step beyond concatenation is required to obtain the final segmentation label for that voxel from multiple candidates. In this work, the “OR logical label fusion method” was employed to obtain the final segmentation results. Briefly, this method was used to fuse all the segmentations from cropped sub-images according to the size of patches and subspaces overlapping into a single final segmentation in MNI space. The space outside each patch was excluded from the label fusion. The final segmentation in the original target image space was then achieved by registering the segmentation to the original space using a rigid registration. Note that if the segments do not overlap, a naive concatenation is employed to obtain a final single segmentation directly without using label fusion.

2.6. Experimental Settings

We evaluated SAU-Net in a series of six experiments to evaluate its accuracy, quality, speed, reproducibility, and sensitivity on a variety of neuroimaging datasets, as summarized in Table 3. In all experiments, SAU-Net was trained as described in Section 2. In the experiment of segmentation performance, we evaluate our segmentation robustness by comparison with other methods. Next, we then performed another three experiments to validate the reliability and consistency of SAU-Net segmentations. To test reliability, we constructed the test dataset as follows:
For a fair comparison, the 30 scans for fine-tuning and the 18 scans for testing in the mixed dataset (Table 1) were consistent with the label definitions of QuickNAT (22 structure labels). Additionally, the existing U-Net [24] and U-Net++ [37], which are well-known architectures, were implemented in the NiftyNet library (based on TensorFlow), which is well optimized for 3D patch-based data loaders and training architectures [60]. The source code and inference model of QuickNAT were downloaded from the author’s GitHub page. [61] Also, the following link is the GitHub page of FastSurfer. [62]

2.7. Metrics for Evaluation

We utilize the Dice overlap [63] and average symmetric surface distance (ASSD) score [64] as an accuracy metric of segmentation, C V s _ a v g [65] and C V t [65] are used as reliability metrics of segmentation.
  • Dice overlap score: A statistical index which measures the similarity between the segmented region ( R o ) and ground truth ( R G ). The higher the estimate, the better the estimator. This score has become arguably the most commonly used index in the evaluation of image segmentation, as follows:
    D i c e   o v e r l a p = 2 | R o   R G | ( | R o | + | R G | )
  • Average symmetric surface distance (ASSD): The average of the distances from every point on the boundary ( B o ) of the segmented region to the boundary ( B G ) of the ground truth. The lower the estimate, the better the estimator.
    A S S D = 1 | B o | + | B G | ( x B o d ( x ,   B G ) + y B G d ( y , B o ) )
    where d is the Euclidean distance [66].
  • C V s _ a v g (Average intra-session coefficient of variation): The measurement of n total average variance for volume estimation was performed between two sessions of one paired scan set. The C V s _ a v g provides the extent of variability in the mean value. The average intra-session coefficient of variation is computed as:
    σ s = i = 2 ( x i μ s ) 2 2
    CV s _ avg = ( i = 0 n σ s / μ s )   n × 100
    where x i and μ s are the i th session measurements and mean of each pair of scan sets.
  • C V t (total coefficient of variation): The coefficient of variation in volume estimates within the total variance over n scans per subject. The C V t is computed as:
    σ t = i = n ( x i μ t ) 2 n
    CV t = σ t μ t × 100
    where x i and μ t are the i th measurements of scans and mean of total scans per subject.

2.8. Statistical Analysis

  • Accuracy: We measured using Dice overlap and ASSD for the selected 18 mixed datasets and used the mean and standard deviation for evaluation.
  • Speed: The atlas-based methods brought the results of the QuickNAT, and the rest of the methods measured the runtime of the final segmentation result after the pre-processing, inference, and post-processing were completed.
  • Device inter-variability: In Table 2, we show the acquired MRI scans according to ten MRI scanners, and conducted training and evaluation. These are necessary for robust training, and it is essential to show consistent performance for different devices. We evaluated the device inter-variability by comparing the overall mean Dice overlap and standard deviation.
  • Intra-session variability: This dataset was released to test the reliability of automated whole brain segmentation algorithms in estimating volumes of brain regions (excluding the corpus callosum) with 32 structure labels using a test–retest protocol. In this dataset, 47 MRI T1 scan pairs were taken on the same device with an interval of time between each scan (that is, two scans per session). All the scans were acquired over a period of six weeks (40.94 ± 4.51 days). We analyzed the C V s _ a v g of the overall 47-pair scan set. Ideally, as volume atrophy is almost insignificant within a period of 30 days, the coefficient of variation in the estimates should be zero.
  • Multi-center variability: In this experiment, we validated the reliability and robustness of estimating volumes across scans acquired from multiple centers. To do this, we used images from the ChinaSet Multi-Center dataset. This dataset includes scans from three patients who traveled to 12 different medical centers in China. Each of the 12 imaging centers used MRI scanners manufactured by the same vendors (different scanning parameters). We analyzed the C V t of the total variance over 12 scans per subject.

2.9. Training Setup

The SAU-Net was implemented using TorchIO library on the PyTorch 1.4 version and trained using Cross Entropy loss function [67] and AdamW optimizer with AMSGrad [68,69]. The learning rate of pre-training was 0.001; that of fine-tuning was 0.0001. The training was conducted for 300,000 iterations. The batch size was set to four, which was the limit that could be handled by the 48 GB RAM of the two QUADRO RTX6000 GPUs.

3. Results

3.1. Results of Segmentation Performance

3.1.1. Accuracy

In this experiment, we compared the performance of SAU-Net with SOTA methods and evaluated the impact of pre-training and fine-tuning. Table 4 shows the results measured using Dice overlap scores and the average symmetric surface distance (ASSD) with 22 structure labels.
Along each row, we observed that for our models, the fine-tuned model yields almost significantly (p < 0.05) better performance than the model pre-trained using the FreeSurfer labels. Table 4 and Figure 4 and Figure 5 show the Dice overlap score and ASSD results for comparison with SOTA brain segmentation methods. Along each row, we observe that for all models, the fine-tuned SAU-Net model yields better performance than U-Net (a difference of 0.5% in Dice score and 0.124 mm in ASSD), FastSurfer (1.3%, 0.131 mm), U-Net++ (1.3%, 0.151 mm), and QuickNAT (5.2%, 0.23 mm).
Additionally, we performed the two-sided t-test between SAU-Net fine-tuned model and the other each model. However, except for SAU-Net pre-trained model since it was trained using the same dataset. The U-Net, U-Net++, and QuickNAT show statistical significance in most of the whole brain ROI regions, but FastSurfer is challenging to confirm statistical significance in the Dice overlap results. (p>0.05 in more than half of the whole brain ROI regions) All the models show statistical significance in most of the whole brain ROI regions in the ASSD score results.

3.1.2. Qualitative Analysis

Sample segmentation results are shown in Figure 6 and Figure 7 for U-Net, U-Net++, FastSurfer, QuickNAT, pre-trained SAU-Net, and fine-tuned SAU-Net along with the radiologist-confirmed FreeSurfer labels. We indicate the important subcortical structures, the left (blue) lateral ventricles, with yellow boxes. We can observe the under-inclusion of the left lateral ventricle in the results obtained by U-Net, U-Net++, and FastSurfer. We also often observe specific classified regions where aliasing appears in the results of U-Net, QuickNAT, and FastSurfer but not in the results of U-Net++ and SAU-Net (indicated by red dashed circles).

3.1.3. Speed

Conventional template-based whole-brain segmentation approaches implemented for 3D deformable volume registration are slow [70]. The results take a long time because a single pair-wise registration takes about 2 h. In the Multi-Atlas Labeling Challenge (MALC) [41] dataset, the approximate segmentation runtime for Penn Image Computing & Science Lab (PICSL) [15] software for multi-atlas segmentation is 30 h per volume. FreeSurfer has its own template and takes about 4–6 h per volume. FastSurfer and QuickNAT are implemented using a 2D multi-view-based segmentation approach, and they take about 50 and 25 s, respectively, to run. Additionally, the 3D patch-based U-Net and U-Net++ take around 18–27 s to run [60]. These methods are substantially faster than the deformable registration-based segmentation approach. In contrast, SAU-Net segments a volume in 28 s, which is comparable to the runtimes of the faster methods. We present the segmentation runtimes in Figure 8 on a logarithmic scale.

3.2. Results of Segmentation Reliability

3.2.1. Device Inter-Variability

In the Dice overlap results, the fine-tuned SAU-Net model outperforms other models in about half of the eight MRI devices included in the dataset in this study (Table 2 and Table 5). In addition, SAU-Net outperforms other models because it has the highest average Dice score (fine-tuned SAU-Net) and lowest standard deviation (pre-trained SAU-Net) on all devices.

3.2.2. Intra-Session Reliability: CoRR-BNU1

We analyzed the C V s _ a v g of the overall 47-pair scan set of CoRR-BNU1. The lower the estimate, the better the estimator. We compare SAU-Net with FastSurfer in this regard. The results of the experiment are reported in Table 6. In the result, fine-tuned SAU-Net performs better than the FastSurfer model (except for accumbens right).

3.2.3. Multi-Center Reliability: ChinaSet Multi-Center

We analyzed the C V t of the total variance over 12 scans per subject (Table 7). We compared SAU-Net with FastSurfer and report the results in Table 7. SAU-Net is more robust for the cerebral white matter (WM) (left/right), lateral ventricle (left/right), caudate (left), accumbens (left), putamen (left/right), amygdala (left), and thalamus (left) structures, whereas FastSurfer performs better on the caudate (right), accumbens (right), amygdala (right), hippocampus (left/right), pallidum (left/right), and thalamus (right) structures. Overall, this challenging experiment demonstrates that SAU-Net and FastSurfer are almost equally robust.

4. Discussion

4.1. Comparison with Deep-Learning Methods

In recent years, CNNs have been proposed for whole-brain segmentation. [26,27,28,29,30,31,32,34,35,71,72] QuickNAT yields lower accuracy than other approaches while requiring only 20–30 s to run. QuickNAT has low accuracy because this model was pre-trained with FreeSurfer labels, but fine-tuned with Neuromorphometrics Inc.’s labels [73], so the overall accuracy is low. FastSurfer is a network proposed for segmenting the structure labels of FreeSurfer based on a 2D multi-view framework, which yields higher accuracy than QuickNAT. Although SLANT could not be directly compared with the SAU-Net model because it has more segmentation labels than our model, the performance seems to have similar or lower accuracy to SAU-Net. However, SLANT is slower in terms of speed (10–15 min). We compared U-Net, U-Net++, QuickNAT, and FastSurfer using an identical experimental setup with the Dice overlap and ASSD metrics. By evaluating SAU-Net on six different datasets and performing a reliability study on two datasets, we have presented a comprehensive evaluation of current CNN-based methods for brain segmentation.

4.2. Pre-Training with Auxiliary Labels

Even if the deep learning-based methods have been highly effective, their cost is high and they require an extensive amount of annotated data for robust training [39]. Access to plentiful annotated data for training is challenging for medical-based applications because of the high cost of creating expert manual annotations. This issue is more critical for deep learning-based methods, where each image slice corresponds to a one-shot data representation, in contrast to patch-based approaches, where many patches can be extracted from a volume [28]. We evaluated a training strategy that leverages extensive unlabeled annotation data and a small amount of manual annotation data to train our SAU-Net model to overcome this issue effectively [29]. We utilized FreeSurfer to automatically segment brain structure labels from unlabeled data, and these labels were used as auxiliary labels to pre-train our model. This pre-trained model, which imitates FreeSurfer’s segmentations, is then fine-tuned with a small amount of manually labeled data to obtain the final model. Our results have demonstrated that a model trained with this training strategy outperforms the same model pre-trained on auxiliary label data only.

4.3. Architectural Design for Deep Learning

The architecture of SAU-Net has been modified to address the challenges associated with multi-label brain segmentation. Our architecture offers faster processing and a broader context than patch-based U-Net and U-Net++. This is because, by adding the split-attention module to the U-Net++ structure, the separation between labels seems to be improved thanks to the split grouping of the feature channels [36]. In the skip pathways with the pyramid levels, concatenating all the layers, which enforces spatial consistency, is an essential aspect for segmenting small subcortical structures. In addition, using EvoNorm [38] instead of a combination of ReLU and batch normalization overcomes the disadvantage that large batch sizes cannot be used because of the property of 3D patches, and shows that sufficient performance is achieved even when small batch sizes are used in training. Our results show that the SAU-Net architecture is a significant improvement on the architectures of the SOTA deep-learning models.

4.4. Segmentation Performance

We demonstrated the high accuracy of SAU-Net in a comprehensive experiment that covers variations in acquisition parameters and neuroanatomy. In an experiment on the mixed dataset, we demonstrated that SAU-Net provides segmentation accuracy and ASSD scores that are slightly higher than those of the SOTA deep-learning-based methods (Table 3, Figure 4 and Figure 5). Our results are similar or slightly better than the performance (for 22 structure labels) of all other methods except for QuickNAT. However, our models show better results in the subcortical areas such as the lateral ventricle. Notably, SAU-Net has not failed on any scan in our datasets.

4.5. Segmentation Reliability

In two additional experiments, we evaluated the reliability of SAU-Net. We observed high consistency for the segmentation of nine representative structures (left and right) on test–retest data with less than 1–4% variation in most brain structures (Table 5). The test–retest data were acquired in the same scanner. We extended the evaluation to a more challenging dataset, where the same subject was scanned in the same machines with various scanning parameters at different sites. As expected, the variation increased in this setup, but the reliability was comparable to that of FastSurfer (Table 6). Since FastSurfer presented the best performance amongst SOTA deep-learning methods (except for NiftyNet-based U-Net style architecture [60]), we have conducted further analysis to compare the reliability of SAU-Net.
However, our results showed that the more unconstrained, deep-learning-based segmentation could achieve higher reliability.

4.6. Limitations

As with any brain segmentation method, SAU-Net has limitations. Our mixed dataset includes various public and patient data from domestic hospitals. However, our labeled data are an improved manual annotation based on FreeSurfer’s segmentation, so it cannot be directly compared with FreeSurfer. In the future, we need to utilize better-defined manual annotations. Additionally, the SAU-Net model, like other SOTA methods, has lower accuracy in smaller subcortical regions such as the accumbens and amygdala than in other brain regions. The accuracy was improved after the model was fine-tuned, however, this issue remains a challenging task.

5. Conclusions

We introduced SAU-Net, a deep-learning-based method for brain segmentation that achieved a great performance with respect to existing methods and is as fast as or orders of magnitude faster than the SOTA deep-learning and template-based approaches. We have demonstrated that SAU-Net generalizes well to other unseen datasets when the training data differ from the testing data and yields high segmentation accuracy. This high segmentation accuracy enhances group analyses by obtaining effect sizes and significance values that better match those of manual segmentations. Furthermore, because of its high test–retest accuracy, it can be effectively used for longitudinal studies. We believe SAU-Net can be a highly impactful segmentation network due to its robustness to neuroanatomical variability, allowing for almost instantaneous access to accurate imaging biomarkers along with its fast processing time.

Author Contributions

Conceptualization, M.L. and D.K.; methodology, M.L.; software, M.L.; validation, M.L. and R.E.K.; formal analysis, M.L. and R.E.K.; investigation, M.L. and Z.R.; resources, D.K.; data curation, J.K., H.G.K., S.W.O., M.K.L. and J.H.Y.; writing—original draft preparation, M.L. and J.K.; writing—review and editing, H.-K.L., R.E.K., J.K., D.W.K., S.-M.W., N.-Y.K., D.K. and Z.R.; visualization, M.L.; supervision, D.K. and H.K.L.; project administration, D.K.; funding acquisition, D.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Research and Business Development Program through the Korea Institute for Advancement of Technology (KIAT) funded by the Ministry of Trade, Industry and Energy (MOTIE) (grant number: P0009507) and the “HPC Support” Project, supported by the Ministry of Science and ICT and National IT Industry Promotion Agency (NIPA).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Giedd, J.N.; Blumenthal, J.; Jeffries, N.O.; Castellanos, F.X.; Liu, H.; Zijdenbos, A.; Paus, T.; Evans, A.C.; Rapoport, J.L. Brain development during childhood and adolescence: A longitudinal MRI study. Nat. Neurosci. 1999, 2, 861–863. [Google Scholar] [CrossRef] [PubMed]
  2. Shaw, P.; Greenstein, D.; Lerch, J.; Clasen, L.; Lenroot, R.; Gogtay, N.; Evans, A.; Rapoport, J.; Giedd, J. Intellectual ability and cortical development in children and adolescents. Nat. Cell Biol. 2006, 440, 676–679. [Google Scholar] [CrossRef] [PubMed]
  3. Alexander-Bloch, A.F.; Giedd, J.N.; Bullmore, E.T. Imaging structural co-variance between human brain regions. Nat. Rev. Neurosci. 2013, 14, 322–336. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Lerch, J.P.; Van Der Kouwe, A.J.; Raznahan, A.; Paus, T.; Johansen-Berg, H.; Miller, K.L.; Smith, S.M.; Fischl, B.; Sotiropoulos, S.N. Studying neuroanatomy using MRI. Nat. Neurosci. 2017, 20, 314–326. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Pham, D.L.; Xu, C.; Prince, J.L. Current Methods in Medical Image Segmentation. Annu. Rev. Biomed. Eng. 2000, 2, 315–337. [Google Scholar] [CrossRef]
  6. Gass, T.; Székely, G.; Goksel, O. Semi-supervised Segmentation Using Multiple Segmentation Hypotheses from a Single Atlas. In International MICCAI Workshop on Medical Computer Vision; Springer: Berlin/Heidelberg, Germany, 2013; Volume 7766, pp. 29–37. [Google Scholar]
  7. Wu, M.; Rosano, C.; Lopez-Garcia, P.; Carter, C.S.; Aizenstein, H.J. Optimum template selection for atlas-based segmentation. NeuroImage 2007, 34, 1612–1618. [Google Scholar] [CrossRef]
  8. Doan, N.T.; De Xivry, J.O.; Macq, B. Effect of inter-subject variation on the accuracy of atlas-based segmentation applied to human brain structures. In Proceedings of the Medical Imaging 2010: Image Processing, San Diego, CA, USA, 12 March 2010. [Google Scholar] [CrossRef]
  9. Iglesias, J.E.; Sabuncu, M.R. Multi-atlas segmentation of biomedical images: A survey. Med. Image Anal. 2015, 24, 205–219. [Google Scholar] [CrossRef] [Green Version]
  10. Heckemann, R.A.; Hajnal, J.V.; Aljabar, P.; Rueckert, D.; Hammers, A. Automatic anatomical brain MRI segmentation combining label propagation and decision fusion. NeuroImage 2006, 33, 115–126. [Google Scholar] [CrossRef]
  11. Rohlfing, T.; Russakoff, D.B.; Maurer, C.R. Performance-Based Classifier Combination in Atlas-Based Image Segmentation Using Expectation-Maximization Parameter Estimation. IEEE Trans. Med. Imaging 2004, 23, 983–994. [Google Scholar] [CrossRef]
  12. Asman, A.J.; Landman, B.A. Non-local statistical label fusion for multi-atlas segmentation. Med. Image Anal. 2013, 17, 194–208. [Google Scholar] [CrossRef] [Green Version]
  13. Bai, W.; Shi, W.; O’Regan, D.P.; Tong, T.; Wang, H.; Jamil-Copley, S.; Peters, N.S.; Rueckert, D. A Probabilistic Patch-Based Label Fusion Model for Multi-Atlas Segmentation with Registration Refinement: Application to Cardiac MR Images. IEEE Trans. Med. Imaging 2013, 32, 1302–1315. [Google Scholar] [CrossRef] [PubMed]
  14. Coupé, P.; Manjón, J.V.; Fonov, V.; Pruessner, J.; Robles, M.; Collins, D.L. Patch-based segmentation using expert priors: Application to hippocampus and ventricle segmentation. NeuroImage 2011, 54, 940–954. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Wang, H.; Yushkevich, P.A. Multi-atlas segmentation with joint label fusion and corrective learning—An open source implementation. Front. Aging Neurosci. 2013, 7, 27. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Wu, G.; Kim, M.; Sanroma, G.; Wang, Q.; Munsell, B.C.; Shen, D. Hierarchical multi-atlas label fusion with multi-scale feature representation and label-specific patch partition. NeuroImage 2015, 106, 34–46. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Eskildsen, S.F.; Coupé, P.; Fonov, V.; Manjón, J.V.; Leung, K.K.; Guizard, N.; Wassef, S.N.; Østergaard, L.R.; Collins, D.L. BEaST: Brain extraction based on nonlocal segmentation technique. NeuroImage 2012, 59, 2362–2373. [Google Scholar] [CrossRef] [PubMed]
  18. Wolz, R.; Aljabar, P.; Hajnal, J.V.; Hammers, A.; Rueckert, D. LEAP: Learning embeddings for atlas propagation. NeuroImage 2010, 49, 1316–1325. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Asman, A.J.; Huo, Y.; Plassard, A.J.; Landman, B.A. Multi-atlas learner fusion: An efficient segmentation approach for large-scale data. Med. Image Anal. 2015, 26, 82–91. [Google Scholar] [CrossRef] [Green Version]
  20. Bai, W.; Shi, W.; Ledig, C.; Rueckert, D. Multi-atlas segmentation with augmented features for cardiac MR images. Med. Image Anal. 2015, 19, 98–109. [Google Scholar] [CrossRef]
  21. Tong, T.; Wolz, R.; Coupé, P.; Hajnal, J.V.; Rueckert, D. Segmentation of MR images via discriminative dictionary learning and sparse coding: Application to hippocampus labeling. NeuroImage 2013, 76, 11–23. [Google Scholar] [CrossRef] [Green Version]
  22. Vakalopoulou, M.; Chassagnon, G.; Bus, N.; Marini, R.; Zacharaki, E.I.; Revel, M.-P.; Paragios, N. AtlasNet: Multi-atlas Non-linear Deep Networks for Medical Image Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2018; pp. 658–666. [Google Scholar]
  23. Wang, H.; Kakrania, D.; Tang, H.; Prasanna, P.; Syeda-Mahmood, T.F. Fast anatomy segmentation by combining coarse scale multi-atlas label fusion with fine scale corrective learning. Comput. Med. Imaging Graph. 2018, 68, 16–24. [Google Scholar] [CrossRef]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  25. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  26. De Brebisson, A.; Montana, G. Deep neural networks for anatomical brain segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015; pp. 20–28. [Google Scholar]
  27. Mehta, R.; Majumdar, A.; Sivaswamy, J. BrainSegNet: A convolutional neural network architecture for automated segmentation of human brain structures. J. Med. Imaging 2017, 4, 024003. [Google Scholar] [CrossRef] [PubMed]
  28. Wachinger, C.; Reuter, M.; Klein, T. DeepNAT: Deep convolutional neural network for segmenting neuroanatomy. NeuroImage 2018, 170, 434–445. [Google Scholar] [CrossRef] [PubMed]
  29. Roy, A.G.; Conjeti, S.; Navab, N.; Wachinger, C. QuickNAT: A fully convolutional network for quick and accurate segmentation of neuroanatomy. NeuroImage 2019, 186, 713–727. [Google Scholar] [CrossRef] [Green Version]
  30. Henschel, L.; Conjeti, S.; Estrada, S.; Diers, K.; Fischl, B.; Reuter, M. FastSurfer—A fast and accurate deep learning based neuroimaging pipeline. NeuroImage 2020, 219, 117012. [Google Scholar] [CrossRef]
  31. Rajchl, M.; Pawlowski, N.; Rueckert, D.; Matthews, P.M.; Glocker, B. Neuronet: Fast and robust reproduction of multiple brain image segmentation pipelines. arXiv 2018, arXiv:1806.04224. [Google Scholar]
  32. Ganaye, P.-A.; Sdika, M.; Benoit-Cattin, H. Semi-supervised Learning for Segmentation Under Semantic Constraint. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2018; pp. 595–602. [Google Scholar]
  33. Brett, M.; Johnsrude, I.S.; Owen, A.M. The problem of functional localization in the human brain. Nat. Rev. Neurosci. 2002, 3, 243–249. [Google Scholar] [CrossRef]
  34. Li, W.; Wang, G.; Fidon, L.; Ourselin, S.; Cardoso, M.J.; Vercauteren, T. On the Compactness, Efficiency, and Representation of 3D Convolutional Networks: Brain Parcellation as a Pretext Task. In International Conference on Information Processing in Medical Imaging; Springer: Berlin/Heidelberg, Germany, 2017; pp. 348–360. [Google Scholar]
  35. Huo, Y.; Xu, Z.; Xiong, Y.; Aboud, K.; Parvathaneni, P.; Bao, S.; Bermudez, C.; Resnick, S.M.; Cutting, L.E.; Landman, B.A. 3D whole brain segmentation using spatially localized atlas network tiles. NeuroImage 2019, 194, 105–119. [Google Scholar] [CrossRef] [Green Version]
  36. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Zhang, Z.; Lin, H.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.J. Resnest: Split-attention networks. arXiv 2020, arXiv:2004.08955. [Google Scholar]
  37. Zhou, Z.; Siddiquee, M.M.R.; Liang, N.T.J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  38. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  39. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  40. Fischl, B. FreeSurfer. Neuroimage 2012, 62, 774–781. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Multi-Atlas Labeling Challenge (MALC). Available online: http://www.neuromorphometrics.com/2012_MICCAI_Challenge_Data.html (accessed on 15 September 2020).
  42. Access Data and Samples. Available online: http://adni.loni.usc.edu/data-samples/access-data/ (accessed on 12 August 2019).
  43. What Is the Connectome Coordination Facility. Available online: https://www.humanconnectome.org/ (accessed on 22 July 2019).
  44. Parkinson’s Progression Markers Initiative. Available online: https://www.ppmi-info.org/access-data-specimens/download-data/ (accessed on 6 March 2019).
  45. News & Publications. Available online: http://adni.loni.usc.edu/aibl-australian-imaging-biomarkers-and-lifestyle-study-of-ageing-18-month-data-now-released/ (accessed on 13 August 2019).
  46. Kang, N.W.; Choi, W.H.; Jung, W.S.; Um, Y.H.; Lee, C.U.; Lim, H.K. Impact of Amyloid Burden on Regional Functional Synchronization in the Cognitively Normal Older Adults. Sci. Rep. 2017, 7, 14690. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Consortium for Reliability and Reproducibility (CoRR). Available online: http://fcon_1000.projects.nitrc.org/indi/CoRR/html/bnu_1.html (accessed on 11 September 2020).
  48. Figshare. Available online: https://figshare.com/articles/Multicenter_dataset_of_multishell_diffusion_magnetic_resonance_imaging_in_healthy_traveling_adults_with_identical_setting/8851955/6 (accessed on 11 September 2020).
  49. Beare, R.; Lowekamp, B.; Yaniv, Z. Image Segmentation, Registration and Characterization in R with SimpleITK. J. Stat. Softw. 2018, 86, 1–35. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Evans, A.C.; Collins, D.L.; Mills, S.R.; Brown, E.D.; Kelly, R.L.; Peters, T.M. 3D statistical neuroanatomical models from 305 MRI volumes. In Proceedings of the 1993 IEEE Conference Record Nuclear Science Symposium and Medical Imaging Conference, San Francisco, CA, USA, 31 October–6 November 1993. [Google Scholar]
  51. Lowekamp, B.C.; Chen, D.T.; Ibáñez, L.; Blezek, D. The design of SimpleITK. Front. Neuroinform. 2013, 7, 45. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Yaniv, Z.R.; Lowekamp, B.C.; Johnson, H.J.; Beare, R. SimpleITK Image-Analysis Notebooks: A Collaborative Environment for Education and Reproducible Research. J. Digit. Imaging 2017, 31, 290–303. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Nyul, L.; Udupa, J.; Zhang, X. New variants of a method of MRI scale standardization. IEEE Trans. Med. Imaging 2000, 19, 143–150. [Google Scholar] [CrossRef] [PubMed]
  54. Pérez-García, F.; Sparks, R.; Ourselin, S. TorchIO: A Python library for efficient loading, preprocessing, augmentation and patch-based sampling of medical images in deep learning. arXiv 2020, arXiv:2003.04696. [Google Scholar]
  55. Simard, P.Y.; Steinkraus, D.; Platt, J.C. Best practices for convolutional neural networks applied to visual document analysis. In Proceedings of the Icdar, Edinburgh, UK, 6 August 2003. [Google Scholar]
  56. Liu, H.; Brock, A.; Simonyan, A.; Le, Q.V. Evolving Normalization-Activation Layers. arXiv 2020, arXiv:2004.02967. [Google Scholar]
  57. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  58. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  59. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 510–519. [Google Scholar]
  60. Gibson, E.; Li, W.; Sudre, C.H.; Fidon, L.; Shakir, D.I.; Wang, G.; Eaton-Rosen, Z.; Gray, R.; Doel, T.; Hu, Y.; et al. NiftyNet: A deep-learning platform for medical imaging. Comput. Methods Programs Biomed. 2018, 158, 113–122. [Google Scholar] [CrossRef]
  61. GitHub. Available online: https://github.com/ai-med/quickNAT_pytorch (accessed on 5 October 2020).
  62. GitHub. Available online: https://github.com/Deep-MI/FastSurfer (accessed on 5 October 2020).
  63. Sorensen, T.A. A method of establishing groups of equal amplitude in plant sociology based on similarity of species content and its application to analyses of the vegetation on Danish commons. Biol. Skar. 1948, 5, 1–34. [Google Scholar]
  64. Heimann, T.; Van Ginneken, B.; Styner, M.A.; Arzhaeva, Y.; Aurich, V.; Bauer, C.; Beck, A.; Becker, C.; Beichel, R.; Bekes, G.; et al. Comparison and Evaluation of Methods for Liver Segmentation from CT Datasets. IEEE Trans. Med. Imaging 2009, 28, 1251–1265. [Google Scholar] [CrossRef] [PubMed]
  65. MacLaren, J.; Han, Z.; Vos, S.B.; Fischbein, N.; Bammer, R. Reliability of brain volume measurements: A test-retest dataset. Sci. Data 2014, 1, 140037. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Arya, S.; Mount, D.; Netanyahu, N.S.; Silverman, R.; Wu, A.Y. An optimal algorithm for approximate nearest neighbor searching fixed dimensions. J. ACM 1998, 45, 891–923. [Google Scholar] [CrossRef] [Green Version]
  67. Rubinstein, R.Y.; Kroese, D.P. The Cross-Entropy Method: A Unified Approach to Combinatorial Optimization, Monte-Carlo Simulation and Machine Learning; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  68. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  69. Reddi, S.J.; Kale, S.; Kumar, S. On the convergence of adam and beyond. arXiv 2019, arXiv:1904.09237. [Google Scholar]
  70. Avants, B.B.; Tustison, N.; Song, G. Advanced normalization tools (ANTS). Insight J. 2009, 2, 1–35. [Google Scholar]
  71. Dey, R.; Hong, Y. CompNet: Complementary Segmentation Network for Brain MRI Extraction. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2018; pp. 628–636. [Google Scholar]
  72. Wong, K.C.L.; Moradi, M.; Tang, H.; Syeda-Mahmood, T. 3D Segmentation with Exponential Logarithmic Loss for Highly Unbalanced Object Sizes. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2018; pp. 612–619. [Google Scholar]
  73. Klein, A.; Ghosh, S.S.; Bao, F.S.; Giard, J.; Häme, Y.; Stavsky, E.; Lee, N.; Rossa, B.; Reuter, M.; Neto, E.C.; et al. Mindboggling morphometry of human brains. PLoS Comput. Biol. 2017, 13, e1005350. [Google Scholar] [CrossRef]
Figure 1. Split-attention U-Net (SAU-Net): (a) overall architecture and (b) modified convolution layer. (EvoNorm-S0: Evolving normalization layer with the Group Normalization and Swish activation function).
Figure 1. Split-attention U-Net (SAU-Net): (a) overall architecture and (b) modified convolution layer. (EvoNorm-S0: Evolving normalization layer with the Group Normalization and Swish activation function).
Brainsci 10 00974 g001
Figure 2. ResNeSt block: (a) overall architecture, (b) split-attention module, and (c) convolution layer. (BN: Batch Normalization, ReLU: Rectified Linear Unit activation function).
Figure 2. ResNeSt block: (a) overall architecture, (b) split-attention module, and (c) convolution layer. (BN: Batch Normalization, ReLU: Rectified Linear Unit activation function).
Brainsci 10 00974 g002
Figure 3. Two-step training strategy for SAU-Net. First, we used FreeSurfer to automatically segment a large unlabeled dataset (a mixed dataset consisting of 250 scans listed in the Table 1). These labels are referred to as auxiliary labels and used for pre-training. Second, we fine-tuned the network on 48 scans selected from the same public dataset that were manually annotated by four radiologists. Fine-tuning does not start from scratch but continues to optimize the pre-trained model to leverage the scarce amount of data with manual annotations.
Figure 3. Two-step training strategy for SAU-Net. First, we used FreeSurfer to automatically segment a large unlabeled dataset (a mixed dataset consisting of 250 scans listed in the Table 1). These labels are referred to as auxiliary labels and used for pre-training. Second, we fine-tuned the network on 48 scans selected from the same public dataset that were manually annotated by four radiologists. Fine-tuning does not start from scratch but continues to optimize the pre-trained model to leverage the scarce amount of data with manual annotations.
Brainsci 10 00974 g003
Figure 4. Comparison of the Dice overlap score results of SAU-Net, U-Net, U-Net++, FastSurfer, and QuickNAT on the 18 test cases from the mixed dataset for all 22 structure labels. Statistical significance (p < 0.05) with respect to SAU-Net fine-tuned model is presented by each colored asterisk. The p-values were estimated using a two-sided t-test. The box represents the confidence interval with 50th percentile, and the line outside the box is 95th percentile, while the whiskers represent the rest of the data except for data it determines to be outliers. (GM: Gray Matter, WM: White Matter).
Figure 4. Comparison of the Dice overlap score results of SAU-Net, U-Net, U-Net++, FastSurfer, and QuickNAT on the 18 test cases from the mixed dataset for all 22 structure labels. Statistical significance (p < 0.05) with respect to SAU-Net fine-tuned model is presented by each colored asterisk. The p-values were estimated using a two-sided t-test. The box represents the confidence interval with 50th percentile, and the line outside the box is 95th percentile, while the whiskers represent the rest of the data except for data it determines to be outliers. (GM: Gray Matter, WM: White Matter).
Brainsci 10 00974 g004
Figure 5. Comparison of ASSD results of SAU-Net, U-Net, U-Net++, FastSurfer, and QuickNAT on the 18 test cases from the mixed dataset for all 22 structure labels. Statistical significance (p < 0.05) with respect to SAU-Net fine-tuned model is presented by each colored asterisk. The p-values were estimated using a two-sided t-test. The box represents the confidence interval with 50th percentile, and the line outside the box is 95th percentile, while the whiskers represent the rest of the data except for data it determines to be outliers. (GM: Gray Matter, WM: White Matter).
Figure 5. Comparison of ASSD results of SAU-Net, U-Net, U-Net++, FastSurfer, and QuickNAT on the 18 test cases from the mixed dataset for all 22 structure labels. Statistical significance (p < 0.05) with respect to SAU-Net fine-tuned model is presented by each colored asterisk. The p-values were estimated using a two-sided t-test. The box represents the confidence interval with 50th percentile, and the line outside the box is 95th percentile, while the whiskers represent the rest of the data except for data it determines to be outliers. (GM: Gray Matter, WM: White Matter).
Brainsci 10 00974 g005
Figure 6. Example results for qualitative evaluation: (a) manually annotated image, (b) U-Net, (c) U-Net++, (d) QuickNAT, (e) FastSurfer, (f) pre-trained SAU-Net, and (g) fine-tuned SAU-Net.
Figure 6. Example results for qualitative evaluation: (a) manually annotated image, (b) U-Net, (c) U-Net++, (d) QuickNAT, (e) FastSurfer, (f) pre-trained SAU-Net, and (g) fine-tuned SAU-Net.
Brainsci 10 00974 g006
Figure 7. Example results with zoom-in for qualitative evaluation: (a) manually annotated image, (b) U-Net, (c) U-Net++, (d) QuickNAT, (e) FastSurfer, (f) pre-trained SAU-Net, and (g) fine-tuned SAU-Net. the red dashed circles indicate low-quality results, such as aliasing, and the yellow boxes show under-inclusion of the left lateral ventricles (blue color).
Figure 7. Example results with zoom-in for qualitative evaluation: (a) manually annotated image, (b) U-Net, (c) U-Net++, (d) QuickNAT, (e) FastSurfer, (f) pre-trained SAU-Net, and (g) fine-tuned SAU-Net. the red dashed circles indicate low-quality results, such as aliasing, and the yellow boxes show under-inclusion of the left lateral ventricles (blue color).
Brainsci 10 00974 g007
Figure 8. Runtime performance of the comparison methods.
Figure 8. Runtime performance of the comparison methods.
Brainsci 10 00974 g008
Table 1. Summary of the public datasets used for training and testing.
Table 1. Summary of the public datasets used for training and testing.
DatasetNo. of SubjectsScanner TypeMatrix SizePixel Spacing (mm)AnnotationsPurpose
IXI17T1256 × 256 × 2561.0 × 1.0 × 1.033 labelsAll
ADNI207
HCP30
PPMI8
AIBL14
CABI24
Total mixed dataset300
CoRR-BNU1 47144 × 256 × 2561.3 × 1.0 × 1.0NoneOnly testing
ChinaSet
Multi-Center
3240 × 256 × 2561.0 × 1.0 × 1.2NoneOnly testing
Total testing only dataset50
33 labels: 33 structure labels of the whole brain. All: training/testing. “: indication that it is the same as the row line above.
Table 2. Summary of the numbers of subjects according to 10 magnetic resonance imaging (MRI) scanners and description.
Table 2. Summary of the numbers of subjects according to 10 magnetic resonance imaging (MRI) scanners and description.
MRI ScannerNo. of SubjectsPre-TrainingFine-TuningTestingDescription
Intera302532Philips Intera 3.0T
Prisma fitSiemens 3.0T Magnetom Prisma Fit
IngeniaPhilips MRI Ingenia 3.0T
Signa HDxtGE Signa HDxt 3.0T MR system
Signa HDxGE Signa HDx 3.0T MRI system
AchievaPhilips Achieva 3.0T MRI system
DISCOVERY MR750GE Discovery™ MR750 3.0T system
TrioTim Siemens 3.0T Trio TIM MRI scanner
VerioSiemens 3.0T MAGNETOM Verio
SKYRA Siemens 3.0T MAGNETOM Skyra
“: indication that it is the same as the row line above.
Table 3. Experimental settings for the evaluation of SAU-Net.
Table 3. Experimental settings for the evaluation of SAU-Net.
EvaluationExperimentTestingPurpose of Experiment
Segmentation PerformanceAccuracySelected mixed dataset for testing only (18) *Comparison of Dice overlap and ASSD scores with other methods
Qualitative AnalysisComparison of segmentation quality with other methods
SpeedComparison of runtime speed with other methods
Segmentation ReliabilityDevice Inter-VariabilityInter-scan reliability with nine MRI scanners
Intra-Session ReliabilityCoRR-BNU1 (47)Intra-session reliability with two scans per session (same scanner)
Multi-Center ReliabilityChinaSet Multi-Center (3)Reliability across 12 locations with same scanners
ASSD: average symmetric surface distance. * Numbers in parentheses: numbers of the corresponding dataset. “: indication that it is the same as the row line above.
Table 4. Comparison of the Dice overlap score and ASSD of SAU-Net with 22 structure labels (mean ± standard deviation).
Table 4. Comparison of the Dice overlap score and ASSD of SAU-Net with 22 structure labels (mean ± standard deviation).
MethodDice overlapASSD (mm)
U-Net0.892 ± 0.049 *0.557 ± 0.326 *
U-Net++0.884 ± 0.061 *0.584 ± 0.334 *
QuickNAT0.845 ± 0.080 *0.663 ± 0.167 *
FastSurfer0.884 ± 0.0620.564 ± 0.332 *
SAU-Net pre-trained0.887 ± 0.0640.479 ± 0.159
SAU-Net fine-tuned0.897 ± 0.0560.433 ± 0.136
Star symbols in bold (*): Statistical significance (p < 0.05) between the SAU-Net fine-tuned model (reference) and the other models. (Except for SAU-Net pre-trained model since the training data is same with the SAU-Net fine-tuned model).
Table 5. Device inter-variability with respect to the Dice overlap of SAU-Net and other State-Of-The-Art (SOTA) methods for 22 structure labels (mean ± standard deviation).
Table 5. Device inter-variability with respect to the Dice overlap of SAU-Net and other State-Of-The-Art (SOTA) methods for 22 structure labels (mean ± standard deviation).
MRI ScannerSAU-Net Fine-TunedSAU-Net Pre-TrainedU-NetU-Net++FastSurferQuickNAT
Intera0.907 ± 0.0550.892 ± 0.0770.901 ± 0.0550.894 ± 0.0680.904 ± 0.0730.851 ± 0.103
Prisma fit0.891 ± 0.0730.898 ± 0.0510.900 ± 0.0390.898 ± 0.0480.902 ± 0.0490.846 ± 0.076
Ingenia0.903 ± 0.0540.877 ± 0.1290.878 ± 0.1210.877 ± 0.1110.876 ± 0.1290.837 ± 0.134
Signa HDxt0.884 ± 0.1100.883 ± 0.0640.890 ± 0.0550.882 ± 0.0620.875 ± 0.0730.842 ± 0.085
Signa HDx0.894 ± 0.0530.875 ± 0.1000.872 ± 0.1000.897 ± 0.0550.910 ± 0.0440.857 ± 0.069
Achieva0.917 ± 0.0440.901 ± 0.0580.900 ± 0.0550.895 ± 0.0570.901 ± 0.0580.852 ± 0.079
DISCOVERY MR7500.899 ± 0.0560.882 ± 0.0720.883 ± 0.0640.873 ± 0.0810.885 ± 0.0760.837 ± 0.102
TrioTim0.899 ± 0.0610.891 ± 0.0610.900 ± 0.0470.885 ± 0.070.896 ± 0.0540.860 ± 0.069
Verio0.898 ± 0.0680.885 ± 0.0760.888 ± 0.0640.879 ± 0.0730.880 ± 0.0740.877 ± 0.074
Total0.897 ± 0.0560.887 ± 0.0640.892 ± 0.0490.884 ± 0.0490.884 ± 0.0620.845 ± 0.080
Table 6. Variation in volume measurement per structure. Fine-tuned SAU-Net is compared with FastSurfer in terms of the C V s _ a v g between them. The mean volume estimates per structure are also reported.
Table 6. Variation in volume measurement per structure. Fine-tuned SAU-Net is compared with FastSurfer in terms of the C V s _ a v g between them. The mean volume estimates per structure are also reported.
StructureVolume (cc) C V s _ a v g   ( % )
SAU-NetFastSurferSAU-NetFastSurfer
Cerebral WM Left224.703238.1690.2942.553
Cerebral WM Right231.598238.0530.3572.496
Lateral Ventricle Left7.1586.6321.7376.509
Lateral Ventricle Right6.7165.5341.6886.622
Caudate Left3.9693.8391.0263.291
Caudate Right3.4573.9961.1963.551
Accumbens Left0.5590.663.9053.599
Accumbens Right0.6680.6542.7934.37
Putamen Left5.6975.5141.1042.537
Putamen Right5.7265.5790.8322.479
Amygdala Left2.0491.751.9672.827
Amygdala Right2.021.8651.5022.872
Hippocampus Left4.1234.2431.1322.524
Hippocampus Right4.5444.4211.0822.355
Pallidum Left2.0632.1743.6482.872
Pallidum Right1.8852.1242.292.766
Thalamus Left8.2557.9721.0571.997
Thalamus Right7.2757.8550.7811.95
SAU-Net: fine-tuned SAU-Net model.
Table 7. C V t in volume estimation for the nine representative structures (left/right) for each subject, using fine-tuned SAU-Net (SAU-Net) and FastSurfer.
Table 7. C V t in volume estimation for the nine representative structures (left/right) for each subject, using fine-tuned SAU-Net (SAU-Net) and FastSurfer.
Structure Subject   ID   with   C V t   ( % )
Subject 1Subject 2Subject 3
SAU-NetFastSurferSAU-NetFastSurferSAU-NetFastSurfer
Cerebral WM Left1.1261.321.9792.3541.8872.105
Cerebral WM Right0.8851.2351.5542.0182.0042.286
Lateral Ventricle Left4.9654.7762.8123.0884.1515.213
Lateral Ventricle Right5.5715.9782.1063.2393.6394.215
Caudate Left1.6011.7891.8072.6742.3231.366
Caudate Right1.7821.6031.391.0811.2621.742
Accumbens Left7.34112.6794.4927.0895.8284.597
Accumbens Right8.4818.0247.646.6888.727.082
Putamen Left2.6362.7071.5621.4011.7372.58
Putamen Right1.7252.0381.0821.2021.5192.042
Amygdala Left8.9433.862.9414.8434.9995.077
Amygdala Right11.018.4743.5993.6937.7193.229
Hippocampus Left3.2881.6691.5771.8612.2731.467
Hippocampus Right3.9583.643.042.1013.5533.295
Pallidum Left4.9894.6575.5153.0467.8113.57
Pallidum Right2.862.6583.7252.9617.265.007
Thalamus Left0.8221.0851.6941.7262.2171.279
Thalamus Right2.5232.5411.7851.0943.1522.438
SAU-Net: fine-tuned SAU-Net model.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, M.; Kim, J.; EY Kim, R.; Kim, H.G.; Oh, S.W.; Lee, M.K.; Wang, S.-M.; Kim, N.-Y.; Kang, D.W.; Rieu, Z.; et al. Split-Attention U-Net: A Fully Convolutional Network for Robust Multi-Label Segmentation from Brain MRI. Brain Sci. 2020, 10, 974. https://doi.org/10.3390/brainsci10120974

AMA Style

Lee M, Kim J, EY Kim R, Kim HG, Oh SW, Lee MK, Wang S-M, Kim N-Y, Kang DW, Rieu Z, et al. Split-Attention U-Net: A Fully Convolutional Network for Robust Multi-Label Segmentation from Brain MRI. Brain Sciences. 2020; 10(12):974. https://doi.org/10.3390/brainsci10120974

Chicago/Turabian Style

Lee, Minho, JeeYoung Kim, Regina EY Kim, Hyun Gi Kim, Se Won Oh, Min Kyoung Lee, Sheng-Min Wang, Nak-Young Kim, Dong Woo Kang, ZunHyan Rieu, and et al. 2020. "Split-Attention U-Net: A Fully Convolutional Network for Robust Multi-Label Segmentation from Brain MRI" Brain Sciences 10, no. 12: 974. https://doi.org/10.3390/brainsci10120974

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop