Next Article in Journal
Comparison of Ultrasound Image Classifier Deep Learning Algorithms for Shrapnel Detection
Next Article in Special Issue
Frontiers in Retinal Image Processing
Previous Article in Journal
A Generic Framework for Depth Reconstruction Enhancement
Previous Article in Special Issue
Towards a Connected Mobile Cataract Screening System: A Future Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intraretinal Layer Segmentation Using Cascaded Compressed U-Nets

by
Sunil Kumar Yadav
1,2,
Rahele Kafieh
1,
Hanna Gwendolyn Zimmermann
1,
Josef Kauer-Bonin
1,2,
Kouros Nouri-Mahdavi
3,
Vahid Mohammadzadeh
3,
Lynn Shi
3,
Ella Maria Kadas
2,
Friedemann Paul
1,4,
Seyedamirhosein Motamedi
1,† and
Alexander Ulrich Brandt
1,5,*,†
1
Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany
2
Nocturne GmbH, 10119 Berlin, Germany
3
Glaucoma Division, Stein Eye Institute, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA 90095, USA
4
Department of Neurology, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 10098 Berlin, Germany
5
Department of Neurology, University of California Irvine, Irvine, CA 92697, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2022, 8(5), 139; https://doi.org/10.3390/jimaging8050139
Submission received: 31 March 2022 / Revised: 23 April 2022 / Accepted: 3 May 2022 / Published: 17 May 2022
(This article belongs to the Special Issue Frontiers in Retinal Image Processing)

Abstract

:
Reliable biomarkers quantifying neurodegeneration and neuroinflammation in central nervous system disorders such as Multiple Sclerosis, Alzheimer’s dementia or Parkinson’s disease are an unmet clinical need. Intraretinal layer thicknesses on macular optical coherence tomography (OCT) images are promising noninvasive biomarkers querying neuroretinal structures with near cellular resolution. However, changes are typically subtle, while tissue gradients can be weak, making intraretinal segmentation a challenging task. A robust and efficient method that requires no or minimal manual correction is an unmet need to foster reliable and reproducible research as well as clinical application. Here, we propose and validate a cascaded two-stage network for intraretinal layer segmentation, with both networks being compressed versions of U-Net (CCU-INSEG). The first network is responsible for retinal tissue segmentation from OCT B-scans. The second network segments eight intraretinal layers with high fidelity. At the post-processing stage, we introduce Laplacian-based outlier detection with layer surface hole filling by adaptive non-linear interpolation. Additionally, we propose a weighted version of focal loss to minimize the foreground–background pixel imbalance in the training data. We train our method using 17,458 B-scans from patients with autoimmune optic neuropathies, i.e., multiple sclerosis, and healthy controls. Voxel-wise comparison against manual segmentation produces a mean absolute error of 2.3 μm, outperforming current state-of-the-art methods on the same data set. Voxel-wise comparison against external glaucoma data leads to a mean absolute error of 2.6 μm when using the same gold standard segmentation approach, and 3.7 μm mean absolute error in an externally segmented data set. In scans from patients with severe optic atrophy, 3.5% of B-scan segmentation results were rejected by an experienced grader, whereas this was the case in 41.4% of B-scans segmented with a graph-based reference method. The validation results suggest that the proposed method can robustly segment macular scans from eyes with even severe neuroretinal changes.

1. Introduction

Optical coherence tomography (OCT) is a noninvasive in vivo imaging modality, which is able to acquire 3D volume scans of the retina with micrometer resolution [1]. The retina is part of the central nervous system, and OCT-derived intraretinal layer thickness or volume measurements, which are typically measured within the macular region, are promising noninvasive biomarkers querying neuronal structures. For instance, the combined macular ganglion cell and inner plexiform layer (GCIPL) shows thinning in eyes affected by optic neuritis (ON) [2,3], multiple sclerosis (MS) [4], neuromyelitis optica spectrum disorders (NMOSD) [5] and Alzheimer’s disease [6]. GCIPL is a potential biomarker predicting disease activity in patients with early MS [7,8] and of cognitive decline in Parkinson’s disease [9]. The macular inner nuclear layer (INL) may be affected by inflammatory activity in MS and related disorders [10] or Müller cell damage in NMOSD [5,11,12]. In glaucoma, thinning of inner retinal layers such as GCIPL and the retinal nerve fiber layer (RNFL) may appear before visual function loss and is a potential biomarker for early diagnosis [13]. Notably, these applications rely on changes in layer thickness, while the overall configuration of the retina is mostly unaltered, which separates their segmentation needs from disorders with clear microscopic retinal pathologies such as age-related macular degeneration (AMD) or diabetic retinopathy (DR) [14]. Indeed, retinal scans with comorbid eye diseases such as AMD and DR are typically excluded from their use as biomarkers in neuronal diseases such as MS [15].
Intraretinal layer segmentation is not a trivial task as layer changes can be small, while texture gradients can be weak and influenced by noise [16]. While many OCT manufacturers provide device-integrated segmentation of some intraretinal layers, results are often noisy and regularly require manual correction, which in itself is challenging for clinical use and difficult to handle in research settings when multiple graders review macular OCT images [16,17]. A reliable and robust algorithm for intraretinal layer segmentation that requires no or minimal manual correction is an unmet need to foster reliable and reproducible research as well as clinical application [18].
In this paper, we propose a cascade of compressed convolutional neural networks with a U-Net-like architecture for fully automated intraretinal layer segmentation (CCU-INSEG). We compress the U-Net models to reduce required resources without losing a significant amount of accuracy in the segmentation process. The cascaded architecture helps the algorithm to increase its robustness with regard to different types of OCT images without adding too many parameters in the overall network. We validate the method using internal and external data on a voxel, measurement parameter and scan level.

State of the Art

Intraretinal layer segmentation of OCT-derived macular volume scans can be divided into two categories, mathematical modeling-based methods and machine learning (ML)-based methods, with the latter recently extending towards deep learning.
Mathematical modeling-based methods follow a set of rules based on well-defined deterministic mathematical and image analysis concepts to segment intraretinal layer boundaries. These methods have been the predominant approach in the previous decade and are not within the scope of this paper.
Data-driven or ML-based methods learn features from image data sets automatically and have produced reliable and accurate measurements in several retinal imaging applications [14]. In the first paper proposing deep learning-derived methods for intraretinal layer segmentation, Fang et al. [19] used a Cifar-style convolutional neural network (CNN) to create class labels and probability maps, followed by a graph theory and dynamic programming step to derive final layer boundaries in OCT images from patients with age-related macular degeneration (AMD). Shortly thereafter, Roy et al. [20] proposed an end-to-end fully convolutional framework to perform intraretinal segmentation, which was also able to delineate fluid collection in diabetic retinopathy. The authors avoided all deterministic pre- and postprocessing and performed the entire segmentation procedure using a fully convolutional network inspired by DeconvNet [21] and U-Net [22] with an encoder–decoder architecture, skip connections and unpooling layers. Kugelman et al. [23] used recurrent neural networks (RNN) instead of CNNs for 2D image segmentation by formulating image patches as a sequential problem. Unlike CNNs, RNNs use interim output as input in the next iteration of computations, thereby utilizing a memory network in a specific classification task on sequential data [24].
Several recent publications make use of the U-Net design, which was introduced to target problems in medical image analysis [22]. U-Net’s symmetrical architecture consists of a contracting path to capture context and an expanding path that enables precise localization, linked by skip connections. Mishra et al. [25] proposed a method similar to Fang et al. but relying on U-Net for the generation of probability maps and shortest path segmentation for final boundary detection. He et al. [26] proposed an intraretinal segmentation method using two cascading U-Nets, the first one preforming the actual intraretinal segmentation task, and the second one then reconstructing 3D surfaces. Combining segmentation with surface reconstruction using CNN was initially suggested by Shah et al. [27] and has the benefit that incorrect segmentation results in 2D planes are corrected when reconstructing a continuous surface. This approach has been recently combined into a single unified framework by the same group [28]. Pekala et al. [29] proposed a method based on DenseNet, which makes use of more extensive skip connections but is otherwise similar in design to U-Net for boundary pixel segmentation, followed by a Gaussian process regression for surface extraction. Another extension of the U-Net concept for retinal segmentation was proposed by Liu et al. [30], who used a nested U-Net shape for multi-scale input and output; this approach performed favorably based on accuracy in comparison to traditional U-Net architectures. Instead of U-Net, Li et al. [31] proposed a method utilizing a combination of the Xception65 network and an atrous spatial pyramid pooling module for retinal boundary detection. Cascaded network architectures provide better generalization and are therefore suitable for further improvements in robustness and efficiency [32]. In the literature, cascaded networks are successfully applied to solve several problems—for example, time-series prediction [33], human–object interaction modeling [34] and medical image segmentation [35]. In this paper, we have also utilized the cascaded network framework with compressed U-Net architecture.
The segmentation accuracy of these previously proposed deep learning-based methods is favorable, with more recent and complex approaches performing better than early methods utilizing simpler network architectures. Hamwood et al. [36] showed that network design choices such as patch size may have a strong influence over network performance in retinal segmentation tasks. In contrast, in real-world medical applications, the selection of training data, meaningful pre- and postprocessing utilizing domain knowledge and computational costs are highly relevant. The latter favors simpler network architectures, because complex architectures with multiple parameters are prohibitive for environments with limited resources. Methods implementing lean networks and specific postprocessing for segmentation errors, e.g., by reconstructing surfaces, can achieve high segmentation efficiency while being performant [28].
One way in which performance optimization may be achieved is by compressing network architectures—for example, by architecture pruning, weight quantization or knowledge distillation [37]. The latter was recently used by Borkovkina et al. [38] to propose a retinal segmentation method, which utilizes a performance-optimized smaller network that was trained on a complex original segmentation network as teacher network.
We here utilize the compression of a U-Net-style network based on architecture pruning, which leads to substantial performance improvements while not suffering in accuracy. We further utilize appropriate pre- and postprocessing to reduce errors and achieve high segmentation accuracy, by introducing an adaptive hole filling process. Of note, our method was also trained on more diverse data than previous approaches, improving its readiness for immediate application rather than merely being a methodological proposal. Together, these contributions highlight important pathways from research-driven proof-of-concept implementations towards application.

2. Materials and Methods

The proposed method starts with preprocessing an input volume scan to remove undesired components from each B-scan (Figure 1). It then performs intraretinal segmentation using a cascaded network, where each stage is a compressed U-Net-like network. Finally, during postprocessing, we derive continuous surfaces and introduce anatomical constraints. The term cascaded network refers to the combination of two networks in a sequential manner and each network is a compressed U-Net-like architecture, which is a lighter version of U-Net, to minimize the computational resources. In the following, we explain each component of the pipeline in detail.

2.1. Segmentation Method

2.1.1. Preprocessing

Let us consider an input OCT volume scan sampled from the macula consisting of several B-scans. During data acquisition, artefacts are inevitable due to various internal and external factors. These artefacts may include noise, which is responsible for a low signal-to-noise (SNR) ratio, cut scans (upper and lower cuts) and improper focus [15]. Additionally, parts of the optic nerve head (ONH) frequently appear in macular scans, as shown in Figure 1 (the input image). These artefacts are not desirable and affect the performance of segmentation algorithms.
To automatically detect and remove these artefacts, we employ a deep learning-based quality control method described and validated in detail previously [39,40]. Briefly, the automatic quality analysis (AQuA) method combines the detection of center, signal quality and image completeness artifacts. These features are evaluated before segmentation, and only if the input volume is free from cuts, focused around the macular center, and have acceptable signal quality, the segmentation operation will be performed. AQuA further allows some correction of quality issues: the scan center is corrected using parabola fitting in the foveal region. Furthermore, based on the computed foveal center, B-scans are cropped in both A-scan and B-scan directions to cover only a 6x6mm region around the center for both training and prediction [40].

2.1.2. Cascaded and Compressed U-Net Networks

Our segmentation method consists of two cascaded U-Net-like architectures, as shown in Figure 1. The first network, termed RS-Net (retinal segmentation network), is responsible for segmentation of the retina from a full OCT B-scan. In its simplest form, this is the retinal tissue between the inner limiting membrane (ILM) and Bruch’s membrane (BM) as boundaries. The trained network takes a preprocessed OCT B-scan as input and produces a three-class output: purple region (above ILM), green region (retinal region) and yellow region (below BM) (Figure 1). The output of RS-Net and the original input serve as the input for the second network. The second network, the intraretinal layer segmentation network (IS-Net), takes the two-channel input and generates a 9-class output providing boundaries for 8 different intraretinal layers and membranes: ILM—inner limiting membrane, mRNFL—macular retinal nerve fiber layer, GCL—ganglion cell layer, IPL—inner plexiform layer, GCIPL—macular ganglion cell/inner plexiform layer, INL—inner nuclear layer, OPL—outer plexiform layer, ONL—outer nuclear layer, ELM—external limiting membrane, MZ—myoid zone, EZ—ellipsoid zone, OS—outer segment, RPE—retinal pigment epithelium, OL—outer layer and BM—Bruch’s membrane (Figure 2).
The original U-Net is a fully CNN that consists of a contracting path (encoder) and an expansive path (decoder) [22]. Paths are connected at multiple places by skip connections, a key aspect of U-Net that allows the network to restore spatial information that was lost during the pooling operation. In general, U-Net is capable of producing high-fidelity segmentation results and most of the state-of-the-art segmentation methods are based on U-Net architecture and focused on high accuracy (see above). However, the deep architecture of U-Net requires considerable computational resources, which restricts the deployment of the model on systems with limited processing power. To find a desirable balance between processing needs and accuracy, we studied compressed variations of this network with regard to channel depth and filter size.
Figure 3 shows the architecture of the proposed compressed U-Net. Similar to U-Net, the proposed method has contracting and expansive paths. The contracting or encoder part has four stages and, at each consecutive stage, the channel depth is doubled and the input size is halved (via max pooling). Additionally, dropout layers are added after the third and the fourth stages to avoid overfitting. In the decoder part, the inverse of the encoder part occurs, except that it has two additional blocks: merging and output convolution layers with a kernel size of 1, as shown in Figure 3. The terms h and w represent the size of the preprocessed image. The number of the input channels is represented by n and it is clear from Figure 1 that n = 1 for RS-Net and n = 2 for IS-Net. Optimal architecture parameter values (h, w, f, c) are explained in Section 2.3.

2.1.3. Loss Function

For end-to-end training of the proposed network, a suitable loss function will be the key to a high-fidelity segmentation outcome. In the case of OCT layer segmentation, one of the main challenges would be to tackle the background–foreground class imbalance. As shown in Table 1, 7 out of 9 classes are highly imbalanced and having less than 10% of foreground pixels with respect to all pixels in the training data. These imbalance factors should be corrected to avoid false negatives, as shown in Figure 4b (red rectangular region). Figure 4b shows the outcome from the combination of Tversky [41] and CCE (categorical cross-entropy) loss functions.
Focal loss is a robust loss function to provide a better segmentation outcome with extremely foreground–background imbalanced training data. It is a dynamically scaled version of cross-entropy to down-weight the background pixels and focus the model on foreground pixels (layers). Focal loss is defined as [42]:
E f l = α t ( 1 p t ( x ) ) γ l o g ( p t ( x ) ) ,
where p t is defined as:
p t ( x ) = p c ( x ) , if g c ( x ) = 1 1 p c ( x ) , otherwise ,
where x is the pixel value from the retinal region Ω . The terms g c ( x ) and p c ( x ) show the ground truth and the predicted probability at x from the class c, respectively. Furthermore, α t and γ are modulating factors and, in our experiment, we used default values of α = 0.25 and γ = 2 . As can be seen from Figure 4c, the outcome improves compared to Tversky loss [41]. However, we can see a significant amount of false negatives. To improve the focal loss outcome, we replace α t with w t . The weighting fact w t and modified focal loss are defined as:
w t = 1 n f n t , E w f l = w t ( 1 p t ( x ) ) γ l o g ( p t ( x ) ) ,
where n f and n t represent the number of pixels in the foreground and in the image, respectively. For our training, we have used weighted focal loss E w f l , where each class is weighted based on the corresponding foreground–background pixel ratio. The improvement in the outcome is visualized in Figure 4.

2.1.4. Postprocessing

At the last stage of the pipeline, we apply postprocessing on the segmentation outcome to remove artefacts. First, adaptive hole filling removes small holes from the background and the foreground of the segmentation outcome as shown in Figure 5. These holes occasionally appear because of the misclassification of individual or a few pixels during the segmentation process. To remove small holes, we use morphological operators with an adaptive threshold. The value of the threshold is computed differently for each class of the network’s outcome. RS-Net generates a three-class output, and for upper and lower classes (above ILM and below BM), the threshold value is n n / f , where n n is the number of non-zero pixels in this class image and f is a constant factor with a default value of 2. Additionally, the value of f will keep increasing by 1 until n n / f < n z , where n z is the number of zero pixels. The same process is repeated for the upper and lower class of the IS-Net outcome. For the internal layer classes of the networks, the default value of f is 10 and it is increased by 1 until n n / f < n z .
After hole filling, different layers from the multi-channel outcome of IS-Net are extracted. To detect the boundary in each class image, we search the first non-zero entry in each column. After layer extraction, we remove outliers from the layer surfaces based on the Laplacian of the thickness values. Furthermore, we recompute missing values using Piecewise Cubic Hermite Interpolating Polynomial (PCHIP) interpolation. Lastly, an isotropic smoothing with 3 × 3 kernel size is applied to each layer surface to remove small ripples and to generate smoother boundaries.

2.1.5. Implementation

The method was fully implemented in Python. Network training was performed with Python (v3.6) using Keras (v2.3) and TensorFlow (v2.2) libraries on a Linux server with two Intel Xeon Gold 6144 CPUs (Intel Corporation, Santa Clara, CA, USA) and two NVIDIA GeForce GTX1080 Ti GPUs (NVIDIA Corporation, Santa Clara, CA, USA).

2.2. Data

To train the proposed method, we selected macular OCT volume scans that were obtained with Spectralis spectral-domain OCT (Heidelberg Engineering, Heidelberg, Germany). All data were selected from the institute’s OCT image database at Charité—Universitätsmedizin Berlin (CUB).
In our previous experience with intraretinal layer segmentation, we noticed that macular OCT volume scans with low, yet acceptable, signal to noise ratios, as well as scans from patients with very thin RNFL, needed extensive manual segmentation correction. Hence, from a pool of macular scans from healthy controls and patients with clinically isolated syndrome (CIS), MS and NMOSD, we selected a total of 445 OCT volume scans of 254 eyes, the majority of which had a low to average signal to noise ratio, thin RNFL and microcystic macular pathology, an occasional OCT finding in eyes from patients with various forms of optic neuropathy such as autoimmune optic neuropathies [43]. All selected volumes were centered on the fovea, covering 25 × 30 with different resolutions (61/25 B-scans and 768/1024/1536 A-scans per B-scan).
All scans underwent automatic quality control to avoid data with insufficient quality for training and validation purposes [39,40]. In the end, we selected 17,458 B-scans for training, 3081 for validation and 208 for testing.

2.2.1. Manual Correction

For the manual correction of segmented boundaries, we used standardized procedures developed during an international study on intraretinal segmentation with several reference centers [17]. Data were manually segmented by multiple graders, who were all trained and experienced in applying these procedures. All grading was supervised by experienced graders (HGZ, EMK). Intraretinal segmentation was performed using the SAMIRIX toolbox [44], which uses OCTLayerSegmentation from AUtomated Retinal Analysis Tools (AURA) developed by [45], and then manually corrected by experts using the same toolbox. A detailed description of this process, including validation, is outlined in [44].

2.2.2. Validation Data

To validate the method in a multicenter format, additional unseen macular volume scans from our center (CUB), as well as two other centers—the University of California, Los Angeles (UCLA) and Johns Hopkins University (JHU)—were added.
CUB: Twenty-five additional unseen macular 3D OCT scans of 24 eyes were included in this study from our center (CUB). The volumes were from HCs and patients with MS and NMOSD, covering 25 × 30 macular area using 61 B-scans and 768 A-scans per B-scan. All volumes underwent segmentation and manual correction using the SAMIRIX toolbox [44]. Two scans were rejected by our quality check because of the low signal to noise ratio.
UCLA: In order to ensure the quality of the developed method on OCT images from patients with other optic neuropathies, which were not part of the training data, we included a second data set for testing from the University of California, Los Angeles (UCLA). The data set consists of 12 OCT volume scans of 12 eyes from glaucoma patients or healthy subjects, each with 61 horizontal B-scans and 768 A-scans per B-scan, covering 25 × 30 centered on the fovea. The OCT volumes in this data set were segmented and then manually corrected using the SAMIRIX toolbox for comparison against the developed method. The volume scans in this data set were cropped to 5.5 × 5.5 mm square in the preprocessing step to ensure complete exclusion of ONH from macular scans.
JHU: To validate our method further, we applied the developed segmentation method on a publicly available OCT data set from Johns Hopkins University (JHU) [26]. Fifteen volume scans of 15 eyes were selected from this data set, which are a mix of HC and MS data. These scans were segmented and manually corrected by its publisher, so we compared our segmentation results against their manual delineation. The volume scans in the JHU data set cover 20 × 20 of the macula using 49 B-scans and 1024 A-scans per B-scan.

2.3. Optimization

To obtain optimal values of filter kernel size f and channel depth c, we performed compression with different values of f and c, as shown in Table 2. The network’s channel depth should be proportional to the complexity of task. From Figure 1, it can be seen that RS-Net produces a three-channel output ( o = 3 ), which is a less complicated task compared to generating a 9-class output (IS-net).
For RS-Net, we start with two different values of the depth channel ( c = { 4 , 8 } ), which increases up to four times the input size, as shown in Figure 3. RS-Net is responsible for retinal segmentation, and it should not focus on minor details. Therefore, we set the kernel size to 5 × 5 ( f = 5 ), which helps the network to focus on significant features. To check the performance of the RS-Net with different values of c, we use a smaller data set (1000 B-scans) for training, validation and testing along with the loss function, which is explained in Section 2.1.3. As can be seen from the Table 2, the performance of RS-Net is similar in terms of the dice similarity coefficient (DSC) for both values of c. However, the number of model parameters is almost four times less with c = 4 in comparison to c = 8 . Therefore, f = 5 and c = 4 are the optimal configuration for RS-Net with respect to model size and performance metrics.
In contrast to RS-Net, IS-Net is responsible for separating 8 layers (9 classes) with the need to detect small details. Different configurations of IS-Net along with the performance metrics are shown in Table 2. The two combinations of channel depth and kernel size ( f , c = { 5 , 16 } & f , c = { 3 , 32 } ) produce the best performance metrics. However, the first configuration ( f = 5 , c = 16 ) has a smaller number of model parameters and, therefore, is the preferable configuration of IS-Net for our purpose.
In contrast to U-Net, the proposed architecture has fewer stages. U-Net consists of 5 stages ranging from 64 to 1024 channel depths, and our networks consist of only 4 stages with variable channel depths. Table 3 shows the direct comparison with U-Net and deep residual U-Net ([46]) in terms of the number of model parameters, DSC and the prediction time. ResU-Net performs slightly better compared to the original U-Net and our proposed method. It is worth mentioning that the above comparison is performed at the architectural level. If U-Net is used in the cascaded setting as in [47], then the computational complexity will be high. Although ResU-Net has not been used in the cascaded setting so far in the state-of-the-art methods, it has better potential because of the low number of parameters and slightly better performance compared to the original U-Net. However, it still has more parameters than the proposed architecture.
As a result, RS-Net is almost 392-times smaller compared to the original U-Net in terms of parameters and model size. Similarly, IS-Net is almost 26-times smaller compared to U-Net. The comparison of prediction time could be challenging because it depends on several factors, including hardware, programming style, libraries, etc. However, it is clear that the smaller network will need smaller computational resources and will be faster compared to U-Net from the running time perspective. Table 3 shows the training time for 1 epoch and the mentioned time is taken on a Linux server with two Intel Xeon Gold 6144 CPUs and two NVIDIA GeForce GTX1080 TI graphical processing units (GPUs). Furthermore, the prediction time is recorded only on a Core-i7 CPU. From Table 3, it is clear that the proposed networks maintain the balance between the need for computational resources (lower running time or both training and prediction) and accuracy. Our method produces a quite similar DSC score to ResU-Net and the original U-Net. Furthermore, the number of model parameters indicates the memory requirement during the training and inference.
Compared to two-stage cascaded state-of-the-art methods, our CCU network is much smaller regarding parameters and model size, with almost similar or better accuracy. For example, Ref. [35] proposed a cascade of two U-Net-like architectures, where each network is a combination of the original U-Net with additional layers. Therefore, the number of parameters in each network is more than 31M (for the input image size of 512 × 512 ), which is much larger compared to our CCU network. Similarly, [26] used a modified U-Net for the first network with a double channel depth compared to our CCU network.

2.4. Statistical Analysis

Intraclass correlation coefficient (ICC) and lower and upper confidence intervals were calculated based on the variance components of a one-way ANOVA, using the ICC package of R. Mean absolute error (MAE), standard deviation and DSC are computed using Python.

3. Results

3.1. Training Analysis

End-to-end training was performed using 17,458, 3081 and 208 B-scans for training, validation and testing, respectively. To minimize the loss function (Equation (1)), the Adam optimizer was used with a fixed learning rate of 0.001. We set the number of iterations to 80 for both networks with early stopping criteria. For early stopping, we monitored the validation accuracy.
As can be seen from Table 4, the optimal and most stable accuracies are reached in 23 and 11 epochs for RS-Net and IS-Net, respectively. Additionally, Table 4 consists of different training- and testing-related metrics of the proposed CCU network. Furthermore, Figure 6 represents a visual outcome with the data from three different centres.

3.2. Comparison against Device-Implemented Segmentation

We then compared the proposed method and standard device segmentation without further correction. Table 5 shows MAE (SD) between manually corrected gold standard segmentation (15 volume scans of 15 eyes) and the device segmentation (first row). Similarly, the second row shows MAEs (SD) between the proposed method’s outcome and gold standard segmentation. Our method has almost 2.5 times lower MAEs and SDs compared to the device segmentation.

3.3. Voxel-Wise Comparison against Multicenter Data

We also validated our method with multi-center data, which included data from Charité—Universitätsmedizin Berlin (CUB), the University of California, Los Angeles (UCLA) and publicly available data from Johns Hopkins University (JHU).
DSC values were computed between manually corrected segmentation using the SAMIRIX toolbox and the proposed method’s outcome for nine classes for CUB and UCLA. The corresponding values are reported in Table 6, where vitreous is the region above ILM and b-BM is the region below BM, as shown in Figure 7. As can be seen from Table 6, the first and the last classes (vitreous, b-BM) yield the highest DSC values (close to 1) as the differences between the numbers of background and foreground pixels are small. Moreover, OPL has the lowest value of DSC since this is one of the thinnest layers and the difference between the numbers of background and foreground pixels can be large. Compared to the state-of-the-art methods, our method produces better overall DSC scores (CUB-0.95, UCLA-0.94 and JHU-0.92), except [38,48], where validations are performed using HC data only.
Table 7 shows layer-based MAE computed between CCU-INSEG’s outcome and the manually corrected gold standard segmentation for the CUB and UCLA data sets and the manually corrected segmentation by the publishers for the JHU data set. As can be seen, the proposed method produces a high-fidelity segmentation with regard to the gold standard segmentation. Errors for different layers are in sub-pixels, except for the INL-OPL boundary in the JHU data.
We would like to emphasize an important point here that our method is trained only on data from one center (CUB). The CCU network was not trained on the data either from UCLA or from JHU. However, the performance of our method is quite similar for all these data in terms of MAE (sub-pixel level). Figure 6 shows the effectiveness of the proposed method against different challenges from these three different data sets. Figure 6 (top) is a noisy scan (CUB data), Figure 6 (middle) is a scan with microcysts from JHU data, and UCLA data have a quite thin RNFL on one side of the macula. As can be seen from Table 7, MAE values for BM and INL-OPL are on the higher side (close to 1 pixel) in the JHU data, where manual segmentation was not performed by the same expert as in CUB and UCLA, and it is quite possible that the segmentations from different experts are not exactly the same [17].

3.4. Reliability and Fidelity of Derived Thickness and Volume Parameters

In this section, we compute the standard OCT parameters based on the segmented layers using the proposed method. Here, we are focusing on five standard thickness parameters within a 5 mm circular region from the foveal center. The parameters are described in Figure 2, and additionally we provide data for the ganglion cell complex (GCC, tissue between ILM and IPL-INL boundary) and total macular thickness (TM, between ILM and BM).
The test–retest reliability of the proposed method is evaluated on a data set that has three repeated measurements of 30 healthy eyes. The intra-class correlation coefficient (ICC) for the standard parameters varied from 0.93 (mRNFL) to 0.99 (GCIPL, GCC and total macula) for both thicknesses and volumes, as shown in Table 8. As shown in Table 8, mRNFL has the lowest ICC because mRNFL is one of the thinnest layers in the retina. Therefore, the measurement of this layer can be easily contaminated by noise components, which leads to the lower value of ICC.
To check the fidelity of our method, we use the same multi-center data sets with the gold standard segmentation. Furthermore, the standard parameters are computed on both segmentations (the gold standard and the outcome of the proposed method). Table 8 (the last three rows) shows MAEs computed between parameter values based on the gold standard segmentation and our outcome. For CUB and UCLA data, the errors are limited up to 2 μ m for thickness parameters. For JHU data, INL-related parameters are not quite similar. It is possible that the higher error is because of the different manual corrections at different centers.

3.5. Comparison against State-of-the-Art Methods

For comparison against state-of-the-art methods, we randomly selected 20 macula volume scans of 20 eyes from patients with NMOSD from a recent publication [44]. Eyes from NMOSD patients regularly show severe neuroaxonal retinal damage, resulting in problematic image quality and/or larger neuroaxonal changes. While this does not reflect performance in a random data set, it does test the segmentation accuracy in difficult data. For comparison, we chose OCTLayerSegmentation version 2.11 from AURA (random forest-based boundary classification, [45]), OCTExplorer version 3.8.0 (OCTExp) from The Iowa Reference Algorithms (Retinal Image Analysis Lab, Iowa Institute for Biomedical Imaging, Iowa City, IA, USA) (graph-cut-based segmentation, [49,50,51]) and the PyTorch implementation of RelayNet (deep learning-based segmentation, [20]), three of the most highly cited segmentation methods, which are also made publicly available by their authors.
Table 9 shows that our method outperforms these methods in this problematic data set. AURA performs weakly for mRNFL and GCIPL, which has been reported before [26]. OCTExp shows weak agreement with the reference segmentation in all investigated layers, likely caused by over-smoothing and disagreement in several boundary positions. RelayNet produced several false positive and false negative pixels in the prediction, leading to overall low DSC values. In Figure 8, we show a sample image, exemplifying the segmentation performance on an exceptionally noisy B-scan. To explore segmentation performance on scans with retinal pathologies, we tested all methods on sample scans with geographic atrophy, Susac syndrome and Drusen (Supplementary Data). All methods failed to produce acceptable results in these scans.

3.6. Manual Correction Timing Comparison

Finally, we compare the correction time required to correct the segmentation obtained from the proposed method and the state-of-the-art methods. An experienced grader quality-controlled the segmentation outcome of 15 volume scans from four different methods: HeyEx (device segmentation), OCTExp [49], AURA [45] and CCU (the proposed method). The same grader performed segmentation correction if needed and recorded the corresponding time needed for manual correction (Table 10). Additionally, MAE (mean absolute error) is computed between the corrected and uncorrected segmentation. As can be seen from Table 10, the proposed method requires the minimum average correction time and MAE is also much smaller compared to the state-of-the-art methods. Moreover, only 2 volumes out of 15 required any correction from our method. On the other hand, the other methods required correction for at least 40% of data.

4. Discussion

In this paper, we propose a method for the automatic intraretinal layer segmentation of macular OCT images (CCU-INSEG). The focus of this work was to develop an efficient and high-fidelity segmentation algorithm, which requires little to no manual correction and can be easily deployed on devices with limited memory and computational power. The proposed method is a two-stage segmentation algorithm, where the retinal region is separated in the first stage and intraretinal layer segmentation is performed in the second stage. We further purposefully optimized the method to detect changes in OCT from neurological disorders, which typically show layer thickness differences, whereas scans with macroscopic retinal findings from primary eye disorders are of little relevance in this application.
In a direct comparison, our method outperforms several state-of-the-art methods in low-quality scans, substantially reducing the need for manual correction of segmentation results, which was the main goal of our study. This is likely a result of the optimized training set rather than the network architecture. Our study further demonstrates that good segmentation accuracy can be achieved using compressed networks, but meaningful domain-specific processing and training data selection. We selected a heterogeneous training data set, which included healthy controls and different autoimmune inflammatory and neurodegenerative diseases of the central nervous system. Thorough quality control was performed on each image, and we selected images representing a range of real-world quality, including low-quality images. As such, our method is specialized towards retinas with thickness changes from neuroinflammation and neurodegeneration, but without structural retinal changes from primary eye diseases.
We performed voxel-wise evaluation of our proposed method on multi-center data, which included data from three different centers (CUB, UCLA and JHU). Here, our method produces almost equally good segmentation (sub-pixel level MAE) as the reference segmented data from these centers. Importantly, one data set was from patients with glaucoma, and this disease was not part of the training data set. Nonetheless, performance was comparable, confirming that our method can extend beyond its training domain. On the other hand, while generally good, noise was higher when validating against JHU data, mainly for the INL-OPL boundary (above a pixel) and BM (close to a pixel). We hypothesize that this was caused by different graders manually correcting each data set. Previously, we have shown that manual grader segmentation can systematically differ within and between centers, which is one of the core motivations for a fully automatic, reproducible segmentation method [17].
While the outcomes of our proposed method are encouraging for applications in neurology and neuro-ophthalmology, ours and other proposed methods are still limited in their generalizability, making each method only robust in a select scenario intended by its investigators. Here, we aimed for excellent performance in scans from patients with autoimmune optic neuropathies such as MS, NMOSD and MOGAD, whereas eyes with primary retinopathies and eye disorders are typically excluded in this scenario [15]. However, generalization towards primary eye disorders presumably constitutes a problem of training data and less the underlying architecture. Consequently, improving the representation of primary eye diseases within the training data, including class expansions, is a promising route to further generalizability and clinical application.

Author Contributions

Conceptualization: A.U.B.; Methodology: S.K.Y., R.K., J.K.-B. and S.M.; Software: S.K.Y. and J.K.-B.; Validation: H.G.Z., S.M., K.N.-M., V.M. and L.S.; Formal analysis: S.K.Y.; Data curation: S.M., H.G.Z., J.K.-B. and E.M.K.; Writing—original draft preparation: S.K.Y. and A.U.B.; Writing—review and editing: all authors; Project administration: F.P., S.M. and A.U.B.; Funding acquisition: K.N.-M., F.P. and A.U.B. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded in part by technology transfer grants from the German Federal Ministry for Economic Affairs and Engery (BMBF Exist I 03EFEBE079 to Charité—Universitätsmedizin Berlin and BMBF Exist II BD_03EUEBE079 to Nocturne GmbH) and by the NeuroCure Clinical Research Center (NCRC), funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy–EXC-2049–390688087, and by research grants from the German Federal Ministry for Education and Research (BMBF 13N15432) to Charité—Universitätsmedizin Berlin and Nocturne GmbH, and the Central Innovation Programme for SMEs (ZIM) of the German Federal Ministry for Economic Affairs and Climate Action (ZF4741901CR9) to Charité—Universitätsmedizin Berlin. The UCLA validation cohort was supported by an NIH R01 grant (EY027929, K.N.-M), an unrestricted Departmental Grant to UCLA Department of Ophthalmology from Research to Prevent Blindness (K.N.-M), an unrestricted grant from Heidelberg Engineering (K.N.-M) and a departmental grant from Research to Prevent Blindness (K.N.-M). We acknowledge financial support from the Open Access Publication Fund of Charité—Universitätsmedizin Berlin and the German Research Foundation (DFG).

Institutional Review Board Statement

The study was approved by the local ethics committee at Charité—Universitätsmedizin Berlin (EA1/041/14 from 4 March 2014, EA1/163/12 from 13 June 2013, EA1/182/10 from 4 October 2010) and was conducted in accordance with the Declaration of Helsinki in its current applicable form and the applicable European and German laws. Data from UCLA were provided under UCLA IRB #11-003602 from 9 February 2012. Data from JHU are freely available and were downloaded on 24 February 2021 from http://iacl.jhu.edu/index.php?title=Resources [52].

Informed Consent Statement

All participants gave written informed consent.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to limitations from the European General Data Protection Regulations and German laws.

Conflicts of Interest

S.K.Y. and E.M.K. are founders and hold shares of Nocturne GmbH, a company interested in commercializing parts of the method described in this study. J.K.-B. is an employee of Nocturne GmbH. A.U.B. and F.P. hold shares in Nocturne. K.N.-M. has received an unrestricted grant from Heidelberg Engineering. All other authors report no potential conflicts of interest.

References

  1. Huang, D.; Swanson, E.; Lin, C.; Schuman, J.; Stinson, W.; Chang, W.; Hee, M.; Flotte, T.; Gregory, K.; Puliafito, C.; et al. Optical coherence tomography. Science 1991, 254, 1178–1181. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Andorra, M.; Alba-Arbalat, S.; Camos-Carreras, A.; Gabilondo, I.; Fraga-Pumar, E.; Torres-Torres, R.; Pulido-Valdeolivas, I.; Tercero-Uribe, A.I.; Guerrero-Zamora, A.M.; Ortiz-Perez, S.; et al. Using Acute Optic Neuritis Trials to Assess Neuroprotective and Remyelinating Therapies in Multiple Sclerosis. JAMA Neurol. 2019, 77, 234–244. [Google Scholar] [CrossRef] [PubMed]
  3. Chen, J.J.; Sotirchos, E.S.; Henderson, A.D.; Vasileiou, E.S.; Flanagan, E.P.; Bhatti, M.T.; Jamali, S.; Eggenberger, E.R.; Dinome, M.; Frohman, L.P.; et al. OCT retinal nerve fiber layer thickness differentiates acute optic neuritis from MOG antibody-associated disease and Multiple Sclerosis: RNFL thickening in acute optic neuritis from MOGAD vs MS. Mult. Scler. Relat. Disord. 2022, 58, 103525. [Google Scholar] [CrossRef] [PubMed]
  4. Petzold, A.; Balcer, L.J.; Calabresi, P.A.; Costello, F.; Frohman, T.C.; Frohman, E.M.; Martinez-Lapiscina, E.H.; Green, A.J.; Kardon, R.; Outteryck, O.; et al. Retinal layer segmentation in multiple sclerosis: A systematic review and meta-analysis. Lancet Neurol. 2017, 16, 797–812. [Google Scholar] [CrossRef] [Green Version]
  5. Oertel, F.C.; Specovius, S.; Zimmermann, H.G.; Chien, C.; Motamedi, S.; Bereuter, C.; Cook, L.; Lana Peixoto, M.A.; Fontanelle, M.A.; Kim, H.J.; et al. Retinal Optical Coherence Tomography in Neuromyelitis Optica. Neurol. Neuroimmunol. Neuroinflamm. 2021, 8, e1068. [Google Scholar] [CrossRef]
  6. Cabrera, D.; Wysocka, M.; Grzybowski, A.; Kanclerz, P. Identification of Retinal Biomarkers in Alzheimer’s Disease Using Optical Coherence Tomography: Recent Insights, Challenges, and Opportunities. J. Clin. Med. 2019, 8, 996. [Google Scholar] [CrossRef] [Green Version]
  7. Zimmermann, H.G.; Knier, B.; Oberwahrenbrock, T.; Behrens, J.; Pfuhl, C.; Aly, L.; Kaminski, M.; Hoshi, M.M.; Specovius, S.; Giess, R.M.; et al. Association of Retinal Ganglion Cell Layer Thickness with Future Disease Activity in Patients with Clinically Isolated Syndrome. JAMA Neurol. 2018, 75, 1071. [Google Scholar] [CrossRef]
  8. Lin, T.Y.; Vitkova, V.; Asseyer, S.; Martorell Serra, I.; Motamedi, S.; Chien, C.; Ditzhaus, M.; Papadopoulou, A.; Benkert, P.; Kuhle, J.; et al. Increased Serum Neurofilament Light and Thin Ganglion Cell-Inner Plexiform Layer Are Additive Risk Factors for Disease Activity in Early Multiple Sclerosis. Neurol. Neuroimmunol. Neuroinflamm. 2021, 8, e1051. [Google Scholar] [CrossRef]
  9. Murueta-Goyena, A.; Del Pino, R.; Galdós, M.; Arana, B.; Acera, M.; Carmona-Abellán, M.; Fernández-Valle, T.; Tijero, B.; Lucas-Jiménez, O.; Ojeda, N.; et al. Retinal Thickness Predicts the Risk of Cognitive Decline in Parkinson Disease. Ann. Neurol. 2021, 89, 165–176. [Google Scholar] [CrossRef]
  10. Balk, L.J.; Coric, D.; Knier, B.; Zimmermann, H.G.; Behbehani, R.; Alroughani, R.; Martinez-Lapiscina, E.H.; Brandt, A.U.; Sánchez-Dalmau, B.; Vidal-Jordana, A.; et al. Retinal inner nuclear layer volume reflects inflammatory disease activity in multiple sclerosis: A longitudinal OCT study. Mult. Scler. J. Exp. Transl. Clin. 2019, 5, 205521731987158. [Google Scholar] [CrossRef] [Green Version]
  11. Oertel, F.C.; Kuchling, J.; Zimmermann, H.; Chien, C.; Schmidt, F.; Knier, B.; Bellmann-Strobl, J.; Korn, T.; Scheel, M.; Klistorner, A.; et al. Microstructural visual system changes in AQP4-antibody–seropositive NMOSD. Neurol. Neuroimmunol. Neuroinflamm. 2017, 4, e334. [Google Scholar] [CrossRef] [Green Version]
  12. Motamedi, S.; Oertel, F.C.; Yadav, S.; Kadas, E.; Weise, M.; Havla, J.; Ringelstein, M.; Aktas, O.; Albrecht, P.; Ruprecht, K.; et al. Altered fovea in AQP4-IgG–seropositive neuromyelitis optica spectrum disorders. Neurol. Neuroimmunol. Neuroinflamm. 2020, 7, e805. [Google Scholar] [CrossRef] [PubMed]
  13. Mohammadzadeh, V.; Fatehi, N.; Yarmohammadi, A.; Lee, J.; Sharifipour, F.; Daneshvar, R.; Caprioli, J.; Nouri-Mahdavi, K. Macular imaging with optical coherence tomography in glaucoma. Surv. Ophthamol. 2020, 65, 597–638. [Google Scholar] [CrossRef] [PubMed]
  14. Ting, D.S.W.; Pasquale, L.R.; Peng, L.; Campbell, J.P.; Lee, A.Y.; Raman, R.; Tan, G.S.W.; Schmetterer, L.; Keane, P.A.; Wong, T.Y. Artificial intelligence and deep learning in ophthalmology. Br. J. Ophthalmol. 2019, 103, 167–175. [Google Scholar] [CrossRef] [Green Version]
  15. Tewarie, P.; Balk, L.; Costello, F.; Green, A.; Martin, R.; Schippling, S.; Petzold, A. The OSCAR-IB consensus criteria for retinal OCT quality assessment. PLoS ONE 2012, 7, e34823. [Google Scholar] [CrossRef] [PubMed]
  16. Oberwahrenbrock, T.; Jost, R.; Zimmermann, H.; Beckers, I.; Paul, F.; Brandt, A.U. Signal Quality Dependency of Intra-Retinal Segmentation Algorithms; ECTRIMS Online Library: Basel, Switzerland, 2016; p. 559. [Google Scholar]
  17. Oberwahrenbrock, T.; Traber, G.L.; Lukas, S.; Gabilondo, I.; Nolan, R.; Songster, C.; Balk, L.; Petzold, A.; Paul, F.; Villoslada, P.; et al. Multicenter reliability of semiautomatic retinal layer segmentation using OCT. Neurol. Neuroimmunol. Neuroinflamm. 2018, 5, e449. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Aytulun, A.; Cruz-Herranz, A.; Aktas, O.; Balcer, L.J.; Balk, L.; Barboni, P.; Blanco, A.A.; Calabresi, P.A.; Costello, F.; Sanchez-Dalmau, B.; et al. The APOSTEL 2.0 Recommendations for Reporting Quantitative Optical Coherence Tomography Studies. Neurology 2021, 97, 68–79. [Google Scholar] [CrossRef]
  19. Fang, L.; Cunefare, D.; Wang, C.; Guymer, H.; Li, S.; Farsiu, S. Automatic segmentation of nine retinal layer boundaries in OCT images of non-exudative AMD patients using deep learning and graph search. Biomed. Opt. Express 2017, 8, 2732–2744. [Google Scholar] [CrossRef] [Green Version]
  20. Roy, A.G.; Conjeti, S.; Karri, S.P.K.; Sheet, D.; Katouzian, A.; Wachinger, C.; Navab, N. ReLayNet: Retinal layer and fluid segmentation of macular optical coherence tomography using fully convolutional networks. Biomed. Opt. Express 2017, 8, 3627–3642. [Google Scholar] [CrossRef]
  21. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  23. Kugelman, J.; Alonso-Caneiro, D.; Read, S.A.; Vincent, S.J.; Collins, M.J. Automatic segmentation of OCT retinal boundaries using recurrent neural networks and graph search. Biomed. Opt. Express 2018, 9, 5759–5777. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Chen, H.; Dou, Q.; Ni, D.; Cheng, J.Z.; Qin, J.; Li, S.; Heng, P.A. Automatic fetal ultrasound standard plane detection using knowledge transferred recurrent neural networks. In Medical Image Computing and Computer-Assisted Intervention—MICCAI, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Springer International Publishing: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  25. Mishra, Z.; Ganegoda, A.; Selicha, J.; Wang, Z.; Sadda, S.R.; Hu, Z. Automated Retinal Layer Segmentation Using Graph-based Algorithm Incorporating Deep-learning-derived Information. Sci. Rep. 2020, 10, 9541. [Google Scholar] [CrossRef] [PubMed]
  26. He, Y.; Carass, A.; Liu, Y.; Jedynak, B.M.; Solomon, S.D.; Saidha, S.; Calabresi, P.A.; Prince, J.L. Deep learning based topology guaranteed surface and MME segmentation of multiple sclerosis subjects from retinal OCT. Biomed. Opt. Express 2019, 10, 5042–5058. [Google Scholar] [CrossRef] [Green Version]
  27. Shah, A.; Zhou, L.; Abrámoff, M.D.; Wu, X. Multiple surface segmentation using convolution neural nets: Application to retinal layer segmentation in OCT images. Biomed. Opt. Express 2018, 9, 4509. [Google Scholar] [CrossRef]
  28. He, Y.; Carass, A.; Liu, Y.; Jedynak, B.; Solomon, S.; Saidha, S.; Calabresi, P.; Prince, J.L. Structured layer surface segmentation for retina OCT using fully convolutional regression networks. Med. Image Anal. 2021, 68, 101856. [Google Scholar] [CrossRef]
  29. Pekala, M.; Joshi, N.; Liu, T.A.; Bressler, N.; DeBuc, D.C.; Burlina, P. OCT Segmentation via Deep Learning: A Review of Recent Work. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 316–322. [Google Scholar]
  30. Liu, W.; Sun, Y.; Ji, Q. MDAN-UNet: Multi-Scale and Dual Attention Enhanced Nested U-Net Architecture for Segmentation of Optical Coherence Tomography Images. Algorithms 2020, 13, 60. [Google Scholar] [CrossRef] [Green Version]
  31. Li, Q.; Li, S.; He, Z.; Guan, H.; Chen, R.; Xu, Y.; Wang, T.; Qi, S.; Mei, J.; Wang, W. DeepRetina: Layer Segmentation of Retina in OCT Images Using Deep Learning. Transl. Vis. Sci. Technol. 2020, 9, 61. [Google Scholar] [CrossRef]
  32. Littmann, E.; Ritter, H. Generalization Abilities of Cascade Network Architecture. In Proceedings of the 5th International Conference on Neural Information Processing Systems (NIPS’92), Denver, CO, USA, 30 November–3 December 1992; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1992; pp. 188–195. [Google Scholar]
  33. Warsito, B.; Santoso, R.; Suparti; Yasin, H. Cascade Forward Neural Network for Time Series Prediction. J. Phys. Conf. Ser. 2018, 1025, 012097. [Google Scholar] [CrossRef]
  34. Zhou, T.; Qi, S.; Wang, W.; Shen, J.; Zhu, S.C. Cascaded Parsing of Human-Object Interaction Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2827–2840. [Google Scholar] [CrossRef]
  35. Ma, D.; Lu, D.; Heisler, M.; Dabiri, S.; Lee, S.; Ding, G.W.; Sarunic, M.V.; Beg, M.F. Cascade Dual-branch Deep Neural Networks for Retinal Layer and fluid Segmentation of Optical Coherence Tomography Incorporating Relative Positional Map. In Proceedings of the 3rd Conference on Medical Imaging with Deep Learning, Montreal, QC, Canada, 6–9 July 2020; Volume 121, pp. 493–502. [Google Scholar]
  36. Hamwood, J.; Alonso-Caneiro, D.; Read, S.A.; Vincent, S.J.; Collins, M.J. Effect of patch size and network architecture on a convolutional neural network approach for automatic segmentation of OCT retinal layers. Biomed. Opt. Express 2018, 9, 3049–3066. [Google Scholar] [CrossRef] [Green Version]
  37. Mangalam, K.; Salzamann, M. On Compressing U-net Using Knowledge Distillation. arXiv 2018, arXiv:1812.00249. [Google Scholar]
  38. Borkovkina, S.; Camino, A.; Janpongsri, W.; Sarunic, M.; Jian, Y. Real-time retinal layer segmentation of OCT volumes with GPU accelerated inferencing using a compressed, low-latency neural network. Biomed. Opt. Express 2020, 11, 3968–3984. [Google Scholar] [CrossRef]
  39. Kauer, J.; Gawlik, K.; Zimmermann, H.G.; Kadas, E.M.; Bereuter, C.; Paul, F.; Brandt, A.U.; Haußer, F.; Beckers, I.E. Automatic quality evaluation as assessment standard for optical coherence tomography. In Proceedings of the Advanced Biomedical and Clinical Diagnostic and Surgical Guidance Systems XVII, San Francisco, CA, USA, 3–5 February 2019; Volume 10868, pp. 76–86. [Google Scholar]
  40. Kauer-Bonin, J.; Yadav, S.K.; Beckers, I.; Gawlik, K.; Motamedi, S.; Zimmermann, H.G.; Kadas, E.M.; Haußer, F.; Paul, F.; Brandt, A.U. Modular deep neural networks for automatic quality control of retinal optical coherence tomography scans. Comput. Biol. Med. 2021, 141, 104822. [Google Scholar] [CrossRef] [PubMed]
  41. Salehi, S.S.M.; Erdogmus, D.; Gholipour, A. Tversky Loss Function for Image Segmentation Using 3D Fully Convolutional Deep Networks. In Machine Learning in Medical Imaging; Wang, Q., Shi, Y., Suk, H.I., Suzuki, K., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 379–387. [Google Scholar]
  42. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Kaufhold, F.; Zimmermann, H.; Schneider, E.; Ruprecht, K.; Paul, F.; Oberwahrenbrock, T.; Brandt, A.U. Optic Neuritis Is Associated with Inner Nuclear Layer Thickening and Microcystic Macular Edema Independently of Multiple Sclerosis. PLoS ONE 2013, 8, e71145. [Google Scholar] [CrossRef]
  44. Motamedi, S.; Gawlik, K.; Ayadi, N.; Zimmermann, H.G.; Asseyer, S.; Bereuter, C.; Mikolajczak, J.; Paul, F.; Kadas, E.M.; Brandt, A.U. Normative Data and Minimally Detectable Change for Inner Retinal Layer Thicknesses Using a Semi-automated OCT Image Segmentation Pipeline. Front. Neurol. 2019, 10, 1117. [Google Scholar] [CrossRef]
  45. Lang, A.; Carass, A.; Hauser, M.; Sotirchos, E.S.; Calabresi, P.A.; Ying, H.S.; Prince, J.L. Retinal layer segmentation of macular OCT images using boundary classification. Biomed. Opt. Express 2013, 4, 1133. [Google Scholar] [CrossRef] [Green Version]
  46. Zhang, Z.; Liu, Q.; Wang, Y. Road Extraction by Deep Residual U-Net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef] [Green Version]
  47. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  48. Ngo, L.; Cha, J.; Han, J.H. Deep Neural Network Regression for Automated Retinal Layer Segmentation in Optical Coherence Tomography Images. IEEE Trans. Image Process. 2019, 29, 303–312. [Google Scholar] [CrossRef]
  49. Garvin, M.K.; Abramoff, M.D.; Wu, X.; Russell, S.R.; Burns, T.L.; Sonka, M. Automated 3-D Intraretinal Layer Segmentation of Macular Spectral-Domain Optical Coherence Tomography Images. IEEE Trans. Med. Imaging 2009, 28, 1436–1447. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Li, K.; Wu, X.; Chen, D.; Sonka, M. Optimal Surface Segmentation in Volumetric Images-A Graph-Theoretic Approach. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 119–134. [Google Scholar] [PubMed] [Green Version]
  51. Abràmoff, M.D.; Garvin, M.K.; Sonka, M. Retinal Imaging and Image Analysis. IEEE Rev. Biomed. Eng. 2010, 3, 169–208. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. He, Y.; Carass, A.; Solomon, S.; Saidha, S.; Calabresi, P.; Prince, J.L. Retinal layer parcellation of optical coherence tomography images: Data resource for multiple sclerosis and healthy controls. Data Brief 2019, 22, 601–604. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The pipeline of the proposed method consists of four steps: preprocessing, retinal region segmentation, intraretinal layer segmentation, and postprocessing.
Figure 1. The pipeline of the proposed method consists of four steps: preprocessing, retinal region segmentation, intraretinal layer segmentation, and postprocessing.
Jimaging 08 00139 g001
Figure 2. Visual representation of different retinal layers and boundaries after segmentation. Our method segments 9 different regions (7 highlighted with yellow background on the left side of the image and other two region are: above ILM and below BM) from 8 different retinal boundaries (highlighted with cyan background on the right side of the image). Layer abbreviations are positioned within their respective layer; boundary abbreviations are positioned on top of their respective segmentation lines. Abbreviations: ILM—inner limiting membrane, mRNFL—macular retinal nerve fibre layer, GCL—ganglion cell layer, IPL—inner plexiform layer, GCIPL—macular ganglion cell/inner plexiform layer, INL—inner nuclear layer, OPL—outer plexiform layer, ONL—outer nuclear layer, ELM—external limiting membrane, MZ—myoid zone, EZ—ellipsoid zone, OS—outer segment, RPE—retinal pigment epithelium, OL—outer layer and BM—Bruch’s membrane.
Figure 2. Visual representation of different retinal layers and boundaries after segmentation. Our method segments 9 different regions (7 highlighted with yellow background on the left side of the image and other two region are: above ILM and below BM) from 8 different retinal boundaries (highlighted with cyan background on the right side of the image). Layer abbreviations are positioned within their respective layer; boundary abbreviations are positioned on top of their respective segmentation lines. Abbreviations: ILM—inner limiting membrane, mRNFL—macular retinal nerve fibre layer, GCL—ganglion cell layer, IPL—inner plexiform layer, GCIPL—macular ganglion cell/inner plexiform layer, INL—inner nuclear layer, OPL—outer plexiform layer, ONL—outer nuclear layer, ELM—external limiting membrane, MZ—myoid zone, EZ—ellipsoid zone, OS—outer segment, RPE—retinal pigment epithelium, OL—outer layer and BM—Bruch’s membrane.
Jimaging 08 00139 g002
Figure 3. The general architecture of the proposed compressed U-Net. RS-net and IS-Net both have the same architecture with different channel depth c, kernel size f, number of input channels n and number of output classes o.
Figure 3. The general architecture of the proposed compressed U-Net. RS-net and IS-Net both have the same architecture with different channel depth c, kernel size f, number of input channels n and number of output classes o.
Jimaging 08 00139 g003
Figure 4. The comparison of different loss functions on a single class (INL). As can be seen, the weighted focal loss is capable of producing the minimum false positives.
Figure 4. The comparison of different loss functions on a single class (INL). As can be seen, the weighted focal loss is capable of producing the minimum false positives.
Jimaging 08 00139 g004
Figure 5. The first column shows small holes, which appear during the classification in both networks. The second column consists of images without holes. The last column shows 3D reconstructed surfaces of different layers.
Figure 5. The first column shows small holes, which appear during the classification in both networks. The second column consists of images without holes. The last column shows 3D reconstructed surfaces of different layers.
Jimaging 08 00139 g005
Figure 6. The visualization of the segmentation lines computed by the proposed method for three different data sets.
Figure 6. The visualization of the segmentation lines computed by the proposed method for three different data sets.
Jimaging 08 00139 g006
Figure 7. All nine different classes segmented by the proposed method.
Figure 7. All nine different classes segmented by the proposed method.
Jimaging 08 00139 g007
Figure 8. Example segmentation with a low-quality scan to show the performance of the proposed method compared to state-of-the-art methods (HeyEx—device segmentation, AURA [45], OCTExp [49]).
Figure 8. Example segmentation with a low-quality scan to show the performance of the proposed method compared to state-of-the-art methods (HeyEx—device segmentation, AURA [45], OCTExp [49]).
Jimaging 08 00139 g008
Table 1. Background and foreground pixel imbalance: the table shows the percentage of foreground pixels (FP) with respect to the total pixels in the training data for each class.
Table 1. Background and foreground pixel imbalance: the table shows the percentage of foreground pixels (FP) with respect to the total pixels in the training data for each class.
DataVitreousmRNFLGCIPLINLOPLONLMZOLb-BM
FP33%5%9%5%3%9%3%8%24%
Table 2. The compression experiments with different architectural parameters for both RS and IS networks. The selected configuration for each network is highlighted in bold.
Table 2. The compression experiments with different architectural parameters for both RS and IS networks. The selected configuration for each network is highlighted in bold.
Network ArchitecturefcValidation DSCTest DSC# Parametersonhw
RS-Net540.99790.997679K31512512
580.99840.99750.31M31512512
5160.99330.99331.2M92512512
IS-Net3160.99170.989148K92512512
380.98260.98280.12M92512512
3320.99330.99361.9M92512512
Table 3. A comparison between the original U-Net and the compressed U-Net (proposed). The prediction time (in seconds) is observed on CPU only for a single B-scan. Abbreviations: VDSC—validation DSC, TDSC—test DSC, MS—model size, TT—training time, PT—prediction time. The entries corresponding to the proposed network are highlighted in bold.
Table 3. A comparison between the original U-Net and the compressed U-Net (proposed). The prediction time (in seconds) is observed on CPU only for a single B-scan. Abbreviations: VDSC—validation DSC, TDSC—test DSC, MS—model size, TT—training time, PT—prediction time. The entries corresponding to the proposed network are highlighted in bold.
Network Arch.Stage# ParametersVDSCTDSCMS(Mb)TTPT
U-Net [22]RS31M0.99860.99851211141.97
IS31M0.99350.99291211292.06
ResU-Net [46]RS13M0.99860.9985521121.08
IS13M0.99430.9944521301.51
CompressedRS79K0.99790.99760.4180.11
U-NetIS1.2M0.99330.99335510.31
Table 4. Training-related metrics. Abbreviations: NetA—network architecture Tr—training, Te—test, Val—validation, Acc—accuracy.
Table 4. Training-related metrics. Abbreviations: NetA—network architecture Tr—training, Te—test, Val—validation, Acc—accuracy.
NetATr AccTr DSCTr LossVal AccVal DSCVal LossTe AccTe DSCTe LossEpoch
RS0.99850.99920.00640.99840.99920.00640.99850.99910.006128
IS0.99530.99740.01750.99520.99740.01790.99500.99730.018511
Table 5. A comparison between HeyEx segmentation and the proposed method (CCU-INSEG) using MAE (standard deviation) in μm. Abbreviations: HeyEx—Heidelberg eye explorer (device integrated segmentation software).
Table 5. A comparison between HeyEx segmentation and the proposed method (CCU-INSEG) using MAE (standard deviation) in μm. Abbreviations: HeyEx—Heidelberg eye explorer (device integrated segmentation software).
MethodILMRNFL-GCLIPL-INLINL-OPLOPL-ONLELMEZ-OSBMTotal
HeyEx3.5 (3.2)6.3 (2.6)6.6 (2.5)7.6 (2.5)4.8 (3.0)3.5 (2.8)5.2 (2.9)4.0 (2.7)5.2 (2.6)
CCU2.1 (0.4)3.0 (0.6)2.6 (0.4)2.8 (0.5)2.6 (0.7)1.7 (0.4)1.7 (0.4)1.9 (0.6)2.3 (0.4)
Table 6. DSC between CCU-INSEG’s outcome and the manually segmented data.
Table 6. DSC between CCU-INSEG’s outcome and the manually segmented data.
DataVitreousmRNFLGCIPLINLOPLONLMZOLb-BMTotal
CUB0.990.920.950.920.890.960.900.970.990.95
UCLA0.990.890.940.920.880.960.880.970.990.94
JHU0.990.900.940.840.830.940.850.950.990.92
Table 7. MAE (standard deviation) in μ m between CCU-INSEG’s outcome and the manually corrected segmentation for each boundary.
Table 7. MAE (standard deviation) in μ m between CCU-INSEG’s outcome and the manually corrected segmentation for each boundary.
DataILMRNFL-GCLIPL-INLINL-OPLOPL-ONLELMEZ-OSBMTotal
CUB2.1 (0.4)3.1 (0.6)2.5 (0.4)2.9 (0.4)2.6 (0.7)1.8 (0.4)1.7 (0.4)1.8 (0.4)2.3 (0.4)
UCLA2.2 (0.2)4.2 (1.0)2.7 (0.4)3.2 (0.5)2.9 (0.4)2.0 (0.3)2.0 (0.4)2.0 (0.5)2.6 (0.3)
JHU3.2 (0.4)4.0 (0.7)3.9 (0.6)5.3 (1.0)3.8 (1.0)2.6 (0.4)2.6 (0.4)4.0 (0.9)3.7 (0.3)
Table 8. The reliability of the standard parameters (top three rows) and fidelity of the computed parameters (MAE in μ m) from CCU-INSEG’s segmentation with respect to the manual segmentation (the last three rows). Abbreviations (The last two letters are either T or V, where T is thickness and V is volume. The other abbreviations are as follows: ICC—intra-class correlation coefficient, LCI—lower confidence interval, UCI—upper confidence interval, mRNFL—macular retinal nerve fiber layer, GCIPL—macular ganglion cell and inner plexiform layer, INL—inner nuclear layer, GCC—ganglion cell complex, TM—total macula).
Table 8. The reliability of the standard parameters (top three rows) and fidelity of the computed parameters (MAE in μ m) from CCU-INSEG’s segmentation with respect to the manual segmentation (the last three rows). Abbreviations (The last two letters are either T or V, where T is thickness and V is volume. The other abbreviations are as follows: ICC—intra-class correlation coefficient, LCI—lower confidence interval, UCI—upper confidence interval, mRNFL—macular retinal nerve fiber layer, GCIPL—macular ganglion cell and inner plexiform layer, INL—inner nuclear layer, GCC—ganglion cell complex, TM—total macula).
ParammRNFL TGCIPL TINL TGCC TTM TmRNFL VGCIPL VINL VGCC VTM V
ICC0.930.990.960.990.990.930.990.960.990.99
LCI0.880.980.930.970.980.880.980.930.970.98
UCI0.960.990.980.990.990.960.990.980.990.99
CUB1.20.90.80.50.70.0240.0180.0160.0100.014
UCLA2.01.60.80.50.60.0390.0320.0160.0090.013
JHU0.91.86.91.42.40.0180.0350.1400.0280.044
Table 9. DSCs are computed using 45 manually delineated B-scan images between AURA, OCTExp, RelayNet and CCU. ★ RelayNet and OCTExp segment ONL + MZ in a single class; therefore, MZ column is empty for both, and ONL column shows DSC for both ONL and MZ. ▲ RelayNet segments OL in two different classes. For this comparison, we combined both classes in OL and computed DSC with respect to manual segmentation. The best outcomes for each layer are highlighted in bold.
Table 9. DSCs are computed using 45 manually delineated B-scan images between AURA, OCTExp, RelayNet and CCU. ★ RelayNet and OCTExp segment ONL + MZ in a single class; therefore, MZ column is empty for both, and ONL column shows DSC for both ONL and MZ. ▲ RelayNet segments OL in two different classes. For this comparison, we combined both classes in OL and computed DSC with respect to manual segmentation. The best outcomes for each layer are highlighted in bold.
MethodVitreousmRNFLGCIPLINLOPLONLMZOLb-BMTotal
AURA [45]0.990.570.620.860.850.960.920.980.990.86
OCTExp [49]0.990.560.760.730.690.89 -0.850.990.80
RelayNet [20]0.870.700.730.640.860.77 -0.90 0.920.80
CCU0.990.880.930.880.840.960.890.980.990.93
Table 10. A comparison between the proposed method (CCU, in bold) and the state-of-the-art methods in terms of the correction time (in seconds) and the deviation (in μ m) between the corrected and uncorrected versions of the selected 15 volumes. MAE: mean absolute error. The best outcomes are highlighted in bold.
Table 10. A comparison between the proposed method (CCU, in bold) and the state-of-the-art methods in terms of the correction time (in seconds) and the deviation (in μ m) between the corrected and uncorrected versions of the selected 15 volumes. MAE: mean absolute error. The best outcomes are highlighted in bold.
MethodAvg Correction Time (Min-Max) (s)Correction Needed (Volumes)MAE (Min–Max) (in μ m)
HeyEx162 (99–320)11 (73.3%)0.016 (0–0.301)
OCTExp [49]209 (86–638)6 (40%)0.034 (0–0.950)
AURA  [45]596 (204–1658)15 (100%)0.164 (0–5.938)
CCU97 (82–150)2 (13.3%)0.001 (0–0.115)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yadav, S.K.; Kafieh, R.; Zimmermann, H.G.; Kauer-Bonin, J.; Nouri-Mahdavi, K.; Mohammadzadeh, V.; Shi, L.; Kadas, E.M.; Paul, F.; Motamedi, S.; et al. Intraretinal Layer Segmentation Using Cascaded Compressed U-Nets. J. Imaging 2022, 8, 139. https://doi.org/10.3390/jimaging8050139

AMA Style

Yadav SK, Kafieh R, Zimmermann HG, Kauer-Bonin J, Nouri-Mahdavi K, Mohammadzadeh V, Shi L, Kadas EM, Paul F, Motamedi S, et al. Intraretinal Layer Segmentation Using Cascaded Compressed U-Nets. Journal of Imaging. 2022; 8(5):139. https://doi.org/10.3390/jimaging8050139

Chicago/Turabian Style

Yadav, Sunil Kumar, Rahele Kafieh, Hanna Gwendolyn Zimmermann, Josef Kauer-Bonin, Kouros Nouri-Mahdavi, Vahid Mohammadzadeh, Lynn Shi, Ella Maria Kadas, Friedemann Paul, Seyedamirhosein Motamedi, and et al. 2022. "Intraretinal Layer Segmentation Using Cascaded Compressed U-Nets" Journal of Imaging 8, no. 5: 139. https://doi.org/10.3390/jimaging8050139

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop