Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Deep learning approaches for segmenting Bruch’s membrane opening from OCT volumes

Open Access Open Access

Abstract

Automated segmentation of the eye’s morphological features in OCT datasets is fundamental to support rapid clinical decision making and to avoid time-consuming manual segmentation of the images. In recent years, deep learning (DL) techniques have become a commonly employed approach to tackle image analysis problems. This study provides a description of the development of automated DL segmentation methods of the Bruch’s membrane opening (BMO) from a series of OCT cross-sectional scans. A range of DL techniques are systematically evaluated, with the secondary goal to understand the effect of the network input size on the model performance. The results indicate that a fully semantic approach, in which the whole B-scan is considered with data augmentation, results in the best performance, achieving high levels of similarity metrics with a dice coefficient of 0.995 and BMO boundary localization with a mean absolute error of 1.15 pixels. The work further highlights the importance of fully semantic methods over patch-based techniques in the classification of OCT regions.

© 2020 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Bruch's membrane, which was first discovered by Karl Bruch [1], is a thin 5-layer structure located between the retinal pigment epithelium and the choroid. It partly regulates the reciprocal exchange of biomolecules, nutrients, oxygen, fluids and metabolic waste products between the retina and the general circulation [2]. This membrane stretches for over half of the eye, and can conform its shape to changes in the intraocular pressure [3]. Around the optic nerve this membrane is not present and that gap is defined as the Bruch's membrane opening (BMO). Anatomical changes in Bruch’s membrane have been observed with age [4] and degenerative changes are associated with the development of age-related macular degeneration (AMD) [5,6,7].

The most commonly used methods to observe the optic nerve head are fundus photography, scanning laser ophthalmoscopy (SLO) and optical coherence tomography (OCT). While the first two provide en-face images, OCT allows cross-sectional imaging of the three-dimensional tissue morphology with high resolution. The BMO-based outer optic nerve head boundary can provide precise estimation of neuroretinal rim tissue and can be used for defining the optic nerve head (ONH) anatomy [8,9]. Additionally, parameters based on BMO can be used for determining nerve damage at the optic disc, supporting the early diagnosis of glaucoma [10,11]. Since glaucoma can lead to irreversible vision loss, early detection is crucial in order to instigate clinical treatment. BMO-based parameters such as horizontal/perpendicular rim width, minimum rim width and area are useful biomarkers for glaucoma diagnosis in addition to the commonly employed thickness of the retinal nerve fiber layer (RNFL) [12,13,14]. Some of the BMO-based parameters correlate more highly with glaucoma-related consequences such visual field loss, than do changes in RNFL thickness [10,12,1517] and can be used in assessing the progression of glaucoma [18]. Additionally, combining BMO-based parameters with RNFL thickness can improve the efficacy in diagnosing glaucoma [13]. The location of BMO can also be used for determining the position and shape of lamina cribrosa [19,20], which also provides information about glaucoma progression.

Accurate, automated, and objective methods to extract the BMO from a series of OCT scans and to estimate the corresponding shape parameters is fundamental for achieving useful diagnostic information that can support clinical decision making. Additionally, given that the retina and choroid layers are not anatomically present in this region, accurate estimation of the BMO location can facilitate segmentation of other nearby retinal features [21,22]. Manual extraction of the BMO, although often treated as the ground truth, is subjective and time consuming. Hence, there are various incentives to develop accurate and precise automated BMO segmentation.

Several approaches to solve the problem of automated segmentation of the BMO using image analysis methods have been proposed. Some of them use information obtained only from the OCT volumetric scan [2325], and some use that information in combination with images from color fundus photography [26,27]. Three of these methods focus on finding a representative set of features from OCT volumes based on previously segmented intraretinal layers and other image-specific parameters such as pixel intensity. Then, these sets of features are used to classify columns or voxels into two classes (background or BMO) using standard machine learning (ML) classifiers such as a random forest or k-NN (nearest neighborhood) classifier. After classification, post-processing steps are applied to optimize the BMO segmentation [23,24,26]. Another method utilizes an image deconvolution approach to find Bruch’s membrane and then based on termination of this membrane and ANN-PCA (artificial neural network-principal component analysis) algorithm, it finds the elliptical shape of BMO path [25]. Further, a method by Chen and colleagues [27] proposed a 2-step deep learning procedure. The first step is a coarse detection of a region of interest (ROI), which is located near the BMO, after which the small ROI image is further processed using a model, with an architecture similar to U-Net fully semantic network, to extract the precise location of the BMO.

Deep learning algorithms have been successfully used in problems of biomedical image segmentation, and have shown state-of-the-art performance in a number of OCT image analysis tasks, including retinal segmentation [2831], choroid segmentation [32,33] and image classification for pathology detection [34,35]. Deep learning OCT image analysis methods can be applied at different levels (i.e., input sizes). Some of them use ANN on a single vector obtained for every image column (A-scan) [36], whereas others use a convolutional neural network (CNN) applied to a set of image columns (patch or collection of A-scan) [37], and others use the entire OCT image (B-scan) applying a fully convolutional network (FCN) with architecture like U-Net [38].

In this study, a range of deep learning techniques to segment BMO in OCT images, without using any feature extraction techniques, are developed and compared. Different ML segmentation methods (ANN, CNN, and FCN) based on a model that can utilize a different input size (image column, group of columns, and whole image) are developed. Particular attention is paid to understand the effect of the network input size on the segmentation task performance. Additionally, the effect of volumetric input data on the segmentation performance for each of these methods is assessed. Although the presented methods are applied for BMO segmentation problems, the findings of this study may be utilized in other OCT image segmentation tasks with deep learning method [3638].

2. Material and methods

2.1 Subjects and data acquisition

Data from a retrospective study [39] including a total of 325 participants (219 females, 106 males, age ranging from 42 to 86 years) were used in this study. The study protocol was approved by the Bioethical Committee of the Wroclaw Medical University (KB–332/2015) and adhered to the tenets of the Declaration of Helsinki. All participants gave written consent to participate in the study. Data from one eye, randomly selected from every patient, was included in the dataset. All details about data acquisition and selection of participants were described elsewhere [39].

2.2 Dataset

The study included imaging of the ONH, using spectral domain optical coherence tomography (SD-OCT) (Spectralis, Heidelberg Engineering GmbH, Heidelberg, Germany). For each participant, a volumetric scan was acquired, comprised of 49 horizontal B-scans. Each B-scan is a single channel grey scale image of size 436 × 384 pixels (height ×width), which corresponds to physical dimensions of about 1.93 × 4.40 mm. An example of a SLO image with the scanning pattern and a B-scan is shown in Fig. 1. For each volumetric scan, the 16 centrally located scans with visible BMO, were manually annotated from the total of 49 scans that were collected. In total, 5200 annotated B-scans were used in this study. Thus, the dataset consists of those horizontal B-scans that intersect the BMO, the location of BMO was manually annotated in each B-scan by a trained ophthalmologist, setting the coordinates of two points that indicate this opening.

 figure: Fig. 1.

Fig. 1. SLO image with the OCT scanning protocol (A), orange lines represent the position of the annotated scans, whereas blue lines represent the additional scans used in 3-D approaches. Pink lines refer to the position of scans that have not been manually annotated. Illustrative SD-OCT B-scan with overlaid coordinates of Bruch’s membrane opening. The termination points of Bruch’s membrane opening are BMO1 and BMO2 (B) and the corresponding ground truth mask (C).

Download Full Size | PDF

The dataset was randomly divided into training and testing parts using a 6 to 4 ratio. Division was made based on patients, which ensured that all images that belong to one patient were assigned to the same group. The training set was further divided into training and validation parts in 8 to 2 ratio and the validation set was used only in the training process to supervise it. The final data spilt was 2495/625/2080 image for training, validation and testing, respectively. To have a fair comparison of the method, this data split was kept constant in the evaluation of each method.

2.3 Preprocessing of the dataset and data augmentation

Given that the dataset used in this study consists of images and their corresponding pair of coordinates that delineate the limits of the BMO, it was necessary to prepare a ground truth mask. For every annotated scan, every pixel that was located between and including the BMO points in horizontal axis was assigned to BMO class (class 1) and every pixel either side of those points was treated as a background (class 0). An example B-scan with its corresponding mask is shown in Fig. 1(B and C).

For all 3-D approaches, additional preprocessing of the images was done to create an input dataset using three neighboring B-scans, where the BMO coordinates were taken as a reference from the middle scan. Two additional B-scans, to those 16 annotated by the ophthalmologist, were used to create 3-D images (one scan before the first annotated scan and one scan after the last annotated scan) (see Fig. 1(A), where those two additional scans are shown in blue color).

Data were collected to ensure both the BMO and the lamina cribrosa were placed within the instrument’s imaging depth. Consequently, the location of the ONH within each volumetric scan was not in the same vertical position. Given this, vertical data augmentation for each of the considered methods (i.e., ANN, CNN, and FCN) was applied. That augmentation includes random vertical shift of images in a range from −90 to 90 pixels (covering about 20% of the image height).

2.4 Comparison of segmentation algorithms

In order to investigate the problem of BMO segmentation using deep learning techniques, it was decided to use different segmentation approaches. Each of them uses the same dataset, but a different input data size. The considered segmentation methods are: one dimensional (A-scan, a column from a B-scan), a two-dimensional patch based (group of consecutive A-scans) and a large two dimensional (entire image). For each of these three methods, their corresponding 3-D approaches, which extend the data to the neighboring B-scans, were also considered. The 3-D approach is an approach which uses a 3-channel image created from 3 neighboring scans, in which each channel represents different spatial information (i.e. different B-scan slice).

The general overview of the utilized methods is presented in Fig. 2. To ensure that the results obtained can be compared across the methods, the same pre-processing and the post-processing steps were used to determine the evaluation metrics. Some minor differences will be noted in the description of each specific method. The entire experiment was run on Python 3.6.4 with following libraries: keras 2.2.4, numpy 1.10.0, scikit-learn 0.22.2, scipy 1.1.0, tensorflow 1.14.0.

 figure: Fig. 2.

Fig. 2. General overview of the proposed segmentation methods. Columns present (left to right): input type with respect to the whole image, method used (network architecture) and direct output of the model (prediction map). Rows present an overview of the operation of the method based on (top to bottom): a single A-scan, a set of columns (column based), and an entire B-scan (whole image).

Download Full Size | PDF

2.5 A-scan segmentation

The A-scan method is based on a one-dimensional input, so the input to the ANN is a single A-scan (image column). The ANN architecture, which was adapted from [36], and follows these layer dimensions: 436-200-75-10-2. Each of the 16 annotated scans per subject was divided into a set of single columns and labeled according to the previously created mask. Additionally, a 3-D version of this method was developed by extending the model input. A single A-scan from each neighboring scan (from the 3-D cube) was taken as the input to the ANN model and three of these models were connected by adding an average layer to combine the outputs of the three models, which resulted in a single model consisting of three ANN models.

2.6 Column-based segmentation

Two variations of the column-based method were developed and compared in this study. To prepare the patches, the same strategy was used for the different input images (2-D or 3-D). Overlapping patches of 16 neighboring columns (A-scans) were taken in steps of 4 pixels. Based on the masks created previously, each patch was labeled to the class that the majority of pixels in the mask belong to. The network architecture was adapted from [37] and extended to allow for 3-D input. Note that the only difference between the 2-D and 3-D approaches is the shape of input (number of input channels). For the purpose of testing, in both methods, patches were evaluated in steps of 1 pixel.

2.7 B-scan fully semantic segmentation

The last method is based on the information contained in the entire B-scan. In this approach, a deep learning model was trained with the entire set of images and ground truth masks. A fully convolutional network with U-Net like architecture [38] was used. The architecture is presented in Fig. 3. Note that some padding was added to the input image to be able to down-sample it five times. At the output, the padding was removed to be able to compare the network-predicted maps with the ground truth masks. For the 3-D approach, the network architecture was the same, but with the number of channels in the input image set to 3.

 figure: Fig. 3.

Fig. 3. Overview of fully convolutional network architecture (U-Net) on the right side. The convolution block is described in the left side. In the figure, n refers to number of input image channels n = 1 in 2D methods and n=3 in 3-D methods. F is the number of filters.

Download Full Size | PDF

Some additional post-processing was done for this specific method. As this approach provides a per-pixel level classification not a column based one, as does the previous two methods. Thus, the predicted maps can have columns in which two classes are present. To standardize the results, the output prediction maps were post-processed. If the mean value of pixels in one column was greater than 0.5, then all pixel values in that column were set to class 1 (BMO). Otherwise, they was set to class 0 (background).

Additionally, to assess the impact on performance a range of models with modifications to the standard U-net architecture were investigated. Three types of modifications were used including: incorporation of residual learning (Residual) [40], the addition of squeeze-excitation blocks (SE)[41] and with a recurrent neural network bottleneck (RNN) [42]. In case of the last method, the standard RNN layers were replaced with CNN-LSTM layers [43].

2.8 Training

Each model was trained with the Adam optimizer with a learning rate equal to 0.001 and binary cross entropy as a loss function. The training process was set to a maximum of 30 epochs, which was deemed sufficient to train all models (determined empirically by examining the trajectory of the loss function). During training, at each step, the accuracy on the validation set was recorded and the highest accuracy model (the best) was used in the evaluation process. Batch size was set to 100 for A-scan and columns-based segmentation, 5 for B-scan fully semantic and 1 for 3-D B-scan fully semantic. For each method, to evaluate its consistency, the experiments were repeated three times (network trained 3 times with random weights initialization) and the mean value of all measures with standard deviation was reported.

2.9 Post-processing and evaluation

After evaluation, each method produced a probability map. The same validation measures were calculated across the different techniques. Two groups of measures were selected: similarity and BMO distance measures. The first group calculates the similarity between the predicted and ground truth mask and includes Dice coefficient (DSC), Jaccard similarity coefficient (JSC) and pixel accuracy (ACC), see equations (1) to (3) respectively. In equations TP, TN, FP, FN are values calculated from the confusion matrix and denote true positive, true negative, false positive and false negative, respectively.

$$DSC\; = {\; }\frac{{2TP}}{{2TP + FP + FN{\; }}}{\; \; \; \; }$$
$$JSC = \; \frac{{TP}}{{TP + FP + FN\; }}$$
$$ACC{\; } = {\; }\frac{{TP + TN}}{{TP + \textrm{TN} + FP + FN{\; }}}$$
The second group of metrics, the distance measures, include: mean absolute error (MAE), mean error (ME), median absolute error (MDAE) and root mean square error (RMSE). In order to calculate them, the difference between annotated BMO points and predicted BMO had to be evaluated. These predicted points needed to be first extracted from output prediction maps.

Since sometimes there was a lack of homogeneity in the predicted maps from the different networks, which means that such maps contained more than one block predicted as class one (BMO region), a post-processing step was included. The main purpose of this step was to determine the coordinates at the beginning and the end of the BMO. This post-processing was conducted using the following procedure and its overview is presented in Fig. 4.

  • 1. BMO1’s position was designated as a location of the first transition from class 0 to class 1, not closer than 50 pixels (determined empirically) away from the beginning of the scan.
  • 2. By checking each subsequent column, the location of the transition from class 1 to class 0 was temporarily marked as BMO2. The number of columns being constantly predicted as class 0 was calculated and saved as the gap width.
  • 3. If any subsequent columns were marked as class 1 in the scan, their width (marked as temporary width in the figure) was compared with the gap width. If larger, the last of these columns was saved as BMO2. This step was repeated until the end of the image was found.

 figure: Fig. 4.

Fig. 4. General overview of the proposed postprocessing. It presents B-scan (A), sample mask with not homogeneous BMO region (B) and a clean mask (C). Width of a gap between two BMO-regions in figure B is marked as gap width. Width of the second BMO region is annotated as temporary width and because it is larger than gap width, temporary BMO2 is moved to the end of the second column and it is marked as BMO2. An arrow indicates processing direction.

Download Full Size | PDF

Since each scan has two BMO points, the mean value of calculated distance measures of those two points is calculated and presented.

3. Results

The mean value and the standard deviation of the similarity measures are shown in Table 1 and of BMO distance measures in Table 2. For each of the methods, the results are presented in the following order: 2-D, 2-D with data augmentation, 3-D and 3-D with data augmentation. Additionally, results from inter-grader annotation is provided in both tables.

Tables Icon

Table 1. Mean values and the corresponding standard deviations of the similarity measures extracted from 3 independent runs of the different deep learning models and results from inter-grader annotations. The highest values (best performance) are shown in bold font.

Tables Icon

Table 2. Mean values and the corresponding standard deviations of the distance errors extracted from 3 independent runs of the different deep learning models and results from inter-grader annotations. The lowest values (best performance) are shown in bold font. All measure errors are computed in pixels.

Examining the similarly measurements (Table 1), it is clear that all methods provide a reasonable performance with metrics of DSC > 0.93, JSC > 0.90 and ACC > 0.96. When comparing the methods, the fully semantic one provides superior performance compared to the other two. The best DSC results from each method are: 0.973 for the A-scan, 0.989 for the column-based and 0.994 for the fully semantic approach. Other similarity measures (i.e., JSC and ACC) also show a superior performance for the fully semantic segmentation with U-NET architecture. The result in Table 1 also shows the impact of data augmentation and the 3-D approach on each segmentation method. For the A-scan segmentation, data augmentation does not seem to improve the segmentation result, but additional information from neighboring scans (3-D approach) does improve the segmentation result from 0.964 to 0.973-pixel accuracy. For the column-based segmentation with CNN architecture, both data augmentation and the 3-D approach marginally improve the segmentation results, whereas for the fully semantic segmentation method, the best result is achieved with data augmentation, but without any additional scans (2-D method). It is worth noting that the changes in similarity metrics are generally relatively small in magnitude.

The BMO distance measurements in Table 2 show a larger difference between the methods than the similarity measurements (Table 1). This is probably an indication that the overall prediction map is accurate, but some of the methods contain more outliers in the map than the others, negatively impacting the correct detection of the BMO points. However, the main findings remain the same. Looking at the absolute errors, the superior performance of the B-scan fully semantic method is evident. Examining the different metrics, it is also worth noting that the low values of the mean error indicate that models are not biased in any direction (i.e., the models make similar errors on right and left side of the BMO point). RMSE values for the A-scan and column-based segmentation method show that these models are not able to segment all of the scans; indicating that there are some outliers. In some of the scans, the BMO has poor visibility because of lower quality of the image and this could be the cause of outliers. However, the segmentation method based on the fully semantic approach, handles those cases well, as shown by the low RMSE result. MAE indicates the best performance (the lowest value) for fully semantic segmentation with U-Net architecture and data augmentation usage of 1.15 pixels. MDAE, which in this case is equal to 1 pixel, also shows that this segmentation method is capable of deriving the BMO with high accuracy.

Table 3 shows comparison between results obtained from model based on the standard U-net architecture and those with model modifications. Results from standard U-net architecture showed the highest performance for the BMO segmentation task. The models based on modified U-net like architectures achieved similar results, exceeding those from the ANN and CNN methods, although overall the performance across the different metrics were slightly inferior to the standard U-net.

Tables Icon

Table 3. Mean values and the corresponding standard deviations of the comparison between models with different architectures. The best values are shown in bold font. All distance errors are computed in pixels.

4. Clinical application

Although the BMO metrics derived from a single OCT scan can be used to capture the cross-sectional shape of the ONH, a 3-D analysis is commonly applied to capture the entire ONH shape. Based on the 2-D results, the best segmentation method was selected (i.e., the fully semantic method with data augmentation approach) for the subsequent 3-D analysis. Because the previous model was trained only on the scans that contained BMO, for further analysis, the selected model was fine-tuned with an additional 10 B-scans for each of the patient datasets, without the BMO present in the image (single class mask). The results obtained after this process are presented in Table 4.

Tables Icon

Table 4. Result from one fine-tuned fully semantic model with U-Net like architecture.

Next, the entire volumetric scan (sets of 49 B-scans) was processed sequentially, extracting two BMO points for every processed scan, if present. The collection of points was used for estimating the BMO outline. In order to compare this result to a ground truth, the testing dataset was manually annotated by a masked research trainee (all 49 B-scans). When comparing the results for the previously annotated 16 scans, the mean absolute difference (± 1 standard deviation) was only 3.1 ± 2.3 pixels. Figure 5 shows an example of the SLO image corresponding to a given OCT volume consisting of 49 B-scans with the segmented BMO outline.

 figure: Fig. 5.

Fig. 5. Illustrative examples of original SLO image (left column) and the same image with overlaid BMO outlines (right column) segmented from the corresponding OCT volume consisting of 49 B-scans. Blue color indicates the BMO outline from manual annotation and the grey one, the BMO outline segmented by the proposed model.

Download Full Size | PDF

Table 5 shows the mean values and the corresponding standard deviations calculated across test subjects of the two measures assessing the performance (DL vs ground-truth): DSC and the unsigned border positioning error (UBPE) [44]. The UBPE is the minimum value of the Euclidean distance for each point in the BMO outline predicted by a model with all points in ground truth annotations. Raw model predictions were checked, without post-processing, as well as with the modeling of the shape of the BMO, using a convex hull algorithm. Both approaches show high similarity values (DSC) of the predicted BMO outline compared to the ground truth of 0.953 and 0.959, and high accuracy of prediction with 11.4 µm and 9.8 µm of UBPE, for raw predictions and with convex hull, respectively.

Tables Icon

Table 5. Mean value and standard deviation (calculated across test subjects) of dice coefficient and unsigned border positioning error for BMO outline compared to the outline calculated with annotated scans.

The main difference between these approaches is the value of the UBPE standard deviation, which is higher for the raw model prediction. This indicates the advantage of using the convex hull approach to model the data and reduce variability.

5. Discussion

The results of this study, in which three different methods were considered (i.e., A-scan, column-based, and full image), show that all considered deep learning techniques can be successfully used to segment the Bruch’s membrane opening from OCT volumes with high similarity measures (i.e., DSC, JSC, and ACC). Nevertheless, the fully semantic approach with data augmentation appears to provide a superior performance over the other methods. This is particularly evident when the distance error measures for the BMO location are considered. This method achieved an MAE in determining the BMO position of 1.15 pixels, which corresponds to 13.17 µm.

Part of the motivation of this work was to better understand the influence that the size of the input has on the methods performance (final results of segmentation) and thus it was decided to use and compare three different methods. The first of them, ANN, considered a single A-scan. The next, CNN, used a 16 column-based input (patch-based) for segmentation and the last method used the entire B-scan to perform the segmentation task. Although a number of similar methods to segment and classify OCT image have been proposed previously [2428], to the best of the authors’ knowledge, none of them explore in detail the effect of input size on the performance across different deep learning techniques. The results demonstrate that the performance (similarity metrics) increase as the size of the input increases, pointing to the positive effect of information from the whole image on the segmentation of the BMO position. Thus, were possible, fully semantic methods may need to be considered for classification/segmentation tasks that have been previously applied in an image subsection [27]. In addition, in this study we explore the benefit of using the neighboring scans (i.e., the 3-D approach) to improve segmentation results. Examining the similarly metrics, only marginal differences are observed across the methods. The A-scan and column-based segmentation show some improvement, whereas in the case of the fully semantic segmentation, the 3-D approach slightly decreases the performance. It is evident from these findings that there is not a clear benefit from the use of 3-D data for this particular application.

To fully assess the potential use of this method as a clinical or research tool, the fully semantic segmentation with data augmentation was evaluated to extract the entire BMO outline from OCT volumes. All 49-scans were assessed by the trained model and it resulted in a BMO outline segmentation with a Dice coefficient value of 0.959. This shows the potential of using the proposed model for this purpose, making it possible to accurately extract the size of ONH. Although the obtained results cannot be directly compared to those reported elsewhere, because of different datasets, it is worth noting that in work [26] the authors reported a DSC value of 0.855 and in [23] of 0.853, which are significantly lower than the results obtained by the proposed method of 0.959. Additionally, the UBPE in those papers was 67 µm and 61 µm, respectively, which is a larger error than the 9.8 µm achieved here. This further highlights the potential of the proposed method.

Of practical interest is also the computational complexity of the developed methods. When a single segmentation task is considered for the entire image (B-scan), an average (mean ± 1 standard deviation) processing time of 0.028 ± 0.003 s, 53.61 ± 2.67 s, and 2.36 ± 0.12 s, was obtain for the ANN, CNN and U-net methods, respectively (simulations ran on a personal computer with CPU Intel Core i5-5200U). This indicates that the superior performance of the fully semantic method comes at a time cost when compared to the A-scan method, and that the column-based method cannot match any of the other two methods in terms of computational efficiency.

It is important to acknowledge that in our comparison of different network architectures the number of trainable parameters is different. It was 270 857 (881 367), 24 564 802 (24 567 490), and 12 589 425 (12 589 713) for ANN (ANN 3-D), CNN (CNN 3-D), and U-net (U-net 3-D), respectively. However, in the comparison, it was intended to ascertain whether a particular type of network architecture provides better segmentation result than the other with the size of the input being adequately matched to the particular type of network architecture to maximize its performance.

The findings of the effect of input size need to be confirmed with other larger datasets and this is a study limitation that needs to be further explored. Additionally, it is likely that the performance, particularly error distance metrics, of the ANN and CNN methods could be improved with more sophisticated BMO boundary detection techniques. However, for consistency across the methods, a relatively simple post-processing rule was applied. Although this could be seen as a limitation of the proposed method, it also further demonstrates the robustness of the fully semantic method.

6. Conclusion

There is a continuing interest in the development of machine learning methods for segmenting morphological features of the eye in OCT volumes. In this work, a range of deep-learning-based automated segmentation methods of the Bruch’s membrane opening were considered, as this particular retinal feature can provide important clinical information supporting early glaucoma diagnosis [10,11]. A systematic approach for developing and selecting the most appropriate deep learning technique was taken, in which the input size governed the type of method used and its architecture. Using retrospective OCT volumetric data from 325 participants, the best segmentation results, in terms of similarity metrics and position errors, have been achieved with a fully sematic approach with standard U-Net like architecture, followed by modified U-net architectures, the column-based (CNN) method and the A-scan (ANN) method. Additionally, the performance of the selected method on segmentation of the entire Bruch’s membrane opening outline was considered for the entire volumetric scan, achieving an unsigned positioning border error of 9.8 µm, which shows the potential of this method.

The contributions of this work can be summarized as follows:

  • • three DL techniques to segment BMO in OCT images were developed and compared,
  • • the effect of the network input size on the segmentation task performance was evaluated, and
  • • the effect of volumetric (3-D) input data on the segmentation performance for each of the DL techniques was assessed.

Although the presented methods are purposely designed for segmenting the Bruch’s membrane opening, the findings of this study may be generalized to other OCT image segmentation tasks using deep learning methods. Additionally, given the BMO is an area where the retinal layers are not present, the proposed model also provides useful information to support retina and choroid segmentation methods [21,22]

Funding

Interdisciplinary Doctoral Studies Projects at WUST; Rebecca L. Cooper Medical Research Foundation; National Health and Medical Research Council (APP1186915); Narodowe Centrum Nauki (2018/29/B/ST7/02451); Narodowa Agencja Wymiany Akademickiej (PPI/APM/2019/1/00085/DEC/1).

Disclosures

The authors declare no conflicts of interest.

References

1. K. Bruch, “Untersuchungen zur kenntniss des körnigen pigments der wirbelthiere in physiologischer und pathologischerhinsicht,” Meyer und Zeller (1884).

2. J. C. Booij, D. C. Baas, J. Beisekeeva, T. G. Gorgels, and A. A. Bergen, “The dynamic nature of Bruch's membrane,” Prog. Retinal Eye Res. 29(1), 1–18 (2010). [CrossRef]  

3. C. A. Curcio, M. Johnsonet, and M. Johnson, “Structure, function, and pathology of Bruch’s membrane,” Retina 1(2), 465–481 (2013). [CrossRef]  

4. J. Johnstone, M. Fazio, K. Rojananuangnit, B. Smith, M. Clark, C. Downs, C. Owsley, M. J. Girard, J. M. Mari, and C. A. Girkin, “Variation of the axial location of Bruch’s membrane opening with age, choroidal thickness, and race,” Invest. Ophthalmol. Visual Sci. 55(3), 2004–2009 (2014). [CrossRef]  

5. M. C. Killingsworth, J. P. Sarks, and S. H. Sarks, “Macrophages related to Bruch's membrane in age-related macular degeneration,” Eye 4(4), 613–621 (1990). [CrossRef]  

6. I. Bhutto and G. Lutty, “Understanding age-related macular degeneration (AMD): relationships between the photoreceptor/retinal pigment epithelium/Bruch’s membrane/choriocapillaris complex,” Mol. Aspects Med. 33(4), 295–317 (2012). [CrossRef]  

7. M. C. Killingsworth, J. P. Sarks, and S. H. Sarks, “Macrophages related to Bruch's membrane in age-related macular degeneration,” Eye 4(4), 613–621 (1990). [CrossRef]  

8. B. C. Chauhan and C. F. Burgoyne, “From clinical examination of the optic disc to clinical assessment of the optic nerve head: a paradigm change,” Am. J. Ophthalmol. 156(2), 218–227.e2 (2013). [CrossRef]  

9. A. S. Reis, N. O’Leary, H. Yang, G. P. Sharpe, M. T. Nicolela, C. F. Burgoyne, and B. C. Chauhan, “Influence of clinically invisible, but optical coherence tomography detected, optic disc margin anatomy on neuroretinal rim evaluation,” Invest. Ophthalmol. Visual Sci. 53(4), 1852–1860 (2012). [CrossRef]  

10. D. R. Muth and C. W. Hirneiß, “Structure–function relationship between Bruch’s membrane opening–based optic nerve head parameters and visual field defects in glaucoma,” Invest. Ophthalmol. Visual Sci. 56(5), 3320–3328 (2015). [CrossRef]  

11. B. C. Chauhan, N. O’Leary, F. A. AlMobarak, A. S. C. Reis, H. Yang, G. P. Sharpe, D. M. Hutchison, M. T. Nicolela, and C. F. Burgoyne, “Enhanced detection of open-angle glaucoma with an anatomically accurate optical coherence tomography–derived neuroretinal tim parameter,” Ophthalmology 120(3), 535–543 (2013). [CrossRef]  

12. L. Reznicek, S. Burzer, A. Laubichler, A. Nasseri, C. P. Lohmann, N. Feucht, M. Ulbig, and M. Maier, “Structure-function relationship comparison between retinal nerve fibre layer and Bruch’s membrane opening-minimum rim width in glaucoma,” Int. J. Ophthalmol. 10(10), 1534 (2017). [CrossRef]  

13. J. M. Gmeiner, W. A. Schrems, C. Y. Mardin, R. Laemmer, F. E. Kruse, and L. M. Schrems-Hoesl, “Comparison of Bruch’s membrane opening minimum rim width and peripapillary retinal nerve fiber layer thickness in early glaucoma assessment,” Invest. Ophthalmol. Visual Sci. 57(9), OCT575–OCT584 (2016). [CrossRef]  

14. P. Enders, W. Adler, F. Schaub, M. M. Hermann, T. Dietlein, C. Cursiefen, and L. M. Heindl, “Novel bruch’s membrane opening minimum rim area equalizes disc size dependency and offers high diagnostic power for glaucoma,” Invest. Ophthalmol. Visual Sci. 57(15), 6596–6603 (2016). [CrossRef]  

15. E. J. Lee, K. M. Lee, H. Kim, and T. W. Kim, “Glaucoma diagnostic ability of the new circumpapillary retinal nerve fiber layer thickness analysis based on Bruch’s membrane opening,” Invest. Ophthalmol. Visual Sci. 57(10), 4194–4204 (2016). [CrossRef]  

16. G. Rebolleda, A. Casado, N. Oblanca, and F. J. Muñoz-Negrete, “The new Bruch’s membrane opening–minimum rim width classification improves optical coherence tomography specificity in tilted discs,” Clin. Ophthalmol. 10, 2417–2425 (2016). [CrossRef]  

17. P. Enders, F. Schaub, W. Adler, R. Nikoluk, M. M. Hermann, and L. M. Heindl, “The use of Bruch’s membrane opening-based optical coherence tomography of the optic nerve head for glaucoma detection in microdiscs,” Br. J. Ophthalmol. 101(4), 530–535 (2017). [CrossRef]  

18. A. Belghith, C. Bowd, F. A. Medeiros, N. Hammel, Z. Yang, R. N. Weinreb, and L. M. Zangwill, “Does the location of Bruch’s membrane opening change over time? Longitudinal analysis using San Diego automated layer segmentation algorithm (salsa),” Invest. Ophthalmol. Visual Sci. 57(2), 675–682 (2016). [CrossRef]  

19. G. Rebolleda, J. García-Montesinos, E. De Dompablo, N. Oblanca, F. J. Muñoz-Negrete, and J. J. González-López, “Bruch’s membrane opening changes and lamina cribrosa displacement in non-arteritic anterior ischaemic optic neuropathy,” Br. J. Ophthalmol. 101(2), 143–149 (2017). [CrossRef]  

20. P. Krzyżanowska-Berkowska, A. Melińska, I. Helemejko, and D. R. Iskander, “Evaluating displacement of lamina cribrosa following glaucoma surgery,” Graefe's Arch. Clin. Exp. Ophthalmol. 256(4), 791–800 (2018). [CrossRef]  

21. D. Alonso-Caneiro, S. A. Read, and M. J. Collins, “Automatic segmentation of choroidal thickness in optical coherence tomography,” Biomed. Opt. Express 4(12), 2795–2812 (2013). [CrossRef]  

22. M. Heisler, M. Bhalla, J. Lo, Z. Mammo, S. Lee, M. J. Ju, M. F. Beg, and M. V. Sarunic, “Semi-supervised deep learning based 3-D analysis of the peripapillary region,” Biomed. Opt. Express 11(7), 3843–3856 (2020). [CrossRef]  

23. K. Lee, M. Niemeijer, G. M. Garvin, Y. H. Kwon, M. Sonka, and M. D. Abramoff, “Segmentation of the optic disc in 3-D OCT scans of the optic nerve head,” IEEE Trans. Med. Imaging 29(1), 159–168 (2010). [CrossRef]  

24. M. S. Miri, M. D. Abràmoff, Y. H. Kwon, M. Sonka, and M. K. Garvin, “A machine-learning graph-based approach for 3-D segmentation of Bruch’s membrane opening from glaucomatous SD-OCT volumes,” Med. Image Anal. 39, 206–217 (2017). [CrossRef]  

25. A. Belghith, C. Bowd, R. N. Weinreb, and L. M. Zangwill, “A hierarchical framework for estimating neuroretinal rim area using 3-D spectral domain optical coherence tomography (SD-OCT) optic nerve head (ONH) images of healthy and glaucoma eyes, “ 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society. 3869–3872 (2014).

26. M. S. Miri, M. D. Abràmoff, K. Lee, M. Niemeijer, J. K. Wang, Y. H. Kwon, and M. K. Garvin, “Multimodal segmentation of optic disc and cup from SD-OCT and color fundus photographs using a machine-learning graph-based approach,” IEEE Trans. Med. Imaging 34(9), 1854–1866 (2015). [CrossRef]  

27. Z. Chen, P. Peng, H. Shen, H. Wei, P. Ouyang, and X. Duan, “Region-segmentation strategy for Bruch’s membrane opening detection in spectral domain optical coherence tomography images,” Biomed. Opt. Express 10(2), 526–538 (2019). [CrossRef]  

28. L. Fang, D. Cunefare, C. Wang, R. H. Guymer, S. Li, and S. Farsiu, “Automatic segmentation of nine retinal layer boundaries in OCT images of non-exudative AMD patients using deep learning and graph search,” Biomed. Opt. Express 8(5), 2732–2744 (2017). [CrossRef]  

29. J. Hamwood, D. Alonso-Caneiro, S. A. Read, S. J. Vincent, and M. J. Collins, “Effect of patch size and network architecture on a convolutional neural network approach for automatic segmentation of OCT retinal layers,” Biomed. Opt. Express 9(7), 3049–3066 (2018). [CrossRef]  

30. S. K. Devalla, P. K. Renukanand, B. K. Speedhar, G. Subramanian, L. Zhang, S. Perera, J. Mari, K. S. Chin, T. A. Tun, N. G. Strouthidis, T. Aung, A. H. Thiery, and M. J. A. Griard, “DRUNET: a dilated-residual U-Net deep learning network to segment optic nerve head tissues in optical coherence tomography images,” Biomed. Opt. Express 9(7), 3244–3265 (2018). [CrossRef]  

31. P. Zang, J. Wang, T. T. Hormel, L. Liu, D. Huang, and Y. Jia, “Automated segmentation of peripapillary retinal boundaries in oct combining a convolutional neural network and a multi-weights graph search,” Biomed. Opt. Express 10(8), 4340–4352 (2019). [CrossRef]  

32. J. Kugelman, D. Alonso-Caneiro, S. A. Read, J. Hamwood, S. J. Vincent, F. K. Chen, and M. J. Collins, “Automatic choroidal segmentation in OCT images using supervised deep learning methods,” Sci. Rep. 9(1), 13298–13 (2019). [CrossRef]  

33. X. Sui, Y. Zheng, B. Wei, H. Bi, J. Wu, X. Pan, Y. Yin, and S. Zhang, “Choroid segmentation from optical coherence tomography with graph-edge weights learned from deep convolutional neural networks,” Neurocomputing 237, 332–341 (2017). [CrossRef]  

34. M. Treder, J. L. Lauermann, and N. Eter, “Automated detection of exudative age-related macular degeneration in spectral domain optical coherence tomography using deep learning,” Graefes Arch. Clin. Exp. Ophthalmol. 256(2), 259–265 (2018). [CrossRef]  

35. M. Shah, A. R. Ledo, and J. Rittscher, “Automated classification of normal and Stargardt disease optical coherence tomography images using deep learning,” Acta Ophthalmologica (2020).

36. K. K. Dansingani, K. K. Vupparaboina, S. T. Devarkonda, S. Jana, J. Chhablani, and K. B. Freund, “Amplitude-scan classification using artificial neural networks,” Sci. Rep. 8(1), 12451 (2018). [CrossRef]  

37. J. Loo, L. Fang, D. Cunefare, G. Jaffe, and S. Farsiu, “Deep longitudinal transfer learning-based automatic segmentation of photoreceptor ellipsoid zone defects on optical coherence tomography images of macular telangiectasia type 2,” Biomed. Opt. Express 9(6), 2681–2698 (2018). [CrossRef]  

38. O. Ronneberger, P. Fischer, and T. Brox, “U-Net: convolutional networks for biomedical image segmentation,” International Conference on Medical image computing and computer-assisted intervention, Springer, Cham, 234–241 (2015).

39. P. Krzyżanowska-Berkowska, K. Czajor, P. Syga, and D. R. Iskander, “Lamina cribrosa depth and shape in glaucoma suspects. Comparison to glaucoma patients and healthy controls,” Curr. Eye Res. 44(9), 1026–1033 (2019). [CrossRef]  

40. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” 2016 IEEE Conference on Computer Vision and Pattern Recognition,” IEEE (2016).

41. A. G. Roy, N. Navab, and C. Wachinger, “Concurrent spatial and channel ‘squeeze & excitation’in fully convolutional networks,” International conference on medical image computing and computer-assisted intervention, Springer, Cham, 421–429 (2018).

42. F. Visin, M. Ciccone, A. Romero, K. Kastner, K. Cho, Y. Bengio, M. Matteucci, and A. Courville, “Reseg: A recurrent neural network-based model for semantic segmentation,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2016).

43. J. Wang, L. C. Yu, K. R. Lai, and X. Zhang, “Dimensional sentiment analysis using a regional CNN-LSTM model,” Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, 2, 225–230 (2016).

44. M. Sonka, H. Vaclav, and B. Roger, “Image processing, analysis, and machine vision,” Cengage Learning (2014).

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (5)

Fig. 1.
Fig. 1. SLO image with the OCT scanning protocol (A), orange lines represent the position of the annotated scans, whereas blue lines represent the additional scans used in 3-D approaches. Pink lines refer to the position of scans that have not been manually annotated. Illustrative SD-OCT B-scan with overlaid coordinates of Bruch’s membrane opening. The termination points of Bruch’s membrane opening are BMO1 and BMO2 (B) and the corresponding ground truth mask (C).
Fig. 2.
Fig. 2. General overview of the proposed segmentation methods. Columns present (left to right): input type with respect to the whole image, method used (network architecture) and direct output of the model (prediction map). Rows present an overview of the operation of the method based on (top to bottom): a single A-scan, a set of columns (column based), and an entire B-scan (whole image).
Fig. 3.
Fig. 3. Overview of fully convolutional network architecture (U-Net) on the right side. The convolution block is described in the left side. In the figure, n refers to number of input image channels n = 1 in 2D methods and n=3 in 3-D methods. F is the number of filters.
Fig. 4.
Fig. 4. General overview of the proposed postprocessing. It presents B-scan (A), sample mask with not homogeneous BMO region (B) and a clean mask (C). Width of a gap between two BMO-regions in figure B is marked as gap width. Width of the second BMO region is annotated as temporary width and because it is larger than gap width, temporary BMO2 is moved to the end of the second column and it is marked as BMO2. An arrow indicates processing direction.
Fig. 5.
Fig. 5. Illustrative examples of original SLO image (left column) and the same image with overlaid BMO outlines (right column) segmented from the corresponding OCT volume consisting of 49 B-scans. Blue color indicates the BMO outline from manual annotation and the grey one, the BMO outline segmented by the proposed model.

Tables (5)

Tables Icon

Table 1. Mean values and the corresponding standard deviations of the similarity measures extracted from 3 independent runs of the different deep learning models and results from inter-grader annotations. The highest values (best performance) are shown in bold font.

Tables Icon

Table 2. Mean values and the corresponding standard deviations of the distance errors extracted from 3 independent runs of the different deep learning models and results from inter-grader annotations. The lowest values (best performance) are shown in bold font. All measure errors are computed in pixels.

Tables Icon

Table 3. Mean values and the corresponding standard deviations of the comparison between models with different architectures. The best values are shown in bold font. All distance errors are computed in pixels.

Tables Icon

Table 4. Result from one fine-tuned fully semantic model with U-Net like architecture.

Tables Icon

Table 5. Mean value and standard deviation (calculated across test subjects) of dice coefficient and unsigned border positioning error for BMO outline compared to the outline calculated with annotated scans.

Equations (3)

Equations on this page are rendered with MathJax. Learn more.

D S C = 2 T P 2 T P + F P + F N
J S C = T P T P + F P + F N
A C C = T P + T N T P + TN + F P + F N
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.