Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Deep learning for quality assessment of retinal OCT images

Open Access Open Access

Abstract

Optical coherence tomography (OCT) is a promising high-speed, non-invasive imaging modality providing high-resolution retinal scans. However, a variety of external factors such as light occlusion and patient movement can seriously degrade OCT image quality, which complicates manual retinopathy detection and computer-aided diagnosis. As such, this study first presents an OCT image quality assessment (OCT-IQA) system, capable of automatic classification based on signal completeness, location, and effectiveness. Four CNN architectures (VGG-16, Inception-V3, ResNet-18, and ResNet-50) from the ImageNet classification task were used to train the proposed OCT-IQA system via transfer learning. The ResNet-50 with the best performance was then integrated into the final OCT-IQA network. The usefulness of this approach was evaluated using retinopathy detection results. A retinopathy classification network was first trained by fine-tuning Inception-V3 model. The model was then applied to two test datasets, created randomly from the original dataset, one of which was screened by the OCT-IQA system and only included high quality images while the other was mixed by high and low quality images. Results showed that retinopathy detection accuracy and area under curve (AUC) were 3.75% and 1.56% higher, respectively, for the filtered data (compared with the unfiltered data). These experimental results demonstrate the effectiveness of the proposed OCT-IQA system and suggest that deep learning could be applied to the design of computer-aided systems (CADSs) for automatic retinopathy detection.

© 2019 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Optical coherence tomography (OCT), which can image retinal structures in vivo [1], has been widely applied in diagnostic ophthalmology due to its ease-of-use, lack of ionizing radiation, and high resolution [2]. There are approximately 30 million OCT procedures performed worldwide each year [3], with hundreds of consecutive B-scans comprising the majority of each procedure. This produces large quantities of data and limits the manual evaluation of individual images. Recent developments in computer-aided diagnostic systems (CADSs) have aided in retinopathy diagnosis and reduced the workload for clinicians. This has reduced processing times by accelerating image evaluation and improving diagnoses [4]. Previous studies have focused on screening retinal diseases using OCT images and an automatic CAD system [59]. Farsiu et al. assessed OCT scan quality by comparing results with manual evaluations performed by experts, excluding poor quality images from the dataset [5]. Wang et al. eliminated poor quality images during pre-processing and Kermany et al. removed half of the original OCT data in an initial quality review, which decreased the sample size [6,9]. Rasti et al. and Treder et al. used only images from public datasets that passed an image quality assessment (IQA) process. As is common, none of these studies used low-quality images in training the CADS. However, in practice, low quality retinal OCT images are inevitably produced due to insufficient contrast, lighting conditions, patient movement, and signal occlusion, which can cause signal void or blurring and prevent images from being suitable for diagnosis. As such, the use of an IQA in a CADS is critical for eliminating low quality images.

IQA methods can be divided into subjective and objective assessments. Subjective IQA methods were applied in the studies listed above, wherein experts manually eliminated unsatisfactory images [5,6,9]. Objective IQA methods can assist clinicians by automatically evaluating image quality, without requiring manual intervention. Several objective IQA algorithms have been proposed in recent years, including reference and blind IQA techniques. Reference methods primarily involve determining image parameters. For example, Ishikawa et al. [10] calculated the signal-to-noise ratio (SNR) and an image quality (IQ) metric to assess OCT scans. Stein et al. [11] introduced the quality index (QI), a new IQA parameter based on histogram information. Liu et al. [12] proposed signal deviation (SD), which considers the standard deviation of measured intensities, and Huang et al. [13] used a maximum tissue contrast index (mTCI) to quantify the image signal from multiple devices. Reference IQA can automatically and objectively eliminate unqualified images but robustness and accuracy suffer from a heavy dependence on image parameters (i.e., intensity, SNR, and signal strength). This can amplify parameter calculation errors in subsequent assessments. On the other hand, these image parameters mentioned above often fail to represent image quality. For example, an image with high SNR could be unacceptable for retinopathy diagnosis because of an off-centered signal, making the reference IQA unsuitable.

Blind IQA has become more generalizable with the development of machine learning and does not require additional information beyond the original data [14]. As such, it is commonly used for fundus images classification [1521] and has proven to be effective in eliminating low-quality images. Few studies have investigated the use of blind IQA methods based on OCT until 2019. Kauer et al. proposed an AQuANet to classify OCT B-scans into ‘good,’ ‘bad,’ ‘upper,’ and ‘lower’ categories using A-scans [22]. However, this study only investigated the position of the signal. Zhang et al. proposed a blind IQA architecture based on ResNet50. In this process, OCT images were partitioned into non-overlapping patches for preprocessing. ResNet50 was then adopted as a feature extractor and support vector regression was used to train the IQA model. However, this study primarily focused on assessing signal intensity [23]. These two blind OCT-IQA techniques neglected both signal effectiveness and completeness, which are critical for retinopathy detection.

In this study, we first develop a multi-class deep neural network for pre-filtering a retinopathy detection CADS. This approach pays attention to signal location, effectiveness and completeness, which can automatically assess the quality of retinal OCT images, including signal occlusion, signal centering, and the position of the region of interest. Deep neural networks are proficient in extracting image features, which become increasingly abstract as the layer depth increases. However, deep neural networks characteristically generate multiple parameters and require large quantities of data to avoid over-fitting during the training process. As such, we adopted transfer learning to develop an OCT-IQA network and avoid high computational costs. Four pre-trained CNN architectures (VGG, ResNet-18, ResNet-50, and Inception-V3) were implemented in the study, each of which won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC). We fine-tuned the learned networks for our specific classification task and adopted the highest performing ResNet-50. In addition, a separate retinopathy detection model was developed by fine-tuning the Inception-V3 to test the influence of unqualified images on further CADS and demonstrate the necessity of IQA. Two test datasets collected from the original dataset were adopted for retinopathy detection, one of which was first fed to the OCT-IQA network to eliminate unqualified images and produce a ‘pure’ dataset. The other was composed of acceptable and unacceptable images with a 1:1 ratio, and labeled the ‘mixed’ set. Test results showed that retinopathy detection accuracy and area under curve (AUC) were 3.75% and 1.56% higher, respectively, for the pure data (compared with the mixed data), which demonstrated that image quality is a vital element in automatic retinopathy detection.

2. Materials and methods

2.1 Database

We acquired 15,379 original retinal OCT images using a Zeiss Cirrus device (Carl Zeiss Meditec, Inc., Dublin, CA) from 710 persons of ages ranging from 47 to 85 years and stored them as JPG files at a resolution of 1536 × 1024 pixels. A minimum of 5 B-scans from this area were selected for each patient, to thoroughly observe the retinal macular area.

Data annotation was conducted after the images were collected. Two medical students manually screened the data and removed unclassifiable images (i.e. complete signal obstruction). As a result, 781 images were excluded in this first step. The remaining images were separately assessed by two specialists with more than five years of clinical experience. This evaluation included four quality categories: ‘good,’ ‘off-center,’ ‘signal-shielded,’ and ‘other,’ as well as two clinical categories: ‘abnormal’ and ‘normal.’ This reference standard complies with the situation detailed in Table 1, as 1063 images were eliminated due to a lack of consensus between the two specialists. Typical examples from each category are shown in Fig. 1 and representative ‘normal’ and ‘abnormal’ retina images are shown in Fig. 2.

 figure: Fig. 1.

Fig. 1. Examples of images with different qualities, including: (a) good, (b) off-center, (c) signal-shielded, and (d) other.

Download Full Size | PDF

 figure: Fig. 2.

Fig. 2. Examples of normal and abnormal images, including: (a) the normal retina, (b) central serous retinopathy (CSR), (c) a macular hole, and (d) diabetic macular edema (DME).

Download Full Size | PDF

Tables Icon

Table 1. A description of quality annotation standards.

Following annotation, the database distribution was as follows: good (11804, 56.27% abnormal and 43.72% normal), off-center (647, 17.16% unrecognizable anomaly, 34% abnormal, and 48.84% normal), signal-shielded (710, 36.06% unrecognizable anomaly, 27.04% abnormal, and 36.9% normal), and other (351, 100% unrecognizable anomaly). All poor quality images (including unrecognizable and recognizable anomaly) and 20% of the randomly selected high quality images were used to train the OCT-IQA classifier to make the image numbers of high and low quality balanced. The included dataset exhibited an imbalance problem as fewer poor quality images (including ‘off-center,’ ‘signal-shielded,’ and ‘other’ categories) were available compared with high quality. This is primarily due to a limited sample size. This uneven sample distribution can have a significant influence on classification results and the number of poor-quality images is too small to train a model without overfitting. As such, data augmentation was applied to resolve these issues.

2.2 Data augmentation

Data augmentation is used to produce new synthetic samples from simple transformations of original images [24]. This study included few poor quality images, which were augmented to compensate for low quantities. Horizontal mirroring, rotations through random angles (${\pm} $ 10 degrees), and a contrast enhancement defined by:

$$I = \log 2(1 + {\textrm {v}}\ast {\textrm{i}})/ \log 2({\textrm{v}} + 1)$$
were applied. Here, I denotes the destination image, i is the source image, and v (=10 in this study) indicates the degree of contrast enhancement. Before augmentation, one hundred images were randomly selected from the original dataset for each category to test the model, the remaining samples were randomly divided into training and validation sets (scans from the same patients were kept in the same set). Augmentation was then applied to each, to produce sufficient samples while preserving the fraction of images in each category (i.e., fewer samples in the ‘other’ class). The final data distribution is shown in Table 2.

Tables Icon

Table 2. Datasets used for training, evaluating, and testing the model

2.3 Convolutional neural network (CNN)

CNNs are a common framework used for deep learning. VGG, ResNet, and Inception are typical CNNs which have been successfully applied to medical image classification problems such as skin cancer classification [25], early diagnosis of Alzheimer’s disease [26], and retinal vessel detection in fundus images [27].

A CNN is typically composed of convolutional, pooling, and fully connected layers. Convolutional layers are used to calculate convolutions between specific kernels and the input data. An activation function is then applied to produce a new feature map. The convolutional operation for a single channel can be expressed as:

$$o_{(x,y)}^{(z + 1)} = \sum\limits_{w = 0}^{{W_{\textrm{z - 1}}}} {\sum\limits_{h = 0}^{{H_{z - 1}}} {\sum\limits_{s = 0}^{{S_{z - 1}}} {k_{w,h,s}^z\cdot i_{x + w,y + h,s}^z} } } ,$$
where k is the convolutional kernel and W, H, and S represent the dimensions of k. The kernel convolves the input ${i^z}$  along its dimensions of width and height to produce the output ${o^{z + 1}}$ at the location $x,y$. The nonlinear activation function F is given by:
$${i^{z + 1}} = F({o^{z + 1}}).$$
The feature map is then sent to the pooling layer for feature selection and information filtering. Finally, every output unit is connected to all units of the previous feature map in the fully connected layer. VGG-16, Inception-V3, and ResNet were adopted in this study, as described below.

VGG-16 [28]: a plain CNN network consisting of 13 convolutional layers, 5 max-pooling layers, and 3 fully connected layers. All convolutional layers in VGG-16 featured a small 3 × 3 kernel. The max-pooling layer included a 2 × 2 kernel to decrease the number of parameters.

Inception-V3 [29]: a deeper network consisting of 22 layers and fewer parameters. An ‘Inception’ module was utilized in this network architecture, which factored 3 × 3 convolutions into two smaller convolutions, such as 1 × 3 and 3 × 1. This was beneficial for reducing parameters, accelerating calculations, and preventing over-fitting.

ResNet (ResNet-18, ResNet-50) [30]: a residual block with a skip connection bypass to resolve the difficulties of training deeper networks, allowing for higher accuracy. This architecture exhibits varying depths of 18, 50, 101, and 152 layers. In this study, we evaluated the performance of ResNet-18 and ResNet-50 because the residual block architecture in ResNet-50 has made some adjustments to improve performance for ImageNet tasks, as shown in Fig. 3. This performance was then tested separately using our dataset. ResNet-18 and ResNet-50 both consist of six modules denoted by conv1, conv2 x, conv3 x, conv4 x, conv5 x, and FC. Conv1 is a convolutional layer. Conv2 x, conv3 x, conv4 x, and conv5 x include residual blocks numbered as 2, 2, 2, and 2 in ResNet-18 and 3, 4, 6, and 3 in ResNet-50, respectively. FC is the fully connected layer.

 figure: Fig. 3.

Fig. 3. Residual blocks for (a) ResNet-18 and (b) ResNet-50

Download Full Size | PDF

2.4. A transfer learning framework for OCT-IQA

Deep CNNs are invariant to shifts, rotations, and scaling. They also include large quantities of trainable parameters, providing them with a high capacity for generalization. However, large amounts of labeled data are necessary to avoid under-fitting and over-fitting, which is a challenge with medical images. Furthermore, adjustments to the training parameters are required to improve the convergence.

Transfer learning is a useful method for resolving these issues, having been used in a variety of fields [9,3132]. The basic idea of transfer learning is that features extracted by the pre-trained model can be reused in specific classification tasks. The architecture proposed in this study, which consists of an OCT-IQA network and a retinopathy network, was trained using transfer learning (see Fig. 4). The workflow for this architecture is described as follows.

  • 1) The OCT-IQA network model was trained to remove unqualified images, including off-center, signal-shielded, and other quality abnormalities. Four CNNs introduced previously I(VGG, ResNet-18, ResNet-50, and Inception-V3) were fine-tuned to identify the most suitable network for OCT-IQA in two steps. The first step was network initialization where convolutional layers in each CNN were initialized by loaded weights, pre-trained on the ImageNet dataset by the corresponding network [33]. The second step included redesigning new top layers for the network because these architectures were originally designed for 1000-category classification task. A final layer and a softmax layer were then added to the pre-trained networks and retrained to recognize specific classes. In this study, images were classified into four categories during OCT-IQA. A testing dataset was randomly selected to evaluate four fine-tuned networks and the architecture with the best performance was selected.
  • 2) The OCT-IQA model was used to classify the entire dataset into unqualified images, where images needing to be retaken, and acceptable images.
  • 3) Images with serious quality issues were removed. The remaining samples, which had been previously labeled by ophthalmologists, were used to train the retinopathy detection network to classify acceptable images into normal and abnormal categories.

 figure: Fig. 4.

Fig. 4. The transfer learning architecture for OCT-IQA (optical coherence tomography image quality assessment) and retinopathy detection.

Download Full Size | PDF

3. Experimental results

3.1. Experimental setup

Four CNNs were fine-tuned during the OCT-IQA process. To reduce computational complexity, input images were resized to 299 × 299 × 3 for Inception-V3 and to 224 × 224 × 3 for VGG, ResNet-18 and ResNet-50. A learning rate of 0.001, stochastic gradient decent [34] with Nesterov momentum 0.9 was applied to fine-tune the four CNNs. The training process was completed when the cross-entropy cost function and the accuracy converged. Corresponding training and loss curves are shown in Fig. 5. This result required 18 epochs for ResNet18 and 24 epochs for the other architectures with batch sizes of 32. The training data were directly fed to the network with pre-trained weight parameters and processed using a desktop computer with an Intel Xeon E5-2620 CPU, 32 GB RAM, and two NVIDIA GeForce GTX 1080Ti GPUs. The cross entropy loss can be expressed as:

$$L ={-} \sum\limits_i {{{g^{\prime}}_i}} \log ({o_i}),$$
where ${g^{\prime}_i}$ is the ground truth and ${o_i}$ is the predicted output.

 figure: Fig. 5.

Fig. 5. Accuracy and loss curves for (a) VGG16, (b) ResNet-18, (c) ResNet-50, and (d) Inception-V3

Download Full Size | PDF

3.2 Experimental results for OCT-IQA

Images labelled as ‘off-center,’ ‘signal-shielded,’ or ‘other’ were considered unqualified, while those with high quality were classified as acceptable. The OCT-IQA classifier was trained using five-fold cross-validation on the training and validation sets, as demonstrated by Table 2. The data were divided into five separate, equally stratified sets. Each subset was used to evaluate the model while the remaining four subsets were used to train the model. The final assessment was made using the test set. Following this five-fold cross-validation, the area under the receiver operating characteristic curve (AUC) was calculated and used to determine the model’s ability to distinguish poor quality images (off-center, signal-shielded, or other) from good quality images. The model with the highest AUC score was selected as the final model and was evaluated using the test set. The AUCs of various architectures used for five-fold validation are shown in Fig. 6. The mean (SD) AUCs for VGG-16, ResNet-18, ResNet-50, and Inception-V3 were 0.99122 (0.0023), 0.9888 (0.0018), 0.9932 (0.0026), and 0.9983 (0.0008), respectively. These results indicate that the standard deviation (SD) of AUC values were small (max: 0.0026, min: 0.0008) for each model, which demonstrate their stability and robustness.

 figure: Fig. 6.

Fig. 6. ROC curves and AUC values for five-fold cross-validation experiments on VGG16, ResNet-18, ResNet-50 and Inception-V3.

Download Full Size | PDF

The highest AUCs model of each architecture were tested using the test dataset. The specificity (SP), recall (Re), and precision (PR) were calculated for the independent classes to assess model performance. In this four-category classification problem, true positive and true negative values were determined by correctly dividing the test images into positive and negative classes. The positive class was the appointed class and the negative class included any samples not in the selected class. False positives and false negatives were determined by incorrectly dividing the test images into negative and positive classes, respectively. The recall indicates the proportion of true positives in the sum to the number of true positives and false negatives. SP is the proportion of true negatives in the sum to the number of true negatives and false positives. PR is defined as the proportion of true positives in the sum to true positives and false positives. In addition, the overall accuracy (OA), overall recall (OR), overall specificity (OS), overall precision (OP), and F1 score (F1) were included, as shown in Table 3. These parameters were calculated using Eqs. (5)–(8), where TP, TN, FP, and FN represent true positive, true negative, false positive, and false negative. Evaluation metrics were defined as follows:

$$Accuracy = \frac{{TP + TN}}{{TP + TN + FP + FN}},$$
$$Specificity = \frac{{TN}}{{FP + TN}},$$
$$Recall = \frac{{TP}}{{TP + FP}},$$
$$Precision = \frac{{TP}}{{TP + FP}}.$$
These results suggest that networks with deeper layers (e.g., ResNet-18, ResNet-50, and Inception-V3) and improved architectures (e.g., residual and inception modules) outperformed networks with basic structures (VGG16). Specifically, ResNet-50 achieved better results for ‘good’ and ‘other’ cases, though its performance was not ideal for ‘signal-shielded’ or ‘off-center’ images. In contrast, Inception-V3 and ResNet-18 achieved better results in these two categories, suggesting these networks may focus on different features. ResNet-50 achieved the highest OA, OR, OS, and F1 scores, while its OP was slightly lower than that of ResNet-18. ROC curves were also used to determine the most suitable architecture for OCT-IQA. These curves plot the true positive rate against the false positive rate, which represents the overall classification abilities of architectures across four classes. Values closer to the upper-left corner indicate better performance. Figure 7 suggests that ResNet-50 performed better than the other four models, achieving the highest AUC (0.9947).

Tables Icon

Table 3. Accuracy, recall, specificity, precision and F1 score for different network architectures used for classification of image quality into “good,” “off-center,” “signal-shielded,” and “other.” The best option within each metric is in bold.

As such, we selected the ResNet-50 model for OCT-IQA, with an OA of 96.25%, an OR of 97%, an OS of 98.98%, and an AUC of 99.47%.

 figure: Fig. 7.

Fig. 7. A comparison of receiver operating characteristic (ROC) curves and AUC values for poor quality images (defined as off-center, signal-shielded, and other) using (a) VGG16 (AUC = 0.9896), (b) ResNet-18 (AUC = 0.9865), (c) ResNet-50 (AUC = 0.9947), and (d) Inception-V3 (AUC = 0.9898)

Download Full Size | PDF

Heat maps were developed during the training phase to determine which regions of an image were being paid the most attention. Figure 8 shows a heat map generated from the retrained ResNet-50 model. The images in Fig. 8(a) and 8(b) were classified as off-center, with regions above and below the retina highlighted. The image in Fig. 8(c) was classified as signal-shielded and the regions with a signal hiatus were highlighted. The image in Fig. 8(d) was classified as ‘other’ and the region around the optic disk was highlighted. These heat map results demonstrate that key features used in OCT-IQA (i.e., signal location and signal appearance) were successfully learned by the network.

 figure: Fig. 8.

Fig. 8. Image quality heat maps showing (a) (b) off-center, (c) signal-shielded, and (d) ‘other’ categories.

Download Full Size | PDF

3.3 Results of retinopathy detection

A second subnetwork was trained to demonstrate the necessity of IQA in retinopathy detection and construct a more robust retinopathy detection model. Inception-V3 has been shown to perform well in retinopathy classification (i.e., drusen, CNV, and DME [9]). As such, we fine-tuned this architecture to train the retinopathy detection model, which was designed to classify lesion images into ‘normal’ and ‘abnormal’ categories. Normal indicated the retina was healthy and abnormal implied the retina suffered from a retinopathy (e.g., drusen, macular edema, neovascularization, etc.), causing retinal distortion. The retinopathy detection dataset was established by eliminating all images belonging to the ‘other’ category and some images in the signal-shielded and off-center categories, since retinopathies in these images could not be recognized during anomaly grading. Afterward, 12,794 images consisting of 7,096 abnormal (6,644 good, 220 off-center, and 192 signal-shielded) and 5,738 normal (5,160 good, 316 off-center, and 262 signal-shielded) samples were selected from 13,535 images. The dataset was separated into training, validation, and testing sets in a process similar to that of the network described above. Only the training data were augmented using contrast enhancement (defined by Eq.(1)) and horizontal flipping. The final dataset consisted of 23,602 images including 22,530 training, 472 validation, and 600 test samples (200 poor and 400 good quality). The experimental setting (i.e. optimizer, batch size and learning rate) was the same as Inception-V3 used in the first sub-network.

Five-fold cross-validation was applied to select hyper-parameters and confirm the robustness and stability of the Inception-V3 model, the training and validation sets were involved in. Results showed slight variability in the performance of the network with a mean AUC of 99.90%. The model with the highest AUC score (99.96%) was selected as the final model. Corresponding five-fold ROC curves are shown in Fig. 9. We have also tested the generalizability of proposed model by public dataset in study [9], which captured from Spectralis OCT (Heidelberg Engineering, Germany) and had different signal distribution and image size from ours. The public dataset has classified images into four categories: choroidal neovascularization, diabetic macular edema, drusen and normal. We tested our retinopathy detection model on their open test dataset (1000 images, 250 from each category). As our model is a two-class model, we put the choroidal neovascularization, diabetic macular edema, drusen together to form the abnormal category, and images of normal form the normal category. Finally, excellent results of 99.5% AUC and 99.87% sensitivity were obtained, which proved that the proposed model can be generalized.

 figure: Fig. 9.

Fig. 9. Five-fold cross-validation receiver operating characteristic (ROC) curves used to differentiate normal and abnormal retinas.

Download Full Size | PDF

The effect of unqualified images on retinopathy detection was evaluated using ‘pure’ and ‘mixed’ test datasets. The dataset was fed into the OCT-IQA first to get quality label. Then the ‘pure’ set consisted only of images in good quality (400 images, 200 from each category), the mixed dataset consisted of 100 poor-quality images (50 off-center and 50 signal-shielded) and 100 acceptable-quality images (400 images). Corresponding test results are shown in Table 4.

Tables Icon

Table 4. Test results for the retinopathy detection model with pure and mixed datasets.

As seen in the table, retinopathy detection accuracy for the pure dataset was 97.5%, higher than that of the mixed dataset at 93.75%. The sensitivity and specificity values were also higher in the pure dataset. ROC (Fig. 10) and confusion matrices (Fig. 11) were generated to evaluate model performance. The resulting AUC values were 99.74% and 98.18% for the pure and mixed sets, respectively, suggesting that model classification performance was compromised by unqualified images. The confusion matrices imply that the true positives and true negatives for the pure dataset improved significantly.

 figure: Fig. 10.

Fig. 10. AUC for retinopathy detection of (a) pure and (b) mixed datasets (right).

Download Full Size | PDF

 figure: Fig. 11.

Fig. 11. Confusion matrices for (a) pure and (b) mixed retinopathy detection datasets.

Download Full Size | PDF

Heat maps were also generated during the test to determine whether lesions were discovered by the network. The images shown in Fig. 12(a) and 12(b) are of good quality, 12(c) is off-center, and 12(d) is signal-shielded. Drusens were highlighted in Fig. 12(a) and retina thickening and pigment epithelial detachment are evident in Fig. 12(b). Figure 12(c) demonstrates macular edema and a retinal atrophy is visible in 12(d). It is evident that the network classified the retina as either normal or abnormal based on the lesion. The network was able to form the correct conclusion, even for poor quality images, as long as the lesion could be observed. As such, lesion visibility is critical in retinopathy detection. If the retina structure or lesion is severely shielded, the model cannot acquire sufficient information to draw accurate conclusions.

 figure: Fig. 12.

Fig. 12. Heat maps for a retinopathy classifier showing: (a) a drusen structure, (b) thickening of the retina and pigment epithelial detachments (PED), (c) macular edema, and (d) retinal atrophy.

Download Full Size | PDF

These experimental results indicate that unqualified images can negatively affect retinopathy detection. Poor quality images often exhibit signal-shielding or signal deficiency, which can cause difficulties for identifying lesions. OCT-IQA can detect unqualified images in real-time and recapture them after they have been eliminated. This reduces processing time and is useful for improving the accuracy of retinopathy detection.

4. Conclusion and discussion

In this study, an architecture based on a deep CNN was proposed to classify the quality of OCT images and identify the existence of retinopathies. The system consisted of an OCT-IQA classifier and a retinopathy detector. Several extant CNNs, including VGG, ResNet-18, ResNet-50, and Inception-V3 were fine-tuned by replacing the final layers and retraining on our dataset to determine the proper CNN for the proposed architecture. ResNet-50 achieved the best performance in classifying OCT image quality, with an overall accuracy of 96.25%, an OR of 97.33%, and an AUC of 99.47%. We confirmed the influence of image quality on retinopathy detection by constructing a separate model, which adopted the Inception-V3 architecture and was tested using ‘pure’ and ‘mixed’ datasets. The pure dataset was classified by OCT-IQA and only included images with good quality. The mixed dataset was mixed with high quality images and unqualified images. The model achieved better classification results with the dataset that passed OCT-IQA, where testing accuracy was 3.75% higher for the pure dataset than the mixed dataset. In summary, a trained IQA model was used to reduce non-beneficial images during CADS preprocessing. In addition, a CNN model was developed and tested, achieving excellent performance for retinopathy detection. This suggests that the models discussed above could be advantageous in the design of a computer-aided diagnostic system (CADS) for automatic lesion detection.

Furthermore, the proposed OCT-IQA model demonstrate ability to classify images into four specified quality categories, which can be further applied to hardware control system to assist the automatic reacquisition of poor-quality images in the future.

Funding

National Basic Research Program of China (973 Program) (2017YFC0108201, 2016YFE0107700); Jiangsu Provincial Key Research and Development Program (BE2019682, BE2018667); National Natural Science Foundation of China ( 61605210, 61675226); Youth Innovation Promotion Association of the Chinese Academy of Sciences (2019320).

Acknowledgment

We would like to thank LetPub (www.letpub.com) for providing linguistic assistance during the preparation of this manuscript.

Disclosures

The authors declare that there are no conflicts of interest related to this article.

References

1. M. E. Velthoven, D. J. Faber, F. D. Verbraak, T. G. van Leeuwen, and M. D. de Smet, “Recent developments in optical coherence tomography for imaging the retina,” Prog. Retinal Eye Res. 26(1), 57–77 (2007). [CrossRef]  

2. A. Lang, A. Carass, M. Hauser, E. S. Sotirchos, P. A. Calabresi, H. S. Ying, and J. L. Prince, “Retinal layer segmentation of macular OCT images using boundary classification,” Biomed. Opt. Express 4(7), 1133–1152 (2013). [CrossRef]  

3. E. A. Swanson and J. G. Fujimoto, “The ecosystem that powered the translation of OCT from fundamental research to clinical and commercial impact,” Biomed. Opt. Express 8(3), 1638–1664 (2017). [CrossRef]  

4. M. R. K. Mookiah, U. R. Acharya, C. K. Chua, C. M. Lim, E. Y. K. Ng, and A. Laude, “Computer-aided diagnosis of diabetic retinopathy: A review,” Comput. Biol. Med. 43(12), 2136–2155 (2013). [CrossRef]  

5. S. Farsiu, S. J. Chiu, R. V. O’ Connell, F. A. Folgar, E. Yuan, J. A. Izatt, and C. A. Toth, “Quantitative classification of eyes with and without intermediate age-related macular degeneration using optical coherence tomography,” Ophthalmology 121(1), 162–172 (2014). [CrossRef]  

6. Y. Wang, Y. Zhang, Z. Yao, R. Zhao, and F. Zhou, “Machine learning based detection of age-related macular degeneration (AMD) and diabetic macular edema (DME) from optical coherence tomography (OCT) images,” Biomed. Opt. Express 7(12), 4928–4940 (2016). [CrossRef]  

7. R. Rasti, H. Rabbani, A. Mehridehnavi, and F. Hajizadeh, “Macular OCT classification using a multi-scale convolutional neural network ensemble,” IEEE Trans. Med. Imaging 37(4), 1024–1034 (2018). [CrossRef]  

8. M. Treder, J. L. Lauermann, and N. Eter, “Automated detection of exudative age-related macular degeneration in spectral domain optical coherence tomography using deep learning,” Graefe’s Arch. Clin. Exp. Ophthalmol. 256(2), 259–265 (2018). [CrossRef]  

9. D. S. Kermany, M. Goldbaum, W. Cai, C. C. S. Valentim, H. Liang, S. L. Baxter, A. McKeown, G. Yang, X. Wu, F. Yan, J. Dong, M. K. Prasadha, J. Pei, M. Y. L. Ting, J. Zhu, C. Li, S. Hewett, J. Dong, I. Ziyar, A. Shi, R. Zhang, L. Zheng, R. Hou, W. Shi, X. Fu, Y. Duan, V. A. N. Huu, C. Wen, E. D. Zhang, C. L. Zhang, O. Li, X. Wang, M. A. Singer, X. Sun, J. Xu, A. Tafreshi, M. Anthony Lewis, H. Xia, and K. Zhang, “Identifying medical diagnoses and treatable diseases by image-based deep learning,” Cell 172(5), 1122–1131.e9 (2018). [CrossRef]  

10. H. Ishikawa, G. Wollstein, M. Aoyama, D. Stein, S. Beaton, J. G. Fujimoto, and J. S. Schuman, “Stratus OCT image quality assessment,” Invest. Ophthalmol. Visual Sci. 45(13), 3317 (2004).

11. D. M. Stein, H. Ishikawa, R. Hariprasad, G. Wollstein, R. J. Noecker, J. G. Fujimoto, and J. S. Schuman, “A new quality assessment parameter for optical coherence tomography,” Br. J. Ophthalmol. 90(2), 186–190 (2006). [CrossRef]  

12. S. Liu, A. S. Paranjape, B. Elmaanaoui, J. Dewelle, H. G. Rylander, M. K. Markey, and T. E. Milner, “Quality assessment for spectral domain optical coherence tomography (OCT) images,” In Multimodal Biomedical Imaging IV (Vol. 7171, p. 71710X). International Society for Optics and Photonics (2009).

13. Y. Huang, S. Gangaputra, K. E. Lee, A. R. Narkar, R. Klein, B. E. K. Klein, S. M. Meuer, and R. P. Danis, “Signal Quality Assessment of Retinal Optical Coherence Tomography Images. Signal quality assessment of retinal optical coherence tomography images,” Invest. Ophthalmol. Visual Sci. 53(4), 2133–2141 (2012). [CrossRef]  

14. X. Gao, F. Gao, D. Tao, and X. Li, “Universal blind image quality assessment metrics via natural scene statistics and multiple kernel learning,” IEEE Trans. Neural Netw. Learning Syst. 24(12), 2013–2026 (2013). [CrossRef]  

15. K. Gu, G. Zhai, X. Yang, and W. Zhang, “Deep learning network for blind image quality assessment,” In 2014 IEEE International Conference on Image Processing (ICIP), (2014, October), pp. 511–515. IEEE.

16. R. Tennakoon, D. Mahapatra, P. Roy, S. Sedai, and R. Garnavi, “Image Quality Classification for DR Screening Using Convolutional Neural Networks,” In: X. Chen, M. K. Garvin, J. Liu, E. Trucco, and Y. Xu, eds. Proceedings of the Ophthalmic Medical Image Analysis Third International Workshop, OMIA 2016, Held in Conjunction with MICCAI 2016, Athens, Greece, (October 21, 2016), 113–120.

17. S. K. Saha, B. Fernando, J. Cuadros, D. Xiao, and Y. Kanagasingam, “Deep Learning for Automated Quality Assessment of Color Fundus Images in Diabetic Retinopathy Screening,” arXiv preprint arXiv:1703.02511. (2017).

18. J. Sun, C. Wan, J. Cheng, F. Yu, and J. Liu, “Retinal image quality classification using fine-tuned CNN,” In Fetal, Infant and Ophthalmic Medical Image Analysis, (2017), pp. 126–133. Springer, Cham.

19. F. Yu, J. Sun, A. Li, J. Cheng, C. Wan, and J. Liu, “Image quality classification for DR screening using deep learning,” In 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) (IEEE, 2017), pp. 664–667.

20. Y. Zhang, L. Wang, Z. Wu, J. Zeng, Y. Chen, R. Tian, J. Zhao, and G. Zhang, “Development of an Automated Screening System for Retinopathy of Prematurity Using a Deep Neural Network for Wide-angle Retinal Images,” IEEE Access 7, 10232–10241 (2019). [CrossRef]  

21. G. T. Zago, R. V. Andreão, B. Dorizzi, and E. O. T. Salles, “Retinal image quality assessment using deep learning,” Comput. Biol. Med. 103, 64–70 (2018). [CrossRef]  

22. J. Kauer, K. Gawlik, H. G. Zimmermann, E. M. Kadas, C. Bereuter, and F. Paul, … & I. E. Beckers, “Automatic quality evaluation as assessment standard for optical coherence tomography,” In Advanced Biomedical and Clinical Diagnostic and Surgical Guidance Systems XVII, (2019, February), Vol. 10868, p. 1086814. International Society for Optics and Photonics.

23. M. Zhang, J. Y. Wang, L. Zhang, J. Feng, and Y. Lv, “Deep residual-network-based quality assessment for SD-OCT retinal images: preliminary study,” In Medical Imaging 2019: Image Perception, Observer Performance, and Technology Assessment, (2019, March), Vol. 10952, p. 1095214. International Society for Optics and Photonics.

24. A. Courville, I. Goodfellow, and Y. Bengio, “Deep Learning Book,” Deep learning, 21(1), 111–124. arXiv:arXiv:1011. (2015).

25. A. Esteva, B. Kuprel, R. A. Novoa, J. Ko, S. M. Swetter, H. M. Blau, and S. Thrun, “Dermatologist-level classification of skin cancer with deep neural networks,” Nature 542(7639), 115–118 (2017). [CrossRef]  

26. S. Liu, W. Cai, S. Pujol, R. Kikinis, and D. Feng, “Early Diagnosis of Alzheimer’s Disease with Deep Learning,” in IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI), Beijing, (2014), pp. 1015–1018.

27. D. Maji, A. Santara, P. Mitra, and D. Sheet, “Ensemble of deep convolutional neural networks for learning to detect retinal vessels in fundus images,” arXiv preprint arXiv:1603.04833. (2016).

28. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556. (2014).

29. C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” In Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), pp. 2818–2826.

30. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” In Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), pp. 770–778.

31. S. P. K. Karri, D. Chakraborty, and J. Chatterjee, “Transfer learning based classification of optical coherence tomography images with diabetic macular edema and dry age-related macular degeneration,” Biomed. Opt. Express 8(2), 579–592 (2017). [CrossRef]  

32. P. M. Burlina, N. Joshi, M. Pekala, K. D. Pacheco, D. E. Freund, and N. M. Bressler, “Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks,” JAMA Ophthalmol. 135(11), 1170–1176 (2017). [CrossRef]  

33. J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” IEEE conference on computer vision and pattern recognition (2009). pp. 248–255.

34. L. Bottou, “Large-scale machine learning with stochastic gradient descent,” In Proceedings of COMPSTAT”2010, (2009, June), pp. 177–186. Physica-Verlag HD.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (12)

Fig. 1.
Fig. 1. Examples of images with different qualities, including: (a) good, (b) off-center, (c) signal-shielded, and (d) other.
Fig. 2.
Fig. 2. Examples of normal and abnormal images, including: (a) the normal retina, (b) central serous retinopathy (CSR), (c) a macular hole, and (d) diabetic macular edema (DME).
Fig. 3.
Fig. 3. Residual blocks for (a) ResNet-18 and (b) ResNet-50
Fig. 4.
Fig. 4. The transfer learning architecture for OCT-IQA (optical coherence tomography image quality assessment) and retinopathy detection.
Fig. 5.
Fig. 5. Accuracy and loss curves for (a) VGG16, (b) ResNet-18, (c) ResNet-50, and (d) Inception-V3
Fig. 6.
Fig. 6. ROC curves and AUC values for five-fold cross-validation experiments on VGG16, ResNet-18, ResNet-50 and Inception-V3.
Fig. 7.
Fig. 7. A comparison of receiver operating characteristic (ROC) curves and AUC values for poor quality images (defined as off-center, signal-shielded, and other) using (a) VGG16 (AUC = 0.9896), (b) ResNet-18 (AUC = 0.9865), (c) ResNet-50 (AUC = 0.9947), and (d) Inception-V3 (AUC = 0.9898)
Fig. 8.
Fig. 8. Image quality heat maps showing (a) (b) off-center, (c) signal-shielded, and (d) ‘other’ categories.
Fig. 9.
Fig. 9. Five-fold cross-validation receiver operating characteristic (ROC) curves used to differentiate normal and abnormal retinas.
Fig. 10.
Fig. 10. AUC for retinopathy detection of (a) pure and (b) mixed datasets (right).
Fig. 11.
Fig. 11. Confusion matrices for (a) pure and (b) mixed retinopathy detection datasets.
Fig. 12.
Fig. 12. Heat maps for a retinopathy classifier showing: (a) a drusen structure, (b) thickening of the retina and pigment epithelial detachments (PED), (c) macular edema, and (d) retinal atrophy.

Tables (4)

Tables Icon

Table 1. A description of quality annotation standards.

Tables Icon

Table 2. Datasets used for training, evaluating, and testing the model

Tables Icon

Table 3. Accuracy, recall, specificity, precision and F1 score for different network architectures used for classification of image quality into “good,” “off-center,” “signal-shielded,” and “other.” The best option within each metric is in bold.

Tables Icon

Table 4. Test results for the retinopathy detection model with pure and mixed datasets.

Equations (8)

Equations on this page are rendered with MathJax. Learn more.

I = log 2 ( 1 + v i ) / log 2 ( v + 1 )
o ( x , y ) ( z + 1 ) = w = 0 W z - 1 h = 0 H z 1 s = 0 S z 1 k w , h , s z i x + w , y + h , s z ,
i z + 1 = F ( o z + 1 ) .
L = i g i log ( o i ) ,
A c c u r a c y = T P + T N T P + T N + F P + F N ,
S p e c i f i c i t y = T N F P + T N ,
R e c a l l = T P T P + F P ,
P r e c i s i o n = T P T P + F P .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.