Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Digital resolution enhancement in low transverse sampling optical coherence tomography angiography using deep learning

Open Access Open Access

Abstract

Optical coherence tomography angiography (OCTA) requires high transverse sampling density for visualizing retinal and choroidal capillaries. Low transverse sampling causes digital resolution degradation, such as the angiograms in wide-field OCTA. In this paper, we propose to address this problem using deep learning. We conducted extensive experiments on converting the centrally cropped 3 × 3 mm2 field of view (FOV) of the 8 × 8 mm2 foveal OCTA images (a sampling density of 22.9 µm) to the native 3 × 3 mm2 en face OCTA images (a sampling density of 12.2 µm). We employed a cycle-consistent adversarial network architecture in this conversion. The quantitative analysis using the perceptual similarity measures shows the generated OCTA images are closer to the native 3 × 3 mm2 scans. Besides, the results show the proposed method could also enhance the signal-to-noise ratio. We further applied our method to enhance diseased cases and calculate vascular biomarkers, which demonstrates its generalization performance and clinical perspective.

© 2020 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

In recent years, optical coherence tomography angiography (OCTA) has drawn tremendous attention in ophthalmology as a novel imaging modality to complement traditional fluorescein angiography (FA) and indocyanine green angiography (ICGA) [1]. OCTA is non-invasive thus avoids the risk of dye. Compared with the overlapping depth information of vasculature in 2D imaging like FA and ICGA, OCTA is capable of resolving the vessels and capillaries in-depth direction with a high axial resolution of $5\sim 10$ $\mu$m. It has been widely used in studies of various ocular diseases, such as glaucoma [2], age-related macular degradation [3], and retinopathy of prematurity [4]. Also, the OCTA imaging of human retina was found to be able to indicate neurodegenerative disorders, such as mild cognitive impairment [5] and Alzheimer’s disease [6].

OCTA employs the temporal decorrelation of repeated B-scans at the same location to distinguish the blood flow from surrounding tissue, so it usually requires several repeated volumetric scans to achieve en face OCTA images [7]. Besides, for minimizing motion noise and artifacts and enhancing the signal-to-noise ratio (SNR) of blood vessels, hardware-based eye-tracking [8] and software-based orthogonal registration [9] further increase the acquisition duration time of OCTA. So for a given field of view (FOV) and A-line rate, OCTA scans spend a much longer time than volumetric OCT scans.

On the other hand, different from OCT that emphasizes the structure information of retinal layers, an important function of OCTA is to visualize en face retinal and choroidal capillaries. So OCTA has higher demands on the transverse resolution and sampling density of the acquisition systems. However, the imaging of the posterior segment suffers from the intrinsic aberrations of the eye, which limits the transverse resolution of the ocular imaging systems to $10\sim 15$ $\mu$m. Incorporating with adaptive optics (AO) could compensate the aberrations and improve the transverse resolution [10], but the AO components will significantly increase the cost and complexity of the acquisition systems. Besides, the resolution is improving in sacrifice of the FOV, which hinders the clinical applications of the AO OCTA.

Enabling a high transverse sampling could avoid the loss of capillary details, namely, the degradation of digital resolution induced by undersampling. However, it will further increase the acquisition time on the basis of the repeated scans. To avoid the discomfort of the imaging subject caused by a long acquisition duration (the human spontaneous eye blink rate is $\sim 15$ blinks per minute [11]), researchers and manufacturers are trending to use high-speed systems for the OCTA imaging. In research fields, the acquisition A-line rate has reached several megahertz [12]. For commercial systems, 100-kHz swept-source (SS) OCT systems have emerged. However, the popularization of SS-OCT systems was significantly impeded by the limited wavelength regime and availability of sources with a broad bandwidth at the moment.

Fig. 1(a) give a simple estimation of the relationship between the sampling density and the A-line speed of the OCT systems. It was derived based on several practical assumptions. (1) An acquisition duration of 4 seconds, which is the upper limit time between two successive spontaneous blinks. (2) A FOV of $3\times 3$ mm$^2$, which is commonly used in OCTA scans. (3) Two repeated scans, which is the minimum for the OCTA algorithms. (4) Evenly-spaced sampling along the horizontal and vertical directions, which is mostly employed. For the sampling density at the Nyquist–Shannon limit (7.5 $\mu$m for an optical transverse resolution of $\sim 15$ $\mu$m), an A-line speed of $\sim 160$ kHz is required. We also listed the sampling density of the $3\times 3$ mm$^2$ and $8\times 8$ mm$^2$ scan protocols in the ZEISS CIRRUS OCTA system. For the $3\times 3$ mm$^2$ scans, each transverse direction has 245 data points which corresponds to a sampling density of 12.2 $\mu$m. So the required A-line speed is $\sim 65$ kHz, which is in accordance with the real-world situation. For the $8\times 8$ mm$^2$ scans, each direction has 350 data points which corresponds to a sampling density of 22.9 $\mu$m. So the demand for the A-line rate is reduced to $\sim 20$ kHz. However, insufficient sampling causes the blur and loss of the capillaries as demonstrated in Fig. 1(b). The left side of Fig. 1(b) shows the $8\times 8$ mm$^2$ superficial vascular plexus (SVP) scan centered on the fovea, the zoom-in view of its central FOV is shown in the upper left of the figure. The lower left of Fig. 1(b) demonstrates the native $3\times 3$ mm$^2$ SVP image.

 figure: Fig. 1.

Fig. 1. (a) Relationship between the sampling density (rate) and the A-line speed of the OCT systems. (b) The comparison of $8\times 8$ mm$^2$ and $3\times 3$ mm$^2$ FOV OCTA images.

Download Full Size | PDF

The insufficient sampling not only affects the visualization and qualitative analysis of the OCTA images, but also brings difficulty and inaccuracy to the extraction of quantitative vascular biomarkers, such as vessel area density, vessel diameter index, and vessel perimeter index. Several studies have reported the discrepancy among the OCTA scan patterns with different FOVs [1315]. Usually, a large FOV scan would cause the underestimation of the vascular biomarkers mentioned above.

Here we propose to use deep learning techniques to remedy the digital resolution degradation caused the low transverse sampling in OCTA. We expect that it will have two major functionalities. (1) To improve the accuracy of extracting the vascular biomarkers using wide-field scans and/or low sampling scans. (2) Lower the speed requirement of the OCTA acquisition thus reduce the cost of the OCTA systems, which will promote the popularization of this technique. We illustrated this idea in Fig. 2. A low speed thus low-cost OCT system is used to acquire the OCTA data with a relatively low sampling density. Then the OCTA images are feed into a deep neural network and generates high-digital-resolution OCTA images. We believe this solution will benefit the development of healthcare in low-resource settings.

 figure: Fig. 2.

Fig. 2. Proposed workflow of using low-cost OCT systems and deep learning to get high digital resolution OCTA images.

Download Full Size | PDF

The idea of using deep learning in this low to high resolution conversion of OCTA is inspired by the success of the deep learning-based methods in enhancing the resolution of various types of microscopy very recently. Wang et al. realized the transformation of ordinary fluorescence microscopic images into super-resolution images using a generative adversarial network (GAN) [16]. Rivenson et al. used a convolutional neural network to convert low-quality mobile-phone-based microscopic images to high-quality table-top microscopic images [17]. de Haan et al. demonstrated the conversion from low to high resolution scanning electron microscopic images using deep learning [18]. Different from the enhancement of optical resolution in the works mentioned above, our work is to address the resolution degradation induced by low digital sampling density.

In this paper, we conducted extensive experiments to evaluate the feasibility of the proposed idea. we tried to convert the centrally cropped $3\times 3$ mm$^2$ field of view (FOV) of the $8\times 8$ mm$^2$ foveal OCTA images to the native $3\times 3$ mm$^2$ en face OCTA images. A cycle-consistent GAN architecture was employed in this conversion. The results demonstrate the proposed deep-learning-based method is capable of significantly improving the digital resolution of the low transverse sampling OCTA images.

2. Materials and methods

2.1 Data preparation

40 eyes from 20 healthy participants (10 males and 10 females; ages ranging from 20 to 35 years old) were recruited for this digital-resolution-enhancement task at our Intelligent Ophthalmic Imaging & Laser Laboratory. The human study protocol was approved by the Institutional Review Board of Cixi Institute of Biomedical Engineering, Chinese Academy of Sciences and followed the tenets of the Declaration of Helsinki. We employed a OCTA system (ZEISS CIRRUS 5000, Zeiss Meditec, USA) for the data collection. Each eye was scanned by two different imaging protocols: $3\times 3$ mm$^2$ and $8\times 8$ mm$^2$ FOV centered on the fovea. The $3\times 3$ mm$^2$ and $8\times 8$ mm$^2$ scans have equivalent samplings of 245 and 350 along the two transverse directions, respectively. Each A-line has 1024 data points. The en face angiograms including the SVP, and deep capillary plexus (DCP) were exported from the ZEISS acquisition and analysis software (AngioPlex v1.0, Zeiss Meditec, USA). We manually checked and corrected the results of automatic retinal layer segmentation. The low transverse sampling image was obtained by cropped $3\times 3$ mm$^2$ are centered on the fovea from wide-field image of $8\times 8$ mm$^2$ protocols. We defined the cropped $3\times 3$ mm$^2$ images as the original images and the GAN-processed images as the generated images below for simplicity. Also, the native $3\times 3$ mm$^2$ OCTA images are referred as high-definition (HD) OCTA images.

2.2 Deep learning network

It’s difficult to collect the paired OCTA images because of the deviation of scanning range adjusted by operators in each acquisition and the influences caused by the motions of living eyes. So the cycle-consistent adversarial network architecture [19] can be used to repair the digital resolution degradation induced by the low transverse sampling with the unpaired images. The overall framework is illustrated in Fig. 3. It aims to learn a mapping $G_{AB}$ (mapping: A to B) such that the images from cropped $3\times 3$ mm$^2$ protocol is indistinguishable from the original $3\times 3$ mm$^2$ at both the pixel-level and feature-level with the adversarial learning. Then we couple it with an inverse mapping $G_{BA}$ (mapping: B to A) and introduce a cycle consistency loss to enforce the generated image is as similar to the reconstructed image as possible (and vice versa). The cycle consistency enforces a transition cycle: $InputA\rightarrow G_{AB}(InputA) \rightarrow G_{BA}[G_{AB}(InputA) ] \approx InputA$. The backward cycle has the same architecture but enforces a inverse transition cycle: $InputB\rightarrow G_{BA}(InputB) \rightarrow G_{AB}[G_{BA}(InputB) ]$ $\approx InputB$. The Euclidean norm measures the distance between the inputs and the outputs. Such a translation does not ensure that the input image and output image are paired up in a meaningful way, so two discriminator network $D_{A}$ and $D_{B}$ are brought out to restraint input images and synthesized images from generators in the same distribution. The network aims to learn the mapping between low-quality images (cropped $3\times 3$ mm$^2$ OCTA) and high-definition images (native $3\times 3$ mm$^2$ OCTA) and convert low-quality input images into high-quality output images. The overall loss is defined by:

 figure: Fig. 3.

Fig. 3. Overview of the framework for the digital resolution enhancement task, there are two generators $G_{AB}$ and $G_{BA}$ and two discriminators $D_{A}$ and $D_{B}$. A denotes the low transverse sampling image domain, B denotes the high-definition image domain, to further regularize the mappings, the network employed three losses: cycle consistency loss, adversarial loss, and identity loss.

Download Full Size | PDF

$$\begin{aligned} L\left(G_{A B}, G_{B A}, D_{A}, D_{B}\right) & =L_{G A N}\left(G_{A B}, D_{B}, A, B\right)+L_{G A N}\left(G_{B A}, D_{A}, B, A\right)+\beta L_{c y c}\left(G_{A B}, G_{B A}\right) \\ & +\gamma L_{i d e}\left(G_{A B}, G_{B A}\right) \end{aligned}$$
where $\beta$ and $\gamma$ control the proportion of each loss, $\ L_{G A N}$, $\ L_{c y c}$, $\ L_{i d e}$ denote the adversarial loss, cycle consistency loss, and identity loss. A denotes the low quality images (cropped $3\times 3$ mm$^2$ OCTA) in source domain, B denotes the high-definition images (native $3\times 3$ mm$^2$ OCTA) in target domain.

Adversarial loss [20] is applied in the mapping function, the objective is expressed as:

$$\begin{aligned} \min _{G_{A B}} \max _{D_{B}} L_{G A N}\left(G_{A B},\right. & \left.D_{B}, A, B\right) \\= & E_{i_{b} \sim P_{d a t a}(b)}\left[\log D_{B}(b)\right]+E_{i_{b} \sim P_{d a t a}(a)}\left[\log \left(1-D_{B}\left(G_{A B}(a)\right)\right)\right] \end{aligned}$$
where $G_{AB}$ is trained to enhance the digital resolution of low transverse sampling images to make it similar to the high-definition OCTA while $D_{B}$ aims to distinguish between images generated by $G_{AB}$ and real images in domain B. $G_{AB}$ aims to minimize this objective against an adversary $D_{B}$ that tries to maximize it. We employ a similar adversarial loss for the mapping function $G_{BA}$ and its discriminator $D_{A}$ as well.

Adversarial training can learn mappings $G_{AB}$ and $G_{BA}$ that produce outputs identically distributed as target domain B and source domain A, respectively. However, the network may map the same set of input images to any image in the target domain, that is to say, if we only employ adversarial loss, it’s hard to learn the function maps with a single input $A_{i}$ to the desired output $B_{i}$. So the cycle consistent loss is used to the network to constrain the mapping, for each image A from source domain, the image translation cycle should be able to bring A back to the original image:

$$\begin{aligned} L_{c y c}\left(G_{A B}, G_{B A}\right)= & E_{i a} \sim P_{d a t a}\left[\left\|G_{B A}\left(G_{A B}\left(i_{a}\right)\right)-i_{a}\right\|_{1}\right] \\ & +E_{i_{b} \sim P_{d a t a}(b)}\left[\left\|G_{A B}\left(G_{B A}\left(i_{b}\right)\right)-i_{b}\right\|_{1}\right]. \end{aligned}$$
where the cycle consistency loss enforces the constraint that $G_{AB}$ and $G_{BA}$ should be inverse of each other, i.e., it encourages
$$G_{B A}\left(G_{A B}\left(i_{A}\right)\right) \approx i_{A} \textrm{ and }G_{A B}\left(G_{B A}\left(i_{B}\right)\right) \approx i_{B}.$$

In order to prevent excessive color composition between input and output, we regularize the input of network provide by target domain to be near an identity mapping:

$$L_{i d e}\left(G_{A B}, G_{B A}\right)=E_{i_{b} \sim P_{d a t a}(b)}\left[\left\|G_{A B}(b)-i_{b}\right\|_{1}\right]+E_{i_{a} \sim P_{d a t a}(a)}\left[\left\|G_{B A}\left(i_{a}\right)-i_{a}\right\|_{1}\right].$$

2.3 Implementation

Our model is implemented based on the deep learning framework PyTorch in the Ubuntu 16.04 LTS operating system and the training of the model was performed with NVIDIA GeForce GTX 1080 Ti GPU with 12 GB RAM. We employed the open-source implementation of the cycle-consistent adversarial network shared by its inventors [19]. The generator and discriminator are trained from scratch using the Adam optimizer [21] with an initial learning rate of $2\times 10^{-4}$ and a batch size of 1. We set $\beta _1=0.5$ and $\beta _2=0.999$ for both of the two Adam optimizers. In the training of the GAN model, the loss curve was shown to be convergent after 180 epochs, so we stopped at 200 epochs for achieving the smallest loss. It took about 4 hours for each training.

In this study, 40 OCTA data sets were split into 35 training sets and 5 testing sets. We did not set up validation sets because of the very limited size of data sets. This splitting strategy has been commonly used both in medical image analysis and computer vision fields [2224]. Besides, as mentioned above, the model converged well using this strategy, so the validation data sets may not be necessary. We also used the data augmentation methods for enhancing the training data sets, including Gaussian blurring, contrast adjustment, adding Gaussian noise, and random rotation. So a total of 280 data sets were used in each training. Then the model was tested on the 5 testing data sets and other diseased data sets. For each of the SVP and DCP, we trained their model separately following the same procedure described above.

2.4 Quantitative evaluation metrics

The vasculature captured the OCTA imaging of the exactly same region using the $8\times 8$ mm$^2$ and $3\times 3$ mm$^2$ scan protocols does not have the same spatial mapping (as shown in Fig. 1(b)), which is the primary obstacle for the quantitative evaluation of the results in this work. The discrepancy in the spatial mapping has also been observed in several OCTA repeatability and reproducibility studies [2528], which may primarily related to eye motion, signal strength difference caused by focusing or alignment, undersampling alias, and segmentation inaccuracy. So we could not use the native $3\times 3$ mm$^2$ scans as the ground truth for calculating the pixel-wise image similarity metrics, such as SSIM, PSNR, Dice coefficient, and Jaccard index (IoU).

2.4.1 Perceptual similarity measures

Instead, we turned to employ the quantitative metrics that are used in the evaluation of GAN models. The GAN- synthetic images also do not have a paired ground truth, which is similar to the situation in this work. Inception score (IS) and Fréchet inception distance (FID) are the two most widely-used metrics in measuring the performance of GAN models. However, the IS may not suit our task because it calculates the KL divergence between the properties of the generated images and the features of the objects among the ImageNet classes [29]. On the other hand, the FID calculates the Wasserstein-2 distance between the features of two images. It could be used here to measure which one between the original image and the generated images is more similar to the paired HD OCTA image. Because the FID has been shown to be consistent with human perception [30], we refer to as the perceptual similarity here corresponding to the pixel-wise similarity.

Using a single metric may lead to bias in the evaluation, so we employed the kernel inception distance (KID) as supplementary, which was proposed very recently [31] and becoming more and more popular in the studies of GAN models. Compared with the FID, the KID measure the squared mean discrepancy between features instead of the Wasserstein-2 distance, which has an unbiased estimator with a cubic kernel while the FID may be empirically biased [32].

We employed an open-source implementation of the FID and KID [30]. Both the FID and KID use the Inception v3 network for feature extraction then employ the extracted features to calculated the FID and KID scores. Lower FID and KID scores mean higher perceptual similarity. In the implementation, we could select the dimensionality of features including 64 (first max pooling features), 192 (second max pooling features), 768 (pre-aux classifier features), and 2048 (final average pooling features). Because different levels of features represent different aspects of perception, e.g., texture and structure, we included all of these feature dimensions in the evaluation.

2.4.2 Vessel caliber change and SNR metrics

To further characterize the performance of the proposed method, we quantified the changes of the vessel caliber and flow SNR before and after the digital resolution enhancement processing. These parameters of the HD OCTA are used as the reference. To quantify these changes, we defined the discrepancy $S$ of the vessel calibers between the HD OCTA and the original or generated OCTA as

$$S = \frac{|W_{O, G}-W_{HD}|}{W_{HD}}\times 100\%,$$
where $W$ is the full width at half maximum (FWHM) width of the intensity profile. $O$ and $G$ refer to the original and generated images, respectively. We calculated this value for all of the intensity profiles from each test data set and using their average and standard deviation as the quantitative measure of the vessel caliber changes. Because the noise is hard to differentiate with the vessels in the DCP, CC, and choroid, we only used the SVP for the evaluation here. We followed the calculation of the SNR in [33] as
$$SNR = \frac{\overline{D}_{parafoveal}-\overline{D}_{FAZ}}{\sigma_{D_{FAZ}}},$$
where $\overline {D}_{parafoveal}$ is the averaged flow signal in the parafovea, which is concentric with the fovea, has an outer diameter of 2.5 mm and an inner diameter of 0.6 mm. $\overline {D}_{FAZ}$ is the averaged flow signal in the FAZ and $\sigma _{D_{FAZ}}$ is its standard deviation.

2.5 Quantitative vascular biomarkers

Quantitative assessment of OCTA is clinically useful for various ocular diseases [34]. We further applied the proposed method in the application of quantifying OCTA biomarkers. Following the methodology in [34], we employed the following vascular biomarkers including vessel area density (VAD), vessel skeleton density (VSD), vessel diameter index (VDI), vessel perimeter index (VPI), vessel complexity index (VCI). These parameters have proven to be effective in the studies of ophthalmic diseases [7,35,36].

Because the DCP, CC, and choroid are severely contaminated by motion noises, we only calculated the vascular biomarkers of the SVP images. We first applied a hard threshold to remove the noises, which is two standard deviations above the mean value in the FAZ as defined in [37]. Then we applied the 2D Frangi filter [38] to the thresholded image with a scale range of [0.8 1.5], a scale ratio of 0.1. The Frangi $\beta _1$ and $\beta _2$ were set as 0.5 and 30, respectively. The vessel area map could further be acquired by using the MATLAB function imbinarize with the adaptive threshold decided by the function adaptthresh. Based on the vessel area map, we could calculate the vessel skeleton map and vessel perimeter map by using the functions bwskel and bwperim, respectively. Finally, the vascular biomarkers mentioned above could be easily calculated from the vessel maps by following the methods in [34].

3. Results

3.1 Visual comparison

We first compare the images generated by the GAN model with the original input and the pair HD OCTA image visually.

Figure 4 shows the results of the SVP. The zoom-in views of three regions of interest (ROIs) labeled as a, b, c (the red boxes) are demonstrated at the right side of the figure. The tag numbers 1, 2, and 3 refer to the original low transverse sampling image, the GAN-generated image, and the HD OCTA, respectively. In the ROI (a), we can see the sharpness of the capillaries around the foveal avascular zone (FAZ) is significantly enhanced. Besides, the bulk motion noise inside the FAZ seems to be suppressed because of the conversion. Inside the ROI (b), large vessel branches are captured by the OCTA imaging, but the capillaries are almost absent even in the HD OCTA image. Because of the digital resolution enhancement, the calibers of the large vessels decrease in the generated image. The sizes of the vessels are similar to those in the HD image. The ROI (c) is located close to the edge of this $3\times 3$ mm$^2$ FOV. We can observe both the vessels and capillaries are in this region. The proposed method successfully converted the original blurry angiogram into the high-resolution image.

 figure: Fig. 4.

Fig. 4. Digital resolution enhancement results of the SVP, the zoom-in views of three regions of interest (ROIs) labeled as a, b, c (the red boxes) are demonstrated at the right side of the figure. The tag numbers 1, 2, and 3 refer to the original low transverse sampling image, the GAN-generated image, and the HD OCTA, respectively.

Download Full Size | PDF

The DCP is a purer and denser capillary plexus compared with the SVP as and we conducted a very similar analysis as demonstrated in Fig. 5. The GAN-based conversion significantly improves the digital resolution and contrast of the original blurry images.

 figure: Fig. 5.

Fig. 5. Digital resolution enhancement results of the DCP. (a1-c1) The zoom-in views of interest (ROIs) labeled in the red boxes of original low transverse sampling image. (a2-c2) Demonstrated the corresponding zoom-in region of the GAN-generated image. (a3-c3) Magnified corresponding area at the red boxes of HD OCTA.

Download Full Size | PDF

As mentioned above, the original image and the HD OCTA do not have the exact same spatial mapping essentially, so we can observe that their capillary vasculatures have discrepancies with that of the HD OCTA image. Fortunately, the spatial mapping between the original and generated images is excellent, which means the proposed method does not create the features that do not exist.

3.2 Quantitative evaluation

3.2.1 Perceptual similarity

As shown in Table 1, for both the SVP and DCP, the FID and KID scores of the generated images are lower than those of the corresponding original images which means they have better perceptual similarity with the HD OCTA images. It should be noted that the discrepancies between the scores of the original and generated images are larger in the lower feature dimensions (64 and 192), which is because the low-level feature maps are closer to the original images. while the high-level feature maps are more abstract.

3.2.2 Vessel caliber change and SNR

For calculating the changes of vessel caliber, we randomly selected 5 vessel positions in the SVP images of each test data set as demonstrated in the left side of Fig. 6. Then we plotted their corresponding intensity profiles as shown in the right side of Fig. 6. We can see the vessel calibers are significantly decreased in the generated and HD OCTA image.

We plotted the variation of vessel calibers in 25 random positions from the 5 testing data sets in the top of Fig. 7. The black, red, and blue lines are from the original, generated, and HD OCTA images, respectively. We also plotted the corresponding discrepancies between the original and HD (black) and the generated and HD (red) at the bottom of Fig. 7. We can see the calibers of the vessel from the generated images are always smaller than those of the original images and closer to the HD images.

 figure: Fig. 6.

Fig. 6. Quantifying the changes of retinal vessel caliber. Left: 5 randomly selected positions of the vessels. Right: their intensity profiles. The black, red, and blue ones are from the original, generated, and HD OCTA images, respectively.

Download Full Size | PDF

 figure: Fig. 7.

Fig. 7. Quantify the changes of vessel caliber. Top: the variation of vessel calibers in 25 random positions from the 5 testing data sets. The black, red, and blue lines are from the original, generated, and HD OCTA images, respectively. Bottom: The discrepancies between the original and HD (black) and the generated and HD (red).

Download Full Size | PDF

Tables Icon

Table 1. Quantitative comparison of the perceptual similarity measures

Table 2 is the quantitative comparison of the SNR. It shows the generated images has significant improvement in SNR, which is in accordance with the visual inspection above.

Tables Icon

Table 2. Quantitative comparison of the SNR

3.3 Application in visualizing pathologies

For validating the generalization performance of the proposed method and evaluating its value in clinical applications, we tested our method on two unseen diseased cases. They are from the pre-installed database inside the ZEISS acquisition and analysis software. The case 1 is from the right eye of a male (age 51), who has branch retinal vein occlusion. The case 2 is from the right eye of a 70-years-old subject (gender unknown), who has diabetic retinopathy. Because the database only stores the $8\times 8$ mm $^2$ scans for these two cases, so we do not have the HD OCTA images for comparison. The $3\times 3$ mm $^2$ FOV images were cropped centered on the fovea as the input of the trained deep network.

As demonstrated in Fig. 8, we can see the original cropped images have blurry vessels and low and unevenly distributed signal strength. After applying the proposed method, the vessels in the generated OCTA images are much sharper and narrower than those of the original images. Besides, we can see the visibility of the capillaries around the FAZ is significantly improved in the generated images. As mentioned above, our model was trained using normal images only, so its performance on the diseased data is an excellent indication of its potential in clinical applications.

 figure: Fig. 8.

Fig. 8. Application of the proposed method in unseen diseased cases. The case 1 is from the right eye of a male (age 51), who has branch retinal vein occlusion. The case 2 is from the right eye of a 70-years-old subject (gender unknown), who has diabetic retinopathy. From left to right are the $8\times 8$ mm $^2$ scans for these two cases, the cropped $3\times 3$ mm $^2$ images referring to original, and the outputted images for the deep network referring to generated. The red boxes give the FOVs for the cropping.

Download Full Size | PDF

 figure: Fig. 9.

Fig. 9. Representative OMAG images to calculate the quantitative indexes of the SVP. (a) Original en face OCTA image. (b) Vessel area map is used for VAD, VDI, and VCI quantification. (c) Vessel skeleton map is used for VSD and VDI quantification. (d) Vessel perimeter map is used for VPI and VCI quantification.

Download Full Size | PDF

3.4 Application in quantifying vascular biomarkers

To further validate the proposed method, we conducted a quantitative analysis of the SVP by comparing the vessel feature maps and their biomarkers from the original, GAN-generated, and HD OCTAs. Figure 9 shows the images and maps of the SVP. (a) are the OCTA images. (b), (c), and (d) are the vessel area map, vessel skeleton map, and vessel perimeter map, respectively. We can see the vessel maps of the generated image and the HD OCTA are quite similar, while the vessel maps calculated from the original image are evidently sparser, especially at the regions away from the FAZ.

Based on the vessel maps, we calculated the vascular biomarkers of these three types of images including the VDI, VAD, VSD, VPI, and VCI as listed in Table 3. The VDI is an indicator of the vessel diameters. We can see the generated SVP image has a low VDI compared with the original one, which confirms the improvement of the digital resolution. The VAD, VSD, VPI, and VCI are all indicators of the abundance of the capillaries from different aspects of the vasculature (length, caliber, and so on). The low transverse sampling will cause the underestimation of these biomarkers. The deep-learning-based conversion, on the other hand, can effectively minimize the underestimation. Besides, the vascular biomarkers extracted from the generated SVP image is in good accordance with those extracted from the HD SVP image.

Tables Icon

Table 3. Comparison of the vascular biomarkers of the SVP.

4. Discussion and conclusions

Enhancing the digital resolution of the low transverse sampling data using deep learning loses the requirement of high-speed acquisition systems in OCTA, which has the potential to reduce the cost of clinical OCTA machines and promote the applications of this technique in a wider population. This paper demonstrates the implementation of this idea by training the low transverse sampling en face OCTA images to learn the features of the high transverse sampling data using a cycle-consistent adversarial deep network. The results have demonstrated the proposed conversion could significantly improve the digital resolution without disrupting the original morphological features of the retinal vessels. Besides, we have found the improvement in digital resolution also benefited the accurate quantification of the vascular biomarkers, such as vessel area density and vessel perimeter index.

However, this work still has several limitations. (1) OCTA is a 3D imaging modality, so the en face projected images can not provide intact depth information. (2) This work only enhanced the digital resolution of the $3\times 3$ mm$^2$ FOV centered on the fovea, which is far from sufficient for the researches and diagnoses of ocular diseases, especially the early symptoms happening at the peripheral retina, such as the early stages of diabetic retinopathy. (3) The patient number used in this study is relatively small, we are keep collecting data at local hospitals for verifying the proposed method in a large population with more diseased cases. In the next step, we will devote to the development of this new technique in three major directions. (1) The digital resolution enhancement of OCTA B-scans (volumetric OCTA). Because it will involve the axial direction, we will consider to introduce compressed sensing techniques to fulfill this task. (2) Enhancing the digital resolution in larger FOVs. For example, including fovea and optical nerve head simultaneously. We will combine the style transfer deep network with other techniques such as deep learning or low-level feature-based registration and impainting. (3) Not only the digital resolution, this deep-learning-based approach could also be used in optical resolution enhancement, by using AO OCTA data as the target domain. Because of the unpaired transition capacity, instead of using the paired OCTA images captured when the AO loop is turned off and on, we may only need some arbitrary AO OCTA images for training.

In summary, We have developed a deep learning-based technique that could enhance the digital resolution of the low transverse sampling OCTA. Because it is difficult to collect the paired OCTA images due to the deviation of scanning range adjusted by operators in each acquisition and the influences caused by the motions of living eyes, we have employed the cycle-consistent adversarial network architecture for this task. Qualitative and quantitative results have demonstrated the proposed technique could not only improve the transverse digital resolution and SNR of the OCTA images but also benefit the quantification of the vascular biomarkers.

Funding

Ningbo "2025 S&T Megaprojects" (2019B10033, 2019B10061); Natural Science Foundation of Zhejiang Province (LQ19H180001); Ningbo Public Welfare Science and Technology Project (2018C50049); Guangdong Provincial Key Laboratory (2020B121201001).

Acknowledgments

We would like to thank the reviewers and editors for the careful reviewing and insightful comments, which significantly benefits the improvement of this manuscript and helps us refine this new research topic. We also acknowledge the contribution of Zhengjie Chai and Qiangjiang Hao in the calculation of the quantitative metrics FID and KID.

Disclosures

The authors declare no conflicts of interest.

References

1. S. S. Gao, Y. Jia, M. Zhang, J. P. Su, G. Liu, T. S. Hwang, S. T. Bailey, and D. Huang, “Optical Coherence Tomography Angiography,” Invest. Ophthalmol. Vis. Sci. 57(9), OCT27–OCT36 (2016). [CrossRef]  

2. L. Liu, Y. Jia, H. L. Takusagawa, A. D. Pechauer, B. Edmunds, L. Lombardi, E. Davis, and J. C. Morrison, “Optical Coherence Tomography Angiography of the Peripapillary Retina in Glaucoma,” JAMA Ophthalmol. 133(9), 1045–1052 (2015). [CrossRef]  

3. Y. Jia, S. T. Bailey, T. S. Hwang, S. M. McClintic, S. S. Gao, M. E. Pennesi, C. J. Flaxel, A. K. Lauer, D. J. Wilson, J. Hornegger, J. G. Fujimoto, and D. Huang, “Quantitative optical coherence tomography angiography of vascular abnormalities in the living human eye,” Proc. Natl. Acad. Sci. 112(18), E2395–E2402 (2015). [CrossRef]  

4. J. P. Campbell, E. Nudleman, J. Yang, O. Tan, R. V. Chan, M. F. Chiang, D. Huang, and G. Liu, “Handheld optical coherence tomography angiography and ultra-wide-field optical coherence tomography in retinopathy of prematurity,” JAMA Ophthalmol. 135(9), 977–981 (2017). [CrossRef]  

5. H. Jiang, Y. Wei, Y. Shi, C. Wright, X. Sun, G. Gregori, F. Zheng, E. Vanner, B. Lam, T. Rundek, and J. Wang, “Altered Macular Microvasculature in Mild Cognitive Impairment and Alzheimer Disease,” J. Neuro-Ophthalmol. 38(3), 292–298 (2018). [CrossRef]  

6. B. E. O’Bryhim, R. S. Apte, N. Kung, D. Coble, and G. P. Van Stavern, “Association of Preclinical Alzheimer Disease With Optical Coherence Tomographic Angiography Findings,” JAMA Ophthalmol. 136(11), 1242–1248 (2018). [CrossRef]  

7. C.-L. Chen and R. K. Wang, “Optical coherence tomography based angiography [invited],” Biomed. Opt. Express 8(2), 1056–1082 (2017). [CrossRef]  

8. B. Braaf, K. V. Vienola, C. K. Sheehy, Q. Yang, K. A. Vermeer, P. Tiruveedhula, D. W. Arathorn, A. Roorda, J. F. de Boer, D. Huang, E. A. Swanson, C. P. Lin, J. S. Schuman, W. G. Stinson, W. Chang, M. R. Hee, T. Flotte, K. Gregory, C. A. Puliafito, J. G. Fujimoto, Y. Zhao, Z. Chen, C. Saxer, S. Xiang, J. F. de Boer, J. S. Nelson, T. E. Milner, and D. Dave, “Real-time eye motion correction in phase-resolved OCT angiography with tracking SLO,” Biomed. Opt. Express 4(1), 51–65 (2013). [CrossRef]  

9. A. Camino, M. Zhang, C. Dongye, A. D. Pechauer, T. S. Hwang, T. Steven, B. Lujan, D. J. Wilson, D. Huang, and Y. Jia, “Automated registration and enhanced processing of clinical optical coherence tomography angiography,” Quant Imaging Med. Surg. 6(4), 391–401 (2016). [CrossRef]  

10. M. Salas, M. Augustin, L. Ginner, A. Kumar, B. Baumann, R. Leitgeb, W. Drexler, S. Prager, J. Hafner, U. Schmidt-Erfurth, and M. Pircher, “Visualization of micro-capillaries using optical coherence tomography angiography with and without adaptive optics,” Biomed. Opt. Express 8(1), 207–222 (2017). [CrossRef]  

11. M. J. Doughty and T. Naase, “Further analysis of the human spontaneous eye blink rate by a cluster analysis-based approach to categorize individuals with ‘normal’ versus ‘frequent’ eye blink activity,” Eye Contact Lens. 32(6), 294–299 (2006). [CrossRef]  

12. T. Klein and R. Huber, “High-speed OCT light sources and systems [Invited],” Biomed. Opt. Express 8(2), 828 (2017). [CrossRef]  

13. J. Lei, M. K. Durbin, Y. Shi, A. Uji, S. Balasubramanian, E. Baghdasaryan, M. Al-Sheikh, and S. R. Sadda, “Repeatability and reproducibility of superficial macular retinal vessel density measurements using optical coherence tomography angiography en face images,” JAMA Ophthalmol. 135(10), 1092–1098 (2017). [CrossRef]  

14. M. Al-Sheikh, K. G. Falavarjani, H. Akil, and S. R. Sadda, “Impact of image quality on OCT angiography based quantitative measurements,” Int. J. Retin. Vitr. 3(1), 13 (2017). [CrossRef]  

15. X. X. Li, W. Wu, H. Zhou, J. J. Deng, M. Y. Zhao, T. W. Qian, C. Yan, X. Xu, and S. Q. Yu, “A quantitative comparison of five optical coherence tomography angiography systems in clinical performance,” Int. J. Ophthalmol. 11(11), 1784–1795 (2018). [CrossRef]  

16. H. Wang, Y. Rivenson, Y. Jin, Z. Wei, R. Gao, H. Günaydin, L. A. Bentolila, C. Kural, and A. Ozcan, “Deep learning enables cross-modality super-resolution in fluorescence microscopy,” Nat. Methods 16(1), 103–110 (2019). [CrossRef]  

17. Y. Rivenson, H. Ceylan Koydemir, H. Wang, Z. Wei, Z. Ren, H. Günaydin, Y. Zhang, Z. Göröcs, K. Liang, D. Tseng, and A. Ozcan, “Deep learning enhanced mobile-phone microscopy,” ACS Photonics 5(6), 2354–2364 (2018). [CrossRef]  

18. K. de Haan, Z. S. Ballard, Y. Rivenson, Y. Wu, and A. Ozcan, “Resolution enhancement in scanning electron microscopy using deep learning,” Sci. Rep. 9(1), 12050 (2019). [CrossRef]  

19. J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix (2017).

20. I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, X. Bing, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in International Conference on Neural Information Processing Systems, (2014).

21. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv: Learning (2014).

22. J. Staal, M. D. Abràmoff, M. Niemeijer, M. A. Viergever, and B. Van Ginneken, “Ridge-based vessel segmentation in color images of the retina,” IEEE Trans. Med. Imaging 23(4), 501–509 (2004). [CrossRef]  

23. W. Liu, W. Luo, D. Lian, and S. Gao, “Future frame prediction for anomaly detection–a new baseline,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (2018), pp. 6536–6545.

24. D. S. Kermany, M. Goldbaum, W. Cai, C. C. S. Valentim, H. Liang, S. L. Baxter, A. McKeown, G. Yang, X. Wu, F. Yan, J. Dong, M. K. Prasadha, J. Pei, M. Y. L. Ting, J. Zhu, C. Li, S. Hewett, J. Dong, I. Ziyar, A. Shi, R. Zhang, L. Zheng, R. Hou, W. Shi, X. Fu, Y. Duan, V. A. N. Huu, C. Wen, E. D. Zhang, C. L. Zhang, O. Li, X. Wang, M. A. Singer, X. Sun, J. Xu, A. Tafreshi, M. A. Lewis, H. Xia, and K. Zhang, “Identifying medical diagnoses and treatable diseases by image-based deep learning,” Cell 172(5), 1122–1131.e9 (2018). [CrossRef]  

25. J. Hong, B. Tan, N. D. Quang, P. Gupta, E. Lin, D. Wong, M. Ang, E. Lamoureux, L. Schmetterer, and J. Chua, “Intra-session repeatability of quantitative metrics using widefield optical coherence tomography angiography (octa) in elderly subjects,” Acta Ophthalmol. (2019). [CrossRef]  

26. J. Lei, M. K. Durbin, Y. Shi, A. Uji, S. Balasubramanian, E. Baghdasaryan, M. Al-Sheikh, and S. R. Sadda, “Repeatability and reproducibility of superficial macular retinal vessel density measurements using optical coherence tomography angiography en face images,” JAMA Ophthalmol. 135(10), 1092–1098 (2017). [CrossRef]  

27. S. Men, C.-L. Chen, W. Wei, T.-Y. Lai, S. Song, and R. K. Wang, “Repeatability of vessel density measurement in human skin by oct-based microangiography,” Skin Res. Technol. 23(4), 607–612 (2017). [CrossRef]  

28. M.-W. Lee, K.-M. Kim, H.-B. Lim, Y.-J. Jo, and J.-Y. Kim, “Repeatability of vessel density measurements using optical coherence tomography angiography in retinal diseases,” Br. J. Ophthalmol. 103(5), 704–710 (2019). [CrossRef]  

29. A. Borji, “Pros and cons of gan evaluation measures,” Comput. Vis. Image Und. 179, 41–65 (2019). [CrossRef]  

30. M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “Gans trained by a two time-scale update rule converge to a local nash equilibrium,” https://github.com/abdulfatir/gan-metrics-pytorch (2017).

31. M. Binkowski, D. J. Sutherland, M. Arbel, and A. Gretton, “Demystifying MMD GANs,” in International Conference on Learning Representations, (2018).

32. Y. Li, X. Chen, F. Wu, and Z.-J. Zha, “Linestofacephoto: Face photo generation from lines with conditional self-attention generative adversarial networks,” in Proceedings of the 27th ACM International Conference on Multimedia, (ACM, 2019), pp. 2323–2331.

33. A. Camino, Y. Jia, G. Liu, J. Wang, and D. Huang, “Regression-based algorithm for bulk motion subtraction in optical coherence tomography angiography,” Biomed. Opt. Express 8(6), 3053–3066 (2017). [CrossRef]  

34. Z. Chu, J. Lin, G. Chen, X. Chen, Q. Zhang, C. L. Chen, L. Roisman, G. Gregori, P. J. Rosenfeld, and R. K. Wang, “Quantitative assessment of the retinal microvasculature using optical coherence tomography angiography,” J. Biomed. Opt. 21(6), 066008 (2016). [CrossRef]  

35. A. Camino, M. Zhang, S. S. Gao, T. S. Hwang, U. Sharma, D. J. Wilson, D. Huang, and Y. Jia, “Evaluation of artifact reduction in optical coherence tomography angiography with real-time tracking and motion correction technology,” Biomed. Opt. Express 7(10), 3905–3915 (2016). [CrossRef]  

36. M. Alam, D. Thapa, J. I. Lim, D. Cao, and X. Yao, “Quantitative characteristics of sickle cell retinopathy in optical coherence tomography angiography,” Biomed. Opt. Express 8(3), 1741–1753 (2017). [CrossRef]  

37. Y. Jia, O. Tan, J. Tokayer, B. Potsaid, Y. Wang, J. Jonathan, M. F. Kraus, H. Subhash, J. G. Fujimoto, J. Hornegger, and D. Huang, “Split-spectrum amplitude-decorrelation angiography with optical coherence tomography,” Opt. Express 20(4), 4710 (2012). [CrossRef]  

38. A. F. Frangi, W. J. Niessen, K. L. Vincken, and M. A. Viergever, “Multiscale vessel enhancement filtering,” in International conference on medical image computing and computer-assisted intervention, (Springer, 1998), pp. 130–137.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (9)

Fig. 1.
Fig. 1. (a) Relationship between the sampling density (rate) and the A-line speed of the OCT systems. (b) The comparison of $8\times 8$ mm$^2$ and $3\times 3$ mm$^2$ FOV OCTA images.
Fig. 2.
Fig. 2. Proposed workflow of using low-cost OCT systems and deep learning to get high digital resolution OCTA images.
Fig. 3.
Fig. 3. Overview of the framework for the digital resolution enhancement task, there are two generators $G_{AB}$ and $G_{BA}$ and two discriminators $D_{A}$ and $D_{B}$. A denotes the low transverse sampling image domain, B denotes the high-definition image domain, to further regularize the mappings, the network employed three losses: cycle consistency loss, adversarial loss, and identity loss.
Fig. 4.
Fig. 4. Digital resolution enhancement results of the SVP, the zoom-in views of three regions of interest (ROIs) labeled as a, b, c (the red boxes) are demonstrated at the right side of the figure. The tag numbers 1, 2, and 3 refer to the original low transverse sampling image, the GAN-generated image, and the HD OCTA, respectively.
Fig. 5.
Fig. 5. Digital resolution enhancement results of the DCP. (a1-c1) The zoom-in views of interest (ROIs) labeled in the red boxes of original low transverse sampling image. (a2-c2) Demonstrated the corresponding zoom-in region of the GAN-generated image. (a3-c3) Magnified corresponding area at the red boxes of HD OCTA.
Fig. 6.
Fig. 6. Quantifying the changes of retinal vessel caliber. Left: 5 randomly selected positions of the vessels. Right: their intensity profiles. The black, red, and blue ones are from the original, generated, and HD OCTA images, respectively.
Fig. 7.
Fig. 7. Quantify the changes of vessel caliber. Top: the variation of vessel calibers in 25 random positions from the 5 testing data sets. The black, red, and blue lines are from the original, generated, and HD OCTA images, respectively. Bottom: The discrepancies between the original and HD (black) and the generated and HD (red).
Fig. 8.
Fig. 8. Application of the proposed method in unseen diseased cases. The case 1 is from the right eye of a male (age 51), who has branch retinal vein occlusion. The case 2 is from the right eye of a 70-years-old subject (gender unknown), who has diabetic retinopathy. From left to right are the $8\times 8$ mm $^2$ scans for these two cases, the cropped $3\times 3$ mm $^2$ images referring to original, and the outputted images for the deep network referring to generated. The red boxes give the FOVs for the cropping.
Fig. 9.
Fig. 9. Representative OMAG images to calculate the quantitative indexes of the SVP. (a) Original en face OCTA image. (b) Vessel area map is used for VAD, VDI, and VCI quantification. (c) Vessel skeleton map is used for VSD and VDI quantification. (d) Vessel perimeter map is used for VPI and VCI quantification.

Tables (3)

Tables Icon

Table 1. Quantitative comparison of the perceptual similarity measures

Tables Icon

Table 2. Quantitative comparison of the SNR

Tables Icon

Table 3. Comparison of the vascular biomarkers of the SVP.

Equations (7)

Equations on this page are rendered with MathJax. Learn more.

L ( G A B , G B A , D A , D B ) = L G A N ( G A B , D B , A , B ) + L G A N ( G B A , D A , B , A ) + β L c y c ( G A B , G B A ) + γ L i d e ( G A B , G B A )
min G A B max D B L G A N ( G A B , D B , A , B ) = E i b P d a t a ( b ) [ log D B ( b ) ] + E i b P d a t a ( a ) [ log ( 1 D B ( G A B ( a ) ) ) ]
L c y c ( G A B , G B A ) = E i a P d a t a [ G B A ( G A B ( i a ) ) i a 1 ] + E i b P d a t a ( b ) [ G A B ( G B A ( i b ) ) i b 1 ] .
G B A ( G A B ( i A ) ) i A  and  G A B ( G B A ( i B ) ) i B .
L i d e ( G A B , G B A ) = E i b P d a t a ( b ) [ G A B ( b ) i b 1 ] + E i a P d a t a ( a ) [ G B A ( i a ) i a 1 ] .
S = | W O , G W H D | W H D × 100 % ,
S N R = D ¯ p a r a f o v e a l D ¯ F A Z σ D F A Z ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.