Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Automatic measurement of anterior chamber angle parameters in AS-OCT images using deep learning

Open Access Open Access

Abstract

The early assessment of angle closure is of great significance for the timely diagnosis and treatment of primary angle-closure glaucoma (PACG). Anterior segment optical coherence tomography (AS-OCT) provides a fast and non-contact way to evaluate the angle close using the iris root (IR) and scleral spur (SS) information. The objective of this study was to develop a deep learning method to automatically detect IR and SS in AS-OCT for measuring anterior chamber (AC) angle parameters including angle opening distance (AOD), trabecular iris space area (TISA), trabecular iris angle (TIA), and anterior chamber angle (ACA). 3305 AS-OCT images from 362 eyes and 203 patients were collected and analyzed. Based on the recently proposed transformer-based architecture that learns to capture long-range dependencies by leveraging the self-attention mechanism, a hybrid convolutional neural network (CNN) and transformer model to encode both local and global features was developed to automatically detect IR and SS in AS-OCT images. Experiments demonstrated that our algorithm achieved a significantly better performance than state-of-the-art methods for AS-OCT and medical image analysis with a precision of 0.941, a sensitivity of 0.914, an F1 score of 0.927, and a mean absolute error (MAE) of 37.1±25.3 µm for IR, and a precision of 0.805, a sensitivity of 0.847, an F1 score of 0.826, and an MAE of 41.4±29.4 µm for SS, and a high agreement with expert human analysts for AC angle parameter measurement. We further demonstrated the application of the proposed method to evaluate the effect of cataract surgery with IOL implantation in a PACG patient and to assess the outcome of ICL implantation in a patient with high myopia with a potential risk of developing PACG. The proposed method can accurately detect IR and SS in AS-OCT images and effectively facilitate the AC angle parameter measurement for pre- and post-operative management of PACG.

© 2023 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

Glaucoma is one of the leading causes of irreversible vision loss worldwide [1], and can be classified into primary angle-closure glaucoma (PACG) and primary open angle glaucoma (POAG). The predominant glaucoma subtype is POAG [2], accounting for 65.2% of all glaucoma, but PACG is the most visually destructive form of the disease [3]. The most conspicuous feature of PACG is an anatomical angle closure caused by appositional contact between trabecular meshwork (TM) and the iris [4]. Angle closure can cause very high intraocular pressure and reduce aqueous humor flow from the posterior to anterior chambers, which are important risk factors for glaucomatous optic neuropathy [3]. In addition to the natural angle closure, implantable collamer lens (ICL) implantation for correcting myopia also carries the risk of acute angle closure, resulting in a remote increase in intraocular pressure and pupillary block [5]. Nevertheless, PACG can be treated by timely diagnosis and intervention to halt the angle closure process at the early stage to prevent irreversible blindness [6]. Treatments to open angle include laser peripheral iridotomy (LPI) and trabeculectomy, as well as cataract extraction with intraocular lens (IOL) implantation [7].

There are three main ways to assess the closure of the anterior chamber (AC) angle: gonioscopy, ultrasound biomicroscopy (UBM), and anterior segment optical coherence tomography (AS-OCT). Gonioscopy is a contact examination method and the current clinical gold-standard for evaluating the AC anatomy and angle closure, allowing for the assessment of the AC closure angle over 360 degrees [8]. However, gonioscopy is highly subjective with poor reproducibility, requiring extensive experiences of ophthalmologists, which limits its clinical use [9]. UBM is a non-invasive imaging tool and can provide high-resolution scanning images with an axial resolution of 25 µm and a lateral resolution of 50 µm, and a depth of penetration of approximately 4 mm [10] [11]. Anterior segment optical coherence tomography (AS-OCT) is an alternative in vivo imaging method for the assessment of angle closure, and can provide detailed two and three-dimensional cross-sectional images for quantitative measurement of AC angle parameters, including angle opening distance (AOD), trabecular iris space area (TISA), trabecular iris angle (TIA), and anterior chamber angle (ACA). Although AS-OCT and UBM both show similar diagnostic values and reproducibility [10], AS-OCT has undoubtedly faster acquisition speed and higher resolution (<15 µm) than UBM [12]. As shown in Fig. 1, iris root (IR) and scleral spur (SS) are two critical anatomical landmarks of AC, where IR refers to the vertex of AC angle recess, and SS represents the junction between the inner wall of TM and the sclera [12]. Accurate detection of IR and SS is crucial for quantifying AC angle parameters. Nevertheless, current commercial SS-OCT systems such as Heidelberg ANTERION [13] only provide semi-automatic measurements of AC angle parameters requiring the users to label IR and SS manually. However, manual identification of these landmarks is time-consuming and subjective [14]. Therefore, an automated algorithm for localization of IR and SS and for quantitative measurement of AC angle parameters can drastically alleviate the workload of ophthalmologists and reduce inter- and intra-observer variability.

 figure: Fig. 1.

Fig. 1. Illustration of AS-OCT imaging and anatomical structures of anterior chamber. Each AS-OCT volume consists of 6 refraction-corrected two-dimensional cross-sectional AS-OCT images acquired along 6 scan lines separated by 30 degrees apart. The shadow region * represents the area that is not refraction corrected. Red/green point: SS/IR.

Download Full Size | PDF

Deep learning has been proven to be a powerful technique for automated medical image analysis [1518]. Prior to this research, several methods based on deep learning were proposed to automatically detect SS in AS-OCT [14,1923] and UBM images [24] with visible SS. Pham et al. [19] firstly generated a small rectangular focal region with SS as its center, and used a modified U-Net [16] to segment SS. Xu et al. [20] adopted a deep neural network based on ResNet-18 [25] with several fully connected layers to directly predict the coordinates of SS in AS-OCT images. Later, the first angle closure glaucoma evaluation challenge (AGE challenge) [14] providing a public AS-OCT dataset was introduced to facilitate the development of automated algorithms for angle closure assessment, including SS localization and angle closure classification. The top-ranked methods [14,21,22] proposed in the AGE challenge have built coarse-to-fine frameworks for heatmap regression of SS. Comparing with AS-OCT images, the method [24] for SS localization in UBM images showed a more significant prediction error due to the low signal-to-noise ratio of UBM images. These studies have demonstrated exemplary performance with a small mean absolute localization error (MAE). However, due to the limitation of AS-OCT, the location of SS may be invisible and could not be determined in approximately 30% of AS-OCT images [26]. Therefore, previously reported algorithms [1923] that were only designed for SS localization could in fact generate errors for quantitative evaluation of AC angle parameters in AS-OCT images without detectable SS. For localization of IR, one previous method [27] extracted the position of IR from iris segmentation map and used IR as a salient landmark to locate the coarse AC angle region. However, the IR in the research [27] was not defined as the vertex of AC angle recess and could be wrongly recognized.

Although CNN-based methods [1518] have shown good performance in some medical image analysis tasks, their intrinsic local operations based on convolutions inevitably neglect the global semantic information which limits their performance in certain applications requiring global context. Recently, the transformer [28] with self-attention mechanism that is widely used in natural language processing (NLP) was introduced into the computer vision domain [29], and has achieved great success in modeling global and long-range image features [30]. Hence, combining CNN with the transformer [28] can potentially improve the performance of SS and IR detection by exploiting both local and global contexts.

The purpose of this study was to fulfill the unmet clinical need by developing a new method to automatically differentiate whether SS is present, and further detect IR and SS on a full spectrum of possible cases, and accurately calculate AC angle parameters including AOD, TISA, TIA, and ACA. To this end, we proposed a hybrid CNN & transformer model to capture both local and global contextual information of IR and SS, and extensively validated the method by comparing the algorithm performance with two experienced analysts as well as with other state-of-the-art methods for AS-OCT and medical image analysis. In addition, we assessed the implications of AC angle parameters during IOL and ICL implantation by applying the proposed method to a clinical case of PACG patient undergoing IOL implantation, and a patient with high myopia undergoing ICL implantation. Finally, we discuss the significance and limitations of our method in a clinical context.

2. Methods

2.1 Datasets

AS-OCT images from 362 eyes of 203 patients who underwent AS-OCT imaging to diagnose cataract and high myopia were collected by a commercially available AS-OCT system (ANTERION, Heidelberg Engineering GmbH, Heidelberg, Germany) that uses swept-source OCT technology with a center wavelength of 1300nm, an axial resolution of less than 10 µm and a transversal resolution of 45 µm [31], at Shanxi Eye Hospital, Taiyuan, China. The mean age of patients was 58.1$\pm$21.6 years. AS-OCT images with severe artifacts such as eyelids that hampered attempts to identify IR and SS were excluded. The final dataset included for image analysis were 1774 refraction-corrected AS-OCT images from 203 patients (an average of 8.7$\pm$7.2 images/patient acquired with different radial scans) and 1531 refraction uncorrected AS-OCT images from a subset of 66 patients to increase the diversity of training. Images with visible IR and SS respectively accounted for approximately 99% and 67% of the dataset, which was consistent with former research [26]. An experienced ophthalmologist with 10 years of clinical experience labeled IR and SS using an open source software (Labelme, MIT, Cambridge, MA) [32]. The dataset was divided into a training set consisting of 2145 images, 205 eyes, and 116 patients to train the model, a validation set with 572 images, 75 eyes, and 41 patients to evaluate the hyperparameters, and a test set containing 588 images from 82 eyes and 46 patients to evaluate the performance of the algorithm. The training and validation data contained both refraction-corrected and uncorrected images, but the test set only contained refraction-corrected images. The percentage of images with visible SS in the training, validation, and test set was 68%, 72%, and 59%, respectively. To evaluate inter-observer variability, the test set was labeled again by another ophthalmologist with 2 years of experience blinded to the results from the first ophthalmologist and the algorithm. Moreover, since it was necessary to extract the boundaries of AC for the measurement of AC angle parameters, we constructed another dataset by randomly sampling 220 AS-OCT images from the entire dataset for developing and validating a separate algorithm for AC segmentation. The dataset was also divided into a training set (60%), a validation set (20%), and a test set (20%). This study was approved by the local IRB offices of the authors’ institutions. All study patients provided informed consent and patients’ information was de-identified before analysis.

2.2 Methodology overview

The overall workflow of our method consists of AC segmentation, IR and SS detection, and AC angle parameter measurement (Fig. 2). We adopted the classic U-Net-like [16] CNN to segment the AC, which consists of a feature encoder module based on Efficientnet-B4 [33] and a feature decoder module with four de-convolutional layers to recover the spatial resolution reduced by the feature encoder module. The anterior iris surface and posterior corneal surface extracted from the AC segmentation map were used as reference boundaries for subsequent AC parameter measurements. At the stage of IR and SS detection, the original AS-OCT image was split along the median of AC into left and right sub-images which were subsequently input into a hybrid CNN & transformer model. With the boundaries of AC and the coordinate information of IR and SS, the AC angle parameters (AOD, TISA, TIA, and ACA) were subsequently calculated.

 figure: Fig. 2.

Fig. 2. Flow diagram illustrating the overall methodology for AC segmentation, IR and SS detection, and AC angle parameter measurement. The CNN model took the full image as the input to segment AC. For IR and SS detection, two sub-images cropped from the full image along the median of AC were individually input to a hybrid CNN & transformer model. Red/green point: SS/IR.

Download Full Size | PDF

2.3 Vision transformer

Over the past decade, CNNs have been widely used and have shown remarkable performance in various medical image analysis tasks [34]. It is universally acknowledged that CNNs can effectively extract local features by virtue of local receptive fields of convolution operations. However, CNN-based methods exhibit inherent limitations to model long-range dependencies because they focus on local patterns instead of global context. Recently, the transformer [28] was introduced into the computer vision field [29] and demonstrated a powerful capability to capture the long-range relation and to model global context by using the self-attention mechanism [28]. We first review the concepts of transformer below.

An input image $x_{in}\in R^{3\times H\times W}$ with height H and width W can be divided into a sequence of flattened 2D image patches {$x^{i}_{p}\in R^{P^2\cdot 3}| i = 1,\ldots, N$}, where the size of each patch is P$\times$P and N = $\frac {H\times W}{P^2}$ is the number of the resulting image patches. To be consistent with the refraction-corrected AS-OCT image with a three-channel input $x_{in}\in R^{3\times H\times W}$, the gray AS-OCT image $x\in R^{1\times H\times W}$ without refraction correction was converted into $x_{in}\in R^{3\times H\times W}$ by duplicating the image along the channel dimension. The flattened image patches are mapped into a latent D-dimensional embedding space using a trainable patch embedding projection E $\in R^{(P^2\cdot 3)\times D}$, generating a sequence of patch embeddings. A learnable position embedding E$_{pos}\in R^{N\times D}$ is then added to the patch embeddings to retain positional information:

$$z_0= [x^{1}_{p}E; x^{2}_{p}E; x^{3}_{p}E;\cdots;x^{N}_{p}E] + E_{pos}$$

The attention mechanism maps a query and a set of key-value pairs to the output using scaled dot-product attention and multi-head self-attention (MSA). Queries, keys, and values are packed into matrices Q, K, and V. The function of scaled dot-product attention with a scaling factor of $\frac {1}{\sqrt {d_k}}$ can be represented as follows:

$$Attention(Q, K, V) = Softmax(\frac{QK^T}{\sqrt{d_k}})V$$

Multi-head self-attention linearly projects the queries, keys, and values h times to $d_k$, $d_k$, and $d_v$ dimensions:

$$\begin{array}{c} Multihead(Q, K, V) = Concat(head_1, head_2, head_3, \ldots, head_h)W^O\\ where\ head_i = Attention(QW^{Q}_{i}, KW^{K}_{i}, VW^{V}_{i}) \end{array}$$
where $W^{Q}_{i}\in R^{d_{model}\times d_k}$, $W^{K}_{i}\in R^{d_{model}\times d_k}$, $W^{V}_{i}\in R^{d_{model}\times d_v}$, $W^O\in R^{hd_v\times d_{model}}$ are parametric matrices of the projection. The transformer encoder contains L transformer layers with a MSA and a multi-layer perceptron (MLP) block. The output of the l-th transformer layer can be described as follows:
$$z^{'}_{l} = MSA(LN(z_{l-1})) + z_{l-1}$$
$$z_l = MLP(LN(z^{'}_{l})) + z^{'}_{l}$$
where LN($\cdot$) represents the layer normalization operator. The final output $z_L\in R^{N\times D}$ of the transformer encoder is reshaped to $Z\in R^{\frac {W}{P}\times \frac {H}{P}\times D}$.

2.4 Hybrid CNN & transformer model

Accurate detection and localization of SS require a combination of global contexts and local fine-grained features. Therefore, we designed a hybrid CNN & transformer model as shown in Fig. 3, where the feature encoder consists of a CNN-based branch and a transformer-based branch. The two branches take the same image as the input and respectively output a feature map encoding both fine-grained local context and long-range global information.

 figure: Fig. 3.

Fig. 3. Framework of the hybrid CNN & transformer model. The encoder module of the network contains a local branch based on Efficientnet-B4 and a global branch based on transformer. The same image is input into the local branch and global branch, respectively outputting a feature map with fine-grained local context and one with global information. The two feature maps are concatenated in channels, and then upsampled by three decoder blocks and a de-convolutional layer, generating two heatmaps with the same width and height to precisely localize IR and SS, respectively.

Download Full Size | PDF

The CNN-based branch is based on the widely used Efficientnet-B4 [33]. The transformer-based branch consists of L transformer layers. The feature maps generated by the CNN-based and the transformer-based branches are concatenated along channels. A 1$\times$1 convolution operation is used to fuse the combined feature maps. Following the U-Net-like architectures [15], the fused feature maps containing the high-level semantic information are upsampled to recover the spatial information of the target landmarks through several decoder blocks. In addition, skip connections with addition operations fusing the feature maps from the encoder and decoder blocks are further added to preserve more detailed information and remedy resolution loss caused by max-pooling and convolutional layers. After a de-convolution operation in the final layer, two heatmaps respectively containing the position information of IR and SS are generated, which are further filtered using a Gaussian kernel with a size of 15$\times$15, and the positions of the peak values are extracted as the precise coordinates of IR and SS. Finally, we exclude the predicted IR and SS with peak values <0.5, which indicates the IR and SS are invisible in the input OCT images.

2.5 AC angle parameter measurement

In this study, we only focused on the measurement of AOD, TISA, and TIA at 500 µm, as shown in Fig. 2, and the corresponding parameters at 750 µm can be similarly calculated. AOD at 500 µm (AOD 500) represents the length of the line perpendicular to the plane of TM from the TM-500 (a point on the TM at 500 µm from the SS) to the opposing iris [35]. TIA at 500 µm (TIA 500) is defined as the angle measured with the apex in the SS and the arms of the angle passing through TM-500 and a perpendicular point on the opposing iris [35]. TISA at 500 µm (TISA 500) is defined as the trapezoidal area enclosed by AOD 500, the inner cornea-scleral wall, iris surface, and a line drawn from the SS perpendicular to the plane of TM [12]. ACA refers to the angle between the iris tangential line and that of the posterior corneal surface with IR [35].

2.6 Experimental details

Our model was implemented on PyTorch and Ubuntu 16.04 with 4 Nvidia Titan XP GPUs and was trained for 100 epochs with the mini-batch size set to 4. The input images were resized to 512$\times$512. For training, we used the Adam optimizer with the mean square error (MSE) loss and we set the initial learning rate to $2\times 10^{-4}$. The learning rate was multiplied by 0.5 if the loss value was not changed over 2 epochs. The training was early stopped if the loss value was not changed over 5 epochs. The hyper-parameters of the transformer-based branch were consistent with ViT [29], where the size P$\times$P of image patches was set to 16$\times$16, D, $d_{model}$, $d_k$, $d_v$, h, and L were set to 768, 768, 64, 64, 12, and 12, respectively. We did not use any image pre-processing or data augmentation.

2.7 Algorithm evaluation and statistical analysis

Detected IR and SS were classified as true positives (TP), false positives (FP), and false negatives (FN) with the human annotations as the ground truth. The prediction was classified as TP if the predicted landmark location was within 80 µm (approximately 20% of the length of TM) distance both in the X and Y-axis with the ground truth, consistent with previous studies [36]. The detection accuracy of IR and SS was evaluated using the metrics of precision, sensitivity, and F1 score computed as follows:

$$\begin{array}{c} Precision= \frac{TP}{TP+FP}\quad Sensitivity= \frac{TP}{TP+FN}\quad \\ F1\ score= \frac{2\times precision\times sensitivity}{precision + sensitivity} \end{array}$$

We also used the mean absolute localization error (MAE) to evaluate the localization accuracy of IR and SS by calculating the distance between the prediction and the ground truth. Given the predicted coordinate (x, y) and the ground truth ($x_g$, $y_g$), the MAE can be calculated as:

$$MAE = \sqrt{{(x-x_g)}^2+{(y-y_g)}^2}$$

Dice coefficient was used to compute the similarity between the AC segmentation map and the corresponding ground truth:

$$Dice = \frac{2\times TP}{2\times TP+FP+FN}$$
where TP refers to the number of pixels that were correctly identified as part of the AC in both the ground truth and the model’s output. FP represents the number of pixels that the model incorrectly labeled as part of the AC but were not part of the ground truth, while FN indicates the number of pixels that were part of the ground truth but were not detected by the model.

We compared our proposed method with the winner of the AGE challenge [14]. The winner of the AGE challenge adopted a coarse-to-fine localization framework which consisted of two CNNs with the same Efficientnet-B5 [33] based architecture to supervise the regression of the landmarks, achieving the best localization accuracy for SS in the AGE challenge. We also compared our method with HRNet [37], U-Net [16], and recently proposed TransUNet [30]. HRNet was a widely used CNN for heatmap regression of key points. The original U-Net and TransUNet for medical semantic segmentation were modified with heatmap regression with the MSE loss.

Bland-Altman analysis was used to compare the difference between automatic and manual AC angle parameter measurement.

2.8 Clinical application of AC angle parameter measurement in IOL and ICL implantation

Eyes with angle closure inherently have higher intraocular pressure caused by a narrow AC angle and a thick and anteriorly vaulted lens than normal eyes [38]. However, the most common treatments for PACG, iridectomy and iridotomy, do not always permanently control intraocular pressure because of released pigment granules and iris tissue fragments in the case of laser iridotomy [39]. It was reported that cataract extraction might have a beneficial clinical outcome [39] in long-term intraocular pressure control. Phacoemulsification and IOL implantation are widely used in cataract extraction [40]. Therefore, measuring the AC parameters before and after cataract extraction with IOL implantation allows for assessing its clinical significance for PACG patients. We demonstrated the application of our method towards this need by comparing the AC angle parameters over 360 degrees of a left eye in a 56-year-old male patient with PACG before and after phacoemulsification and IOL implantation.

ICL is a toric posterior chamber phakic intraocular lens to correct myopia, hyperopia, and astigmatism [41]. After ICL insertion, acute angle closure may be caused by elevated intraocular pressure secondary to anterior vaulting of the ICL [42]. Moreover, a previous research [43] reported that the intraocular pressure elevation following ICL implantation may fluctuate over time and reach a peak several days postoperatively. Although ICL implantation is currently considered as a safe and effective treatment for patients with high myopia, its long-term safety and stability warrant further investigation [44]. We aimed to assess the variation of intraocular pressure elevation by measuring the AC angle parameters in a right eye from a 30-year-old female patient undergoing ICL implantation to treat high myopia.

3. Results

3.1 Algorithm performance on IR and SS detection and localization

With 588 test AS-OCT images from 82 eyes and 46 patients, our proposed method achieved a precision of 0.941, a sensitivity of 0.914, a F1 score of 0.927, and a MAE of 37.1$\pm$25.3 µm for IR, and a precision of 0.805, a sensitivity of 0.847, a F1 score of 0.826, and a MAE of 41.4$\pm$29.4 µm for SS (Table 1). Comparing with the winner of AGE and other state-of-the-art methods in the field of medical image analysis, our method achieved superior performance in both detection and localization accuracy. The results of our proposed IR and SS detection model for representative cases are presented in Fig. 4, alongside a comparison with other methods. Our proposed method exhibited accurate detection of IR and SS, achieving a smaller MAE compared to other methods in the general case (Fig. 4(a)), as well as in challenging scenarios, such as partial blockage by the eyelid (Fig. 4(b)) and angle closure (Fig. 4(c)). Furthermore, the proposed method automatically excluded invisible SS in AS-OCT images (Fig. 4(d)).

 figure: Fig. 4.

Fig. 4. Results of IR and SS detection of our proposed algorithm and other methods in four typical cases. (a): Normal AC angle. (a) Partial blockage by eyelids. (c): Severe AC angle closure. (d): Invisible SS. Red/green points: SS/IR.

Download Full Size | PDF

Tables Icon

Table 1. Performance evaluation of the proposed method

To validate the effectiveness of including both the local and global branches in the feature encoder for IR and SS detection, we conducted two ablation studies by eliminating each branch of the proposed model and compared the resulting performance with the full model. As shown in Table 1, ablating either of the branches would deteriorate all the classification and localization metrics, which proved the importance of including both the CNN and transformer modules for IR and SS detection and localization.

To evaluate inter-observer variability, we compared the agreements between the algorithm and the two analysts, and the results are shown in Fig. 5. The precision, sensitivity, and F1 score of [analyst A VS. analyst B, analyst A VS. algorithm, analyst B VS. algorithm] for detecting IR are [0.941, 0.949, 0.924], [0.946, 0.959, 0.918], and [0.943, 0.954, 0.921], respectively. The precision, sensitivity, and F1 score of [analyst A VS. analyst B, analyst A VS. algorithm, analyst B VS. algorithm] for detecting SS are [0.933, 0.832, 0.830], [0.839, 0.889, 0.894] and [0.883, 0.860, 0.861], respectively. The MAE of [analyst A VS. analyst B, analyst A VS. algorithm, analyst B VS. algorithm] of IR and SS are [35.9$\pm$19.8 µm, 35.2$\pm$18.8 µm, 36.8$\pm$19.5 µm] and [31.2$\pm$18.3 µm, 35.8$\pm$20.1 µm, 34.3$\pm$18.1 µm] respectively. These results demonstrated that the algorithm achieved similar agreements with human analysts for both detection and localization of IR and SS.

 figure: Fig. 5.

Fig. 5. Comparison between the algorithm and different analysts for the detection (Left) and localization (Right) of IR and SS.

Download Full Size | PDF

3.2 Evaluation of quantitative measurements

We calculated the four AC angle parameters using the coordinates of the detected IR and SS and the extracted AC boundaries. The AC segmentation model achieved a Dice coefficient of 0.991$\pm$0.002 on the test set. Figure 6 shows that the mean differences of AOD 500, TISA 500, TIA 500, and ACA between automatic and manual measurement are −0.005$\pm$0.393 mm, 0.000$\pm$0.199 mm$^2$, −0.1$\pm$9.3°, and −0.2$\pm$5.0°, respectively, and there are slightly worse agreements between the automatic and manual measurement for the large open angle of AC. This is because the large open angle amplified the prediction error of SS localization.

 figure: Fig. 6.

Fig. 6. Bland-Altman analysis between automatic and manual measurement of AC angle parameters including (a) AOD 500, (b) TISA 500, (c) TIA 500, and (d) ACA.

Download Full Size | PDF

3.3 Clinical application of AC angle parameter measurement in IOL and ICL implantation

We analyzed the AC angle parameters from an eye of a patient with PACG before and after cataract extraction (Fig. 7). Figure 7(a)-(d) show that IOL implantation resulted in increased AC angle parameters. The mean pre- and post-operative [AOD 500, TISA 500, TIA 500, ACA] are [0.392$\pm$0.071 mm, 0.136$\pm$0.038 mm$^2$, 37.5$\pm$5.1$^{\circ }$, 32.8$\pm$5.7$^{\circ }$] and [0.669$\pm$0.145 mm, 0.228$\pm$0.049 mm$^2$, 52.2$\pm$5.4$^{\circ }$, 47.9$\pm$7.1$^{\circ }$], where the post-operative parameters were increased by [70.2$\pm$16.1%, 72.5$\pm$30.3%, 40.0$\pm$9.7%, 47.2$\pm$13.5%] respectively compared to the pre-operative values (p<0.0001).

 figure: Fig. 7.

Fig. 7. AC angle parameter measurement before and after IOL (a-d) and ICL (e-h) implantation. The AC angle parameters over 360 degrees were interpolated by applying cubic-spline interpolation. Blue/orange points: measured AC angle parameters of pre- and post-operation. Green points: interpolated AC angle parameters. Blue and orange curves respectively represent fitted AC angle parameter curves of pre- and post-operation.

Download Full Size | PDF

By contrast, as shown in Fig. 7(e)-(h), the mean pre- and post-operative [AOD 500, TISA 500, TIA 500, ACA] for the patient undergoing ICL implantation to treat high myopia are [0.655$\pm$0.169 mm, 0.241$\pm$0.057 mm$^2$, 51.6$\pm$6.3$^{\circ }$, 42.6$\pm$5.8$^{\circ }$] and [0.325$\pm$0.128 mm, 0.101$\pm$0.056 mm$^2$, 31.9$\pm$10.8$^{\circ }$, 24.8$\pm$4.7$^{\circ }$], where the post-operative AC parameters were reduced by [50.4$\pm$17.9%, 58.2$\pm$23.5%, 38.5$\pm$19.0%, 41.9$\pm$5.2%] respectively compared to the pre-operative values (p<0.0001). This indicated that the ICL implantation caused acute angle closure, which may elevate intraocular pressure that is one of the main risk factors for PACG.

4. Discussion

In this paper, we proposed a novel hybrid CNN & transformer model for automated detection and localization of IR and SS in AS-OCT images. The model incorporates the knowledge of fine-grained local contexts and long-range global dependencies into its feature encoder and has been proven to be effective and robust by testing a full spectrum of possible cases in the clinic (Table 1). The proposed method is advantageous over existing methods [14,16,30,37] both in terms of performance, and by taking into consideration of the visibility of SS in AS-OCT images for the first time. Previous studies [20] [24] for SS detection reported that the CNN model trained on the reference labels provided by one human analyst would not generalize to other analysts. However, we demonstrated that our proposed method achieved similar performance compared with different analysts (Fig. 5).

The automatic detection and localization of IR and SS are preliminary steps for AC angle parameter measurement which are essential for assessing acute angle closure caused by ICL implantation. Implantation of the ICL narrows the angle width and depth [45] and high intraocular pressure can be developed postoperatively after ICL implantation [43] in some eyes, which are important risk factors for PACG. One previous clinical study [46] demonstrated that AC angle parameters measured by AS-OCT have a strong correlation with intraocular pressure. Automatic detection and localization of IR and SS can help physicians quickly evaluate the difference of AC angle parameters between pre- and post-operation, providing the clinical reference for managing the acute PACG, as demonstrated in this study shown in Fig. 7(e)-(h).

Contrary to ICL implantation, phacoemulsification and IOL implantation used in cataract treatment can effectively eliminate the pupillary block and well control the intraocular pressure in PACG patients [47], as was also demonstrated by the increase of AC angle parameters between pre- and post-operation in Fig. 7(a)-(d) of this study. The post-operative intraocular pressure peak is a significant problem that needs to be detected early and treated vigorously [47]. Therefore, automatic AC angle parameter measurement facilitates the monitoring of the progression of angle narrowing and angle closure over time, which can also help find the intraocular pressure peak.

The automatic detection and localization of IR and SS have great significance for standardizing the OCT-based evaluation of angle closure in patients with PACG. The determination of narrow angle is based on the space between the TM and the iris, which is interpreted qualitatively in AS-OCT images in current clinical practice [12] [35] [48]. Nevertheless, there is no consensus on which parameters or cut-off values to use to identify a narrow angle [49]. The previous clinical research [50] determined the thresholds of AOD and TISA for distinguishing open vs. narrow angle eyes with gonioscopy examination as the reference standard, and found that AOD had a better discriminative ability for detecting narrow angles than TISA. Automatic and quantitative measurement of AC angle parameters allows for developing and validating OCT-based diagnostic criteria for narrow angles in large clinical studies.

The automatic measurement of AC parameters is also crucial for determining the anatomical risk factors contributing to the pathogenesis of PACG. In addition to the static anatomic factors, such as shallow anterior chamber depth, greater lens thickness, and larger lens vault, a recent study [51] demonstrated that dynamic changes in AC angle parameters with light-to-dark transition are correlated with the disease progression of PACG. However, the clinical application of this finding is currently limited by the manual analysis of AC angle parameters in AS-OCT images. Automatic analysis of AC angle parameters could contribute to monitoring the dynamic changes of AC angle anatomy and identifying the rick factors associated with the progression of angle narrowing in patients with PACG.

This study has limitations. AS-OCT images in this research were collected from one AS-OCT imaging system (ANTERION). Transfer learning [52] can be implemented for generalizing our method to AS-OCT images from other commercial systems with different imaging configurations and image quality. Therefore, future multicenter study is required to improve the performance and to test the generalization ability of the proposed deep learning method.

5. Conclusion

In summary, we proposed a novel hybrid CNN & transformer model to automatically detect and localize IR and SS, and to automatically compute the AC angle parameters in AS-OCT images. The results demonstrated that our method achieved significantly better performance than state-of-the-art methods for detection and localization of IR and SS, and is highly accurate for AC angle parameter measurement, and has a great potential to be used in clinic for the screening, prevention, and treatment of PACG.

Funding

National Natural Science Foundation of China (62075033, 62135002, 81971697); Shanxi Eye Hospital (B201804); Shanxi Scholarship Council of China (2021-174); Sichuan Province Science and Technology Support Program (23ZDYF1600); University of Electronic Science and Technology of China (ZYGX2021YGCX004); Fundamental Research Funds for the Central Universities (ZYGX2021J009).

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. Y.-C. Tham, X. Li, T. Y. Wong, H. A. Quigley, T. Aung, and C.-Y. Cheng, “Global prevalence of glaucoma and projections of glaucoma burden through 2040: a systematic review and meta-analysis,” Ophthalmology 121(11), 2081–2090 (2014). [CrossRef]  

2. E. W. Chan, X. Li, Y.-C. Tham, J. Liao, T. Y. Wong, T. Aung, and C.-Y. Cheng, “Glaucoma in Asia: regional prevalence variations and future projections,” Br. J. Ophthalmol. 100(1), 78–85 (2016). [CrossRef]  

3. W. Song, L. Shan, F. Cheng, P. Fan, L. Zhang, W. Qu, Q. Zhang, and H. Yuan, “Prevalence of glaucoma in a rural northern china adult population: a population-based survey in Kailu county, inner Mongolia,” Ophthalmology 118(10), 1982–1988 (2011). [CrossRef]  

4. C. Wright, M. A. Tawfik, M. Waisbourd, and L. J. Katz, “Primary angle-closure glaucoma: an update,” Acta Ophthalmol. 94(3), 217–225 (2016). [CrossRef]  

5. A. Frost, D. J. Ritter, A. Trotter, and M. S. Pulia, “Acute angle-closure glaucoma secondary to a phakic intraocular lens, an ophthalmic emergency,” Clin. Pract. Cases Emerg. Medicine 3(2), 137–139 (2019). [CrossRef]  

6. A. Azuara-Blanco, J. Burr, C. Ramsay, D. Cooper, P. J. Foster, D. S. Friedman, G. Scotland, M. Javanbakht, C. Cochrane, and J. Norrie, “Effectiveness of early lens extraction for the treatment of primary angle-closure glaucoma (eagle): a randomised controlled trial,” Lancet 388(10052), 1389–1397 (2016). [CrossRef]  

7. X. Sun, Y. Dai, Y. Chen, D.-Y. Yu, S. J. Cringle, J. Chen, X. Kong, X. Wang, and C. Jiang, “Primary angle closure glaucoma: what we know and what we don’t know,” Prog. Retinal Eye Res. 57, 26–45 (2017). [CrossRef]  

8. M. Rigi, N. P. Bell, D. A. Lee, L. A. Baker, A. Z. Chuang, D. Nguyen, V. R. Minnal, R. M. Feldman, and L. S. Blieden, “Agreement between gonioscopic examination and swept source fourier domain anterior segment optical coherence tomography imaging,” J. Ophthalmol. 2016, 1–8 (2016). [CrossRef]  

9. R. N. Weinreb, T. Aung, and F. A. Medeiros, “The pathophysiology and treatment of glaucoma: a review,” Jama 311(18), 1901–1911 (2014). [CrossRef]  

10. I. Riva, E. Micheletti, F. Oddone, C. Bruttini, S. Montescani, G. De Angelis, L. Rovati, R. N. Weinreb, and L. Quaranta, “Anterior chamber angle assessment techniques: a review,” J. Clin. Med. 9(12), 3814 (2020). [CrossRef]  

11. T. Mansoori and N. Balakrishna, “Anterior segment morphology in primary angle closure glaucoma using ultrasound biomicroscopy,” J. Curr. Glaucoma Pract. 11(3), 86–91 (2017). [CrossRef]  

12. S. Radhakrishnan, D. Huang, and S. D. Smith, “Optical coherence tomography imaging of the anterior chamber angle,” Ophthalmol. Clin. North Am. 18(3), 375–381 (2005). [CrossRef]  

13. J. S. Asam, M. Polzer, A. Tafreshi, N. Hirnschall, and O. Findl, “Anterior segment OCT,” High resolution imaging in microscopy and ophthalmology: new frontiers in biomedical optics, 285–299 (2019).

14. H. Fu, F. Li, X. Sun, et al., “Age challenge: angle closure glaucoma evaluation in anterior segment optical coherence tomography,” Med. Image Anal. 66, 101798 (2020). [CrossRef]  

15. G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, J. A. Van Der Laak, B. Van Ginneken, and C. I. Sánchez, “A survey on deep learning in medical image analysis,” Med. Image Anal. 42, 60–88 (2017). [CrossRef]  

16. O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (Springer, 2015), pp. 234–241.

17. D. Shen, G. Wu, and H.-I. Suk, “Deep learning in medical image analysis,” Annu. Rev. Biomed. Eng. 19(1), 221–248 (2017). [CrossRef]  

18. S. Suganyadevi, V. Seethalakshmi, and K. Balasamy, “A review on deep learning in medical image analysis,” Int. J. Multimed. Inf. Retr. 11(1), 19–38 (2022). [CrossRef]  

19. T. H. Pham, S. K. Devalla, A. Ang, Z.-D. Soh, A. H. Thiery, C. Boote, C.-Y. Cheng, M. J. Girard, and V. Koh, “Deep learning algorithms to isolate and quantify the structures of the anterior segment in optical coherence tomography images,” Br. J. Ophthalmol. 105(9), 1231–1237 (2021). [CrossRef]  

20. B. Y. Xu, M. Chiang, A. A. Pardeshi, S. Moghimi, and R. Varma, “Deep neural network for scleral spur detection in anterior segment oct images: the chinese american eye study,” Transl. Vis. Sci. Technol. 9(2), 18 (2020). [CrossRef]  

21. X. Tao, C. Yuan, C. Bian, Y. Li, K. Ma, D. Ni, and Y. Zheng, “The winner of age challenge: Going one step further from keypoint detection to scleral spur localization,” in 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI) (IEEE, 2021), pp. 1284–1287.

22. P. Li, L. Geng, W. Zhu, F. Shi, and X. Chen, “Automatic angle-closure glaucoma screening based on the localization of scleral spur in anterior segment oct,” in 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI) (IEEE, 2020), pp. 1387–1390.

23. K. Bolo, G. Apolo, M. Chiang, A. Pardeshi, B. Burkemper, X. Xie, A. Huang, M. Simonovsky, and B. Xu, “Automated expert-level scleral spur detection and quantitative biometric analysis on the anterion anterior segment OCT system,” Invest. Ophthalmol. Vis. Sci. 2022, 3351 (2022).

24. W. Wang, L. Wang, T. Wang, X. Wang, S. Zhou, J. Yang, and S. Lin, “Automatic localization of the scleral spur using deep learning and ultrasound biomicroscopy,” Transl. Vis. Sci. Technol. 10(9), 28 (2021). [CrossRef]  

25. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition (IEEE, 2016), pp. 770–778.

26. L. M. Sakata, R. Lavanya, D. S. Friedman, H. T. Aung, S. K. Seah, P. J. Foster, and T. Aung, “Assessment of the scleral spur in anterior segment optical coherence tomography images,” Arch. Ophthalmol. 126(2), 181–185 (2008). [CrossRef]  

27. H. Hao, Y. Zhao, Q. Yan, R. Higashita, J. Zhang, Y. Zhao, Y. Xu, F. Li, X. Zhang, and J. Liu, “Angle-closure assessment in anterior segment oct images via deep learning,” Med. Image Anal. 69, 101956 (2021). [CrossRef]  

28. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems30, 1 (2017).

29. A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929 (2020). [CrossRef]  

30. J. Chen, Y. Lu, Q. Yu, X. Luo, E. Adeli, Y. Wang, L. Lu, A. L. Yuille, and Y. Zhou, “Transunet: transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306 (2021). [CrossRef]  

31. Y. Feng, D. Z. Reinstein, T. Nitter, T. J. Archer, C. McAlinden, X. Chen, G. Bertelsen, T. P. Utheim, and A. Stojanovic, “Heidelberg anterion swept-source OCT corneal epithelial thickness mapping: Repeatability and agreement with optovue avant,” J. Refract. Surg. 38(6), 356–363 (2022). [CrossRef]  

32. B. C. Russell, A. Torralba, K. P. Murphy, and W. T. Freeman, “Labelme: a database and web-based tool for image annotation,” Int. J. Comput. Vis. 77(1-3), 157–173 (2008). [CrossRef]  

33. M. Tan and Q. Le, “Efficientnet: rethinking model scaling for convolutional neural networks,” in International Conference on Machine Learning (PMLR, 2019), pp. 6105–6114.

34. Z. Li, F. Liu, W. Yang, S. Peng, and J. Zhou, “A survey of convolutional neural networks: analysis, applications, and prospects,” in IEEE transactions on neural networks learning systems (IEEE, 2021).

35. C. J. Pavlin, K. Harasiewicz, and F. S. Foster, “Ultrasound biomicroscopy of anterior segment structures in normal and glaucomatous eyes,” Am. J. Ophthalmol. 113(4), 381–389 (1992). [CrossRef]  

36. R. J. Cumba, S. Radhakrishnan, N. P. Bell, K. S. Nagi, A. Z. Chuang, S. C. Lin, K. A. Mankiewicz, and R. M. Feldman, “Reproducibility of scleral spur identification and angle measurements using fourier domain anterior segment optical coherence tomography,” J. Ophthalmol. 2012, 1–14 (2012). [CrossRef]  

37. K. Sun, B. Xiao, D. Liu, and J. Wang, “Deep high-resolution representation learning for human pose estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (IEEE, 2019), pp. 5693–5703.

38. W.-W. Su, P. Y.-F. Chen, C.-H. Hsiao, and H. S.-L. Chen, “Primary phacoemulsification and intraocular lens implantation for acute primary angle-closure,” PLoS One 6(5), e20056 (2011). [CrossRef]  

39. P. C. Jacobi, T. S. Dietlein, C. Lüke, B. Engels, and G. K. Krieglstein, “Primary phacoemulsification and intraocular lens implantation for acute angle-closure glaucoma,” Ophthalmology 109(9), 1597–1603 (2002). [CrossRef]  

40. R. J. Olson, N. Mamalis, L. Werner, and D. J. Apple, “Cataract treatment in the beginning of the 21st century,” Ophthalmology 136(1), 146–154 (2003). [CrossRef]  

41. J. M. Vetter, M. Tehrani, and H. B. Dick, “Surgical management of acute angle-closure glaucoma after toric implantable contact lens implantation,” J. Cataract Refractive Surg. 32(6), 1065–1067 (2006). [CrossRef]  

42. K. C. Chan, W. Birchall, T. B. Gray, and A. P. Wells, “Acute angle closure after implantable contact lens insertion unresponsive to surgical peripheral iridectomy,” J. Cataract Refractive Surg. 34(4), 696–699 (2008). [CrossRef]  

43. S. Almalki, A. Abubaker, N. A. Alsabaani, and D. P. Edward, “Causes of elevated intraocular pressure following implantation of phakic intraocular lenses for myopia,” Int. Ophthalmol. 36(2), 259–265 (2016). [CrossRef]  

44. L. Niu, H. Miao, T. Han, L. Ding, X. Wang, and X. Zhou, “Visual outcomes of visian icl implantation for high myopia in patients with shallow anterior chamber depth,” BMC Ophthalmol. 19(1), 121 (2019). [CrossRef]  

45. Y. S. Chun, I. K. Park, H. I. Lee, J. H. Lee, and J. C. Kim, “Iris and trabecular meshwork pigment changes after posterior chamber phakic intraocular lens implantation,” J. Cataract Refractive Surg. 32(9), 1452–1458 (2006). [CrossRef]  

46. B. Y. Xu, B. Burkemper, J. P. Lewinger, X. Jiang, A. A. Pardeshi, G. Richter, M. Torres, R. McKean-Cowdin, and R. Varma, “Correlation between intraocular pressure and angle configuration measured by oct: the chinese American eye study,” Ophthalmol. Glaucoma 1(3), 158–166 (2018). [CrossRef]  

47. C.-H. Yang and T. Hung, “Intraocular lens position and anterior chamber angle changes after cataract extraction in eyes with primary angle-closure glaucoma,” J. Cataract Refractive Surg. 23(7), 1109–1113 (1997). [CrossRef]  

48. M. Kim, K. H. Park, T.-W. Kim, and D. M. Kim, “Changes in anterior chamber configuration after cataract surgery as measured by anterior segment optical coherence tomography,” Korean J. Ophthalmol. 25(2), 77–83 (2011). [CrossRef]  

49. A. Jindal, I. Ctori, G. Virgili, E. Lucenteforte, and J. G. Lawrenson, “Non-contact methods for the detection of people at risk of primary angle closure glaucoma,” The Cochrane Database Syst. Rev. 2018(2), 1 (2018). [CrossRef]  

50. E. K. Melese, J. D. Chan, L. S. Blieden, A. Z. Chuang, L. A. Baker, N. P. Bell, and R. M. Feldman, “Determination and validation of thresholds of anterior chamber parameters by dedicated anterior segment optical coherence tomography,” Am. J. Ophthalmol. 169, 208–217 (2016). [CrossRef]  

51. J. Lin, Z. Wang, C. Chung, J. Xu, M. Dai, and J. Huang, “Dynamic changes of anterior segment in patients with different stages of primary angle-closure in both eyes and normal subjects,” PLoS One 12(5), e0177769 (2017). [CrossRef]  

52. K. Weiss, T. M. Khoshgoftaar, and D. Wang, “A survey of transfer learning,” J. Big Data 3(1), 9–40 (2016). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (7)

Fig. 1.
Fig. 1. Illustration of AS-OCT imaging and anatomical structures of anterior chamber. Each AS-OCT volume consists of 6 refraction-corrected two-dimensional cross-sectional AS-OCT images acquired along 6 scan lines separated by 30 degrees apart. The shadow region * represents the area that is not refraction corrected. Red/green point: SS/IR.
Fig. 2.
Fig. 2. Flow diagram illustrating the overall methodology for AC segmentation, IR and SS detection, and AC angle parameter measurement. The CNN model took the full image as the input to segment AC. For IR and SS detection, two sub-images cropped from the full image along the median of AC were individually input to a hybrid CNN & transformer model. Red/green point: SS/IR.
Fig. 3.
Fig. 3. Framework of the hybrid CNN & transformer model. The encoder module of the network contains a local branch based on Efficientnet-B4 and a global branch based on transformer. The same image is input into the local branch and global branch, respectively outputting a feature map with fine-grained local context and one with global information. The two feature maps are concatenated in channels, and then upsampled by three decoder blocks and a de-convolutional layer, generating two heatmaps with the same width and height to precisely localize IR and SS, respectively.
Fig. 4.
Fig. 4. Results of IR and SS detection of our proposed algorithm and other methods in four typical cases. (a): Normal AC angle. (a) Partial blockage by eyelids. (c): Severe AC angle closure. (d): Invisible SS. Red/green points: SS/IR.
Fig. 5.
Fig. 5. Comparison between the algorithm and different analysts for the detection (Left) and localization (Right) of IR and SS.
Fig. 6.
Fig. 6. Bland-Altman analysis between automatic and manual measurement of AC angle parameters including (a) AOD 500, (b) TISA 500, (c) TIA 500, and (d) ACA.
Fig. 7.
Fig. 7. AC angle parameter measurement before and after IOL (a-d) and ICL (e-h) implantation. The AC angle parameters over 360 degrees were interpolated by applying cubic-spline interpolation. Blue/orange points: measured AC angle parameters of pre- and post-operation. Green points: interpolated AC angle parameters. Blue and orange curves respectively represent fitted AC angle parameter curves of pre- and post-operation.

Tables (1)

Tables Icon

Table 1. Performance evaluation of the proposed method

Equations (8)

Equations on this page are rendered with MathJax. Learn more.

z 0 = [ x p 1 E ; x p 2 E ; x p 3 E ; ; x p N E ] + E p o s
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d k ) V
M u l t i h e a d ( Q , K , V ) = C o n c a t ( h e a d 1 , h e a d 2 , h e a d 3 , , h e a d h ) W O w h e r e   h e a d i = A t t e n t i o n ( Q W i Q , K W i K , V W i V )
z l = M S A ( L N ( z l 1 ) ) + z l 1
z l = M L P ( L N ( z l ) ) + z l
P r e c i s i o n = T P T P + F P S e n s i t i v i t y = T P T P + F N F 1   s c o r e = 2 × p r e c i s i o n × s e n s i t i v i t y p r e c i s i o n + s e n s i t i v i t y
M A E = ( x x g ) 2 + ( y y g ) 2
D i c e = 2 × T P 2 × T P + F P + F N
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.