Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Phase unwrapping based on a residual en-decoder network for phase images in Fourier domain Doppler optical coherence tomography

Open Access Open Access

Abstract

To solve the phase unwrapping problem for phase images in Fourier domain Doppler optical coherence tomography (DOCT), we propose a deep learning-based residual en-decoder network (REDN) method. In our approach, we reformulate the definition for obtaining the true phase as obtaining an integer multiple of 2π at each pixel by semantic segmentation. The proposed REDN architecture can provide recognition performance with pixel-level accuracy. To address the lack of phase images that are noise and wrapping free from DOCT systems for training, we used simulated images synthesized with DOCT phase image background noise features. An evaluation study on simulated images, DOCT phase images of phantom milk flowing in a plastic tube and a mouse artery, was performed. Meanwhile, a comparison study with recently proposed deep learning-based DeepLabV3+ and PhaseNet methods for signal phase unwrapping and traditional modified networking programming (MNP) method was also performed. Both visual inspection and quantitative metrical evaluation based on accuracy, specificity, sensitivity, root-mean-square-error, total-variation, and processing time demonstrate the robustness, effectiveness and superiority of our method. The proposed REDN method will benefit accurate and fast DOCT phase image-based diagnosis and evaluation when the detected phase is wrapped and will enrich the deep learning-based image processing platform for DOCT images.

© 2020 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Fourier domain optical coherence tomography (FD-OCT) is a high-resolution non-invasive 3D imaging modality, which has been widely used for biomedical research and clinical studies. The measurement of blood flow with Doppler OCT (DOCT) systems plays great important roles in both disease diagnosis and clinical intraoperative evaluation [14]. Phase images from DOCT that can provide important underlying biophysical information often suffer from phase wrapping problems when detected phase range is out to the dynamic range of 2π.

Phase unwrapping is a basic signal processing issue to resolve true phase value from detected phase value wrapped in the range of (-π, π]. Ideally phase unwrapping can be achieved by adding an appropriate integer multiple of 2π to each pixel based on the phase difference between the adjacent pixels. However, phase unwrapping becomes very challenging when severe noises, shadow effect and phase discontinuities are existent. Specially, DOCT phase images carrying the feature information of the blood flow suffers from severe speckle noise and random noise. High noise level in DOCT phase images increases the complexity of the phase unwrapping procedure.

Many ingenious phase unwrapping methods have been proposed over the past years. The first category includes those path-following methods, which contains quality-guided algorithm [5,6], cut algorithm [7,8], and minimum discontinuity algorithm [9,10]. Performance of these algorithms are effective but not robust to severe noises. The second category is based on the framework of minimum-norm. The least-squares is the most representative algorithm [11,12] with fast Fourier transformation and discrete cosine transform operations. L1-Norm envelope sparsity theorem has been proposed [13], which gives a sufficient condition to achieve phase unwrapping. The modified network programming (MNP) algorithm [14,15] assumed that the mapping relation between wrapped phase and true phase can be not necessarily the integer multiple of 2π. The limitation of the minimum-norm methods is that the output is dependent on the computation path. It may introduce distortion in the regular regions without noise. The third category is featured with phase noise filtering, such as unscented Kalman filter [16,17] and the recursive phase unwrapping (RPU) system [1819], which can achieve accurate and noise-immune unwrapping. However, it produces over smoothness on phase image.

Recently, deep learning methods based on convolutional neural network (CNN) have been proposed to address signal phase unwrapping issues. The idea is to transform it into semantic segmentation, which is also referred as pixel-wise classification,” aims at classifying each pixel into one of the pre-determined classes [2022]. Utilization of hundreds of hidden layers, consisting of millions of parameters in CNN methods makes it possible to discover intricate structures in large datasets [2326]. Recently PhaseNet based on U-net architecture was proposed to address simulated phase data with Gaussian function for general purpose study [27]. Deep learning based phase unwrapping (DLPU) based on U-net architecture was further evaluated on holographic interferometry imaging [28]. Zhang et al used the DeepLabV3+ architecture to address phase unwrapping for interferometric metrology almost contemporarily [29].

Currently, no deep learning based methods have been reported nor evaluated on addressing DOCT phase image unwrapping issues. Incorporating the contextual information and adding short-term memory to each layer, residual neural network architecture has demonstrated its superior performance in semantic segmentation [3035]. To solve phase unwrapping problem in DOCT phase images, we propose a noise-immune and robust residual en-decoder network (REDN) by uniting multi-scale context with pixel-level accuracy. The REDN includes a residual stream and a pooling stream. The residual stream carries feature maps at full image resolution, which can be combined with classical residual block (RB) and full resolution residual block (FRRB) from the pooling stream. The pooling stream performs the pooling operations and plays an important role in capturing high-level information through the network. These two streams are concatenated at the full image resolution. Without additional pre-processing or post-processing procedure, our method can obtain accurate phase unwrapping results for Doppler OCT phase images. Compared to DeepLabV3+, PhaseNet and MNP methods, our method demonstrates its superiority.

The remainder of this paper is organized as follows. Details of our method are illustrated in Section 2. Experimental results and comparison with other methods on simulated images, phantom plastic tube flow images and mouse artery flow images are presented in Section 3. Discussion are shown in Section 4. Finally, the main conclusions are presented in Section 5.

2. Methods

Phase unwrapping is the process to get true phase value from detected phase value wrapped in the range of (-π, π]. When the detected phase exceeds the range of 2π, phase wrapping becomes an issue. The relation between the true phase and wrapped phase is given below:

$$\phi (x,y) = \varphi (x,y) + 2\pi \ast k(x,y)$$
where $\phi (x,y)$ is the true phase and $\varphi (x,y)$ is the wrapped phase, $(x,y)$ represents the spatial coordinates of a pixel and $k(x,y)$ denotes integer multiple of 2π referred to as phase jump-count to be added to the wrapped phase to get the true phase.

In our framework, the phase unwrapping process regarded as the semantic segmentation task is learned through the architecture that takes $\varphi (x,y)$ as the input and gives the output as jump-count $k(x,y)$. The ground truth of the training process can be computed using Eq. (2):

$$k(x,y) = \textrm{round}(\frac{{\phi (x,y) - \varphi (x,y)}}{{2\pi }})$$

2.1 Data sources

2.1.1 Simulated data

One of the basic requisites for deep learning is sufficient amount of labeled training data. For DOCT phase images, there is limited number of training data due to the complexity of in-vivo experiment to get sufficient sample images. In addition, the ground truth DOCT phase image with no wrapping issue is hard to obtain. Thus, simulation data is an alternate approach to take. In this work, we generated the training dataset and validation dataset by the definitive relation between the wrapped phase image and the true phase image as shown in Fig. 1.

 figure: Fig. 1.

Fig. 1. Process of the simulated data generation: (a) typical DOCT phase image, (b) selected background region marked by white box in (a), (c) simulated true phase image, (d) simulated phase image with noise from (b), (e) wrapped phase image with noise as architecture input, and (f) ground truth jump-count map.

Download Full Size | PDF

It is worth mentioning that noises need to be considered to make the dataset faithful to the phase images from DOCT. A typical Doppler OCT phase image is shown in Fig. 1(a). We extracted the noise from the background of the DOCT phase image including mainly shot noise as shown in Fig. 1(b). The abundant simulation data were generated using linear combination of multiple Gaussian phase distributions located at different locations with both forward and backward flow directions as shown in Fig. 1(c). The mean and variance of each Gaussian phase distribution were randomly varied. This enables the network to learn phase continuities for broad general shapes rather than limiting it to certain definitive patterns. Figure 1(b) and Fig. 1(c) are combined to get the phase image with noise, shown in Fig. 1(d). Wrapped image Fig. 1(e) is used as the input of the neural networks, whose minimal value is shifted to zero. The ground truth jump-count map of simulation data is shown in Fig. 1(f) obtained according to Eq. (2). Background region without the target information is thresholded with an absolute value of 0.05 to avoid feature interference and save training time.

2.1.2 Experimental data

To validate the performance and robustness of REDN method on real sample phase images, two types of experimental data were used. One is the transparent plastic tube (referred as phantom, 0.5 mm inner diameter, 0.9 mm outer diameter) with different velocities of flowing milk and the other is mouse artery.

The phantom images were obtained from a home-built spectral domain OCT system with a central wavelength of 1300 nm and a bandwidth of 60 nm. The system ran at 70 fps with 1000 A-scans per frame and had a measured axial resolution of 14 µm, imaging range of 6.7 mm. The sensitivity and phase stability of the OCT system were 92 dB and 70 mrad, respectively. The Doppler flow imaging speed range thus is calculated to be from 0.316 mm/s to 14.2 mm/s in both directions parallel to the scanning beam with adjacent A-scan time difference of 14.2 µs [15].

The mouse artery was imaged with a MEMS-based handheld probe. The system ran at 36 fps with each frame size of 1000 (lateral)×512 (axial) pixels using a swept source with a central wavelength of 1310 nm and tuning range from 1260 nm to 1360 nm. The axial and lateral resolution of OCT was to be 12.6 µm in air and 17.5 µm respectively. The system had a sensitivity of 84 dB and sensitivity roll-off of 5.7 dB/mm over an imaging range of 5 mm. The phase stability of the OCT system is 70 mrad. The Doppler flow imaging speed range thus is calculated to be from 0.363 mm/s to 16.3 mm/s in both directions parallel to the scanning beam with adjacent A-scan time difference of 20 µs [36].

2.2 Network architecture

Our REDN architecture, illustrated in Fig. 2, is derived from the architecture described in [37], which is a very deep network. The input is a 256×256-pixel image. The output is the predicted jump-count map obtained by the softmax. We adopted two threads to combine the high-level features for recognition and low-level features for localization. In the pooling thread, we reduced the size of the features and increased the receptive field of the networks with the max pooling layers. The residual thread can compute the residuals at full image resolution with FRRB and RB that make the high-level features to go through the network [3840]. Meanwhile, it has been proven that it is much easier to optimize a residual mapping than the original plain network [41].

 figure: Fig. 2.

Fig. 2. Architecture of the residual en-decoder network (REDN). The subscript n in RBn and FRRBn represents the number of convolution channels. The parameter c denotes the number of classes to predict.

Download Full Size | PDF

The full pre-activation RB was adopted in our network architecture, as shown in Fig. 3(a). The 1×1 convolution layer is essentially a linear projection onto the space of the same dimensionality and an additional non-linearity is presented by the rectification function [42]. The FRRB acting as a residual unit for residual thread includes two inputs and outputs. Figure 3(b) shows the detail of the full resolution residual blocks. At first, Input1 from the residual thread goes through pooling layers first to concatenate with Input2 from the pooling thread. Then the concatenated features are flowed into two convolution blocks. Before the 3×3 convolution layers, we add the batch normalization layer and Relu activation function as the non-linearity transformation. And a second convolution block on one hand forms Output2 for the next FRRB and on the other hand is followed by a 1×1 convolution layer and up-sampling layer to make concatenation with the residual thread.

 figure: Fig. 3.

Fig. 3. Architecture of the residual block (a) and full resolution residual block (b) in the network.

Download Full Size | PDF

2.3 System implementation

All the experiments were performed on a Sitonholy (Beijing, China) IW4200-4G workstation (Xeon CPU E5-2650 v4, 2.2GHz-2.9 GHz) with one NVIDIA (Santa Clara, California, USA) Telsa P100 graphics processing units (GPU) and 128GB of RAM. The models were implemented with Python (v3.5.2) based on Keras (v2.1.6) with NVIDIA CUDA (v8.0) and cu-dnn (v6.1) libraries.

3. Results

We compared our results with recently proposed deep learning based DeepLabV3+ and PhaseNet methods for signal phase unwrapping and traditional modified networking programming (MNP) method. Firstly, quantitative evaluation metrics are presented. Then, we exhibit the phase unwrapping results on the simulated image, phantom tube image and mouse artery image.

3.1 Quantitative evaluation

The evaluation metrics were applied to evaluate both the segmentation and classification performance, which contained sensitivity (SE), specificity (SP) and accuracy (AC). We computed the average of the test dataset to get the final results. The evaluation metrics are defined as:

$$AC = \frac{{{N_{tp}} + {N_{tn}}}}{{{N_{tp}} + {N_{fn}} + {N_{fp}} + {N_{tn}}}}$$
$$SE = \frac{{{N_{tp}}}}{{{N_{tp}} + {N_{fn}}}},\,SP = \frac{{{N_{tn}}}}{{{N_{fp}} + {N_{tn}}}},$$
where Ntp, Ntn, Nfp and Nfn represent the number of true positive, true negative, false negative and false positive, which are defined on the pixel level. A predicted ‘jump pixel’ is regarded as a true positive if its ground truth is ‘jump’. Otherwise, it is regarded as a false negative. A predicted ‘non-jump pixel’ is considered as a true negative if its ground truth is ‘non-jump’. Otherwise, it’s regarded as a false positive.

For the simulated image, root mean square error (RMSE) can be used to describe the similarity between the predicted phase image $P(k,j)$ and the true phase image:

$$RMSE = \sqrt {\frac{{\sum\nolimits_{k = 1}^m {\sum\nolimits_{j = 1}^n {{{[P(k,j) - \phi (k,j)]}^2}} } }}{{m \times n}}}$$
where k and j denote pixel position, m and n are the height and width of the image, respectively. However, the true phase image of the DOCT real experimental data cannot be obtained, the RMSE is not applicable. Here, we calculated the total variation (TV) of the predicted phase image to evaluate the unwrapping effectiveness. The TV is defined as the integral of the magnitude of the image gradient, which can be expressed as:
$${J_{{T_0}}}(u) = \int\limits_{{D_u}} {\sqrt {{u_x}^2 + {u_y}^2} dxdy}$$
where ${u_x} = \frac{{\partial u}}{{\partial x}},{u_y} = \frac{{\partial u}}{{\partial y}}$, Du is the image domain. TV as a reference parameter can measure the phase discontinuity of reconstructed phase image based on the fact that unwrapped phase will have smooth phase transition.

3.2 Training procedure

The training dataset consists of 12000 images with the size of 256×256. The validation dataset consists of 3000 images. The wrapped phase image is used as input of the architecture and the jump counts of the training images are distributed in the range of -5 to 5 consisting 11 classes in total. Cross-entropy loss and stochastic gradient decent with momentum 0.9 and initial learning rate of 10−4 were used in our REDN model training process. Cross-entropy loss and Adam optimizer and initial learning rate of 10−4 were used in PhaseNet model training process. Cross-entropy loss and stochastic gradient decent with momentum 0.9 and initial learning rate of 10−6 were used in DeepLabV3+ model training process. In order to reduce overfitting, we used the dropout ratio 0.25, 0.2 and 0.2 corresponding to REDN, PhaseNet and DeepLabV3+ architecture during the training procedure. In Fig. 4, the convergence of the learning curves is illustrated in terms of the training loss and validation loss over the epochs.

 figure: Fig. 4.

Fig. 4. Learning loss during the training of FRRN, PhaseNet and DeepLabV3+ architecture.

Download Full Size | PDF

3.3 Experiments on simulated data

The unwrapping results of the simulated wrapped image for different methods are shown in Fig. 5. Figure 5(a) is the wrapped image and Fig. 5(b) denotes the true phase image of Fig. 5(a). The first row of the Fig. 5(c) denotes the predicted unwrapped phase images with our REDN method, DeepLabV3+, PhaseNet and MNP respectively. The second row of the Fig. 5(c) denotes the residual error between the predicted unwrapped phase image and the true phase image Fig. 5(b). Both our REDN method and MNP method can obtain a good performance even though the wrapped points are almost difficult to identify with naked eyes. The DeepLabV3+ method cannot obtain the discrete jump pixels at the edge of the wrapped region. From the residual error, we can see PhaseNet method performs better than DeepLabV3+ method, but worse than our method and MNP method.

 figure: Fig. 5.

Fig. 5. Phase unwrapping results of the simulated images. (a) simulated wrapped image, (b) true phase image, (c) the first row denotes the unwrapped phase images with different methods, (c) the second row denotes the residual error between the predicted unwrapped phase image and the true phase image.

Download Full Size | PDF

Quantitative parameter comparison including AC, SE, SP, RMSE, and time consumption based on 400 test simulated images is shown in Table 1. Please note for MNP method AC, SE and SP are not applicable. From Table 1, we can see that among all deep learning methods, our method is better than PhaseNet method, which is better than DeepLabV3+ method for simulated data. RMSE value of MNP is very close to our method. However, its time consumption is much longer than deep learning based methods. The reason is that deep learning based methods are GPU accelerated while MNP is processed using CPU only.

Tables Icon

Table 1. Accuracy, sensitivity, specificity, RMSE and Time calculated with our method, PhaseNet, DeepLabV3+ and MNP

3.4 Experiment on real data

To compare the phase unwrapping capability on real DOCT images of four methods, Fig. 6 shows four phantom images with different flux levels. It is clear from Fig. 6 that our method and MNP method can achieve satisfactory results with good continuity and contrast. These comparison result illustrates that our method has the outstanding performance compared to the other two deep learning methods for DOCT phantom images. When the flow velocity is high, the predicted results with DeepLabV3+ are more accurate than PhaseNet. The PhaseNet method cannot unwrap the phase image precisely at central pixels and the DeepLabV3+ method has low accuracy near the tube wall region. However, as the flow velocity decreases, PhaseNet performs better than the DeepLabV3 + . It implies that the DeepLabV3+ model can recognize the general region of the wrapped phase but lack the good pixel-accuracy localization. And the PhaseNet can localize the wrapped phase pixel-by-pixel while the accuracy is not high enough. Please note motion effect caused by movement of sample or instrument, assuming its continuous variation over a single phase image, will change the detected flow profile. However, phase wrapping boundaries with value of integer multiple of 2π are independent of motion effect. Bulky motion correction can be applied for phase images after unwrapping process when it is necessary.

 figure: Fig. 6.

Fig. 6. Phase unwrapping results of phantom images with four methods at different flux levels. The first column denotes phantom intensity image. The second column denotes corresponding phase image. Third to sixth column denotes the unwrapped phase image with our method, DeepLabV3+, PhaseNet and MNP respectively. (Scale bar: 250 µm)

Download Full Size | PDF

Figure 7(a-d) show the unwrapping results on mouse femoral artery with four methods at different vessel locations. We can see that four methods can unwrap the real DOCT blood flow images to some extent. Particularly, the noise increases as the imaging depth increases due to the blood scattering. The bottom of the vessel is blurred with severe noise and the blood flow information cannot be obtained clearly. The MNP method and our proposed method both have similar good unwrapping result at most flow region. Large differences come from high noise region. Proposed REDN method has a better performance than MNP method on vessel phase image in high noise region especially at the bottom of vessel. Meanwhile, the other two deep learning methods cannot obtain accurate result near the vessel wall area.

 figure: Fig. 7.

Fig. 7. Phase unwrapping results of in-vivo vessel images with our REDN method, PhaseNet, DeepLabV3+ and MNP methods: (a) - (d) rows show four group vessel images at different positions of the vessel, (e) profiles of dashed red line marked in (d). (Scale bar: 250 µm)

Download Full Size | PDF

Figure 7(e) plots the wrapped phase profile and the reconstructed flow profile for different methods over the dashed red line marked in Fig. 7(d). We can see the difference of unwrapping results between these four methods clearly. The unwrapping results of our method are consistent with the MNP’s to some extent. Our unwrapping result looks more reasonable intuitively, while DeepLabV3+ and PhaseNet generate erroneous results that are of large discontinuity.

To evaluate the performance of these four methods on DOCT real phase images quantitatively, we calculated the TV in Table 2. Our method has approximate effectiveness with MNP. The MNP has the smallest TV value on phantom image while our method has the smallest value on vessel image. However, our method has a better performance than those two deep learning methods. Please note the reason for larger TV value of phantom image compared to vessel image is due to the larger cross section pixel numbers in phantom image. TV can serve as a reference parameter but not a definitive parameter. It should be combined with visual image inspection.

Tables Icon

Table 2. Comparison of total variation calculated with predicted unwrapped phase image by Our method, DeepLabV3+, PhaseNet, and MNP

4. Discussions

In the model training process, the noise level of the training data is fixed, and the corresponding network was trained under this fixed noise level. In practice, it is impossible to train different parameters subject to different noise levels. Here, we make a phase unwrapping comparison on the noise effect in test simulated data. The noise level of training data is regarded as 1, and we added different noise to the same simulated data with levels from ratio 0.1 to 1.2, corresponding SNR of 35.9dB to 0.24dB based on SNR (dB) = 10log10(Psignal/ Pnoise). Figure 8 shows the mean RMSE values and its stand deviation. We can see that as the SNR increases, the mean RMSE value gets smaller, which is as expected. The MNP method has the best performance at high SNR region while our method has the best performance than MNP in low SNR region. As SNR gets smaller, the RMSE value of PhaseNet method and DeepLabV3+ method becomes larger quickly. And the maximal RMSE value of our method is still small, which means that our method is more robust than those two deep learning methods.

 figure: Fig. 8.

Fig. 8. Comparison of RMSE and its standard deviation using our REDN method, DeepLabV3+, PhaseNet and MNP for simulated data with different SNR.

Download Full Size | PDF

5. Conclusions

We proposed a deep learning based REDN phase unwrapping method for DOCT phase images by regarding it as a semantic segmentation. To address the insufficiency of qualified training data, we trained the REDN model using the simulated data synthesized with DOCT phase image noise. We combined residual thread and pooling thread at two resolution levels to achieve consistently robust performance confirmed by experiments on simulated, phantom and vessel phase images. Comparison with recently proposed deep learning based DeepLabV3+ and PhaseNet methods and traditional modified networking programming method shows the superiority of our method. The proposed REDN method can retrieve unwrapped phase information from DOCT systems for accurate quantitative diagnosis and evaluation when detected phase is wrapped. It could be integrated with other deep learning-based image processing models for future OCT image analysis.

Funding

National Natural Science Foundation of China (61505006); National Key Research and Development Program of China (2017YFC0107801, 2017YFC0107900); Youth Talent Innovation Program China Association for Science and Technology; Overseas Expertise Introduction Project for Discipline Innovation (B18005); Beijing Institute of Technology (2018CX01018).

Disclosures

The authors declare that there are no conflicts of interest.

References

1. Y. Huang, Z. Ibrahim, D. Tong, S. Zhu, Q. Mao, J. Pang, W. P. Andree Lee, G. Brandacher, and J. U. Kang, “Microvascular anastomosis guidance and evaluation using real-time three-dimensional Fourier-domain Doppler optical coherence tomography,” J. Biomed. Opt. 18(11), 111404 (2013). [CrossRef]  

2. Y. Wang, B. A. Bower, J. A. Izatt, O. Tan, and D. Huang, “Retinal blood flow measurement by circumpapillary Fourier domain Doppler optical coherence tomography,” J. Biomed. Opt. 13(6), 064003 (2008). [CrossRef]  

3. V. Doblhoff-Dier, L. Schmetterer, W. Vilser, G. Garhöfer, M. Gröschl, R. A. Leitgeb, and R. M. Werkmeister, “Measurement of the total retinal blood flow using dual beam Fourier-domain Doppler optical coherence tomography with orthogonal detection planes,” Biomed. Opt. Express 5(2), 630–642 (2014). [CrossRef]  

4. V. J. Srinivasan, S. Sakadzić, I. Gorczynska, S. Ruvinskaya, W. Wu, J. G. Fujimoto, and D. A. Boas, “Quantitative cerebral blood flow with optical coherence tomography,” Opt. Express 18(3), 2477–2494 (2010). [CrossRef]  

5. Y. Zhang, S. Wang, G. Ji, and Z. Dong, “An improved quality guided phase unwrapping method and its applications to MRI,” Prog. Electromagn. Res. 145, 273–286 (2014). [CrossRef]  

6. M. Arevalillo-Herráez, F. R. Villatoro, and M. A. Gdeisat, “A Robust and Simple Measure for Quality-Guided 2D Phase Unwrapping Algorithms,” IEEE Trans. on Image Process. 25(6), 2601–2609 (2016). [CrossRef]  

7. J. Dong, F. Chen, D. Zhou, T. Liu, Z. Yu, and Y. Wang, “Phase unwrapping with graph cuts optimization and dual decomposition acceleration for 3D high-resolution MRI data,” Magn. Reson. Med. 77(3), 1353–1358 (2017). [CrossRef]  

8. D. Gao and F. Yin, “Mask cut optimization in two-dimensional phase unwrapping,” IEEE Geosci. Remote Sensing Lett. 9(3), 338–342 (2012). [CrossRef]  

9. J. Xu, D. An, X. Huang, and P. Yi, “An efficient minimum-discontinuity phase-unwrapping method,” IEEE Geosci. Remote Sensing Lett. 13(5), 666–670 (2016). [CrossRef]  

10. Y. Liu, Y. Han, F. Li, and Q. Zhang, “Speedup of minimum discontinuity phase unwrapping algorithm with a reference phase distribution,” Opt. Commun. 417, 97–102 (2018). [CrossRef]  

11. M. D. Pritt and J. S. Shipman, “Least-squares two-dimensional phase unwrapping using FFT’s,” IEEE Trans. Geosci. Electron. 32(3), 706–708 (1994). [CrossRef]  

12. S. Xing and H. Guo, “Temporal phase unwrapping for fringe projection profilometry aided by recursion of Chebyshev polynomials,” Appl. Opt. 56(6), 1591–1602 (2017). [CrossRef]  

13. H. Yu, Y. Lan, J. Xu, D. An, and H. Lee, “Large-Scale L0-Norm and L1-Norm 2-D Phase Unwrapping,” IEEE Trans. Geosci. Electron. 55(8), 4712–4728 (2017). [CrossRef]  

14. L. Grady, “Random walks for image segmentation,” IEEE Trans. Pattern Anal. Mach. Intell. 28(11), 1768–1783 (2006). [CrossRef]  

15. S. Xia, Y. Huang, S. Peng, Y. Wu, and X. Tan, “Robust phase unwrapping for phase images in Fourier domain Doppler optical coherence tomography,” J. Biomed. Opt. 22(3), 036014 (2017). [CrossRef]  

16. Z. Cheng, D. Liu, Y. Yang, T. Ling, X. Chen, L. Zhang, J. Bai, Y. Shen, L. Miao, and W. Huang, “Practical phase unwrapping of interferometric fringes based on unscented Kalman filter technique,” Opt. Express 23(25), 32337–32349 (2015). [CrossRef]  

17. Y. Wang, D. Huang, Y. Su, and X. S. Yao, “Two-dimensional phase unwrapping in Doppler Fourier domain optical coherence tomography,” Opt. Express 24(23), 26129–26145 (2016). [CrossRef]  

18. M. A. Navarro, J. C. Estrada, M. Servin, J. A. Quiroga, and J. Vargas, “Fast two-dimensional simultaneous phase unwrapping and low-pass filtering,” Opt. Express 20(3), 2556–2561 (2012). [CrossRef]  

19. J. C. Estrada, M. Servin, and J. Vargas, “2D simultaneous phase unwrapping and filtering: A review and comparison,” Opt. Lasers Eng. 50(8), 1026–1029 (2012). [CrossRef]  

20. V. Badrinarayanan, A. Kendall, and R. Cipolla, “Segnet: A deep convolutional encoder-decoder architecture for image segmentation,” IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481–2495 (2017). [CrossRef]  

21. F. G. Venhuizen, B. V. Ginneken, B. Liefers, F. V. Asten, V. Schreur, S. Fauser, C. Hoyng, T. Theelen, and C. I. Sanchez, “Deep learning approach for the detection and quantification of intra retinal cystoid fluid in multivendor optical coherence tomography,” Biomed. Opt. Express 9(4), 1545–1569 (2018). [CrossRef]  

22. G. S. Liu, M. H. Zhu, J. Kim, P. Raphael, B. E. Applegate, and J. S. Oghalai, “ELHnet: a convolutional neural network for classifying cochlear endolymphatic hydrops imaged with optical coherence tomography,” Biomed. Opt. Express 8(10), 4579–4594 (2017). [CrossRef]  

23. O. Oktay, E. Ferrante, K. Kamnitsas, M. Heinrich, W. Bai, J. Caballero, S. A. Cook, A. de Marvao, T. Dawes, D. P. O’Regan, B. Kainz, B. Glocker, and D. Rueckert, “Anatomically Constrained Neural Networks (ACNNs): Application to Cardiac Image Enhancement and Segmentation,” IEEE Trans. Med. Imaging 37(2), 384–395 (2018). [CrossRef]  

24. C. Wu, Y. Xie, L. Shao, J. Yang, D. Ai, H. Song, Y. Wang, and Y. Huang, “Automatic boundary segmentation of vascular Doppler optical coherence tomography images based on cascaded U-net architecture,” OSA Continuum 2(3), 677–688 (2019). [CrossRef]  

25. S. Devalla, P. K. Renukanand, B. K. Sreedhar, G. Subramanian, L. Zhang, S. Perera, J. Mari, K. Chin, T. A. Tun, N. G. Strouthidis, T. Aung, A. H. Thiéry, and M. J. A. Girard, “DRUNET: a dilated-residual U-Net deep learning network to segment optic nerve head tissues in optical coherence tomography images,” Biomed. Opt. Express 9(7), 3244–3265 (2018). [CrossRef]  

26. A. Shah, L. Zhou, M. D. Abramoff, and X. Wu, “Multiple surface segmentation using convolution neural nets: application to retinal layer segmentation in OCT images,” Biomed. Opt. Express 9(9), 4509–4526 (2018). [CrossRef]  

27. G. E. Spoorthi, S. Gorthi, and R. K. S. S. Gorthi, “PhaseNet: A Deep Convolutional Neural Network for Two-Dimensional Phase Unwrapping,” IEEE Signal Processing Letters 26(1), 54–58 (2019). [CrossRef]  

28. K. Wang, Y. Li, K. Qian, J. Di, and J. Zhao, “One-step robust deep learning phase unwrapping,” Opt. Express 27(10), 15100–15115 (2019). [CrossRef]  

29. T. Zhang, S. Jiang, Z. Zhao, K. Dixit, X. Zhou, J. Hou, Y. Zhang, and C. Yan, “Rapid and robust two-dimensional phase unwrapping via deep learning,” Biomed. Opt. Express 27(16), 23173–23185 (2019). [CrossRef]  

30. L. Yu, H. Chen, Q. Dou, J. Qin, and P. A. Heng, “Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks,” IEEE Trans Med Imaging. 36(4), 994–1004 (2017). [CrossRef]  

31. L. V. Fulton, D. Dolezel, J. Harrop, Y. Yan, and C. P. Fulton, “Classification of Alzheimer's Disease with and without Imagery using Gradient Boosted Machines and ResNet-50,” Brain Sci. 9(9), 212 (2019). [CrossRef]  

32. H. Chen, Q. Dou, L. Yu, J. Qin, and P. A. Heng, “VoxResNet: Deep voxel wise residual networks for brain segmentation from 3D MR images,” Neuroimage. 170(170), 446–455 (2018). [CrossRef]  

33. Y. Rivenson, Z. Gorocs, H. Gunaydin, Y. Zhang, H. Wang, and A. Ozcan, “Deep learning microscopy,” Optica 4(11), 1437–1443 (2017). [CrossRef]  

34. V. A. Santos, L. Schmetterer, H. Stegmann, M. Pfister, A. Messner, G. Schmidinger, G. Garhofer, and R. M. Werkmeister, “CorneaNet: fast segmentation of cornea OCT scans of healthy and keratoconic eyes using deep learning,” Biomed. Opt. Express 10(2), 622–641 (2019). [CrossRef]  

35. A. Abdolmanafi, L. Duong, N. Dahdah, I. R. Adib, and F. Cheriet, “Characterization of coronary artery pathological formations from OCT imaging using deep learning,” Biomed. Opt. Express 9(10), 4936 (2018). [CrossRef]  

36. Y. Huang, G. J. Furtmüller, D. Tong, S. Zhu, W. P. Lee, G. Brandacher, and J. U. Kang, “MEMS-based handheld Fourier domain Doppler optical coherence tomography for intraoperative microvascular anastomosis imaging,” PLoS One 9(12), e114215 (2014). [CrossRef]  

37. T. Pohlen, A. Hermans, M. Mathias, and B. Leibe, “Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes,” Computer vision and pattern recognition, arXiv: 1611.08323 (2016).

38. K. He, X. Zhang, S. Ren, and J. Sun, “Identity Mappings in Deep Residual Networks,” arXiv: 1603.05027 (2016).

39. K. He, X. Zhang, S. Ren, and J. Sun, “Deep Residual Learning for Image Recognition,” arXiv: 1512.03385 (2015).

40. K. He, X. Zhang, S. Ren, and J. Sun, “Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification,” arXiv: 1502.01852 (2015).

41. T. Kepp, C. Droigk, M. Casper, M. Evers, G. Hüttmann, N. Salma, D. Manstein, M. P. Heinrich, and H. Handels, “Segmentation of mouse skin layers in optical coherence tomography image data using deep convolutional neural networks,” Biomed. Opt. Express 10(7), 3484–3496 (2019). [CrossRef]  

42. K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,” arXiv: 1409.1556 (2015).

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (8)

Fig. 1.
Fig. 1. Process of the simulated data generation: (a) typical DOCT phase image, (b) selected background region marked by white box in (a), (c) simulated true phase image, (d) simulated phase image with noise from (b), (e) wrapped phase image with noise as architecture input, and (f) ground truth jump-count map.
Fig. 2.
Fig. 2. Architecture of the residual en-decoder network (REDN). The subscript n in RBn and FRRBn represents the number of convolution channels. The parameter c denotes the number of classes to predict.
Fig. 3.
Fig. 3. Architecture of the residual block (a) and full resolution residual block (b) in the network.
Fig. 4.
Fig. 4. Learning loss during the training of FRRN, PhaseNet and DeepLabV3+ architecture.
Fig. 5.
Fig. 5. Phase unwrapping results of the simulated images. (a) simulated wrapped image, (b) true phase image, (c) the first row denotes the unwrapped phase images with different methods, (c) the second row denotes the residual error between the predicted unwrapped phase image and the true phase image.
Fig. 6.
Fig. 6. Phase unwrapping results of phantom images with four methods at different flux levels. The first column denotes phantom intensity image. The second column denotes corresponding phase image. Third to sixth column denotes the unwrapped phase image with our method, DeepLabV3+, PhaseNet and MNP respectively. (Scale bar: 250 µm)
Fig. 7.
Fig. 7. Phase unwrapping results of in-vivo vessel images with our REDN method, PhaseNet, DeepLabV3+ and MNP methods: (a) - (d) rows show four group vessel images at different positions of the vessel, (e) profiles of dashed red line marked in (d). (Scale bar: 250 µm)
Fig. 8.
Fig. 8. Comparison of RMSE and its standard deviation using our REDN method, DeepLabV3+, PhaseNet and MNP for simulated data with different SNR.

Tables (2)

Tables Icon

Table 1. Accuracy, sensitivity, specificity, RMSE and Time calculated with our method, PhaseNet, DeepLabV3+ and MNP

Tables Icon

Table 2. Comparison of total variation calculated with predicted unwrapped phase image by Our method, DeepLabV3+, PhaseNet, and MNP

Equations (6)

Equations on this page are rendered with MathJax. Learn more.

ϕ ( x , y ) = φ ( x , y ) + 2 π k ( x , y )
k ( x , y ) = round ( ϕ ( x , y ) φ ( x , y ) 2 π )
A C = N t p + N t n N t p + N f n + N f p + N t n
S E = N t p N t p + N f n , S P = N t n N f p + N t n ,
R M S E = k = 1 m j = 1 n [ P ( k , j ) ϕ ( k , j ) ] 2 m × n
J T 0 ( u ) = D u u x 2 + u y 2 d x d y
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.