Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Fourier ptychography multi-parameter neural network with composite physical priori optimization

Open Access Open Access

Abstract

Fourier ptychography microscopy(FPM) is a recently developed computational imaging approach for microscopic super-resolution imaging. By turning on each light-emitting-diode (LED) located on different position on the LED array sequentially and acquiring the corresponding images that contain different spatial frequency components, high spatial resolution and quantitative phase imaging can be achieved in the case of large field-of-view. Nevertheless, FPM has high requirements for the system construction and data acquisition processes, such as precise LEDs position, accurate focusing and appropriate exposure time, which brings many limitations to its practical applications. In this paper, inspired by artificial neural network, we propose a Fourier ptychography multi-parameter neural network (FPMN) with composite physical prior optimization. A hybrid parameter determination strategy combining physical imaging model and data-driven network training is proposed to recover the multi layers of the network corresponding to different physical parameters, including sample complex function, system pupil function, defocus distance, LED array position deviation and illumination intensity fluctuation, etc. Among these parameters, LED array position deviation is recovered based on the features of brightfield to darkfield transition low-resolution images while the others are recovered in the process of training of the neural network. The feasibility and effectiveness of FPMN are verified through simulations and actual experiments. Therefore FPMN can evidently reduce the requirement for practical applications of FPM.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

Corrections

Delong Yang, Shaohui Zhang, Chuanjian Zheng, Guocheng Zhou, Lei Cao, Yao Hu, and Qun Hao, "Fourier ptychography multi-parameter neural network with composite physical priori optimization: publisher’s note," Biomed. Opt. Express 14, 5709-5709 (2023)
https://opg.optica.org/boe/abstract.cfm?uri=boe-14-11-5709

3 October 2023: A correction was made to the title.

1. Introduction

In conventional microscopic systems, there is often a trade-off between field of view(FOV) and spatial resolution limited by the objective numerical aperture. The spatial resolution in a microscope is determined by numerical aperture (NA) of the objective and the illumination wavelength. Conventional method to solve the contradiction between large FOV and high resolution is to combine large NA objective and high-precision two-dimensional mechanical scanning, which increases the complexity and cost of the whole microscopic system simultaneously. Unlike the conventional way of relying entirely on hardware modification, Fourier ptychography microscopy (FPM), as one typical example of computational imaging methods, can achieve large FOV and high resolution at the same time with a cost-effective platform and without any mechanical scanning devides. Instead of scanning in the spatial domain, FPM scan and stitch the sample in the Fourier domain by illuminating it from different directions with a LED array [15]. The prerequisite of FPM to successfully achieve resolution enhancement is the phase retrieval of each sub-aperture in frequency domain, and a certain degree of sub-aperture overlap is the key to ensure that phase retrieval can be achieved. In fact, the information redundancy caused by the sub-spectrum overlap can not only ensure the recovery of sample complex information, but also some system parameters or experimental parameters, such as coherent transfer function of system, LED array position deviation, defocus distance of sample and illumination intensity fluctuation from different LEDs(hereafter refer to as other factors) [6].

There are many method using characteristics of data redundancy to recover other factors, Ou, Xiaoze et al proposed Embedded pupil function recovery method to recover the pupil function of the system. [7], Sun Jiasong et al. and Eckert Regina et al. proposed method to recover LED array position deviation [810]. However, the conventional optimization algorithm for reconstruction needs to mathematically find analytical differentiation of captured low-resolution images (LR images) on the target to be recovered [11,12]. However, it is difficult to find the analytical differentiation of LR images on other factors, such as defocus distance of sample and illumination intensity fluctuation of LEDs, which also makes it difficult for conventional algorithms to incorporate the other factors into optimization variables. And because the other factors are coupled in the imaging process, it is difficult to optimize multiple parameters simultaneously [13]. Due to the difficulty of finding the analytical differentiation of some factors at present stage [14,15], we turn to numerical differential method [16]. The training process of neural network [17] is optimizing the parameters of network according to the numerical differentiation. And there is no need to find the analytical differentiation of loss on network parameters [18,19]. As the structure of network is established, it can automatically find the numerical differentiation. Moreover, thanks to rapid development of deep learning, many loss functions can be constructed to improve the speed of solving optimization problems and achieve global optimum [20,21]. When solving inverse problems through neural network, we can make full use of neural network training tools to realize optimal solution of multi-parameters [22]. In 2018, SHAOWEI JIANG et al. first proposed to model FPM forward propagation process with convolutional neural network [23]. They get the complex object information through training process of network. However, their method take the process of lightwave through microscope is ideal, and no other factors are taken into account.

To improve the robustness of neural network reconstruction algorithm, and achieve multi-parameter reconstruction, we model the other factors in the system as network layers. Since the convolution neural network will lead to rapid growth in computing volume with the size of image increasing [17], we use element-by-element multiplication between layers representing different factors to increase the computation efficiency. But at the same time, the increase of parameters to be optimzed will cause pathological degree of problem to increase [24]. Zhao, Ming et al. have tried modeling the LED array position deviation into neural network, but the optimization process is very time-consuming [25]. Since using physical means to reduce the morbidity of inverse problems is a very effective method [2628], we have reduced the pathologicality of phase recovery problems through inserting a wedge angle in front of the microscope to improve the speed and quality of reconstruction [29]. In this paper, in order to reduce the pathological degree of reconstruction problem, we turn to find physical method to correct LED array position deviation. We choose four brightfield to darkfield transition LR images which located on orthogonal direction, using boundary of bright-field to dark-field transition on the LR images to calculate LED array position deviation inspired by the method proposed in reference [3]. We named the neural network proposed in this paper for FPM reconstruction algorithm as Fourier Ptychographic Multiparamater Net(FPMN), the physics method for correcting LED array position deviation as Array Correction Fourier ptychography(ACFP). In FPMN, the lightwave propagation process is modeled as fixed parameters of network layer (discrete reverse Fourier transformation). We implement model construction with deep learning framework pytorch, using automatic differentiation of pytorch to optimize network parameters.

This paper is structutred as follows. The principle of standard FPM framework and system setup are presented in section 2.1. The working principle and structure of FPMN are presented in section 2.2$\sim$2.3. The definition of LED array position deviation and the principle of ACFP are presented in 2.4. In section 3, we conduct simulation experiments with different defocus distance and mixed deviations(defocus aberration and LED array position deviation), comparing the reconstruction quality achieved through EPRY and our method. The USAF chart and biological samples are used to demonstrate the effectiveness of FPMN in section 4.1$\sim$4.2. We discuss the influence of different pupil function model method on reconstruction quality in 4.3. Conclusions and discussions are given in Section 5.

2. Principles

2.1 FPM principle

Unlike conventional imaging, which relies more on hardware systems, computational imaging refers to joint optimization of physical models, imaging pipelines, and inverse problem algorithms [30]. Therefore, clear physical model and accurate model parameters during imaging process are essential for the success of FPM, As a representative computational imaging method shown in Fig. 1(a), a typical FPM system consists of a LED array providing angular varied illuminations, an microscopic imaging module collecting the diffracted light emitted from the sample, and a image sensor to receive and record the two-dimensional intensity images. For a small sample segment, whose size is much smaller than the distance between the LED array and sample, the illumination wave can be approximately treated as a parallel plane wave. Thus, according to the Fourier optics, illuminating the sample by an oblique plane wave with a wavevector $(k_{x},k_{y})$ is equivalent to shifting the sensor of the sample’s spectrum by $(k_{x},k_{y})$ in the Fourier domain. The microscopic objective acts as the combination of a Fourier transformer and a filter in Fourier domain. Therefore, the forward light field propagating and imaging process of FPM can be expressed as

$$\mathbf{I}_{n}(x, y) = |\mathscr{F}^{{-}1}\{\mathscr{F}\{\mathbf{t}(x, y)\}\cdot\mathbf{P}(k_{x}, k_{y})\}|^2$$
where $\mathbf {t}(x, y) = s(x, y) \cdot e^{i(x k_{xn} + y k_{yn})}$ denotes the exit wave distribution of the sample $s(x, y)$ that is illuminated by an oblique illumination with a wavevector $(k_{xn}, k_{yn})$. ’$\mathscr {F}$’ and ’$\mathscr {F}^{-1}$’ indicate the Fourier and inverse Fourier transform respectively. $\mathbf {P}(k_{x}, k_{y})$ is the pupil function of the objective, $(x, y)$ is the 2D spatial coordinates in the spatial domain and $(k_{x}, k_{y})$ is the corresponding spatial frequencies in the frequency domain. $I_{n}(x, y)$ is the intensity image acquired by the camera.

 figure: Fig. 1.

Fig. 1. Imaging principle of FPM and actual system.

Download Full Size | PDF

2.2 Principle and structure of the Fourier ptychography multi-parameters neural network

According to the Eq. (1), the forward propagation of light field in FPM can be modeled as an element-wise neural network. In order to simplify the network and reduce the amount of computation, we rewrite Eq. (1) as Eq. (2) to directly model the sample in frequency domain. Since complex differentiation is not supported in the advanced Neural network framework Pytorch, we model the sample complex function as two sub-channel layers. We model the the complex pupil function $\mathbf {P}(k_{x},k_{y})$ in Eq. (2) corresponding network layer with the top ten Zernike coefficients, a classic two-dimensional phase distribution representation approach. This method of incorporating classical physical models into the network can evidently reduce the complexity of the network and the ill-conditioning degree of the network training. The computational relationship between layers can be express asma

$$I_{n}(x, y) = |\mathscr{F}^{{-}1}(\hat{\mathbf{O}}(k-k_{n}) \cdot \mathbf{P}(k_{x}, k_{y}))|^{2} ,n = 1, 2, \ldots N^{2}$$
$$\hat{\mathbf{O}}(k) =\ \hat{\mathbf{O}}_{r}(k) + \hat{\mathbf{O}}_{i}(k), \quad \mathbf{P}(k_{x}, k_{y}) = T_{Zernike}(z_{1}, z_{2}, \; \cdots , \; z_{10})$$
$$\left\{ \begin{array}{l} \mathbf{E}_{nr} = \hat{\mathbf{O}}_{r}(k-k_{n})\cdot cos(\mathbf{P}(k_{x}, k_{y})) - \hat{\mathbf{O}}_{i}(k-k_{n})\cdot sin(\mathbf{P}(k_{x}, k_{y})) \\ \mathbf{E}_{ni} = \hat{\mathbf{O}}_{r}(k-k_{n})\cdot sin(\mathbf{P}(k_{x}, k_{y})) + \hat{\mathbf{O}}_{i}(k-k_{n})\cdot cos(\mathbf{P}(k_{x}, k_{y})) ,n = 1, 2\ \cdots , N^{2} \end{array} \right.$$
$$\mathbf{I}_{n}(x, y) = |\mathscr{F}^{{-}1}\{\mathbf{E}_{nr} + \mathbf{E}_{ni}\}|^2 ,n = 1, 2\ \cdots , \; N^{2}$$
where $\mathbf {O}(k)$ denotes the Fourier transform of the object function, $k_{n}$ denotes the illumination vector, $\mathbf {P}(k_{x},k_{y})$ represents the pupil function(radian of phase modulation in Eq. (3)), $N$ represents the number of rows or columns of LEDs on LED array, $\hat {\mathbf {O}}_{r}(k)$ and $\hat {\mathbf {O}}_{i}(k)$ are the real part and imaginary part respectively, $\mathbf {E}_{nr}$ and $\mathbf {E}_{ni}$ are the real part and imaginary part of exit wave distribution through the pupil. The inverse Fourier transformation in Eq. (5) are then modeled as a four-channel layer with fixed parameters. Then the light field impinging on the image detector surface can be expressed as
$$\mathscr{F}^{{-}1}(\mathbf{E}_{n}) = \mathbf{G}_{a} \cdot \mathbf{E}_{n} \cdot \mathbf{G}_{b} \rightarrow \mathscr{F}^{{-}1}(\mathbf{E}_{n}) = (\mathbf{G}_{ar}+\mathbf{G}_{ai}) \cdot (\mathbf{E}_{nr}+\mathbf{E}_{ni}) \cdot (\mathbf{G}_{br}+\mathbf{G}_{bi})$$
where $\mathbf {G}_{ar}$ and $\mathbf {G}_{ai}$ denote the real part and imainary part of $\mathbf {G}_{a}$, $\mathbf {G}_{br}$ and $\mathbf {G}_{bi}$ denote the real part and imainary part of $\mathbf {G}_{b}$, $\mathbf {G}_{a}$ and $\mathbf {G}_{b}$ can be express as
$$\mathbf{G_{a}} = \begin{pmatrix} e^{{-}j2\pi\frac{0\cdot0}{M}} & e^{{-}j2\pi\frac{0\cdot1}{M}} & \cdots & e^{{-}j2\pi\frac{0\cdot(M-1)}{M}} \\ e^{{-}j2\pi\frac{1\cdot0}{M}} & e^{{-}j2\pi\frac{1\cdot1}{M}} & \cdots & e^{{-}j2\pi\frac{1\cdot(M-1)}{M}} \\ \vdots & \vdots & \ddots & \vdots \\ e^{{-}j2\pi\frac{(M-1)\cdot0}{M}} & e^{{-}j2\pi\frac{(M-1)\cdot1}{M}} & \cdots & e^{{-}j2\pi\frac{(M-1)\cdot(M-1)}{M}} \end{pmatrix}$$
$$\mathbf{G_{b}} = \begin{pmatrix} e^{{-}j2\pi\frac{0\cdot0}{N}} & e^{{-}j2\pi\frac{0\cdot1}{N}} & \cdots & e^{{-}j2\pi\frac{0\cdot(N-1)}{N}} \\ e^{{-}j2\pi\frac{1\cdot0}{N}} & e^{{-}j2\pi\frac{1\cdot1}{N}} & \cdots & e^{{-}j2\pi\frac{1\cdot(N-1)}{N}} \\ \vdots & \vdots & \ddots & \vdots \\ e^{{-}j2\pi\frac{(N-1)\cdot0}{N}} & e^{{-}j2\pi\frac{(N-1)\cdot1}{N}} & \cdots & e^{{-}j2\pi\frac{(N-1)\cdot(N-1)}{N}} \end{pmatrix}$$
Since we have split the real-imaginary part of the complex light field, we break $\mathbf {G_{a}}$ down into $\mathbf {G_{ar}}$ and $\mathbf {G_{ai}}$ in Eq. (6) according to the Euler formula. Similarly, $\mathbf {G_{b}}$ can also be decomposed into $\mathbf {G_{br}}$ and $\mathbf {G_{bi}}$. In addition to these hidden layers, we model the sequential angular varied illumination plane waves as the input layer, and the two-dimensional intensity images as the output layer, respectively. Up to now, as shown in Fig. 2(a), the neural network corresponding to the ideal FPM imaging process has been constructed successfully.

 figure: Fig. 2.

Fig. 2. (a)Structure of FPN(Fourier Ptychography Network) with ideal FPM imaging process.(b)Structure of FPMN.

Download Full Size | PDF

As mentioned previously, the HR complex function of the sample can be recovered through by training of the neural network, with the help of the advanced various network optimizers. There are several ways to construct the loss functions optimizing the neural network, such as L0,L1, and L2, etc. We choose L1-norm as the loss function in our approach [23], as shown in the following equation.

$$\mathbf{loss} = diff(\mathbf{I}^{gt}_{n} - \mathbf{I}^{predict}_{n}) = \sum_{n=1}^{N}|\mathbf{I}^{gt}_{n} - \mathbf{I}^{predict}_{n}|\quad(n = 1, 2 \; \cdots \, N^{2})$$

It is worth noting that the accurate position of each sub-spectrum needs to be known when updating the HR spectrum for high recovery quality. In other words, the illumination wave vectors determined by the position of each LED element and the distance between LED array and sample need to be precisely known. The LED array position deviation is defined in Fig. 3. Since the pitch and yaw of the LED board are easily zeroed with a spirit level, we only take the horizontal movement and rotation angle about the optical axis into consideration. Where $\Delta x$ and $\Delta y$ in Fig. 3 respectively denotes the translation of LED array in two orthogonal directions, $\theta$ denotes the rotation of the LED array.

 figure: Fig. 3.

Fig. 3. The rotation and translation of LED array in the horizontal plane.

Download Full Size | PDF

Unlike the ideal FPM model that has no errors, there are several critical system parameter errors in the system construction and raw data acquisition process. As shown in Fig. 2(b), we add three dominal system errors, the LED array position deviation, the sample defocus distance, and the illumination intensity fluctuation into the ideal FPM forward network. In the original FPM model, we suppose each LED element is a point source that emits quasi-monochromatic light, and the whole LED array should be aligned. Nevertheless, the positional deviation of the LED board is unavoidable during the process of constructing or modifying FPM systems. So we add a LED array position correct layer, we use $\Delta x, \Delta y, \theta$ in Fig. 3 as the parameters in this layer and the corrected sub-aperture position can be express as

$$x_{n}^{correct} = cos(\theta) \cdot x_{n} + \Delta x, \; y_{n}^{correct} = sin(\theta) \cdot y_{n} + \Delta y, \; n = 1,2 \; \cdots \,N^{2}$$
$$\theta_{xn}^{correct} = arctan(\frac{x_{n}^{correct}}{h}), \; \theta_{yn}^{correct} = arctan(\frac{y_{n}^{correct}}{h}), \; n = 1,2\ \cdots \, N^{2}$$
$$k_{n} = (k_{xn}, k_{yn}), \; k_{xn} = 2\pi \frac{sin(\theta_{xn}^{correct})}{\lambda}, \; k_{yn} = 2\pi \frac{sin(\theta_{yn}^{correct})}{\lambda} , \; n = 1,2 \; \cdots \,N^{2}$$

The ideal FPM model assume that the sample is a flat thin simple in focus, while the non-planar distribution and improper focusing will cause different defocus distances for parts or the whole sample. Therefore, in FP imaging process, we should take the defocus distance of sample into account, modeling effect of defocus in FPMN, so that the recovery of the complex object can be free from the defocus aberration. According to the defocus distance of sample, the distribution of the light field after free-space propagation can be expressed as

$$\mathbf{E}_{1}(x, y) = \mathscr{F}^{{-}1}\{\mathscr{F}\{\mathbf{E}_{0}(x, y)\}\cdot \mathbf{H}(k_{x}, k_{y}, z)\}$$
$$\mathbf{H}(k_{x}, k_{y}, z) = exp(j \frac{2 \pi}{\lambda} \cdot z \cdot \sqrt{1-k_{x}^{2}-k_{y}^2})$$
where $\mathbf {E}_{0}(x, y)$ is a known complex light field in a focusing plane, $E_{1}(x, y)$ is the field in a plane with a distance of $z$. According to Eq. (10), we model the free-space propagation as an defocus layer in FPMN. We set defocus distance z as the parameters to be optimized in defocus layer. The defocus layer is set behind the pupil function layer and before the IFFT fixed parameter layer. Similarly, the defocus parameter can also be recovered by training the FPMN simultaneously. Since pytorch does not support complex differentiation, free-space propagation for light field in FPMN with defocus taken into consideration can be expressed as
$$\mathbf{E}_{nr}^{d}(x, y) = \mathscr{F}^{{-}1}\{\mathscr{F}\{{\mathbf{E}}_{nr}(x, y)\}\cdot \mathbf{H}_{r}(k_{x}, k_{y}, z) - \mathscr{F}\{\mathbf{E}_{ni}(x, y)\}\cdot \mathbf{H}_{i}(k_{x}, k_{y}, z)\}$$
$$\mathbf{E}_{ni}^{d}(x, y) = \mathscr{F}^{{-}1}\{\mathscr{F}\{\mathbf{E}_{nr}(x, y)\}\cdot \mathbf{H}_{i}(k_{x}, k_{y}, z) - \mathscr{F}\{\mathbf{E}_{ni}(x, y)\}\cdot \mathbf{H}_{r}(k_{x}, k_{y}, z)\}$$
where $E_{nr}^{d}$, $E_{ni}^{d}$ denotes real and imaginary part of the light field passing through defocus layer. $H_{r}$, $H_{i}$ denotes real and imaginary part of $H(k_{x},k_{y},z)$.

In FPM system, there are inevitable illumination intensity fluctuation from different angle LEDs, Due to manufacturing limitations, there will be differences in the luminous brightness of LEDs in different positions. According to the lighting model in FPM, the LED illumination distance and illumination angle of different LEDs are different. Under the influence of above mentioned factors, it is difficult to explicitly calculate the intensity fluctuation of different angle LEDs. We think that further correction of intensity fluctuation can improve recovery quality of the complex sample. We compensate intensity deviation with intensity correction coefficient $\gamma$. After adding defocus correction, intensity correction, forward propagation process of FPMN can be expressed as

$$\mathbf{I}_{n}(x, y) = |\mathscr{F}^{{-}1}(\gamma_{n} \cdot \hat{\mathbf{O}}(k-k_{n}) \cdot\mathbf{P}(k_{x}, k_{y}) \cdot \mathbf{H}(k_{x}, k_{y}, z)|^2 \quad (\gamma_{n} = \gamma_{1}, \gamma_{2}, \; \cdots \; \gamma_{N})$$

We model the process of intensity correction as FPMN network layer, which has $N^{2}$ parameters to optimize, each parameter corresponds to one LED illumination intensity fluctuation. The structure of the complete FPMN with all these parameters taken into consideration is illustrated in Fig. 2(b).

2.3 LED array correction method (ACFP) through physics model

Since the optimization of the LED array position deviation in FPMN will be very time-consuming [25]. In this subsection, we proposed a correction method ACFP to correct the LED array position deviation through physics model.

As shown in Fig. 4(a) and Fig. 4(b), the bright-field to dark-field transition feature will appear in the image under some specific oblique angle illumination. The circular transition feature is caused by the circular shape of the circular pupil function of the imaging system. Since the relationship between the illumination direction and the bright-field image position is linear, we propose a framework to calculate the LED pose parameters according to the position characteristics of the bright-field image positions. Fig. 4(a-d) indicate four LR images corresponding to four symmetrical LED illumination with known serial numbers. Symmetrical circles can form a fusiform zone whose location can be used for the recovery of the LED pose correctly. For convenience, we fit a bright-field circle by using the arc-shaped bright-field to dark-field transition boundary line, and use the center of the circle to indicate the position of the bright-field area.

 figure: Fig. 4.

Fig. 4. (a)LR images illuminated by LED in top, bottom, right and down of LED array center when there is no LED array position deviation. Fusiform zone formed by fitted circle. (b)(c)(d)There are translation and rotation on the LED array. (b)LR images illuminated by LED in left and bottom of LED array center. Fusiform zone formed by fitted circle. (c)LR images illuminated by two adjacent LEDs in the right of LED array center and fitted circle. (d)LR images illuminated by two adjacent LEDs in the top of LED array center and fitted circle.

Download Full Size | PDF

As shown in Fig. 4, $M_{v}$ denotes the midpoint of the upper vertices and lower vertices. $M_{h}$ denotes the midpoint of the left vertices and right vertices. $C$ denotes the center of the full field of view LR image. When there is no LED array position deviation, as shown in Fig. 4(a), $M_{v}$, $M_{h}$, $C$ coincident with one point. When the position of LED array is inaccurate, as shown in Fig. 4(b), the deviation of $M_{h}$ and $C$ in y direction, the deviation of $M_{v}$ and $C$ in x direction, are respectively related to the LED array position deviation in y direction and x direction.

Subsequently, we calibrate the scaling factor between the translation of the bright-field image and the LED array horizontal deviation with the help of bright-field lateral offset corresponding two adjacent LED units. As shown in Fig. 4(c) and Fig. 4(d), the pixel offset $\Delta \hat {x}_{p}$ and $\Delta \hat {y}_{p}$ correspond to the LED spacing, so $\Delta x_{p}$ and $\Delta y_{p}$ correspond to the translation of LED array can be express as

$$\Delta x = \frac{\Delta x_{p}}{\Delta\hat{x}_{p}} \times P$$
$$\Delta y = \frac{\Delta y_{p}}{\Delta\hat{y}_{p}} \times P$$
where P denotes the LED spacing in LED array, $\theta$ in Fig. 4(c) can directly represent the rotation angle of LED array.

2.4 Training process of FPMN

The framework of multi-parameters reconstruction algorithm of the proposed method is shown in Fig. 5. We choose four brightfield to darkfield LR images in two orthogonal directions to calculate the LED array position deviation by ACFP. The LED array position deviation is used as a supplementary input to FPMN. Initialize $\hat {O}_{x, y}=Ae^{j\varphi }$ as the complex sample parameters, with using the intensity distribution of central LR image as $A$ and $\varphi$. Initialized pupil function without aberration, defocus distance with zero, intensity correction coefficients with ones.

 figure: Fig. 5.

Fig. 5. Algorithm outline of our proposed method.

Download Full Size | PDF

3. Simulations

The simulation parameters were chosen to realistically model an FPM platform, with an incident illumination wavelength of 470nm, an image sensor with pixel size of 2.4um, and an 2X objective with NA of 0.454$\times$2. We simulated the use of the central $17\times 17$ LEDs in the array placed 83mm beneath the sample, and the distance between adjacent LEDs is 2.5mm. The raw low-resolution data is limited to a small region with only $256\times 256$ pixel resolution and the final high-resolution complex field with $1024\times 1024$ pixel is recovered by different approaches. To generate the LR images in simulation, we initial the ground truth of different physical parameters into the FPMN layer corresponding to different physical definitions, then we conduct one time inference to get the LR images. We compared the results of our method (FPMN) and EPRY at different defocus state in Fig. 6. And the result has improved after we activate the optimization of the intensity correction coefficients layer. Although the pupil function looks different from the ground truth, we think it is caused by the coupling of multiple parameters such as defocus aberration since the recover defocus distance also has 3% error from the ground truth. As we can see in (d), (e) of Fig. 6 that the tendency of the zero diffraction light to the secondary diffraction light is gradually decreasing.

 figure: Fig. 6.

Fig. 6. a(1), b(1)Results of EPRY in defocus state. a(2), b(2)Results of FPMN without ICC(intensity correction coefficients) in defocus state. a(3), b(3)Results of complete FPMN in defocus state. (c)Groundtruth of the sample and pupil function. The left of (d), (e) of ICC and ICC divide groundtruth. (f)Groundtruth of ICC.

Download Full Size | PDF

To verify the robustness for mixed deviation(defocus aberration and LED array position deviation), we also conduct a simulation experiment. As shown in Fig. 7, amplitude and phase of reconstruction object were terrible when there are LED array position deviation in the system, but we can see the trend of translation in the recover result of intensity correction coefficients(ICC). As a comparison, when we obtain the LED array position deviation through ACFP, and enter it to LED array position correction layer of the FPMN, FPMN can still obtain high-quality reconstructed intensity image and phase image even with such large positional deviations.

 figure: Fig. 7.

Fig. 7. simulation results with defocus aberration and LED array position deviation

Download Full Size | PDF

After introducing the different defocus aberration and intensity correction coefficients(adding base on exposure time and random intensity error) into the imaging process of simulation experiment, we plot the SSIM curve to compare the results with different method in Fig. 8. As we can see, our method have a robust result with the defocus that in the range of 400um, and the results with intensity correction coefficients present best among all.

 figure: Fig. 8.

Fig. 8. SSIM curve between optim result and GT. Legend FPMN represents the complete optim net. Legend FPMN WOIC represents FPMN without intensity correction coefficients layer. Legend FPN represents FPMN without defocus layer and intensity correction coefficients layer. Legend EPRY represents Embedded pupil function recovery method.

Download Full Size | PDF

4. Experiment

4.1 Experimental results of an USAF resolution target with FPMN

We build a FPM system as shown in Fig. 1(b) to acquire LR images. The experimental setup generally followed our simulation parameters, the distance between the LED array and the sample plane is set to 83 mm. We use a LED array (CMN, P 2.5, 17$\times$17) for varied-angles illuminations. The wavelength of illuminations is 470 nm and the bandwidth of illuminations is 20nm, the parameters of objective(Olympus, magnification 2x, NA$\approx$0.1) we used in the experiment are same with simulation. A camera (FLIR, BFS-U3-200S6M-C, sensor size 1", dynamic range 71.89 dB, pixel size 2.4 $\mathrm{\mu}m$) is used for recording LR images. To demonstrate the feasibility of FPMN, we use an amplitude-only USAF chart as the sample. The USAF chart is placed at the sample plane that adjusted to different defocus distance by a high-precision moving device.

First there is no LED array position deviation in the system, we manually adjust the samples in different defocus positions and compare the results of EPRY and results of our method. The results shown in Fig. 9 corresponding to different defocus planes are 180um, 100um, 20um, −60um and −140um respectively. Sequentially, we use the FPMN to realize digital refocusing. The EPRY is used as a comparative experiment. As we can see, when the sample is placed at the focus plane, both the EPRY and FPMN can recover the HR amplitude successfully. But with the defocus distance increase, the performance of EPRY has a degradation, which is consistent with the simulation result shown in Fig. 6. In contrast, the reconstructed quality of FPMN can be evidently improved compared with EPRY.

 figure: Fig. 9.

Fig. 9. High resolution reconstruction results of EPRY and the proposed method with different defocus distances

Download Full Size | PDF

Then we further introduce the LED array position deviation into the system. The results are shown in Fig. 10. When the mixed deviation of FPM system exists, the reconstruction results of EPRY has been seriously distorted, but the reconstruction results of FPMN still maintain a high quality with ACFP. These results directly prove the effectiveness of FPMN and ACFP.

 figure: Fig. 10.

Fig. 10. optim results with mixed deviation. (a1)(a2)Results with only LED array position deviation $\Delta x=4mm,\Delta y=4mm,\theta =0.07rad,z=0um$. (b1)(b2)Results with mixed deviation including LED array position deviation and defocus aberration $\Delta x=4mm,\Delta y=4mm,\theta =0.07rad,z=100um$

Download Full Size | PDF

4.2 Experimental results of the biological sample with FPMN

Generally, a biological sample is in a 3D distribution form, the non-planar distribution characteristic will cause different defocus distance between different subregions. We usually assume that the biological sample is placed at the focus plane in conventional FPM, which may cause degradation of recovery. In this section, a biological sample Paramecium is used to demonstrate the feasibility of the FPMN. The results are shown in Fig. 11.

 figure: Fig. 11.

Fig. 11. Optim results of biological sample with the EPRY and FPMN. (*1)Reconstructed amplitude with EPRY, (*2)Reconstructed phase with EPRY,(*3)Reconstructed amplitude with FPMN and ACFP. (*4)Reconstructed phase with FPMN and ACFP.

Download Full Size | PDF

Because of the non-planar distribution characteristic of the biological sample, different subregions of the sample correspond to different defocus distances. The whole FOV image is divided into several subregions to recover by FPMN.

The subregion of Fig. 10(a*) is located in the center of the field of view, the subregion of Fig. 10(b*) and Fig. 10(c*) are located in the edge of the field of view. The object in the edge of the field of view have a relative LED array position deviation, so we use ACFP to obtain the deviation. We compared the reconstruction results of EPRY and our method. As we can see, FPMN with ACFP can improve the reconstructed quality to a certain extent, especially for the phase part, even bring the contrast improve. The results prove the effectiveness of the proposed method.

4.3 Different model method of pupil function in FPMN

In the above experiments, we find that the constraint of prior knowledge can effectively improve the reconstruction quality of the algorithm and reduce the pathologicality of the inverse problem, then achieve the precise decoupling of multiple parameters.

We used three different modeling methods for pupil function layer in FPMN. The first method is same with Fourier obj layer, separates the pupil function layer into real and imaginary parts, The modulation can be express as

$$\left\{ \begin{array}{l} \mathbf{E}_{nr} = \;\;\hat{\mathbf{O}}_{nr}\cdot \mathbf{P}_{r} - \hat{\mathbf{O}}_{ni}\cdot \mathbf{P}_{i} \\ \mathbf{E}_{ni} = \;\;\hat{\mathbf{O}}_{nr}\cdot \mathbf{P}_{i} + \hat{\mathbf{O}}_{ni}\cdot \mathbf{P}_{r} \end{array} \right.$$

Considering that coherent transfer function is usually phase modulation for light field in FPM systems. The second method model the pupil function layer as a phase modulation radian value. The modulation can be express as

$$\left\{ \begin{array}{l} \mathbf{E}_{nr} = \;\;\hat{\mathbf{O}}_{nr}\cdot cos(\mathbf{P}) - \hat{\mathbf{O}}_{ni}\cdot sin(\mathbf{P}) \\ \mathbf{E}_{nr} = \;\;\hat{\mathbf{O}}_{nr}\cdot sin(\mathbf{P}) + \hat{\mathbf{O}}_{ni}\cdot cos(\mathbf{P}) \end{array} \right.$$

In order to introduce more priori constraints to FPMN, the pupil function can be fitted with the top ten of the Zenike polynomial [31] in optics. In the third modeling method, the pupil function layer uses only the top ten of the zenike polynomial as the parameters to be optimized. which can be express as Eq. (3) and Eq. (4).

Under three different modeling methods, the results for the sample and pupil function are shown in Fig. 12. As we can see, the reconstruction quality of the samples are better under the latter two modeling methods. And when the defocus aberration is introduced, the last modeling method works better. Therefore, the introduction of prior knowledge constraints can improve the decoupling efficiency of multi-parameters

 figure: Fig. 12.

Fig. 12. (a1)The optimization results of the modeling method that is same with the Fourier obj layer. (b1)(b2)The optimization results of the modeling method with phase modulation radian value. (c1)(c2)The optimization results of the modeling method with zenike polynomial

Download Full Size | PDF

5. Conclusion and discussion

In this paper, we propose a recovery method for FPM, which uses neural networks to model Fourier ptychography imaging processes for optimization. We introduced LED array position deviation correction, defocus aberration correction, and intensity deviation correction into the neuarl network. And we call that neural network FPMN. The LED array position deviation correction method is called ACFP. With ACFP, the FPMN can achieve robustness to mixed deviation and also can obtain the value of each deviation.

In order to get better optimization results and faster speed, we should use more priori constraints and more accurate imaging models, as we discussed in this chapter on the way of pupil function modeling. We are considering that using the defocus distance calculated by the geometric relationship in the imaging process [26] as the initial parameter of FPMN, and calculating the fluctuation of LEDs at different angles by wave optics knowledge as the initial parameter of FPMN. That may speed up the optimization [32] of FPMN and get better decoupling effect.

Funding

National Natural Science Foundation of China (61735003, 61805011); Funding of foundation enhancement program under Grant (2021-JCJQ-JJ-0823).

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. G. Zheng, R. Horstmeyer, and C. Yang, “Wide-field, high-resolution fourier ptychographic microscopy,” Nat. Photonics 7(9), 739–745 (2013). [CrossRef]  

2. P. Song, S. Jiang, H. Zhang, Z. Bian, C. Guo, K. Hoshino, and G. Zheng, “Super-resolution microscopy via ptychographic structured modulation of a diffuser,” Opt. Lett. 44(15), 3645–3648 (2019). [CrossRef]  

3. G. Zheng, C. Shen, S. Jiang, P. Song, and C. Yang, “Concept, implementations and applications of fourier ptychography,” Nat. Rev. Phys. 3(3), 207–223 (2021). [CrossRef]  

4. X. Ou, R. Horstmeyer, C. Yang, and G. Zheng, “Quantitative phase imaging via Fourier ptychographic microscopy,” Opt. Lett. 38(22), 4845–4848 (2013). [CrossRef]  

5. P. C. Konda, L. Loetgering, K. C. Zhou, S. Xu, A. R. Harvey, and R. Horstmeyer, “Fourier ptychography: current applications and future promises,” Opt. Express 28(7), 9603–9630 (2020). [CrossRef]  

6. J. Li, J. Garfinkel, X. Zhang, D. Wu, Y. Zhang, K. De Haan, H. Wang, T. Liu, B. Bai, Y. Rivenson, G. Rubinstein, P. O. Scumpia, and A. Ozcan, “Biopsy-free in vivo virtual histology of skin using deep learning,” Light: Sci. Appl. 10(1), 233 (2021). [CrossRef]  

7. X. Ou, G. Zheng, and C. Yang, “Embedded pupil function recovery for Fourier ptychographic microscopy,” Opt. Express 22(5), 4960–4972 (2014). [CrossRef]  

8. J. Sun, Q. Chen, Y. Zhang, and C. Zuo, “Efficient positional misalignment correction method for fourier ptychographic microscopy,” Biomed. Opt. Express 7(4), 1336–1350 (2016). [CrossRef]  

9. R. Eckert, Z. F. Phillips, and L. Waller, “Efficient illumination angle self-calibration in Fourier ptychography,” Appl. Opt. 57(19), 5434–5442 (2018). [CrossRef]  

10. P. Dwivedi, A. Konijnenberg, S. Pereira, and H. Urbach, “Lateral position correction in ptychography using the gradient of intensity patterns,” Ultramicroscopy 192, 29–36 (2018). [CrossRef]  

11. C. Zuo, J. Sun, and Q. Chen, “Adaptive step-size strategy for noise-robust Fourier ptychographic microscopy,” Opt. Express 24(18), 20724–20744 (2016). [CrossRef]  

12. S. Konijnenberg, “An introduction to the theory of ptychographic phase retrieval methods,” Adv. Opt. Technol. 6(6), 423–438 (2017). [CrossRef]  

13. Y. Zhang, Y. Liu, X. Li, S. Jiang, K. Dixit, X. Zhang, and X. Ji, “PGNN: physics-guided neural network for fourier ptychographic microscopy,” arXiv preprint arXiv:1909.08869 (2019).

14. A. S. Jurling and J. R. Fienup, “Applications of algorithmic differentiation to phase retrieval algorithms,” J. Opt. Soc. Am. A 31(7), 1348–1359 (2014). [CrossRef]  

15. K. Kreutz-Delgado, “The complex gradient operator and the cr-calculus,” arXiv preprint arXiv:0906.4835 (2009).

16. G. Barbastathis, A. Ozcan, and G. Situ, “On the use of deep learning for computational imaging,” Optica 6(8), 921–943 (2019). [CrossRef]  

17. Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature 521(7553), 436–444 (2015). [CrossRef]  

18. F. Wang, Y. Bian, H. Wang, M. Lyu, G. Pedrini, W. Osten, G. Barbastathis, and G. Situ, “Phase imaging with an untrained neural network,” Light: Sci. Appl. 9(1), 77 (2020). [CrossRef]  

19. N. Thanh, Y. Xue, Y. Li, L. Tian, and G. Nehmetallah, “Deep learning approach to fourier ptychographic microscopy,” Opt. Express 26(20), 26470–26484 (2018). [CrossRef]  

20. S. Ruder, “An overview of gradient descent optimization algorithms,” arXiv preprint arXiv:1609.04747 (2016).

21. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980 (2014).

22. X. Lin, Y. Rivenson, N. T. Yardimci, M. Veli, Y. Luo, M. Jarrahi, and A. Ozcan, “All-optical machine learning using diffractive deep neural networks,” Science 361(6406), 1004–1008 (2018). [CrossRef]  

23. S. Jiang, K. Guo, J. Liao, and G. Zheng, “Solving fourier ptychographic imaging problems via neural network modeling and tensorflow,” Biomed. Opt. Express 9(7), 3306–3319 (2018). [CrossRef]  

24. A. N. Tikhonov, “On the solution of ill-posed problems and the method of regularization,” in Doklady Akademii Nauk, vol. 151 (Russian Academy of Sciences, 1963), pp. 501–504.

25. M. Zhao, X. Zhang, Z. Tian, and S. Liu, “Neural network model with positional deviation correction for fourier ptychography,” J. Soc. Inf. Disp. 29(10), 749–757 (2021). [CrossRef]  

26. S. Zhang, G. Zhou, C. Zheng, T. Li, Y. Hu, and Q. Hao, “Fast digital refocusing and depth of field extended fourier ptychography microscopy,” Biomed. Opt. Express 12(9), 5544–5558 (2021). [CrossRef]  

27. G. Zhou, S. Zhang, Y. Hu, and Q. Hao, “Adaptive high-dynamic-range fourier ptychography microscopy data acquisition with a red-green-blue camera,” Opt. Lett. 45(17), 4956–4959 (2020). [CrossRef]  

28. R. Kalita, W. Flanagan, J. Lightley, S. Kumar, Y. Alexandrov, E. Garcia, M. Hintze, M. Barkoulas, C. Dunsby, and P. M. French, “Single-shot phase contrast microscopy using polarisation-resolved differential phase contrast,” bioRxiv (2021).

29. S. Zhang, G. Zhou, Y. Hu, and Q. Hao, “Asymmetric constraint fourier ptychography microscopy,” IEEE Photonics Technol. Lett. 33(6), 309–312 (2021). [CrossRef]  

30. M. Kellman, K. Zhang, E. Markley, J. Tamir, E. Bostan, M. Lustig, and L. Waller, “Memory-efficient learning for large-scale computational imaging,” IEEE Trans. Comput. Imaging 6, 1403–1414 (2020). [CrossRef]  

31. F. Zernike, “Diffraction theory of the knife-edge test and its improved form, the phase-contrast method,” Mon. Not. R. Astron. Soc. 94(5), 377–384 (1934). [CrossRef]  

32. K. He, X. Zhang, S. Ren, and J. Sun, “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification,” in Proceedings of the IEEE international conference on computer vision, (2015), pp. 1026–1034.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (12)

Fig. 1.
Fig. 1. Imaging principle of FPM and actual system.
Fig. 2.
Fig. 2. (a)Structure of FPN(Fourier Ptychography Network) with ideal FPM imaging process.(b)Structure of FPMN.
Fig. 3.
Fig. 3. The rotation and translation of LED array in the horizontal plane.
Fig. 4.
Fig. 4. (a)LR images illuminated by LED in top, bottom, right and down of LED array center when there is no LED array position deviation. Fusiform zone formed by fitted circle. (b)(c)(d)There are translation and rotation on the LED array. (b)LR images illuminated by LED in left and bottom of LED array center. Fusiform zone formed by fitted circle. (c)LR images illuminated by two adjacent LEDs in the right of LED array center and fitted circle. (d)LR images illuminated by two adjacent LEDs in the top of LED array center and fitted circle.
Fig. 5.
Fig. 5. Algorithm outline of our proposed method.
Fig. 6.
Fig. 6. a(1), b(1)Results of EPRY in defocus state. a(2), b(2)Results of FPMN without ICC(intensity correction coefficients) in defocus state. a(3), b(3)Results of complete FPMN in defocus state. (c)Groundtruth of the sample and pupil function. The left of (d), (e) of ICC and ICC divide groundtruth. (f)Groundtruth of ICC.
Fig. 7.
Fig. 7. simulation results with defocus aberration and LED array position deviation
Fig. 8.
Fig. 8. SSIM curve between optim result and GT. Legend FPMN represents the complete optim net. Legend FPMN WOIC represents FPMN without intensity correction coefficients layer. Legend FPN represents FPMN without defocus layer and intensity correction coefficients layer. Legend EPRY represents Embedded pupil function recovery method.
Fig. 9.
Fig. 9. High resolution reconstruction results of EPRY and the proposed method with different defocus distances
Fig. 10.
Fig. 10. optim results with mixed deviation. (a1)(a2)Results with only LED array position deviation $\Delta x=4mm,\Delta y=4mm,\theta =0.07rad,z=0um$. (b1)(b2)Results with mixed deviation including LED array position deviation and defocus aberration $\Delta x=4mm,\Delta y=4mm,\theta =0.07rad,z=100um$
Fig. 11.
Fig. 11. Optim results of biological sample with the EPRY and FPMN. (*1)Reconstructed amplitude with EPRY, (*2)Reconstructed phase with EPRY,(*3)Reconstructed amplitude with FPMN and ACFP. (*4)Reconstructed phase with FPMN and ACFP.
Fig. 12.
Fig. 12. (a1)The optimization results of the modeling method that is same with the Fourier obj layer. (b1)(b2)The optimization results of the modeling method with phase modulation radian value. (c1)(c2)The optimization results of the modeling method with zenike polynomial

Equations (21)

Equations on this page are rendered with MathJax. Learn more.

I n ( x , y ) = | F 1 { F { t ( x , y ) } P ( k x , k y ) } | 2
I n ( x , y ) = | F 1 ( O ^ ( k k n ) P ( k x , k y ) ) | 2 , n = 1 , 2 , N 2
O ^ ( k ) =   O ^ r ( k ) + O ^ i ( k ) , P ( k x , k y ) = T Z e r n i k e ( z 1 , z 2 , , z 10 )
{ E n r = O ^ r ( k k n ) c o s ( P ( k x , k y ) ) O ^ i ( k k n ) s i n ( P ( k x , k y ) ) E n i = O ^ r ( k k n ) s i n ( P ( k x , k y ) ) + O ^ i ( k k n ) c o s ( P ( k x , k y ) ) , n = 1 , 2   , N 2
I n ( x , y ) = | F 1 { E n r + E n i } | 2 , n = 1 , 2   , N 2
F 1 ( E n ) = G a E n G b F 1 ( E n ) = ( G a r + G a i ) ( E n r + E n i ) ( G b r + G b i )
G a = ( e j 2 π 0 0 M e j 2 π 0 1 M e j 2 π 0 ( M 1 ) M e j 2 π 1 0 M e j 2 π 1 1 M e j 2 π 1 ( M 1 ) M e j 2 π ( M 1 ) 0 M e j 2 π ( M 1 ) 1 M e j 2 π ( M 1 ) ( M 1 ) M )
G b = ( e j 2 π 0 0 N e j 2 π 0 1 N e j 2 π 0 ( N 1 ) N e j 2 π 1 0 N e j 2 π 1 1 N e j 2 π 1 ( N 1 ) N e j 2 π ( N 1 ) 0 N e j 2 π ( N 1 ) 1 N e j 2 π ( N 1 ) ( N 1 ) N )
l o s s = d i f f ( I n g t I n p r e d i c t ) = n = 1 N | I n g t I n p r e d i c t | ( n = 1 , 2 N 2 )
x n c o r r e c t = c o s ( θ ) x n + Δ x , y n c o r r e c t = s i n ( θ ) y n + Δ y , n = 1 , 2 N 2
θ x n c o r r e c t = a r c t a n ( x n c o r r e c t h ) , θ y n c o r r e c t = a r c t a n ( y n c o r r e c t h ) , n = 1 , 2   N 2
k n = ( k x n , k y n ) , k x n = 2 π s i n ( θ x n c o r r e c t ) λ , k y n = 2 π s i n ( θ y n c o r r e c t ) λ , n = 1 , 2 N 2
E 1 ( x , y ) = F 1 { F { E 0 ( x , y ) } H ( k x , k y , z ) }
H ( k x , k y , z ) = e x p ( j 2 π λ z 1 k x 2 k y 2 )
E n r d ( x , y ) = F 1 { F { E n r ( x , y ) } H r ( k x , k y , z ) F { E n i ( x , y ) } H i ( k x , k y , z ) }
E n i d ( x , y ) = F 1 { F { E n r ( x , y ) } H i ( k x , k y , z ) F { E n i ( x , y ) } H r ( k x , k y , z ) }
I n ( x , y ) = | F 1 ( γ n O ^ ( k k n ) P ( k x , k y ) H ( k x , k y , z ) | 2 ( γ n = γ 1 , γ 2 , γ N )
Δ x = Δ x p Δ x ^ p × P
Δ y = Δ y p Δ y ^ p × P
{ E n r = O ^ n r P r O ^ n i P i E n i = O ^ n r P i + O ^ n i P r
{ E n r = O ^ n r c o s ( P ) O ^ n i s i n ( P ) E n r = O ^ n r s i n ( P ) + O ^ n i c o s ( P )
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.