Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Automated alignment of a reconfigurable optical system using focal-plane sensing and Kalman filtering

Open Access Open Access

Abstract

Automation of alignment tasks can provide improved efficiency and greatly increase the flexibility of an optical system. Current optical systems with automated alignment capabilities are typically designed to include a dedicated wavefront sensor. Here, we demonstrate a self-aligning method for a reconfigurable system using only focal plane images. We define a two lens optical system with 8 degrees of freedom. Images are simulated given misalignment parameters using ZEMAX software. We perform a principal component analysis on the simulated data set to obtain Karhunen–Loève modes, which form the basis set whose weights are the system measurements. A model function, which maps the state to the measurement, is learned using nonlinear least-squares fitting and serves as the measurement function for the nonlinear estimator (extended and unscented Kalman filters) used to calculate control inputs to align the system. We present and discuss simulated and experimental results of the full system in operation.

© 2016 Optical Society of America

1. INTRODUCTION

Advanced optical systems are widely used in today’s technology, including observing and tracing biological and chemical compounds with microscopes, creating 3D scenes with virtual reality (VR) displays [1], detecting and imaging exoplanets and disks with ground-based and space telescopes [2], and sensing and correcting wavefront aberrations for medical purposes [3]. An automated optical alignment system can save the time and energy spent on manual alignment. This makes the assembly process of many optical devices, including microscopes, medical sensing devices, and camera systems, more efficient. Self-aligning techniques can also improve the alignment between lenses of a virtual reality headset and human eyes. Most current VR headset models only allow a manual adjustment of the interpupillary distance. Moreover, automated alignment is important for space optical systems. Many satellites and space telescopes cannot be serviced after their launch. A slight inaccuracy in the engineering design or disturbance during launch or on orbit can easily cause optical misalignment [4]. The importance and benefits of automatically aligning an optical system increase with the complexity and flexibility of the instruments themselves. Of particular interest is the ability for complex instruments to automatically align using existing internal imaging sensors, without requiring the addition of dedicated wavefront sensors or other large changes to their basic beam paths.

Many static components in optical systems (such as reimaging and collimating optics) are bolted down after being carefully aligned the first time. In these cases, the manual alignment procedures are time consuming, and optical misalignment caused by environmental disturbances cannot be fixed. A reconfigurable system, which has multiple filters or other components in pupil and focal planes, needs the ability to self-align and may be made more flexible if internal components are allowed to move. For example, the Gemini Planet Imager (GPI) [5], a ground-based instrument, which includes a coronagraph and an extreme adaptive optics system for direct imaging of extrasolar planets, has automated alignment features on coronagraph components using computer vision algorithms [6]. The closed-loop control process allows GPI to achieve high precision alignment in the presence of a continuously changing gravity gradient and thermal flexure. A distributed optical system, such as an optical communication system, needs to be accurately aligned within limited space and setup time. Finally, there are cases where allowing for motion degrees of freedom creates new sensing capabilities as in interferometric devices and self-coherent imaging systems [7].

The most widely used alignment methods relate misalignment parameters to optical wavefront error as measured by various wavefront sensing devices. One of these methods involves mapping misalignments to Zernike terms using a sensitivity table [8,9]. Sensitivity tables, however, are limited in their accuracy when the misalignments are large and the nonlinearity increases. Merit function regression solves this problem and is presented in Kim et al. [10]. This method estimates the misalignment by performing damped least-squares optimization with merit function values defined as the difference between the measured and ideal Zernike coefficients of the optical system wavefronts. Lee et al. [11] proposed a differential wavefront sampling (DWS) method for the efficient alignment of optical systems. By perturbing optical elements, this technique generates a set of linear equations used to solve for the misalignment of a system. Oh et al. [12] integrated a revised DWS sampling method with MFR nonlinear optimization on a three-mirror anastigmat optical system. The integrated alignment method results in better alignment accuracies than standard MFR and DWS methods. Instead of using a numerical approach, Gu et al. [13] presented a method for aligning a three-mirror anastigmatic telescope using nodal aberration theory. These methods all require measuring the wavefront error of the system. When detailed knowledge of a wavefront is required for alignment, optical systems are designed to include a dedicated wavefront sensor, such as a Shack–Hartmann wavefront sensor. This increases the complexity of the system and, more importantly, can introduce non-common path errors.

There exist, however, various wavefront sensing schemes employing the primary system sensor and eliminating the need for dedicated wavefront sensors [14,15]. These techniques are already being applied to current scientific instrumentation [16]. Some focal-plane wavefront sensing methods use pupil plane masking or multiple detectors to introduce phase diversity and reconstruct wavefront errors [1719]. The image-moment-based wavefront sensing method [20] uses the image moment of the measured point spread function for alignment correction. Focus diversity is introduced to break the nonlinearity and allows the system to sense full-field wavefront aberrations. These methods can greatly improve wavefront sensing and optical alignment techniques and can be applied to automated optical alignment.

We propose a method that corrects the misalignment of an optical system with existing internal imaging instruments in the system, such as a focal plane camera, thereby saving the extra resources and space needed for splitting the beam and avoiding throughput loss and non-common path error. Moreover, this approach makes it easier to retrofit an existing optical system to perform self-alignment because the major difference is changing the static optical components to kinematic ones. We present a new sensing and control method for aligning a reconfigurable optical system. We demonstrate the ability to align a two lens system using only a focal plane camera. An optical model of a monochromatic beam, two moving lenses, and a science camera is connected to a closed-loop control system. We implement an iterated extended Kalman filter (IEKF) and unscented Kalman filter (UKF) to estimate the states in the control process. Our current alignment methodology is focused on narrow field of view (FOV) systems and on the on-axis signal. However, the basic approach can be extended to also consider off-axis sources and be made relevant for systems with larger FOVs. Examples of small FOV systems currently in use include light detection and ranging (LiDAR) systems for detail local mapping [21], high-contrast imaging system for imaging exoplanets near bright stars [22], and high-resolution satellites [23,24]. In Section 2, we define the optical model and control scheme and discuss the methods used in modeling and estimation, including Karhunen–Loève modal reconstruction, model fitting, and state estimation. We present the simulation result and introduce the experimental setup. In Section 3, the experimental result is presented. Both image reconstruction and closed-loop state estimation are shown. In Section 4, we discuss the result of our current system and lay out our next steps.

2. MODEL AND METHODOLOGY

In this section, we define the optical and control model, discuss image processing methods, model fitting, closed-loop control system, and experimental setup.

A. Model

Figure 1 shows a two lens optical system. The two moving lenses are represented as gray ellipses, the collimated laser beam is represented by the red line along the z axis, the xy plane is normal to the beam path, and the x axis is along the vertical. The collimated Gaussian beam passes through two moving lenses A and B and is focused on a CCD camera. The focal lengths of lens A and B are set as 200 and 100 mm, respectively, the image plane has pixel size 4.54 μm, and the laser beam has wavelength 635 nm. The distance between the collimated laser beam and lens A is 50 mm, and the distance between lens A and lens B is 400 mm. The CCD camera is placed at a distance of 212 mm after lens B. Our goal is to calibrate the moving lenses, which have a total of 8 degrees of freedom shift in the x and y directions, tip, and tilt for each lens. This is a simple model where the despace misalignments of lens A and B are assumed to have smaller influence on the system compared with lateral motions, and the shift along the z axis of both lenses are not included in the model. A column vector x=[SxA,SyA,SxB,SyB,TxA,TyA,TxB,TyB]T is used to describe the state of our system.

 figure: Fig. 1.

Fig. 1. Two lens optical system. A collimated Gaussian beam is passed through two moving lenses A and B and focuses on a CCD camera.

Download Full Size | PDF

Figure 2 shows a schematic of the self-aligning control system. The optical model in Fig. 1 is in the upper dashed block (Plant), and the lower dashed block represents a Kalman filter in closed-loop control system. The images captured from the camera are projected onto KL modes obtained from principal component analysis (PCA), which will be discussed in the next section, with the corresponding KL weights serving as the measurements of the control system. The measurements are sent to the Kalman filter to compare with the measurement predicted from our measurement model function (Section 2.C). The state estimate predicted by the Kalman filter is fed back to correct the misalignment (Section 2.D).

 figure: Fig. 2.

Fig. 2. Schematic of control system. Upper dashed block represents the optical system in Fig. 1. Lower dashed block represents a Kalman filter.

Download Full Size | PDF

B. Image Processing

We simulate images given misalignment parameters using ZEMAX software. The prescription in ZEMAX is set as the optical system described in Section 2.A. Thorlabs lenses LB1945 and LB1676 are imported as lens A and lens B, respectively. The laser, lenses, and camera parameters are chosen to model the conditions in the experiment, as shown in Table 2. Misalignment of the lenses introduces wavefront aberrations into the optical system, resulting in motion and shape changes to the nominally axisymmetric Gaussian spot in focal plane images. Figure 3 shows a sample misaligned image in 250×250 pixels. Our first image processing step is to fit a 2D Gaussian to the image to obtain the center position of the Gaussian spot of the image and then perform PCA to decompose the image data set into KL modes.

 figure: Fig. 3.

Fig. 3. Subframe of the simulated image from Zemax and its 2D Gaussian fitting in contour plot. Gaussian center Cx and Cy will be used as our measurements.

Download Full Size | PDF

1. Gaussian Fitting and COM

We apply a Gaussian fit to the subframe:

T(x,y)=G1+G2exp((xa)2(yb)22)x=(xCx)cosφ(yCy)sinφy=(xCx)sinφ+(yCy)cosφ.
The unknown parameters (Cx, Cy) represent the center position of the Gaussian spot, (G1, G2) represent the Gaussian coefficients, (a, b) represent the semi-major and minor axis, and φ represents the rotational angle of the ellipse. Figure 3 shows the Gaussian fitting of the simulated image in a contour plot. The Gaussian center Cx and Cy can be obtained and will be used as our measurements.

2. PCA and Karhunen–Loève Modes

We perform a PCA using the Karhunen–Loève transform (KLT) to create an orthogonal basis of eigenimages [25,26]. The KLT method decomposes observed signals into a combination of linearly independent modes called principal components. The observed signals are the image data set we collected from ZEMAX. In this section, we call the linearly independent modes KL modes.

We collect the image data set by scanning through eight state variables: SxA, SyA, SxB, SyB, TxA, TyA, TxB, and TyB. Each state is perturbed with three misaligned values δ, 0, and δ, where δ is a small misalignment for each state. We perturb the shift and tip–tilt by 0.4 mm and 4 deg, respectively. The collected images include all combinations of the perturbation on the states. This results in a total number of scanned images n=38. We capture a fixed size subframe (Np×Np) around the Gaussian center Cx and Cy. The subframe image matrix is reshaped into a p element column vector vi, where p=Np2 and i indicates the image number. We use the vector-mean-subtracted value of the image vector,

v¯i=viμ(vi),
where μ(·) is a mean operator. A large matrix containing all the scanned data can be obtained as
V¯=[v¯1,v¯2,,v¯n].
The image-to-image covariance matrix of the scanned data set is given by
S=1p1V¯TV¯,
where S is an n×n matrix. We perform an eigendecomposition of covariance S and obtain matrices Φ and Λ such that
SΦ=ΦΛ,
where Φ is an n×n matrix whose columns are the eigenvectors of S, and Λ is an n×n diagonal matrix whose entries are the corresponding eigenvalues. The KL transform matrix is then
Z=V¯Φ,
where Z is a p×n matrix whose columns are the KL modes.

Figure 4 shows the first 12 KL modes (in order of decreasing eigenvalue) and the corresponding eigenvalue in log scale. Each image has a frame size of 250×250 pixels, and all images are independently stretched to show details of the mode shapes.

 figure: Fig. 4.

Fig. 4. First 12 KL modes obtained from PCA decomposition with subframe 250×250 pixels. Each image is plotted under different intensity scale, and its corresponding eigenvalue is shown under the mode number in log scale.

Download Full Size | PDF

The image v¯i can be reconstructed as a weighted sum of the first m KL modes with coefficients,

wi=Zmv¯i,
where (·) is the pseudoinverse of a matrix, and Zm contains the first m KL modes. The reconstructed image c¯i can be calculated as
c¯i=Zmwi,
where wi is an m element column vector of the coefficients calculated above. Figure 5(a) shows the reconstructed image c¯i of the simulated image in Fig. 3 using the first six KL modes. The subtracted image in Fig. 5(b) shows the difference (v¯ic¯i) between the simulated image in Fig. 3 and the reconstructed image.

 figure: Fig. 5.

Fig. 5. Reconstruction of simulated image in Fig. 3 using the first six KL modes. Subtracted image shows the difference (v¯ic¯i) between the simulated image in Fig. 3 and reconstructed images. (a) Reconstructed image. (b) Subtracted image.

Download Full Size | PDF

Figure 6 shows the reconstruction error using the first eight modes to reconstruct the image, with all of the images plotted on the same intensity scale. As expected, we can see that the reconstruction error decreases gradually as the number of modes used increases. The RMS pixel errors ϵi of the reconstruction of image i is defined as

ϵi=(v¯ic¯i)T(v¯ic¯i)p
and is used as a metric for the quality of the image reconstruction. The blue circle markers in Fig. 7 shows the RMS error ϵi after the first 10 modes used. The RMS pixel error shown is the average over all of the training data.

 figure: Fig. 6.

Fig. 6. Residual error with modes 1–8 used in image reconstruction. Images are plotted under the same intensity scale. The reconstruction error decreases gradually as the number of modes used increases.

Download Full Size | PDF

 figure: Fig. 7.

Fig. 7. Reconstruction RMS pixel error using the first 10 modes. Blue-circle markers and red-square markers represent errors in the simulation and experiment, respectively.

Download Full Size | PDF

C. Measurement Function

The weights of KL modes 2–6, normalized by mode 1, together with the Gaussian center Cx and Cy, are used as the measurements. The measurement y can be written as

y=[w2w1,w3w1,w4w1,w5w1,w6w1,Cx,Cy].
The simulated image set described in Section 2.B.2 was obtained by scanning through the eight system states and generating 6561 images for the KL mode decomposition. Now we generate 60,000 images given random misaligned states to train the measurement function. Two-thirds of these images (40,000) are used as our training set, and the remaining one-third (20,000) as the test set.

We perform a nonlinear least-squares fitting on the training set using the Levenberg–Marquardt algorithm [27]. The nonlinear measurement model function h is learned to predict measurement y, computed as y^=h(x). Each nonlinear function hj is a second-order polynomial, which maps the misaligned state x to predict measurement y^, where j is the number of measurement from 1 to 7. The coefficients of the second-order polynomial is provided in Table 4 in Appendix A. The error between the simulated and predicted measurements is e=y^y, and the normalized root mean square error (NRMSE) of element j in e can be calculated as

NRMSE(ej)=(i=1nej,i)/nmaxyjminyj,
where i ranges from 1 to n for n points in the data set. Table 1 shows the NRMSE of the prediction on measurement 1 to 7. The NRMSE is calculated to ensure the model does not suffer from overfitting.

Tables Icon

Table 1. NRMSE of Measurement y1 to y7a

We plot the histogram of error e, as shown in Fig. 8. The blue and red lines are the best-fit normal distribution functions of the error distributions. The more Gaussian the distribution of the error, the better our measurement model performs in Kalman filtering. The error covariance matrix Rmodel is calculated as

Rmodel=eeT
and used as part of the measurement covariance matrix in Kalman filtering.

 figure: Fig. 8.

Fig. 8. Histogram of the residual e1 to e7 and their best fitted normal distribution. The closer the error distribution is to normal distribution, the more reliable the model is in Kalman filtering.

Download Full Size | PDF

D. Kalman Filtering

Our state space representation is

xk=Fkxk1+Bkuk+qk,yk=gk(xk)+rk,
where the state transition matrix Fk and the control input matrix Bk are taken to be identity matrices. The process noise qk is a zero mean Gaussian with covariance Qk, qkN(0,Qk). The observation function gk maps the misaligned states xk to the measurement yk, and rkN(0,Rk) is the measurement noise. The observation function is modeled by the nonlinear measurement function h learned in Section 2.C.

Kalman filtering is a sequential estimation algorithm, using a series of observations, together with statistical models of noise, to predict partially observed variables in a dynamic system. It iterates over two steps: dynamic propagation and measurement update. Kalman filtering is widely used in optical state estimation and wavefront control, including linear [28,29] and nonlinear [30] filters. In this section, an iterated extended Kalman filter [31] (IEKF) and an unscented Kalman filter [32,33] (UKF) are used to estimate the misaligned states. Extended Kalman filter (EKF) is a nonlinear version of a Kalman filter, which approximates the mean and covariance of current estimate using local linearization of the nonlinear function. IEKF is an iterative version of EKF, which ensures convergency in the measurement update step [34,35]. Because our state transition (Fk, Bk) is linear, the nonlinear approximation only occurs in the measurement update step where we calculate the Jacobian of nonlinear function h. UKF is also a nonlinear Kalman filter, which uses an unscented transform to estimate Gaussian distribution. The mean and covariance of state estimates are approximated by sigma points generated in the algorithm. The general process of Kalman filtering is shown as the lower dashed block in Fig. 2.

We generate process noise with covariance Qk and measurement noise with covariance Rmeas in the simulation. The processing noise is generated to model the stages in the experiment and has standard deviation [0.005,0.005,0.005,0.005,0.04,0.04,0.04,0.04]T. The measurement noise Rmeas is estimated by collecting many stationary images at multiple positions in the experiment and has standard deviation [0.0029,0.0082,0.0050,0.0219,0.0354,0.00025,0.00026]T. The total measurement noise covariance is Rk=Rmodel+Rmeas. The state estimate and estimate covariance in Kalman filtering are denoted as x^ and P^. We implement an IEKF and a UKF given initial guesses of the state estimate x^0 and state estimate covariance P^0.

Figure 9 shows the RMS state residual plot of IEKF and UKF given random control input in the simulation. The lines with blue-diamond red-circle markers represent the IEKF and UKF estimation, respectively. Both IEKF and UKF achieve an approximately 6 μm error in shift and 0.02 deg error in tip and tilt. The lines with green squares and magenta cross markers show the state residuals with full state feedback (uk=x^k1) after the 25th step in the IEKF and UKF estimation. Instead of feeding back the state estimate as the control input, we collect information by giving random inputs away from the center in the beginning of the closed-loop process. The reason we need a random walk for our system is to produce phase diversity, which is not available using a single focal plane image. The perturbation needs to be above a certain value to have diversity for us to track the state, and the number of perturbations needed depends on the initial guesses of the state estimate. The step at which we start feedback is decided by the state covariance obtained in IEKF or UKF. We start full state feedback after the state estimate covariance converges and remain stable for a few steps. In the simulation we are showing one of the worst cases where the state residual decreases gradually. This indicates the initial misalignment is at a position from which our algorithm takes a long time to converge. In most of the cases, the residual will drop down quickly in the first 2–10 steps. Twenty-five steps were taken to get to a point where both our simulation and experiment have state covariance converging to a stable value, for presenting our results.

 figure: Fig. 9.

Fig. 9. RMS state residuals of IEFK and UKF in the simulation. Blue-diamond line and red-circle line represent IEKF and UKF estimation. Green-square line and the magenta-cross line represent the state residuals with full state feedback after the 25th step in IEKF and UKF, respectively.

Download Full Size | PDF

Figure 10 shows the RMS standard deviation of state estimation using IEFK and UKF in the simulation. The standard deviation is the square root of each diagonal element of the state estimate covariance matrix Pk in IEKF and UKF. The lines with blue-diamond and red-circle markers represent IEKF and UKF estimation, respectively. In the initial estimation-only phase, the uncertainty in the state estimate can be seen to decrease rapidly in the first few steps. When we start feeding back the state estimate as a control input, the uncertainties become more stable, as shown by the green square and magenta cross marked lines.

 figure: Fig. 10.

Fig. 10. RMS standard deviation (STD) of state estimation using IEFK and UKF in the simulation. Blue-diamond line and red-circle line represent IEKF and UKF estimation. Green-square line and the magenta-cross line represent the STD of state estimate with full state feedback after the 25th step in IEKF and UKF, respectively.

Download Full Size | PDF

E. Experiment Setup

Figure 11 shows the experimental setup. The input 635 nm laser beam is passed through a customized collimator to produce a collimated beam as in the simulation. A neutral density (ND) filter is installed after the collimator to reduce the power of the laser beam. The moving lens A is placed after the ND filter, and the moving lens B is placed 400 mm after lens A. The CCD camera is 212 mm away from lens B. Both lens A and B are mounted on motorized tip–tilt and translation stages. The stages and CCD camera are connected to a local computer, which performs all data processing and can send actuation commands to the stages. Table 2 shows the detailed information of the components and devices in the experiment. The experimental results are presented in the next section.

 figure: Fig. 11.

Fig. 11. Experiment setup of optical model shown in Fig. 1. A collimated laser beam passes through a ND filter, two moving lenses A and B, and focuses on a CCD camera. The optical system after the ND filter is set up as the one simulated in ZEMAX.

Download Full Size | PDF

Tables Icon

Table 2. List of Components and Devices in the Experiment

3. RESULT

In this section we present the image reconstruction in the experiment and the result of closed-loop control with Kalman filtering.

A. Experiment Image Reconstruction

Figure 12 shows the residual error after reconstructing a single experimental image using the first eight KL modes derived in simulation. The corresponding sum of RMS pixel errors is shown as the red line with square markers in Fig. 7. The sum of RMS errors shown is the average over 500 images with random state inputs collected in the experiment. The reconstruction error in the experiment has the same trend as a function of KL mode number as in the simulation. The higher pixel error in the experiment is caused by the additional noise sources in the system, such as variations in the laser source, imperfections in the optics, and undamped vibrations in the lens stages.

 figure: Fig. 12.

Fig. 12. Residual error of the reconstruction of a single image acquired with the experimental setup shown in Fig. 11. The first eight KL modes derived in simulation are used. Images are plotted using the same intensity scale. As in simulation, the reconstruction error decreases gradually as the number of modes used increases.

Download Full Size | PDF

B. State Estimation Result

In the experiment, we give random input to the stages for the first 25 iteration steps, and then feedback the full state estimate as the control input uk=x^k1 from steps 26 through 50. The state residual cannot be obtained in the experiment because the true state is unknown. Figure 13 shows the stage position as a function of iteration using the IEKF. Stages 1–4 correspond to state elements x1 to x4, and stages 5–8 are the tip and tilt stages driven by translation motors. We decrease the process noise covariance matrix to Qk/4 from step 26 on, as the noise should be relatively small when the motor is moving in a small range. The stage positions converge to steady-state values, as shown in Fig. 13. The UKF experiment has similar results to the IEKF in terms of stage position.

 figure: Fig. 13.

Fig. 13. Stage positions from step 1 to 50 using IEKF. Stages 1–4 correspond to the shift stages, and stages 5–8 are the tip and tilt stages driven by translation motors.

Download Full Size | PDF

Figure 14 shows the initial and final experimental images, before and after closed-loop control. The left image shows the subframe before the state feedback, and the right image is the subframe after the state feedback converges. As expected, the image shifts to the center and becomes significantly more axisymmetric after state feedback converges.

 figure: Fig. 14.

Fig. 14. Experimental image before and after state feedback. The left image shows the 300×300 subframe before the correction, and the right image is the subframe after the correction. The intersection of the green lines represents the center of the camera.

Download Full Size | PDF

Figure 15 shows the RMS standard deviation of state estimation averaging over 20 executions of the closed-loop experiment. The lines with blue-diamond and red-circle markers represent the IEKF and UKF estimation, respectively. Similar to the simulation, the uncertainly drops down rapidly in the first few steps. The STD of tilt estimate increases slightly when we start feeding back the state estimates but drops down gradually after a few steps.

 figure: Fig. 15.

Fig. 15. RMS standard deviation of state estimate averaging over 20 experiments. Blue-diamond line and red-cross line represent IEKF and UKF, respectively.

Download Full Size | PDF

Although the stage positions converge to stable values in a single test, the final stable values vary somewhat between different runs. When the states are close to zero, it is likely to converge to a local minimum where the shifts, tip, and tilt of the two lenses compensate with each other. Table 3 shows the standard deviation of stage convergence value in 20 tests. The variance of the final convergence values is of the same order. There is no evidence showing that UFK outperforms IEKF on our system and vice versa. This indicates that the local linearization approximation in the IEKF is reasonable, as UFK would capture the nonlinearity better than IEKF.

Tables Icon

Table 3. STD of Stage Convergence Value in 20 Runsa

4. CONCLUSION AND FUTURE WORK

The automated alignment algorithm described here has two fundamental steps: 1. image processing and reconstruction; 2. state estimation and control. In the first step, the reconstruction error can be separated into the reconstruction bias and the detector noise. The bias is denoted as (v¯ic¯i) in Section 2.B.2 and is neglected in the calculations for this step because it does not have a direct effect on the measurement error and is considered as redundant information for correcting the system. In contrast, the noise in the image will be projected together with the image into the measurement y and becomes measurement noise (Rmeas) in the control and estimation step.

There are two additional error sources in the control and estimation step, which are the modeling error and the process noise (Qk). The modeling error includes the measurement function fitting error (Rmodel) and inconsistencies between the simulated optical system and the experimental setup. The inconsistencies can include errors in component placement as well as unmodeled effects such as thermal drifts; these effects cannot be corrected with the current implementation of Kalman filtering. The process noise (Qk) includes the actuator repeatability, backlash, stage hysteresis, etc. The overall experiment error is a combination of all of these error sources. The uncertainty of these factors results in the variation of the final stage position shown in Table 3. The less sensitive our lenses are to misalignment, the greater the variation of the stage positions will be. Different local minima are found in various iterations in the experiment. The variation of the pixel value between these iterations has an average standard deviation 24 in a 216 dynamic range. The pixel with the highest STD (600) occurs around the center of the each image.

Although z axis misalignments of lenses A and B are ignored in the current model, the despace parameter (z axis) is important in optical alignment. The component placement error might decrease if the z axis movement is included in state variables. The movement corresponds to focus motion and will occur mostly on axisymmetric KL modes such as Mode 3 (similar to defocus). A nonlinear function mapping from the despace parameter to the weight of KL modes can be learned. We expect the misalignment can be calibrated using an IEKF as long as the modeled nonlinear function has good performance using the local linearization approximation. If the learned measurement model function is highly nonlinear, we expect UKF will outperform IEKF and should be used for state estimation. The misalignment of the focus motions will be included in future applications.

In simulation, we have state residuals in multiple tests that show that the method is able to correct the misalignment to below a certain threshold. However, the method utilizes only an on-axis point source and does not consider the alignment effects on optical aberrations across full FOV. Because many optical instruments use wide FOVs, a particularly important set of future tests will be to evaluate the effects of the variability of the final converged stage positions on off-axis point sources and images of extended sources. To the extent that this variability represents the insensitivity of the whole imaging system to this level of misalignments, we would expect similar results for on-axis and off-axis sources. However, if the different results actually represent truly different local minima where the lenses compensate for each others’ misalignments in different ways, then we may expect different levels of distortion throughout the final image.

Our future work will also focus on extending the concept of our self-aligning method into various systems and applications. One of the extensions is to apply the method to reflective optical systems. For any system, model misalignments can be added into reflective and refractive moving components and similar steps to the ones described here can be performed. Some specific examples of possible applications are aligning components in a two-mirror telescope, aligning optics with human eyes in virtual reality headsets, and adjusting subsystems in a two triplet system. While aligning these optical systems, moving components in optical systems in practice are often restricted by the performance of the actuators. The actuators are limited by their repeatability and backlash. Further investigation on the limitation of actuators affecting the system to meet the performance specification is required.

APPENDIX A: MEASUREMENT FUNCTION

Tables Icon

Table 4. Coefficients of Nonlinear Measurement Function h(x)a

Funding

Cornell University.

REFERENCES

1. J. Rolland and H. Hua, “Head-mounted display systems,” in Encyclopedia of Optical Engineering (Taylor & Francis, 2005), pp. 1–13.

2. R. F. Dantowitz, S. W. Teare, and M. J. Kozubal, “Ground-based high-resolution imaging of mercury,” Astron. J. 119, 2455–2457 (2000). [CrossRef]  

3. N. Maeda, T. Fujikado, T. Kuroda, T. Mihashi, Y. Hirohara, K. Nishida, H. Watanabe, and Y. Tano, “Wavefront aberrations measured with Hartmann-Shack sensor in patients with keratoconus,” Ophthalmology 109, 1996–2003 (2002). [CrossRef]  

4. G. F. Hartig, J. E. Krist, A. R. Martel, H. C. Ford, and G. D. Illingworth, “On-orbit alignment and imaging performance of the HST advanced camera for surveys,” Proc. SPIE 4854, 532–543 (2003). [CrossRef]  

5. B. Macintosh, J. R. Graham, P. Ingraham, Q. Konopacky, C. Marois, M. Perrin, L. Poyneer, B. Bauman, T. Barman, A. S. Burrows, A. Cardwell, J. Chilcote, R. J. De Rosa, D. Dillon, R. Doyon, J. Dunn, D. Erikson, M. P. Fitzgerald, D. Gavel, S. Goodsell, M. Hartung, P. Hibon, P. Kalas, J. Larkin, J. Maire, F. Marchis, M. S. Marley, J. McBride, M. Millar-Blanchaer, K. Morzinski, A. Norton, B. R. Oppenheimer, D. Palmer, J. Patience, L. Pueyo, F. Rantakyro, N. Sadakuni, L. Saddlemyer, D. Savransky, A. Serio, R. Soummer, A. Sivaramakrishnan, I. Song, S. Thomas, J. K. Wallace, S. Wiktorowicz, and S. Wolff, “First light of the Gemini Planet Imager,” Proc. Natl. Acad. Sci. USA 111, 12661–12666 (2014). [CrossRef]  

6. D. Savransky, S. J. Thomas, L. A. Poyneer, and B. A. Macintosh, “Computer vision applications for coronagraphic optical alignment and image processing,” Appl. Opt. 52, 3394–3403 (2013). [CrossRef]  

7. R. Galicher, P. Baudoz, G. Rousset, J. Totems, and M. Mas, “Self-coherent camera as a focal plane wavefront sensor: simulations,” Astron. Astrophys. 509, A31 (2010). [CrossRef]  

8. J. W. Figoski, T. E. Shrode, and G. F. Moore, “Computer-aided alignment of a wide-field, three-mirror, unobscured, high-resolution sensor,” Proc. SPIE 1049, 166–177 (1989). [CrossRef]  

9. Z. Gao, L. Chen, S. Zhou, and R. Zhu, “Computer-aided alignment for a reference transmission sphere of an interferometer,” Opt. Eng. 43, 69–74 (2004). [CrossRef]  

10. S. Kim, H.-S. Yang, Y.-W. Lee, and S.-W. Kim, “Merit function regression method for efficient alignment control of two-mirror optical systems,” Opt. Express 15, 5059–5068 (2007). [CrossRef]  

11. H. Lee, G. B. Dalton, I. A. Tosh, and S.-W. Kim, “Computer-guided alignment ii: optical system alignment using differential wavefront sampling,” Opt. Express 15, 15424–15437 (2007). [CrossRef]  

12. E.-S. Oh, S. Kim, Y. Kim, H. Lee, S.-W. Kim, and H.-S. Yang, “Integration of differential wavefront sampling with merit function regression for efficient alignment of three-mirror anastigmat optical system,” Proc. SPIE 7793, 77930F (2010). [CrossRef]  

13. Z. Gu, C. Yan, and Y. Wang, “Alignment of a three-mirror anastigmatic telescope using nodal aberration theory,” Opt. Express 23, 25182–25201 (2015). [CrossRef]  

14. L. Pueyo, J. Kay, N. J. Kasdin, T. Groff, M. McElwain, and R. Belikov, “Optimal dark hole generation via two deformable mirrors with stroke minimization,” Appl. Opt. 48, 6296–6312 (2009). [CrossRef]  

15. T. Groff, N. J. Kasdin, and M. A. Peters, “Focal plane wavefront estimation using an integral field spectrograph,” in IEEE Aerospace Conference (IEEE, 2013), pp. 1–8.

16. D. Savransky, S. J. Thomas, L. A. Poyneer, D. W. Palmer, R. J. De Rosa, and M. Hartung, “Focal plane wavefront sensing and control for ground-based imaging,” Proc. SPIE 8447, 84476S (2012).

17. R. A. Gonsalves and R. Chidlaw, “Wavefront sensing by phase retrieval,” Proc. SPIE 0207, 32–39 (1979). [CrossRef]  

18. X. Wang, J. K. Wallace, and F. Shi, “Zernike wavefront sensor modeling development for LOWFS on WFIRST-AFTA,” Proc. SPIE 9605, 960528 (2015). [CrossRef]  

19. F. Martinache, “The asymmetric pupil Fourier wavefront sensor,” Publ. Astron. Soc. Pac. 125, 422–430 (2013). [CrossRef]  

20. H. Lee, G. J. Hill, S. E. Tuttle, and B. L. Vattiat, “Fine optical alignment correction of astronomical spectrographs via in-situ full-field moment-based wavefront sensing,” Proc. SPIE 8450, 84500V (2012). [CrossRef]  

21. U. Wandinger, “Introduction to lidar,” in Lidar (Springer, 2005), pp. 1–18.

22. O. Guyon, E. Pluzhnik, M. Kuchner, B. Collins, and S. Ridgway, “Theoretical limits on extrasolar terrestrial Planet detection with coronagraphs,” Astrophys. J. Suppl. Ser. 167, 81–99 (2006). [CrossRef]  

23. G. Dial, H. Bowen, F. Gerlach, J. Grodecki, and R. Oleszczuk, “IKONOS satellite, imagery, and products,” Remote Sens. Environ. 88, 23–36 (2003). [CrossRef]  

24. T. Toutin and P. Cheng, “Quickbird-a milestone for high-resolution mapping,” Earth Obser. Mag. 11, 14–18 (2002).

25. K. R. Rao and P. C. Yip, The Transform and Data Compression Handbook (CRC Press, 2000).

26. R. Soummer, L. Pueyo, and J. Larkin, “Detection and characterization of exoplanets and disks using projections on Karhunen-Loève eigenimages,” Astrophys. J. Lett. 755, L28 (2012). [CrossRef]  

27. J. J. Moré, “The Levenberg-Marquardt algorithm: implementation and theory,” in Numerical Analysis (Springer, 1978), pp. 105–116.

28. D. C. Redding, N. Sigrist, J. Z. Lou, Y. Zhang, P. D. Atcheson, D. S. Acton, and W. L. Hayden, “Optical state estimation using wavefront data,” Proc. SPIE 5523, 212–224 (2004). [CrossRef]  

29. J. Z. Lou, D. Redding, N. Sigrist, Y. Zhang, and S. Basinger, “JWST on-orbit multi-field wavefront control with a Kalman filter,” Proc. SPIE 5867, 58670T (2005). [CrossRef]  

30. A. E. Riggs, N. J. Kasdin, and T. D. Groff, “Wavefront correction with Kalman filtering for the WFIRST-AFTA coronagraph instrument,” Proc. SPIE 9605, 960507 (2015). [CrossRef]  

31. S. J. Julier and J. K. Uhlmann, “Unscented filtering and nonlinear estimation,” Proc. IEEE 92, 401–422 (2004). [CrossRef]  

32. E. Wan and R. Van Der Merwe, “The unscented Kalman filter for nonlinear estimation,” in Adaptive Systems for Signal Processing, Communications, and Control Symposium (AS-SPCC) (IEEE, 2000), pp. 153–158.

33. S. J. Julier and J. K. Uhlmann, “New extension of the Kalman filter to nonlinear systems,” in AeroSense’97 (International Society for Optics and Photonics, 1997), pp. 182–193.

34. R. F. Stengel, Optimal Control and Estimation (Dover, 2012).

35. J. Hedrick and A. Girard, “Control of nonlinear dynamic systems: theory and applications,” in Controllability and Observability of Nonlinear Systems (2005), p. 48.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (15)

Fig. 1.
Fig. 1. Two lens optical system. A collimated Gaussian beam is passed through two moving lenses A and B and focuses on a CCD camera.
Fig. 2.
Fig. 2. Schematic of control system. Upper dashed block represents the optical system in Fig. 1. Lower dashed block represents a Kalman filter.
Fig. 3.
Fig. 3. Subframe of the simulated image from Zemax and its 2D Gaussian fitting in contour plot. Gaussian center C x and C y will be used as our measurements.
Fig. 4.
Fig. 4. First 12 KL modes obtained from PCA decomposition with subframe 250 × 250 pixels. Each image is plotted under different intensity scale, and its corresponding eigenvalue is shown under the mode number in log scale.
Fig. 5.
Fig. 5. Reconstruction of simulated image in Fig. 3 using the first six KL modes. Subtracted image shows the difference ( v ¯ i c ¯ i ) between the simulated image in Fig. 3 and reconstructed images. (a) Reconstructed image. (b) Subtracted image.
Fig. 6.
Fig. 6. Residual error with modes 1–8 used in image reconstruction. Images are plotted under the same intensity scale. The reconstruction error decreases gradually as the number of modes used increases.
Fig. 7.
Fig. 7. Reconstruction RMS pixel error using the first 10 modes. Blue-circle markers and red-square markers represent errors in the simulation and experiment, respectively.
Fig. 8.
Fig. 8. Histogram of the residual e 1 to e 7 and their best fitted normal distribution. The closer the error distribution is to normal distribution, the more reliable the model is in Kalman filtering.
Fig. 9.
Fig. 9. RMS state residuals of IEFK and UKF in the simulation. Blue-diamond line and red-circle line represent IEKF and UKF estimation. Green-square line and the magenta-cross line represent the state residuals with full state feedback after the 25th step in IEKF and UKF, respectively.
Fig. 10.
Fig. 10. RMS standard deviation (STD) of state estimation using IEFK and UKF in the simulation. Blue-diamond line and red-circle line represent IEKF and UKF estimation. Green-square line and the magenta-cross line represent the STD of state estimate with full state feedback after the 25th step in IEKF and UKF, respectively.
Fig. 11.
Fig. 11. Experiment setup of optical model shown in Fig. 1. A collimated laser beam passes through a ND filter, two moving lenses A and B, and focuses on a CCD camera. The optical system after the ND filter is set up as the one simulated in ZEMAX.
Fig. 12.
Fig. 12. Residual error of the reconstruction of a single image acquired with the experimental setup shown in Fig. 11. The first eight KL modes derived in simulation are used. Images are plotted using the same intensity scale. As in simulation, the reconstruction error decreases gradually as the number of modes used increases.
Fig. 13.
Fig. 13. Stage positions from step 1 to 50 using IEKF. Stages 1–4 correspond to the shift stages, and stages 5–8 are the tip and tilt stages driven by translation motors.
Fig. 14.
Fig. 14. Experimental image before and after state feedback. The left image shows the 300 × 300 subframe before the correction, and the right image is the subframe after the correction. The intersection of the green lines represents the center of the camera.
Fig. 15.
Fig. 15. RMS standard deviation of state estimate averaging over 20 experiments. Blue-diamond line and red-cross line represent IEKF and UKF, respectively.

Tables (4)

Tables Icon

Table 1. NRMSE of Measurement y 1 to y 7 a

Tables Icon

Table 2. List of Components and Devices in the Experiment

Tables Icon

Table 3. STD of Stage Convergence Value in 20 Runs a

Tables Icon

Table 4. Coefficients of Nonlinear Measurement Function h ( x ) a

Equations (13)

Equations on this page are rendered with MathJax. Learn more.

T ( x , y ) = G 1 + G 2 exp ( ( x a ) 2 ( y b ) 2 2 ) x = ( x C x ) cos φ ( y C y ) sin φ y = ( x C x ) sin φ + ( y C y ) cos φ .
v ¯ i = v i μ ( v i ) ,
V ¯ = [ v ¯ 1 , v ¯ 2 , , v ¯ n ] .
S = 1 p 1 V ¯ T V ¯ ,
S Φ = Φ Λ ,
Z = V ¯ Φ ,
w i = Z m v ¯ i ,
c ¯ i = Z m w i ,
ϵ i = ( v ¯ i c ¯ i ) T ( v ¯ i c ¯ i ) p
y = [ w 2 w 1 , w 3 w 1 , w 4 w 1 , w 5 w 1 , w 6 w 1 , C x , C y ] .
NRMSE ( e j ) = ( i = 1 n e j , i ) / n max y j min y j ,
R model = ee T
x k = F k x k 1 + B k u k + q k , y k = g k ( x k ) + r k ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.