Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Measuring laser beams with a neural network

Open Access Open Access

Abstract

A deep neural network (NN) is used to simultaneously detect laser beams in images and measure their center coordinates, radii, and angular orientations. A dataset of images containing simulated laser beams and a dataset of images with experimental laser beams—generated using a spatial light modulator—are used to train and evaluate the NN. After training on the simulated dataset the NN achieves beam parameter root mean square errors (RMSEs) of less than 3.4% on the experimental dataset. Subsequent training on the experimental dataset causes the RMSEs to fall below 1.1%. The NN method can be used as a stand-alone measurement of the beam parameters or can compliment other beam profiling methods by providing an accurate region-of-interest.

Published by Optica Publishing Group under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

1. INTRODUCTION

Profiling multiple laser beams on a single image sensor has become increasingly important due to the growing number of multi-beam applications. Spatial light modulators (SLMs) [1], for example, can create multiple, dynamically controlled laser beams—used for optical tweezer arrays in cold atom experiments [24] and multi-site neuron activation in two-photon microscopy [5]—while diffractive optical elements allow multiple beams to be created for machining applications [6,7] and can also form laser beam arrays used in medical skin treatment procedures [8,9].

In recent years, deep neural networks (NNs) have been applied with great success to the analysis of scientific image data. Convolutional neural networks (CNNs) [10,11] often form the basis for image analysis NNs and have been used within optics for tasks such as laser beam mode classification [12,13], modal decomposition [1416], and determination of a beam’s center coordinates [17]. Object detection neural networks (ODNNs) [18,19], which are based on CNNs, can detect objects in images, classify the objects [20], and determine regions-of-interest (ROIs), which bound the objects. In this work, an ODNN [21] that returns rotated regions-of-interest (RROIs) is used to identify multiple ${{\rm{TEM}}_{00}}$ Gaussian laser beams in images and simultaneously measure all their spatial parameters.

The intensity distribution $I(x,y)$ for a ${{\rm{TEM}}_{00}}$ Gaussian beam (see Fig. 1) in a plane orthogonal to the beam’s axis of propagation is given by

$$I(x,y) = {I_0}{e^{- 2\left[{\frac{{{{\left[{\left({x - {x_0}} \right)\cos \theta + \left({y - {y_0}} \right)\sin \theta} \right]}^2}}}{{w_x^2}} + \frac{{{{\left[{\left({y - {y_0}} \right)\cos \theta - \left({x - {x_0}} \right)\sin \theta} \right]}^2}}}{{w_y^2}}} \right]}}{{,}}$$
where ${I_0}$ is the peak intensity of the beam, ${x_0}$ and ${y_0}$ are the beam’s center coordinates, ${w_x}$ and ${w_y}$ are the major and minor radii, and $\theta$ is the angular orientation. Although higher-order modes (e.g., Hermite–Gaussian) have experimental applications [22], we exclusively focus on the ${{\rm{TEM}}_{00}}$ Gaussian beams—henceforth referred to simply as Gaussian beams. The majority of laser beams used in both research and industrial applications have a mode content composed primarily of the ${{\rm{TEM}}_{00}}$ mode, thus making it a good approximation for a beam’s intensity distribution.
 figure: Fig. 1.

Fig. 1. Simulated images. (a), (b) Images sampled from the simulated dataset with a rotated region-of-interest (RROI) box plotted around each laser beam. The tops of the RROI boxes lie parallel to the major axis and are denoted with dashed lines.

Download Full Size | PDF

There are several standarized ways to measure Gaussian beams including scanning slit [23], knife edge [24], and camera-based methods [25]. Within the camera-based methods, the second moment measurement [26] is the industry standard [27] as it allows for fast calculation of multi-modal beams. However, the second moment method is prone to statistical error from image noise [28] and several standardized methods are used to combat this including thresholding low-intensity pixels and performing calculations within a ROI centered on the beam [27]. When profiling Gaussian beams, a two-dimensional (2D) fit of the Gaussian beam to Eq. (1) can also be used due to a priori knowledge of the beam mode. A properly chosen ROI increases the 2D fit accuracy by removing portions of the image without relevant data and also decreases the calculation time by fitting a smaller area. Even with an appropriate ROI, the 2D fit is significantly slower than the second moment, but allows for higher accuracy—particularly in noisy images.

For both the second moment and 2D fit methods, the ROI is generally found via an iterative method such as calculating the beam width within a ROI, recalculating the ROI using this value, and then repeating the process until the beam width converges [27]. However, iterative methods are computationally expensive, have difficulty converging if the image noise is too high, and are generally applicable when only a single laser beam is present in the image.

We present a deep NN-based method that allows for an arbitrary number of beams on a single image to be detected and their spatial parameters {${x_0}$, ${y_0}$, ${w_x}$, ${w_y}$, $\theta$} determined simultaneously, which significantly simplifies the laser beam analysis pipeline. If either the second moment or 2D fit of the beam is still required (e.g., for an ISO 11146-compliant beam measurement [27]), the spatial parameters returned by the NN can be used to determine, ROIs, RROIs, or elliptical RROIs in which these calculations can be performed. Furthermore, this method allows for measurement of Gaussian beams with overlapping edges—which cannot be done with the second moment method and would require prior knowledge of the number of beams for the 2D fit method.

This paper is organized as follows: Section 2 describes the NN model used to detect the laser beams and measure their spatial parameters. Section 3 and Section 4 explain how the simulated and experimental datasets are created. Finally, Section 5 discusses training the NN and the accuracy achieved for both detection and determination of the beams’ spatial parameters.

2. ROTATED REGION PROPOSAL NEURAL NETWORK

Although the Gaussian equation includes an intensity parameter (amplitude ${I_0}$), beam profiling is generally only concerned with the shape and location of the laser beam, which can be described by the spatial parameters {${x_0}$, ${y_0}$, ${w_x}$, ${w_y}$, $\theta$}. The goal is, therefore, to first detect each laser beam (object) in the image and then measure (regress) their geometric parameters.

ODNNs have been heavily researched in the last decade with several different architectures developed. Region-CNN (RCNN) [29] class NNs are extremely popular and utilize a CNN base followed by a region proposal network (RPN), which returns rough ROIs where objects are likely located. The CNN’s output is cropped and pooled using a ROI pooling/alignment stage and passed into one or more classification/regression branches—one of which regresses the ROI coordinates to yield a more accurate value. Although this could be useful for detecting beams, the ROI is aligned along the image axes and only yields information about the center coordinates {${x_0}$, ${y_0}$} and the projection of the beam radii on the image axes.

To regress all the beam’s geometric parameters, we use the rotated region proposal network (RRPN) [21], which was initially developed for detecting rotated text in images but is well suited for detecting laser beams. RRPN is similar to other RCNNs [18] but returns RROIs rather than ROIs. RROIs are rotated rectangles that are defined via the center coordinates {${x_{0r}}$, ${y_{0r}}$}, widths {${d_{\textit{xr}}}$, ${d_{\textit{yr}}}$}, and angular orientation ${\theta _r}$ of the rectangle. For a RROI centered and aligned on a laser beam in an image, the RROI parameters {${x_{0r}}$, ${y_{0r}}$, ${d_{\textit{xr}}}$, ${d_{\textit{yr}}}$, ${\theta _r}$} directly correspond to the laser beam’s geometric parameters and can, thus, be rewritten {${x_0}$, ${y_0}$, $\alpha {w_x}$, $\alpha {w_y}$, $\theta$}, where $\alpha$ is a scale factor relating the RROI widths and beam radii. Thus, RRPN can be used to simultaneously detect and measure laser beams.

RRPN (see Fig. 2) begins with a CNN base (we use ResNet50 [11]), which outputs a feature map [30] for a given input. The feature map is then fed into the RRPN (single node within RRPN), which is similar to the RPN in Faster-RCNN [18] but returns rough RROIs—rather than ROIs—where objects are likely to be located. The RROIs and feature map are both passed into the RROI alignment stage [31], which returns fixed-size feature maps via bi-linear interpolation as successive layers require a fixed input. The fixed-size feature maps are then fed into two separate branches: the first classifies the object within the RROI and assigns a score to its prediction, whereas the second does further regression of the RROI parameters.

 figure: Fig. 2.

Fig. 2. Rotated region proposal networks (RRPNs) model. The neural network begins with a convolutional neural network (CNN) base, which returns a feature map of its input. The feature map is passed into the RRPN, which gives a set of rough rotated regions-of-interest (RROIs) where beams are likely located. These RROIs are used to crop and align the CNN’s feature map to fixed dimensions in the RROI alignment stage after which the fixed-size feature maps are passed into two parallel branches. The first branch classifies the object and assigns a score to its prediction, whereas the second returns a more accurate RROI.

Download Full Size | PDF

To train RRPN, two datasets are created: the first dataset is composed of images with simulated Gaussian beams, whereas the second is composed of experimental images with the beams generated using a SLM. For both the simulated and experimental datasets, the images contain between one and five laser beams, although RRPN could easily be trained to detect a larger number of beams on a single image.

3. SIMULATED DATASET

CNN’s require diverse image training data to allow them to generalize to new data during inference. However, supervised-learning dataset sizes are often limited due to practical considerations such as the time it takes to manually annotate images. Simulated data allow the annotation bottleneck to be circumvented [3234] as the annotations are calculated directly from the simulation parameters. Since Gaussian beams are relatively easy to simulate, we can create an arbitrarily large dataset filled with unique images by randomizing each beam’s Gaussian parameters (see Fig. 1).

When randomizing a beam’s parameters, an initial beam radius is first drawn from a uniform distribution with a minimum bound of 5 pixels and a maximum value of 1/6 the 512 pixel image width. An ellipticity value is then drawn from a normal distribution and multiplied by the initial radius to create the second radius value. The larger of the two radii is the major radius ${w_x}$, the smaller is the minor radius ${w_y}$, and the angular orientation $\theta$ defines the angle between ${w_x}$ and the $x$ axis. The angular orientations are randomly chosen between $-\frac{\pi}{2}$ and $\frac{\pi}{2}$—which gives a strict definition the NN can learn to regress while still covering the full range of possible orientations.

The beam’s center coordinates are randomly drawn from a uniform distribution, but they are subject to the constraint that the beam’s entire RROI must lie on the simulated sensor surface. Additionally, when more than one beam is present, the overlap between beams is restricted to the edges of the distributions. Beam amplitudes are randomly chosen between 0.1 and 1 for all beams and simulated Gaussian noise—with a standard deviation ${\sigma _n}$ randomly chosen—is added to the image. The background intensity is set to ${I_{{b}}} = 2.5{\sigma _n}$ to prevent the Gaussian noise from being substantially clipped.

Using the process above, a simulated dataset with 5000 simulated images ($512 \times 512$ pixels; see Fig. 1) is generated—1000 for each beam class (number of beams on the image). The initially monochrome images are normalized and mapped to RGB using the Viridis colormap as the pre-trained NN we use (see Section 5) expects RGB input. The ground truth RROI annotations are calculated from the simulated beam parameters and defined as {${x_0}$, ${y_0}$, $\alpha {w_x}$, $\alpha {w_y}$, $\theta$}, where we choose $\alpha = 3$. Finally, the dataset is randomly split into a training set with 4000 images and validation dataset with 1000 images.

4. EXPERIMENTAL DATASET

The experimental dataset is created using a SLM [1], which allows structured light [35,36] to be created using holograms [see Fig. 3(a)]. Our setup begins with a MSquared frequency doubled Ti-sapphire laser, which produces 370 nm light and is coupled into a single-mode optical fiber. The beam exits the fiber and is collimated and reflected off the SLM’s surface—after which the beam passes through a $f=400\,\,\rm mm$ lens placed a distance $f$ away from the SLM. An aperture is used to select the first-order diffracted light, which is subsequently imaged by a camera placed at the lens’ focus (Fourier plane).

 figure: Fig. 3.

Fig. 3. Experimental setup. (a) A 370 nm laser beam exits a single-mode fiber and is collimated with a 100 mm lens before being reflected from a Hamamatsu X13267-05 spatial light modulator (SLM)—which features an ${{800}}\;{{\times}}\;{{600}}$ grid with a pixel pitch of $12.5\,\,\unicode{x00B5}\rm m$. A $\lambda /2$ wave plate between the fiber coupler and collimating lens sets the beam’s polarization parallel to the SLM’s vertical axis as the SLM is polarization sensitive. The reflected beam is focused using a $f = 400\,\,\rm mm$ lens placed a distance $f$ from the SLM and passes through an aperture placed directly (1 cm) before the lens’ focus—where the different diffraction orders can be resolved. The aperture allows through only the first-order diffracted light, which is then imaged by a camera at the Fourier plane. (b) A complex amplitude modulation hologram used to generate Gaussian beams in the Fourier plane for the experimental dataset. (c) Image generated using the hologram in (b).

Download Full Size | PDF

Complex amplitude modulation holograms (CAMs) [37] allow both the amplitude and phase of the electric field to be modulated using a phase-only SLM [see Fig. 3(b)]. The amplitude and phase of the laser beam are encoded into a single hologram [38] along with a blazed grating so that the beam parameters {${I_0}$, ${x_0}$, ${y_0}$, ${w_x}$, ${w_y}$, $\theta$} can be dynamically set in the Fourier plane. Adding multiple CAM holograms together—each with different blazed grating frequencies—creates multiple beams in the Fourier plane [see Fig. 3(c)], which can have different sizes, orientations, and positions.

The experimental images contain between one and five beams with the CAM parameters for each beam randomly drawn from a uniform distribution—within physically realizable bounds. The ground truth beam parameters {${x_0}$, ${y_0}$, ${w_x}$, ${w_y}$, $\theta$} at the Fourier plane were found for each beam by performing a 2D fit within a ROI $2\times$ the $1/{e^2}$ radii; however, for overlapping beams, a multi-Gaussian fit was performed. After fitting all the beams in the experimental dataset, the fits were manually inspected   [39] and subsequently used to calculate the RROIs as {${x_0}$, ${y_0}$, $\alpha {w_x}$, $\alpha {w_y}$, $\theta$}, where again $\alpha = 3$.

The 1050 images in the experimental dataset—210 images for each number of beams—are split into a training dataset with 800 images and a validation dataset with 250 images. Although the camera has a sensor size of $1280 \times 1024$ pixels, the beams in the Fourier plane are incident on a small area of the sensor, and the images are cropped to $256 \times 256$ pixels. As with the simulated dataset, the image intensities are normalized and mapped to RGB.

5. TRAINING AND EVALUATION

Rather than building the NN model from scratch, Facebook Artificial Intelligence Research’s (FAIR) Detectron2 [40] framework is utilized, which implements common machine vision models and is written for speedy training and inference. Since Detectron2 only has pre-trained weights for the CNN base, transfer learning [41] can only be partially implemented and significantly more images are needed to train the NN. Our strategy is therefore to first pre-train RRPN on the larger, more diverse simulated dataset before (optionally) doing final training on the experimental dataset.

RRPN is trained on the simulation dataset for 120 epochs using a stochastic gradient descent optimizer with an initial learning rate, which is decayed four times. The initial learning rate, learning rate decay scalar, and the epochs at which the learning rate is decayed are all used as hyperparameters—along with the momentum and batch size. Nominally either random search [42] or Bayesian optimization (BO) would be used to tune the hyperparameters; however, due to the large size of the simulated dataset and limited computational resources (we train our NN in a Google Colab [43] notebook), we manually set the hyperparameters to sensible values: batch size of 4, learning rate of 0.01, momentum of 0.9, learning rate decay of 0.1, and learning rate decay at epochs {80, 100, 110, 115}.

After each training epoch, the NN is evaluated on both the simulated and experimental validation datasets using the mean average precision (mAP) metric [44]. The mAP is a standard object detection metric where intersection-over-union (IoU) scores [45] between the ground truth and NN predicted RROIs are used to form precision-recall curves [46] for different IoU thresholds—which are integrated and averaged to give the mAP value [44].

At the beginning of training, the loss quickly decays while the mAPs climb, whereas toward the end of training both the loss and mAP values become asymptotic [see Fig. 4(a)]. Maximum mAPs of 96.8% and 93.9% are achieved on the simulation and experimental validation datasets, respectively, which correspond to the NN correctly finding 2996/3000 beams [47] on the simulated validation dataset and 750/750 beams on the experimental validation dataset (correct prediction threshold is set as an ${\rm{IoU}} \gt {0.5}$ between the ground truth and predicted RROIs).

 figure: Fig. 4.

Fig. 4. Training on the simulated dataset. (a) The loss on the simulated training dataset and the mean average precision (mAP) on the simulated and experimental validation datasets plotted against the training epoch. (b) The root mean square errors (RMSEs) for the simulation validation dataset beam parameters {${x_0}$, ${y_0}$, ${w_x}$, ${w_y}$, $\theta$ } versus the training epoch. The spatial parameters are normalized by the beam size (see text), and the angular orientation is normalized by $\pi$.

Download Full Size | PDF

Along with the mAP, the beam parameter errors are calculated after each training epoch. For both the ground truth and NN, the parameters are normalized before calculating the error. The center coordinates {${x_0}$, ${y_0}$} are normalized by dividing by the ground truth beam radii along the $x$ and $y$ laboratory axes, whereas the major and minor radii {${w_x}$, ${w_y}$} are divided by the ground truth major and minor radii. For the angular orientation error, $\theta$ is normalized by dividing by the range of angles, $\pi$. Beams with ellipticities below ${w_x}/{w_y} = 1.15$ are considered radially symmetric [27] and removed from the further angular error calculations as the angle for radially symmetric beams is arbitrary.

The spatial parameter root mean square errors (RMSEs) are calculated for each validation dataset. As the training epoch increases, the RMSEs decrease [see Fig. 4(b)], and for the training epochs with the best mAPs the simulated validation dataset RMSEs are all less than 2.6% (see Table 1), whereas the experimental validation dataset RMSEs are less than 3.4%. Note that the simulated dataset contains beams that are difficult to detect/measure such as highly elliptical beams or beams with a low signal-to-noise ratio. These help the NN learn a general definition of the Gaussian beam for inference on unseen experimental data. The high accuracy of the NN on the experimental validation dataset demonstrates the validity of this approach.

Tables Icon

Table 1. Gaussian Beam Parameter Root Mean Square Errorsa at the Training Epoch with the Highest Mean Average Precision

After training on the simulated dataset, the NN model weights are retained, and the NN is trained on the experimental dataset. BO [48,49] is used to tune the hyperparameters within sensible bounds (see Table 2)—this time with a single learning rate decay—using Facebook’s Ax/BoTorch [50] package. Five Sobol [51] evaluations are used to initialize the BO loop, after which a Gaussian process iteratively determines the hyperparameters for the remaining 10 evaluations. For each BO evaluation, the NN is trained for 30 epochs, which still allows high accuracies to be reached due to pre-training on the simulated dataset.

Tables Icon

Table 2. Bounds and Scaling Used During Bayesian Optimization of the Neural Network Hyperparameters for Training on the Experimental Dataset, Along with the Best Set of Hyperparameters Found

Similar to the simulated training run, both the mAP and RMSEs are calculated after every training epoch on the experimental validation dataset. The NN trained with the best set of hyperparameters achieves a mAP of 97.7% and successfully detects all 750 beams. Furthermore, the Gaussian parameter RMSEs are all below 1.1% (see Table 1), which are lower than the NN’s experimental validation dataset RMSEs when trained on the simulated dataset alone; however, the accuracy gain is not substantial.

6. CONCLUSION

The method developed uses a deep NN to detect an arbitrary number of Gaussian laser beams in an image and simultaneously measure their spatial parameters. The NN requires a single pass on an image, which significantly simplifies the beam analysis pipeline compared to other methods. Since training the NN on simulated data alone results in high accuracies on experimental data, this method can be applied in a wide range of experimental settings.

The NN can be used alone or can compliment other beam measurement methods by detecting laser beams in an image and determining ROIs for further calculations, such as a 2D fit or second moment measurement. This removes the need for iterative ROI algorithms, which can generally only find a single beam. Furthermore, if 2D fitting is implemented, the beam parameters extracted with the NN can be used to seed the fit, which increases fitting speed and the likelihood of fit convergence.

Although this method is applied to ${{\rm{TEM}}_{00}}$ beams, it can be extended to higher-order Gaussian modes since RRPN natively handles multiple object types. However, this would require the training datasets to include higher-order beams with a label for each beam mode. Furthermore, for higher-order and multi-modal beams, the radius is determined numerically rather than analytically—which would need to be accounted for when generating the datasets’ RROI annotations.

Funding

John Fell Oxford University (OUP) Research Fund; The Royal Society; Engineering and Physical Sciences Research Council (EP/P009565/1, EP/TO19913/1).

Acknowledgment

L. H. thanks Maximilian Pflüger for helpful discussions.

Disclosures

L. H. was previously employed at DataRay Inc. The authors declare no conflicts of interest.

Data availability

The data that support the findings of this study are openly available in [52]. We additionally make code available in [53].

REFERENCES AND NOTES

1. N. Konforti, E. Marom, and S.-T. Wu, “Phase-only modulation with twisted nematic liquid-crystal spatial light modulators,” Opt. Lett. 13, 251–253 (1988). [CrossRef]  

2. D. Barredo, S. de Léséleuc, V. Lienhard, T. Lahaye, and A. Browaeys, “An atom-by-atom assembler of defect-free arbitrary two-dimensional atomic arrays,” Science 354, 1021–1023 (2016). [CrossRef]  

3. D. Ohl de Mello, D. Schäffner, J. Werkmann, T. Preuschoff, L. Kohfahl, M. Schlosser, and G. Birkl, “Defect-free assembly of 2D clusters of more than 100 single-atom quantum systems,” Phys. Rev. Lett. 122, 203601 (2019). [CrossRef]  

4. M. Endres, H. Bernien, A. Keesling, H. Levine, E. R. Anschuetz, A. Krajenbrink, C. Senko, V. Vuletic, M. Greiner, and M. D. Lukin, “Atom-by-atom assembly of defect-free one-dimensional cold atom arrays,” Science 354, 1024–1027 (2016). [CrossRef]  

5. V. Nikolenko, B. Watson, R. Araya, A. Woodruff, D. Peterka, and R. Yuste, “SLM microscopy: scanless two-photon imaging and photostimulation using spatial light modulators,” Front. Neural Circuits 2, 5 (2008). [CrossRef]  

6. P. Hauschwitz, B. Stoklasa, J. Kuchařík, H. Turčičová, M. Písařík, J. Brajer, D. Rostohar, T. Mocek, M. Duda, and A. Lucianetti, “Micromachining of invar with 784 beams using 1.3 ps laser source at 515 nm,” Materials 13, 2962 (2020). [CrossRef]  

7. S. Katz, N. Kaplan, and I. Grossinger, “Using diffractive optical elements: does for beam shaping–fundamentals and applications,” Opt. Photon. 13, 83–86 (2018). [CrossRef]  

8. E. A. Tanghetti, “The histology of skin treated with a picosecond alexandrite laser and a fractional lens array,” Laser Surg. Med. 48, 646–652 (2016). [CrossRef]  

9. H. C. Lee, J. Childs, H. J. Chung, J. Park, J. Hong, and S. B. Cho, “Pattern analysis of 532-and 1,064-nm picosecond-domain laser-induced immediate tissue reactions in ex vivo pigmented micropig skin,” Sci. Rep. 9, 4186 (2019). [CrossRef]  

10. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Comm ACM 60, 84–90 (2017). [CrossRef]  

11. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (2016), pp. 770–778, https://doi.org/10.1109/CVPR.2016.90.

12. T. Doster and A. T. Watnik, “Machine learning approach to OAM beam demultiplexing via convolutional neural networks,” Appl. Opt. 56, 3386–3396 (2017). [CrossRef]  

13. L. R. Hofer, L. W. Jones, J. L. Goedert, and R. V. Dragone, “Hermite–Gaussian mode detection via convolution neural networks,” J. Opt. Soc. Am. A 36, 936–943 (2019). [CrossRef]  

14. S. Lohani, E. M. Knutson, M. O’Donnell, S. D. Huver, and R. T. Glasser, “On the use of deep neural networks in optical communications,” Appl. Opt. 57, 4180–4190 (2018). [CrossRef]  

15. Y. An, T. Hou, J. Li, L. Huang, J. Leng, L. Yang, and P. Zhou, “Fast modal analysis for Hermite–Gaussian beams via deep learning,” Appl. Opt. 59, 1954–1959 (2020). [CrossRef]  

16. M. G. Schiworski, D. D. Brown, and D. J. Ottaway, “Modal decomposition of complex optical fields using convolutional neural networks,” J. Opt. Soc. Am. A 38, 1603–1611 (2021). [CrossRef]  

17. C.-S. Lin, Y.-C. Huang, S.-H. Chen, Y.-L. Hsu, and Y.-C. Lin, “The application of deep learning and image processing technology in laser positioning,” Appl. Sci. 8, 1542 (2018). [CrossRef]  

18. S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: towards real-time object detection with region proposal networks,” in Twenty-ninth Conference on Neural Information Processing Systems (2015) pp. 91–99, https://proceedings.neurips.cc/paper/2015/file/14bfa6bb14875e45bba028a21ed38046-Paper.pdf.

19. J. Redmon and A. Farhadi, “YOLOv3: an incremental improvement,” arXiv:1804.02767 (2018), https://arxiv.org/abs/1804.02767.

20. L. R. Hofer, M. Krstajić, P. Juhász, A. L. Marchant, and R. P. Smith, “Atom cloud detection and segmentation using a deep neural network,” Mach. Learn. Sci. Technol. 2, 045008 (2021). [CrossRef]  

21. J. Ma, W. Shao, H. Ye, L. Wang, H. Wang, Y. Zheng, and X. Xue, “Arbitrary-oriented scene text detection via rotation proposals,” IEEE Trans. Multimedia 20, 3111–3122 (2018). [CrossRef]  

22. A. L. Gaunt, T. F. Schmidutz, I. Gotlibovych, R. P. Smith, and Z. Hadzibabic, “Bose-Einstein condensation of atoms in a uniform potential,” Phys. Rev. Lett. 110, 200406 (2013). [CrossRef]  

23. R. L. McCally, “Measurement of Gaussian beam parameters,” Appl. Opt. 23, 2227 (1984). [CrossRef]  

24. A. E. Siegman, M. Sasnett, and T. Johnston, “Choice of clip levels for beam width measurements using knife-edge techniques,” IEEE J. Quantum Electron. 27, 1098–1104 (1991). [CrossRef]  

25. A. E. Siegman, “How to (maybe) measure laser beam quality,” in DPSS (Diode Pumped Solid State) Lasers: Applications and Issues (1998), paper MQ1, https://doi.org/10.1364/DLAI.1998.MQ1.

26. T. S. Ross, Laser Beam Quality Metrics (SPIE, 2013), https://doi.org/10.1117/3.1000595.

27. “Lasers and laser-related equipment-test methods for laser beam widths, divergence angles and beam propagation ratios—part 1: stigmatic and simple astigmatic beams,” ISO 11146-1 (2005).

28. L. R. Hofer, R. V. Dragone, and A. D. MacGregor, “Scale factor correction for Gaussian beam truncation in second moment beam radius measurements,” Opt. Eng. 56, 043110 (2017). [CrossRef]  

29. R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in IEEE Conference on Computer Vision and Pattern Recognition (2014), pp. 580–587, https://doi.org/10.1109/CVPR.2014.81.

30. M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks,” in Computer Vision (ECCV) (2014), pp. 818–833, https://doi.org/10.1007/978-3-319-10590-1.

31. J. Huang, V. Sivakumar, M. Mnatsakanyan, and G. Pang, “Improving rotated text detection with rotation region proposal networks,” arXiv:1811.07031 (2018), https://arxiv.org/abs/1811.07031.

32. E. Wood, T. Baltrušaitis, L.-P. Morency, P. Robinson, and A. Bulling, “Learning an appearance-based gaze estimator from one million synthesised images,” in Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications (2016), pp. 131–138, https://doi.org/10.1145/2857491.2857492.

33. D. Yoo, N. Kim, S. Park, A. S. Paek, and I. S. Kweon, “Pixel-level domain transfer,” in European Conference on Computer Vision (2016), pp. 517–532, http://doi.org/10.1007/978-3-319-46484-8_31.

34. X. Zhang, Y. Fu, A. Zang, L. Sigal, and G. Agam, “Learning classifiers from synthetic data using a multichannel autoencoder,” arXiv:1503.03163 (2015), https://arxiv.org/abs/1503.03163.

35. A. Forbes, “Structured light from lasers,” Laser Photon. Rev. 13, 1900140 (2019). [CrossRef]  

36. A. Forbes, A. Dudley, and M. McLaren, “Creation and detection of optical modes with spatial light modulators,” Adv. Opt. Photon. 8, 200–227 (2016). [CrossRef]  

37. V. Arrizón, U. Ruiz, R. Carrada, and L. A. González, “Pixelated phase computer holograms for the accurate encoding of scalar complex fields,” J. Opt. Soc. Am. A 24, 3500–3507 (2007). [CrossRef]  

38. C. Rosales-Guzmán and A. Forbes, How to Shape Light with Spatial Light Modulators (SPIE, 2017), https://doi.org/10.1117/3.2281295.

39. We also take a histogram of the reduced ${\chi ^2}$ values for all the fitted beams, which has a mean of 1.1 and standard deviation of 0.2. This indicates the experimental beams’ intensity distributions are well approximated by the 2D Gaussian.

40. Y. Wu, A. Kirillov, F. Massa, W.-Y. Lo, and R. Girshick, “Detectron2,” GitHub (2019), https://github.com/facebookresearch/detectron2.

41. J. Yosinski, J. Clune, Y. Bengio, and H. Lipson, “How transferable are features in deep neural networks?” in Twenty-eighth Conference on Neural Information Processing Systems (2014), pp. 3320–3328, https://proceedings.neurips.cc/paper/2014/file/375c71349b295fbe2dcdca9206f20a06-Paper.pdf.

42. J. Bergstra and Y. Bengio, “Random search for hyper-parameter optimization,” J. Mach. Learn. Res. 13, 281–305 (2012). [CrossRef]  

43. E. Bisong, “Python,” in Building Machine Learning and Deep Learning Models on Google Cloud Platform (Springer, 2019), pp. 59–64, https://link.springer.com/book/10.1007/978-1-4842-4470-8.

44. T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft COCO: common objects in context,” in Computer Vision (ECCV) (2014), pp. 740–755, https://doi.org/10.1007/978-3-319-10602-1.

45. M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman, “The Pascal visual object classes (VOC) challenge,” Int. J. Comput. Vis. 88, 303–338 (2010). [CrossRef]  

46. K. Boyd, K. H. Eng, and C. D. Page, “Area under the precision-recall curve: point estimates and confidence intervals,” in Joint European Conference on Machine Learning and Knowledge Discovery in Databases (Springer, 2013), pp. 451–466, http://doi.org/10.1007/978-3-642-40994-3.

47. For 35 beams, the NN predicts two RROIs, but the extra predictions are easily filtered using the IoU between predictions and the NN’s prediction score.

48. J. Snoek, H. Larochelle, and R. P. Adams, “Practical Bayesian optimization of machine learning algorithms,” in Advances in Neural Information Processing Systems 25 (NIPS 2012) (Morgan Kaufmann Publishers Inc., 2012), pp. 2951–2959, https://proceedings.neurips.cc/paper/2012/file/05311655a15b75fab86956663e1819cd-Paper.pdf.

49. P. I. Frazier, “A tutorial on Bayesian optimization,” arXiv:1807.02811 (2018), https://arxiv.org/abs/1807.02811.

50. M. Balandat, B. Karrer, D. R. Jiang, S. Daulton, B. Letham, A. G. Wilson, and E. Bakshy, “BoTorch: programmable Bayesian optimization in PyTorch,” arXiv:1910.06403 (2020), https://arxiv.org/abs/1910.06403.

51. I. M. Sobol’, “On the distribution of points in a cube and the approximate evaluation of integrals,” USSR Computational Mathematics and Mathematical Physics 7, 86–112 (1967). [CrossRef]  

52. L. Hofer, M. Krstajić, and R. P. Smith, “Measuring Laser Beams with a Neural Network (Data),” University of Oxford Research Archive (2022), https://doi.org/10.5287/bodleian:JbDXrnQN1

53. L. Hofer, M. Krstajić, and R. P. Smith, “Measuring Laser Beams with a Neural Network,” GitHub (2022), https://github.com/Dipolar-Quantum-Gases/nn-beam-profiling.

Data availability

The data that support the findings of this study are openly available in [52]. We additionally make code available in [53].

52. L. Hofer, M. Krstajić, and R. P. Smith, “Measuring Laser Beams with a Neural Network (Data),” University of Oxford Research Archive (2022), https://doi.org/10.5287/bodleian:JbDXrnQN1

53. L. Hofer, M. Krstajić, and R. P. Smith, “Measuring Laser Beams with a Neural Network,” GitHub (2022), https://github.com/Dipolar-Quantum-Gases/nn-beam-profiling.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (4)

Fig. 1.
Fig. 1. Simulated images. (a), (b) Images sampled from the simulated dataset with a rotated region-of-interest (RROI) box plotted around each laser beam. The tops of the RROI boxes lie parallel to the major axis and are denoted with dashed lines.
Fig. 2.
Fig. 2. Rotated region proposal networks (RRPNs) model. The neural network begins with a convolutional neural network (CNN) base, which returns a feature map of its input. The feature map is passed into the RRPN, which gives a set of rough rotated regions-of-interest (RROIs) where beams are likely located. These RROIs are used to crop and align the CNN’s feature map to fixed dimensions in the RROI alignment stage after which the fixed-size feature maps are passed into two parallel branches. The first branch classifies the object and assigns a score to its prediction, whereas the second returns a more accurate RROI.
Fig. 3.
Fig. 3. Experimental setup. (a) A 370 nm laser beam exits a single-mode fiber and is collimated with a 100 mm lens before being reflected from a Hamamatsu X13267-05 spatial light modulator (SLM)—which features an ${{800}}\;{{\times}}\;{{600}}$ grid with a pixel pitch of $12.5\,\,\unicode{x00B5}\rm m$. A $\lambda /2$ wave plate between the fiber coupler and collimating lens sets the beam’s polarization parallel to the SLM’s vertical axis as the SLM is polarization sensitive. The reflected beam is focused using a $f = 400\,\,\rm mm$ lens placed a distance $f$ from the SLM and passes through an aperture placed directly (1 cm) before the lens’ focus—where the different diffraction orders can be resolved. The aperture allows through only the first-order diffracted light, which is then imaged by a camera at the Fourier plane. (b) A complex amplitude modulation hologram used to generate Gaussian beams in the Fourier plane for the experimental dataset. (c) Image generated using the hologram in (b).
Fig. 4.
Fig. 4. Training on the simulated dataset. (a) The loss on the simulated training dataset and the mean average precision (mAP) on the simulated and experimental validation datasets plotted against the training epoch. (b) The root mean square errors (RMSEs) for the simulation validation dataset beam parameters {${x_0}$, ${y_0}$, ${w_x}$, ${w_y}$, $\theta$ } versus the training epoch. The spatial parameters are normalized by the beam size (see text), and the angular orientation is normalized by $\pi$.

Tables (2)

Tables Icon

Table 1. Gaussian Beam Parameter Root Mean Square Errorsa at the Training Epoch with the Highest Mean Average Precision

Tables Icon

Table 2. Bounds and Scaling Used During Bayesian Optimization of the Neural Network Hyperparameters for Training on the Experimental Dataset, Along with the Best Set of Hyperparameters Found

Equations (1)

Equations on this page are rendered with MathJax. Learn more.

I(x,y)=I0e2[[(xx0)cosθ+(yy0)sinθ]2wx2+[(yy0)cosθ(xx0)sinθ]2wy2],
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.