Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Real-time particle pollution sensing using machine learning

Open Access Open Access

Abstract

Particle pollution is a global health challenge that is linked to around three million premature deaths per year. There is therefore great interest in the development of sensors capable of precisely quantifying both the number and type of particles. Here, we demonstrate an approach that leverages machine learning in order to identify particulates directly from their scattering patterns. We show the capability for producing a 2D sample map of spherical particles present on a coverslip, and also demonstrate real-time identification of a range of particles including those from diesel combustion.

Published by The Optical Society under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

1. Introduction

Airborne pollution particles are a global health challenge since they are linked to illnesses such as asthma, cancer, heart disease and dementia, causing around three million premature deaths a year [1]. Particle characterisation is therefore invaluable in atmospheric pollution monitoring, where aerosols [2] and black carbon [3,4], along with other pollutants such as particles from diesel combustion [5] (referred to here as diesel soot), are understood to influence human health [6–9]. Harmful contaminants such as polystyrene microspheres that can be present in the marine environment and have a negative effect on marine life [10,11], and thus human life, would also be invaluable to monitor. Therefore, the development of techniques that can be used to characterise a broad range of particles is a key strategy in reducing the number of deaths caused by pollution. To date, environmental monitoring through object classification direct from observation of the scattered light has been demonstrated in both water and using aerosol particles [12,13].

Light that is scattered from an object encodes rich information about the shape and structure of that object [14,15]. However, without specialised equipment, only the intensity of the scattered light can be recorded and the phase information is lost. Therefore, an exact description of the object is not directly possible, as both amplitude and phase of the scattered light is required. The challenge of producing the inverse function that maps a scattering pattern to the object has led to lensless imaging approaches such as phase retrieval [16–20] and ptychography [21–23]. Phase retrieval offers a solution by requiring that the object is oversampled, which is equivalent to ensuring that the object has zero padding (i.e. no intensity contribution) outside a well-defined region. Therefore, this approach is generally only applicable to objects that are smaller than the size of the illuminating light source. Ptychography enables imaging over a continuous object, but requires the collection of scattering patterns that correspond to overlapping regions of the object, where, in general, the degree of overlap must also be measured, along with the illumination function. A general solution to this inverse problem, which places no limitations on the sample or the illumination function, would therefore be of significant value.

Neural networks (NNs) [24–26] are a computing paradigm inspired by the interconnected network of neurons present in biological intelligence. Rather than being programmed with specific instructions, a NN learns an algorithmic process from processing input-output data. With appropriate training data, a NN therefore has the potential to learn a direct transfer function from scattering pattern to quantification of the sample, without any need for understanding the physical laws of light-matter interaction. NNs can therefore forgo the need to develop complex light-interaction models [27,28]. An additional advantage of using NNs for sensing is the recently shown ability to update the NN for additional particle type determination or accuracy when required, without retraining on the whole previous data set [29].

Recent advances in NN has allowed lensless computational imaging [30] and phase recovery [31–33], as well as object classification through scattering media [34–36] and through diffraction imaging [37]. Theoretical work by Ulanowski et al. [38] described applying NNs to determining particle size and refractive index. Although the determination of particle refractive index and size of colloidal spherical particles [39–41] has also been demonstrated experimentally using NNs [42], our results presented here demonstrate the application of machine learning to determine the size, material and number of microspheres in a cluster, as well as the real-time identification of a range of real-world airborne pollution particles.

Specifically, we demonstrate firstly a NN that can predict the material and number of microspheres at each position on a substrate, hence producing a sample map, and secondly a NN that runs in real-time for identification of airborne pollutants such as pollen and particles from diesel combustion. The real-time NN receives live camera images of scattering patterns and processes them in real-time, returning its determination as the illuminating laser spot is scanned over the particles. These two experiments offer a proof-of-principle demonstration of the applicability of our technique for both marine and airborne pollution identification. The proposed sensing concept, as shown in Fig. 1, is that a NN can directly quantify the material and number of particles from a single scattering pattern, once provided with appropriate training data.

 figure: Fig. 1

Fig. 1 Proposed application of a NN to quantify an object directly from the scattering pattern. A NN, which can approximate any function if provided with appropriate training data has the potential to directly determine specific sample parameters from the associated scattering pattern.

Download Full Size | PDF

2. Method

2.1 Sample fabrication

Polystyrene microspheres of size 5 µm ± 70 nm and silicon dioxide microspheres of size 5 µm ± 200 nm from Sigma Aldrich were deposited onto a substrate (a 22 mm by 50 mm, 0.15 mm thick borosilicate coverslip) in a dilution of 1 part per 1000 in deionized water. The two solutions were deposited onto the substrate approximately 3 mm apart, each covering a region approximately 3 mm in diameter. The aqueous solution was evaporated by placing the substrate on a hotplate for two minutes at 90 °C.

The airborne pollution particles used were Narcissus pollen grains, Iva xanthiifolia pollen grains, diesel soot (particles from diesel combustion) collected from a car exhaust, and wood ash, which were deposited on separate substrates. This allowed labelling and recording of scattering patterns for training the NN. Subsequently, the four types of pollutants were deposited onto four separate quadrants on a new substrate, for real-time measurements. Images of the pollution particles were recorded using a Zeiss Evo SEM.

2.2 Experimental setup

Light from a HeNe laser operating at 632.8 nm, was focused onto the substrate using a 5 cm focal length lens, producing a spot size with diameter approximately 20 µm. The scattered light from the particles was projected onto a screen at a distance of 5 cm from the substrate, and subsequently imaged by a CMOS camera (Thorlabs, DCC1545M, 1280 x 1024), placed 30 cm away, with an integration time of 100 ms. Owing to the housing of the camera preventing the detector being placed in close proximity to the particle, by imaging the screen, we were able to capture signals from higher scattering angles than that possible with the CMOS camera alone. The camera was connected to a computer to allow real-time capturing of images, and synchronisation with the programmable XYZ movement stages. To image the substrate, the screen was removed and replaced with a 20x microscope objective (Nikon 20x NA = 0.4) connected to a CMOS camera (Thorlabs, DCC1545M, 1280 x 1024).

2.3 Data collection and training

Figure 2 shows the schematic for data collection and NN training. A region of a substrate (a borosilicate coverslip) was sparsely coated with 5 µm diameter polystyrene microspheres and an adjacent region of 5 µm diameter silicon dioxide microspheres, both of total area approximately 3 mm by 3 mm. The substrate was mounted on an XYZ translation stage in order for the focussed light from a HeNe laser to be scanned over the substrate regions of interest. The HeNe light was focussed to a spot size of approximately 20 µm diameter, and the movement steps in X and Y on the substrate, between recording each scattering pattern, was 30 µm.

 figure: Fig. 2

Fig. 2 Collecting the training data set and training the neural network. The schematic shows the collection of scattering patterns and associated microscope images for 568 positions across a substrate, which was used to form the training data set. Each scattering pattern was used as an input for the NN, which then predicted the material and number of particles. The feedback from the comparison function was used to optimize the weighting between the neurons in the NN, via an automatic process known as backpropagation [43], in order to increase the prediction accuracy.

Download Full Size | PDF

Since a NN learns by the processing of training data, rather than being hard-coded with algorithms, the resulting efficacy is generally dependent on the quality and appropriateness of the training data set. Here, each item in the training data set consisted of a single scattering pattern (the input) and two parameters corresponding to the material and the number of particles (the output). For the training data set, for each scattering pattern, the material was known from prior knowledge of the particle distribution and the number of particles was determined from a microscope image of the sample that was visually inspected and human-counted. The training data set consisted of 568 items, collected from both material regions under identical conditions, and contained a distribution of number of particle clusters.

2.4 Neural network

The type of NN used was a convolutional neural network (CNN) [44,45], and the CNN framework used was Tensorflow [46]. CNNs are particularly effective at processing image data, as they can include a series of convolutional processes that, during training, are optimized in order to recognise specific features in the images. Both the NNs used in this work (one for microspheres analysis and one for real-time pollution analysis) consisted of an input layer, two convolutional layers with a fully connected layer (1024 neurons) before the categorisation output. Training took 10000 and 1000 epochs (iterations through the entire data set) for the microspheres and real-time NNs, respectively.

Figure 3 displays a diagram of the NN used for both the microspheres and real-time experiments. The training data was pre-processed by cropping the images to 112 by 112 pixels and normalising to a maximum value of 255, before being passed into the NN. The first two layers following the input layer consisted of 64 and 32 convolutions, for microspheres NN and real-time NN, respectively, with sizes of 112 by 112 and 56 by 56 for microspheres, and 112 by 112 and 56 by 56 for real-time, for the first and second layers. A kernel size of 3 and 15 was used for each layer for microspheres and real-time NN, respectively, with a pooling factor of 2 and stride of 2. The output of the second layer was passed to a fully connected layer of 1024 neurons. A dropout [47] of 0.1 was used for the fully connected layer and ReLU activation function [48] was used at each layer. The training protocol was the adaptive moment estimation (ADAM) optimizer [49], which was used to minimize the softmax cross entropy cost function. A learn rate of 0.0003 was used for both NNs.

 figure: Fig. 3

Fig. 3 Diagram of the NN used for the microspheres and real-time experiments.

Download Full Size | PDF

The output of the microspheres NN had 13 categories, 1-9 corresponding to 0-8 particles of polystyrene, and 10-13 categories corresponding to 0-3 particles of silicon dioxide, with each category assigned a confidence percentage, with the combined value for all categories for each analysed scattering pattern equalling 100%. The output of the real-time NN was 4 categories with each category assigned a confidence percentage for each tested scattering pattern, with the combined value for all 4 categories equalling 100%.

3. Results and discussion

3.1 Particulate mapping

Since marine particle pollution can consist of plastic microspheres, such as polystyrene, the application to the detection of such microspheres is important, and hence is demonstrated here. Experimental scattering patterns from either a single microsphere or a cluster of many microspheres are shown in Fig. 4, with the associated microscope images displayed as an inset to each image. Each image is split in half to emphasise the difference in the scattering patterns for different sample parameters. In general, as the number of particles increased in the clusters, the angle between adjacent interference peaks decreased, owing to interference from multiple particles.

 figure: Fig. 4

Fig. 4 Highlighting the differences between scattering patterns for different materials and numbers of microspheres. (a) Contrasts the scattering patterns from a single (left) polystyrene and (right) silicon dioxide microsphere. (b) Shows that two particles of polystyrene for two orientations had similar but rotated scattering patterns. (c) Shows three polystyrene microspheres in a line (left) and in a triangle (right). In all cases, the associated microscope image is inset.

Download Full Size | PDF

Figure 4(a) shows a single particle of (left) polystyrene and (right) silicon dioxide. Whilst the microscope images are almost indistinguishable, the scattering patterns have distinctly different angles for the interference rings. Figure 4(b) shows the scattering pattern for two different orientations of two polystyrene particles. The two scattering patterns are almost identical, except for the relative rotation. The pattern recognition process in the NN therefore had to become rotationally invariant, as here the scattering patterns have a uniform distribution of rotation. Figure 4(c) shows the difference between three polystyrene particles that are arranged (left) in a line and (right) in a triangle. In this case, even though the material and the number of particles are the same, the scattering patterns have substantial differences, therefore highlighting the challenge of correctly identifying the number when three or more microspheres are present.

Both training and validation data sets contained identical types of data, namely scattering patterns and the associated material and number of particles. However, the two data sets served completely different purposes. The training data set was processed repeatedly, where during each iteration an automatic optimization process gradually changed the weighting of the neural connections to reduce the error in the NN prediction. The validation data set was used to test the accuracy of the trained NN, as the data set was not used during training and hence acted as unseen data. Here, the validation data set corresponded to two separate 600 µm by 600 µm regions (20 by 20 scattering patterns) on the substrate, corresponding to regions of polystyrene and silicon dioxide microspheres. For each position in the validation regions, the scattering pattern was processed by the NN and the material and number was predicted, and subsequently compared to the known material and number.

Figure 5 shows the predicted sample maps for (a) polystyrene and (b) silicon dioxide microspheres, where the number at each point corresponds to the predicted number of particles. For 100% of the validation scattering patterns, the NN correctly predicted the material type, and hence the predicted material type is not labelled on the sample maps. In cases where the predicted number was incorrect, a red cross covers the number. Squares with no numbers correspond to scattering patterns that the NN correctly predicted as zero particles, i.e. positions where the microscope image of the substrate either showed nothing or debris. The colour map corresponds to the confidence percentage that the NN assigned to each prediction.

 figure: Fig. 5

Fig. 5 Particle maps created via application of the trained NN to scattering patterns from a range of positions across the substrate. The material and number at each position, as predicted by the NN, was used to form the sample maps. The maps correspond to regions of (a) polystyrene, and (b) silicon dioxide spheres. The colour map corresponds to the NN confidence percentage for each prediction.

Download Full Size | PDF

For the validation data set, which contained 238 unseen scattering patterns, the NN had a prediction accuracy of 97.1% (7 incorrect). The scattering patterns with zero particles had a 100% prediction accuracy. For the polystyrene particles, for one and two particles, there was a prediction accuracy of 100%. For the silicon dioxide particles, for one and two particles, there were prediction accuracies of 100% and 83% respectively. For three or more particles, across both materials, 40% of the scattering patterns were predicted correctly. The reduced accuracy for a higher number of particles is attributed predominantly to the relative rarity of such scattering patterns in the training data set, particularly given the increase in possible permutations (i.e. rotations and arrangements) of clusters for higher numbers of particles.

3.2 Real-time sensing

To determine the ability of our approach to work in real-time for sensing, such as for airborne pollution detection, a NN was trained on four different particle types, Narcissus pollen grains, Iva xanthiifolia pollen grains, diesel soot and wood ash. The airborne pollutants were deposited onto separate substrates and placed in the setup described earlier. A total of 238 scattering patterns across the four pollutant types (68 Narcissus, 56 Iva xanthiifolia, 59 diesel soot and 55 wood ash) were recorded and used in the training of the NN, while an additional 39 scattering patterns not used in the training were used in the validation of the NN.

To test the rigorousness of our sensing approach for real-world application, after a time period of over one month the trained NN was used to perform real-time measurements on a new substrate. Each pollutant was deposited onto a quadrant of a single, new substrate, for ease of checking correctness of predictions, which was then moved randomly under the laser light focus. Since the training of the particles was carried out on separate slides for each particle, then there is no weighting given to each of the quadrants on the test sample, hence, as far as the NN is concerned, the particles are randomly dispersed as a mixture on the slide.

Each scattering pattern was recorded and processed by the NN (which took 2.8 seconds to load) in less than 50 milliseconds. Figure 6(a) shows the confidence percentage for each measurement, and whether the prediction was correct or incorrect. Out of 50 measurements taken, the NN was correct 43 times. Figure 6(b), which displays a confusion matrix, illustrates the prediction accuracy for each pollutant type. Figure 6(c) shows examples of scanning electron microscope (SEM) images of, and scattering patterns from, the four types of pollutant.

 figure: Fig. 6

Fig. 6 Real-time sensing of pollen, diesel and wood ash particulates. (a) Prediction of the pollutant type as a function of measurement number. The y-axis corresponds to the NN confidence percentage for each prediction. (b) Confusion matrix showing the performance of classification. (c) Examples of SEM images of, and scattering patterns from, the four types of pollutant.

Download Full Size | PDF

The average confidence percentage when correctly identifying Narcissus pollen grains, Iva xanthiifolia pollen grains, diesel soot and wood ash, was 86%, 58%, 40% and 54%, respectively. We attribute these differences to the ease of identification of features in the scattering patterns that are associated with specific pollutant structures. If the majority of features that are associated with a specific pollutant type are not identified, then, for example, the scattering from the sharp edges of the wood ash could be misidentified as the scattering from the sulcus from the pollen grains, due to the similarity in the resultant scattering pattern. Hence, this could explain why wood ash has been misidentified as Narcissus and Iva xanthiifolia. The high confidence percentage of the Narcissus pollen grains is likely due to the lack of variation in the pollen grain sizes and similar structure, whereas the smaller confidence percentage for Iva xanthiifolia pollen grains is likely due to the lager variation in size and structure of the pollen grains. Additionally, as discussed, the NN was trained on a slightly higher number of Narcissus pollen grain scattering patterns than any other scattering patterns, which perhaps led to a greater ability for the NN to predict Narcissus. The average confidence percentage was lower for diesel soot, likely due to the variety of pollutant shape, however, the diesel soot predominantly consisted of features not present in the other pollutants (e.g. specifically sized aggregates), and hence was identified correctly 100% of the time. In order to quantify the quality of the NN’s prediction ability that can include false positives and false negatives as well as true positives and true negatives, the Matthews correlation coefficient (a measure of the quality for a multiclass classification [50]), was calculated. This was found to be 0.81, compared with the maximum potential value of 1.

To determine what output the real-time NN would give for an unfamiliar object (a sample type that it had not been trained on), we tested the real-time NN on a 5 µm polystyrene microsphere scattering pattern. The real-time NN allocated the following probabilities, 29% to Narcissus, 7% to Iva xanthiifolia, 18% to diesel and 46% to wood ash. This is perhaps understandable due to the fewer higher spatial frequencies present in wood ash and polystyrene scattering patterns, compared with the pollen and diesel particle scattering patterns. Avoiding such false identification could be achieved ensuring a sufficiently high cut-off confidence percentage for correct identification, which could be enabled by increasing the accuracy of the NN and training on more data, as well as through expanding the number of objects trained on and through potentially using multiple NNs [51].

The application of this sensor technique to real-world scenarios would therefore require increasing the number of object types to accommodate the many and various particle pollutants present in the air. Such a scaling up would require robotic automation, and potentially data augmentation to reduce training time [52]. Also, the particle scattering patterns would need to be captured when the particles were present and dispersed in air or in water, to simulate real-world scenarios. Exploration of how the accuracy of the NN scales with object type will be necessary to aid in the maintaining of the quality of the NN, which could be achieved by checking the accuracy of the NN as additional particle types are continually added to the network.

4. Conclusion

In conclusion, a particle pollution sensor that uses a NN as a general solution to the scattering inverse problem was presented. The NN was able to determine the material and number of microspheres from a single scattering pattern, and this approach was used to construct a map of their location on the substrate. Additionally, real-time detection of airborne pollutants was demonstrated. The accuracy of our approach could be improved via use of additional training data and optimization of the NN architectures. This method could lead to applications in biological cellular analysis and quantification of marine and airborne pollution particulates.

Funding

EPSRC Early Career Fellowship (EP/N03368X/1); EPSRC Doctoral Prize (EP/N509747/1); BBSRC Future Leader Fellowship (BB/P011365/1).

Acknowledgements

We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan X GPU used for this research, donated through NVIDIA GPU Grant Program. Data supporting this manuscript is available at https://doi.org/10.5258/SOTON/D0436.

References

1. J. Lelieveld, J. S. Evans, M. Fnais, D. Giannadaki, and A. Pozzer, “The contribution of outdoor air pollution sources to premature mortality on a global scale,” Nature 525(7569), 367–371 (2015). [CrossRef]   [PubMed]  

2. M. Kulmala, T. Petäjä, T. Nieminen, M. Sipilä, H. E. Manninen, K. Lehtipalo, M. Dal Maso, P. P. Aalto, H. Junninen, P. Paasonen, I. Riipinen, K. E. Lehtinen, A. Laaksonen, and V. M. Kerminen, “Measurement of the nucleation of atmospheric aerosol particles,” Nat. Protoc. 7(9), 1651–1667 (2012). [CrossRef]   [PubMed]  

3. L. Fierce, T. C. Bond, S. E. Bauer, F. Mena, and N. Riemer, “Black carbon absorption at the global scale is affected by particle-scale diversity in composition,” Nat. Commun. 7, 12361 (2016). [CrossRef]   [PubMed]  

4. V. Ramanathan and G. Carmichael, “Global and regional climate changes due to black carbon,” Nat. Geosci. 1(4), 221–227 (2008). [CrossRef]  

5. J. I. Levy, D. H. Bennett, S. J. Melly, and J. D. Spengler, “Influence of traffic patterns on particulate matter and polycyclic aromatic hydrocarbon concentrations in Roxbury, Massachusetts,” J. Expo. Anal. Environ. Epidemiol. 13(5), 364–371 (2003). [CrossRef]   [PubMed]  

6. M. Loxham, “Harmful effects of particulate air pollution: identifying the culprits,” Respirology 20(1), 7–8 (2015). [CrossRef]   [PubMed]  

7. F. R. Cassee, M.-E. Héroux, M. E. Gerlofs-Nijland, and F. J. Kelly, “Particulate matter beyond mass: recent health evidence on the role of fractions, chemical constituents and sources of emission,” Inhal. Toxicol. 25(14), 802–812 (2013). [CrossRef]   [PubMed]  

8. F. J. Kelly and J. C. Fussell, “Size, source and chemical composition as determinants of toxicity attributable to ambient particulate matter,” Atmos. Environ. 60, 504–526 (2012). [CrossRef]  

9. Z. D. Ristovski, B. Miljevic, N. C. Surawski, L. Morawska, K. M. Fong, F. Goh, and I. A. Yang, “Respiratory health effects of diesel particulate matter,” Respirology 17(2), 201–212 (2012). [CrossRef]   [PubMed]  

10. R. Sussarellu, M. Suquet, Y. Thomas, C. Lambert, C. Fabioux, M. E. Pernet, N. Le Goïc, V. Quillien, C. Mingant, Y. Epelboin, C. Corporeau, J. Guyomarch, J. Robbens, I. Paul-Pont, P. Soudant, and A. Huvet, “Oyster reproduction is affected by exposure to polystyrene microplastics,” Proc. Natl. Acad. Sci. U.S.A. 113(9), 2430–2435 (2016). [CrossRef]   [PubMed]  

11. M. Cole, P. Lindeque, E. Fileman, C. Halsband, and T. S. Galloway, “The impact of polystyrene microplastics on feeding, function and fecundity in the marine copepod Calanus helgolandicus,” Environ. Sci. Technol. 49(2), 1130–1137 (2015). [CrossRef]   [PubMed]  

12. L. A. Philips, D. B. Ruffner, F. C. Cheong, J. M. Blusewicz, P. Kasimbeg, B. Waisi, J. R. McCutcheon, and D. G. Grier, “Holographic characterization of contaminants in water: Differentiation of suspended particles in heterogeneous dispersions,” Water Res. 122, 431–439 (2017). [CrossRef]   [PubMed]  

13. Y. Wu and A. Ozcan, “Lensless digital holographic microscopy and its applications in biomedicine and environmental monitoring,” Methods 136, 4–16 (2018). [CrossRef]   [PubMed]  

14. C. F. Bohren and D. R. Huffman, Absorption and Scattering of Light by Small Particles (John Wiley & Sons, 2008).

15. B. Mills, C. F. Chau, E. T. F. Rogers, J. Grant-Jacob, S. L. Stebbings, M. Praeger, A. M. de Paula, C. A. Froud, R. T. Chapman, T. J. Butcher, J. J. Baumberg, W. S. Brocklesby, and J. G. Frey, “Direct measurement of the complex refractive index in the extreme ultraviolet spectral region using diffraction from a nanosphere array,” Appl. Phys. Lett. 93(23), 231103 (2008). [CrossRef]  

16. J. R. Fienup, “Phase retrieval algorithms: a comparison,” Appl. Opt. 21(15), 2758–2769 (1982). [CrossRef]   [PubMed]  

17. F. Pfeiffer, T. Weitkamp, O. Bunk, and C. David, “Phase retrieval and differential phase-contrast imaging with low-brilliance X-ray sources,” Nat. Phys. 2(4), 258–261 (2006). [CrossRef]  

18. M. R. Teague, “Deterministic phase retrieval: a Green’s function solution,” J. Opt. Soc. Am. 73(11), 1434–1441 (1983). [CrossRef]  

19. R. Trebino and D. J. Kane, “Using phase retrieval to measure the intensity and phase of ultrashort pulses: frequency-resolved optical gating,” J. Opt. Soc. Am. A 10(5), 1101–1111 (1993). [CrossRef]  

20. J. Miao, D. Sayre, and H. N. Chapman, “Phase retrieval from the magnitude of the Fourier transforms of nonperiodic objects,” J. Opt. Soc. Am. A 15(6), 1662–1669 (1998). [CrossRef]  

21. A. M. Maiden, M. J. Humphry, F. Zhang, and J. M. Rodenburg, “Superresolution imaging via ptychography,” J. Opt. Soc. Am. A 28(4), 604–612 (2011). [CrossRef]   [PubMed]  

22. K. Giewekemeyer, P. Thibault, S. Kalbfleisch, A. Beerlink, C. M. Kewish, M. Dierolf, F. Pfeiffer, and T. Salditt, “Quantitative biological imaging by ptychographic x-ray diffraction microscopy,” Proc. Natl. Acad. Sci. U.S.A. 107(2), 529–534 (2010). [CrossRef]   [PubMed]  

23. H. M. L. Faulkner and J. M. Rodenburg, “Movable aperture lensless transmission microscopy: a novel phase retrieval algorithm,” Phys. Rev. Lett. 93(2), 023903 (2004). [CrossRef]   [PubMed]  

24. G. E. Hinton and R. R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” Science 313(5786), 504–507 (2006). [CrossRef]   [PubMed]  

25. H. A. Rowley, S. Baluja, and T. Kanade, “Neural network-based face detection,” IEEE Trans. Pattern Anal. Mach. Intell. 20(1), 23–38 (1998). [CrossRef]  

26. D. F. Specht, “A general regression neural network,” IEEE Trans. Neural Netw. 2(6), 568–576 (1991). [CrossRef]   [PubMed]  

27. B. Mills, D. J. Heath, J. A. Grant-Jacob, and R. W. Eason, “Predictive capabilities for laser machining via a neural network,” Opt. Express 26(13), 17245–17253 (2018). [CrossRef]   [PubMed]  

28. D. J. Heath, J. A. Grant-Jacob, Y. Xie, B. S. Mackay, J. A. G. Baker, R. W. Eason, and B. Mills, “Machine learning for 3D simulated visualization of laser machining,” Opt. Express 26(17), 21574–21584 (2018). [CrossRef]  

29. A. Achille, T. Eccles, L. Matthey, C. P. Burgess, N. Watters, A. Lerchner, and I. Higgins, “Life-Long Disentangled Representation Learning with Cross-Domain Latent Homologies,” arXiv Prepr. arXiv1808.06508 (2018).

30. A. Sinha, J. Lee, S. Li, and G. Barbastathis, “Lensless computational imaging through deep learning,” Optica 4(9), 1117–1125 (2017). [CrossRef]  

31. Y. Wu, Y. Rivenson, Y. Zhang, Z. Wei, H. Günaydin, X. Lin, and A. Ozcan, “Extended depth-of-field in holographic imaging using deep-learning-based autofocusing and phase recovery,” Optica 5(6), 704–710 (2018). [CrossRef]  

32. Y. Rivenson, Y. Zhang, H. Günaydin, D. Teng, and A. Ozcan, “Phase recovery and holographic image reconstruction using deep learning in neural networks,” Light Sci. Appl. 7(2), 17141 (2018). [CrossRef]  

33. Y. Jo, S. Park, J. Jung, J. Yoon, H. Joo, M. H. Kim, S.-J. Kang, M. C. Choi, S. Y. Lee, and Y. Park, “Holographic deep learning for rapid optical screening of anthrax spores,” Sci. Adv. 3(8), e1700606 (2017). [CrossRef]   [PubMed]  

34. S. Li, M. Deng, J. Lee, A. Sinha, and G. Barbastathis, “Imaging through glass diffusers using densely connected convolutional networks,” arXiv Prepr. arXiv1711.06810 (2017).

35. G. Satat, M. Tancik, O. Gupta, B. Heshmat, and R. Raskar, “Object classification through scattering media with deep learning on time resolved measurement,” Opt. Express 25(15), 17466–17479 (2017). [CrossRef]   [PubMed]  

36. E. Valent and Y. Silberberg, “Scatterer recognition via analysis of speckle patterns,” Optica 5(2), 204–207 (2018). [CrossRef]  

37. K. Dong, Y. Feng, K. M. Jacobs, J. Q. Lu, R. S. Brock, L. V. Yang, F. E. Bertrand, M. A. Farwell, and X.-H. Hu, “Label-free classification of cultured cells through diffraction imaging,” Biomed. Opt. Express 2(6), 1717–1726 (2011). [CrossRef]   [PubMed]  

38. Z. Ulanowski, Z. Wang, P. H. Kaye, and I. K. Ludlow, “Application of neural networks to the inverse light scattering problem for spheres,” Appl. Opt. 37(18), 4027–4033 (1998). [CrossRef]   [PubMed]  

39. S.-H. Lee, Y. Roichman, G.-R. Yi, S.-H. Kim, S.-M. Yang, A. van Blaaderen, P. van Oostrum, and D. G. Grier, “Characterizing and tracking single colloidal particles with video holographic microscopy,” Opt. Express 15(26), 18275–18282 (2007). [CrossRef]   [PubMed]  

40. R. W. Perry, G. Meng, T. G. Dimiduk, J. Fung, and V. N. Manoharan, “Real-space studies of the structure and dynamics of self-assembled colloidal clusters,” Faraday Discuss. 159(1), 211–234 (2012). [CrossRef]  

41. C. Wang, F. C. Cheong, D. B. Ruffner, X. Zhong, M. D. Ward, and D. G. Grier, “Holographic characterization of colloidal fractal aggregates,” Soft Matter 12(42), 8774–8780 (2016). [CrossRef]   [PubMed]  

42. A. Yevick, M. Hannel, and D. G. Grier, “Machine-learning approach to holographic particle characterization,” Opt. Express 22(22), 26884–26890 (2014). [CrossRef]   [PubMed]  

43. M. Riedmiller and H. Braun, “A direct adaptive method for faster backpropagation learning: The RPROP algorithm,” in Neural Networks, 1993., IEEE International Conference on (1993), pp. 586–591.

44. S. Lawrence, C. L. Giles, A. C. Tsoi, and A. D. Back, “Face recognition: A convolutional neural-network approach,” IEEE Trans. Neural Netw. 8(1), 98–113 (1997). [CrossRef]   [PubMed]  

45. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems (2012), pp. 1097–1105.

46. M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, and others, “Tensorflow: a system for large-scale machine learning,” in OSDI (2016), Vol. 16, pp. 265–283.

47. N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” J. Mach. Learn. Res. 15(1), 1929–1958 (2014).

48. V. Nair and G. E. Hinton, “Rectified linear units improve restricted boltzmann machines,” in Proceedings of the 27th International Conference on Machine Learning (ICML-10) (2010), pp. 807–814.

49. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv Prepr. arXiv1412.6980 (2014).

50. J. Gorodkin, “Comparing two K-category assignments by a K-category correlation coefficient,” Comput. Biol. Chem. 28(5-6), 367–374 (2004). [CrossRef]   [PubMed]  

51. S.-B. Cho and J. H. Kim, “Combining multiple neural networks by fuzzy integral for robust classification,” IEEE Trans. Syst. Man Cybern. 25(2), 380–384 (1995). [CrossRef]  

52. L. Perez and J. Wang, “The effectiveness of data augmentation in image classification using deep learning,” arXiv Prepr. arXiv1712.04621 (2017).

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1
Fig. 1 Proposed application of a NN to quantify an object directly from the scattering pattern. A NN, which can approximate any function if provided with appropriate training data has the potential to directly determine specific sample parameters from the associated scattering pattern.
Fig. 2
Fig. 2 Collecting the training data set and training the neural network. The schematic shows the collection of scattering patterns and associated microscope images for 568 positions across a substrate, which was used to form the training data set. Each scattering pattern was used as an input for the NN, which then predicted the material and number of particles. The feedback from the comparison function was used to optimize the weighting between the neurons in the NN, via an automatic process known as backpropagation [43], in order to increase the prediction accuracy.
Fig. 3
Fig. 3 Diagram of the NN used for the microspheres and real-time experiments.
Fig. 4
Fig. 4 Highlighting the differences between scattering patterns for different materials and numbers of microspheres. (a) Contrasts the scattering patterns from a single (left) polystyrene and (right) silicon dioxide microsphere. (b) Shows that two particles of polystyrene for two orientations had similar but rotated scattering patterns. (c) Shows three polystyrene microspheres in a line (left) and in a triangle (right). In all cases, the associated microscope image is inset.
Fig. 5
Fig. 5 Particle maps created via application of the trained NN to scattering patterns from a range of positions across the substrate. The material and number at each position, as predicted by the NN, was used to form the sample maps. The maps correspond to regions of (a) polystyrene, and (b) silicon dioxide spheres. The colour map corresponds to the NN confidence percentage for each prediction.
Fig. 6
Fig. 6 Real-time sensing of pollen, diesel and wood ash particulates. (a) Prediction of the pollutant type as a function of measurement number. The y-axis corresponds to the NN confidence percentage for each prediction. (b) Confusion matrix showing the performance of classification. (c) Examples of SEM images of, and scattering patterns from, the four types of pollutant.
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.