
- Photonics Research
- Vol. 10, Issue 1, 104 (2022)
Abstract
1. INTRODUCTION
Single-pixel imaging (SPI) is an emerging computational imaging modality that utilizes the second-order correlation of quantum or classical light to reconstruct a two-dimensional (2D) image from a one-dimensional (1D) bucket signal [1–4]. As most of the photons that interact with the object are collected by the bucket detector, SPI has significant advantages in terms of the detection sensitivity, dark counts, and spectral range. Thus it has received increasing attentions over the past decade for the people working in the divergent fields of remote sensing [5,6], 3D imaging [7,8], spectral imaging [9,10], microscopy [11], and of the sort [3,12]. However, in SPI, each single-pixel measurement contains highly compressed information about the object, and one needs a large amount of such measurements to reconstruct an image with good resolution. This leads to a trade-off between the acquisition time and the image quality that hinders the practical application of SPI. Many studies have been carried out to address this issue. The solutions proposed so far can be categorized into two mainly aspects of strategies. The first one is to design the encoding patterns that ensures each single-pixel measurement contains as more information as possible [13–15]. The second one is to develop an optimization algorithm to obtain better reconstruction using a smaller number of measurements [16,17].
Owing to its capability of solving various challenging problems in divergent fields [18,19], deep learning (DL) has also been adopted for SPI recently. Previous studies have shown that the DL-based SPI methods can dramatically reduce the sampling ratio, promising real-time performance [17,20,21]. Specifically, Lyu
In this work, we report a physics enhanced deep learning technique for SPI. The physics prior we exploit mainly contains two aspects that rely on the forward propagation model of the SPI system,
Sign up for Photonics Research TOC. Get the latest issue of Photonics Research delivered right to you!Sign up now
2. METHODS
As schematically presented in Fig. 1, the proposed method consists of two main steps: a physics-informed autoencoder DNN that generates a set of optimal encoding patterns
Figure 1.Schematic diagram of the physics enhanced deep learning approach for SPI. (a) The physics-informed DNN. (b) The SPI system. (c) The model-driven fine-tuning process. The face images were taken from CelebAMask-HQ [28].
As shown in Fig. 1(a), the physics-informed autoencoder DNN contains three parts. The first part is a set of
Apparently, both the DNN model
Encoding a real-world target by using a typical SPI system shown in Fig. 1(b), one can acquire a 1D raw bucket signal
The network architecture we used to implement
Figure 2.Diagram of the DNN structure we designed. It consists of an encoder path that takes the low-quality image reconstructed by DGI as its input and a decoder path that outputs an enhanced one.
3. RESULTS AND DISCUSSION
Here we perform a comparative study on the effectiveness of the proposed method. For the sake of quantitative evaluation, we first examine its performance by using simulation data. Then we demonstrate its practical applications in laboratory and outdoor experiments.
A. Simulations
First let us examine the effectiveness of the physics-informed layer that we add to the DNN. The results are plotted at the fifth column in Fig. 3(a). It is clearly seen that the DGI reconstructed image with the learned patterns is far better than the one with random illumination. This conclusion retains even if Gaussian white noise (with the variance of
Figure 3.Comparative study of the proposed method with some other fast SPI algorithms with a low sampling ratio (
One can also see that the two gray curves are fairly flat with respect to noise level. This suggests that the DGI reconstruction algorithm is immune to the additive noise [29,30], no matter whether the physics-informed layer is used or not. This robustness is important for the downstream decoding DNN, as it takes the DGI reconstructed image as its input.
Now we proceed to compare the performance of the proposed methods with some of the widespread SPI methods, namely, DCAN [20], the reordering Hadamard SPI (HSI) [14], the compressed sensing based total variation (TV) regularization [31], and the Fourier domain regularized inversion (FDRI) methods [35,36]. The results are plotted in Fig. 3.
As a learning-based end-to-end SPI method, DCAN [20] outperforms the other existing methods (i.e., HSI, TV, and FDRI) except for some high-noise-level cases. The proposed physics-informed method has a similar performance to DCAN when the SNR of the bucket signal is 20 dB and higher, but is much better when the noise level increases. As the DNN parts of the proposed physics-informed method and DCAN do not have much difference, it must be the physics-informed layer described by Eq. (1) that contributes to the high performance [see, for example, the reconstructed images at row 3, columns 4 and 6, in Fig. 3(a)]. The reconstruction is quite time efficient. It takes only 0.32 s to reconstruct a
The results shown in Fig. 3 suggest that fine-tuning the trained DNN model
Figure 4.Convergence behavior of different error functions that measure (a) the objective function, (b) the prediction error, and (c) the error between the estimated bucket signal and the ideal one.
B. Experiments
Now we proceed to demonstrate the proposed method with in-house experiments. We built a typical passive modulated SPI system as the one schematically shown in Fig. 1(b). Three real-world objects were used in our proof-of-principle experiments. They were illuminated by a thermal light source and imaged by an imaging optic with the focal length of 85 mm to a digital micromirror device (DMD, DLP7000, TI). On the DMD, the learned binary patterns
We reconstruct the images following the aformentioned pipeline. First we correlated
Figure 5.Experimental results. The images reconstructed by DGI alone, DGI with physics-informed DNN, and the fine-tuning method. The sampling ratio
Next we endeavor to demonstrate that the proposed fine-tuning method outperforms the other widespread SPI algorithms such as DGI [29,30], HSI [14], DCAN [20]; the total variation minimization by augmented Lagrangian and alternating direction algorithms (TVAL3) [38]; and randomly initialized fine-tuning on the same set of experimental data. The data were acquired with the same SPI system we built. This time we replaced the previous objects with the badge of our institute printed on a white paper for the sake of quantitative analysis. For this purpose, we took the image reconstructed by HSI with full sampling (
Figure 6.Experimental results: images of the badge of our institute reconstructed by (a) HSI with
To demonstrate the practical application of the proposed method, we incorporated the proposed method into a single-pixel LiDAR system upgraded from the one we built previously [5]. As schematically shown in Fig. 7(a), the upgrade was mainly done by replacing the active modulation module based on a rotating ground glass in Ref. [5] by a DMD-based passive one. The light source is a solid-state pulsed laser with the center wavelength of 532 nm and the pulse width of 8 ns at the repetition rate of 10 kHz. The laser light was first collimated and expanded, and then it was sent out to illuminate a remote target. The echo light scattered back from the target was collected by an imaging optic (
Figure 7.Experimental results for single-pixel LiDAR. (a) Schematic diagram of the single-pixel LiDAR system. (b) Satellite image of our experiment scenario. The inset in the top left is the target imaged by a telescope, whereas the one in the bottom right is one of the echoed light signals. (c) Six typical 2D depth slices of the 3D object reconstructed by DGI with the learned patterns illumination, GISC [5], and the proposed fine-tuning method. (d) 3D images of the object reconstructed by the three aforementioned methods.
To obtain a more general model for the remote sensing task, we retrained the same decoding DNN on a training set composed of 90,000 images (
For each measurement, the PMT was triggered with a time delay of 3700 ns with respect to that of the laser emission so that the echoed light contains the reflectivity information of the object within the FOV. The echoed signal measured by the PMT has the dimension of
For comparison, we also plot the images reconstructed by DGI with learned pattern illumination, ghost imaging via sparsity constraint (GISC) [5] side by side in Figs. 7(c) and 7(d). These two images were post-processed by use of median filtering and non-negative constraint. It is apparent that the proposed method has the best performance as evidenced by the clean background, high contrast, the fine details of the reconstructed image.
Finally, let us analyze the time efficiency. First we note that the time period to display all the 1024 learned patterns on DMD, and the DGI reconstruction for each depth-slice image is at the scale of tens of milliseconds. It is therefore in principle possible to perform 3D LiDAR imaging in real time. Comparing with the scanning imaging LiDAR [40], the proposed method has the potential to operate in a more time-efficient way.
4. CONCLUSION
We have proposed a physics enhanced deep learning framework for SPI. The incorporation of physics mainly brings two aspects of advantages. First, the physics informed decoding layer allows us to optimize the illumination patterns and improve the performance of the decoding DNN. Second, the model-driven fine-tuning process imposes an interpretable constraint to the DNN output, so that it is not restricted by the issue of generalization.
We have demonstrated the proposed methods with simulation, in-house, and outdoor experimental data. In particular, we have shown that it allows high quality SPI with
In comparison to conventional data-driven deep learning [20,21] and physics-driven [26,27] optimization approaches, the proposed fine-tuning process takes advantage of both of them, making it possible to use data prior information, i.e., characteristics of the objects, for solving ill-posed inverse problems. Besides, the issue of generalization in conventional learning-based methods can be eliminated at the cost of iterative calculations. As a result, the proposed framework should be applicable for diverse computational imaging systems, not just limited to the SPI we discussed here.
However, it is worth pointing out that the proposed method relies on the accurate model of the forward propagation, making it difficult to use in the cases that the physical model cannot be accurately modeled, e.g., imaging through optically thick scattering media. Further efforts should be made to solve this problem.
References
[1] T. B. Pittman, Y. H. Shih, D. V. Strekalov, A. V. Sergienko. Optical imaging by means of two-photon quantum entanglement. Phys. Rev. A, 52, R3429-R3432(1995).
[2] B. I. Erkmen, J. H. Shapiro. Ghost imaging: from quantum to classical to computational. Adv. Opt. Photon., 2, 405-450(2010).
[3] M. P. Edgar, G. M. Gibson, M. J. Padgett. Principles and prospects for single-pixel imaging. Nat. Photonics, 13, 13-20(2019).
[4] M. F. Duarte, M. A. Davenport, D. Takhar, J. N. Laska, T. Sun, K. F. Kelly, R. G. Baraniuk. Single-pixel imaging via compressive sampling. IEEE Signal Process Mag., 25, 83-91(2008).
[5] W. Gong, C. Zhao, H. Yu, M. Chen, W. Xu, S. Han. Three-dimensional ghost imaging lidar via sparsity constraint. Sci. Rep., 6, 26133(2016).
[6] C. Wang, X. Mei, L. Pan, P. Wang, W. Li, X. Gao, Z. Bo, M. Chen, W. Gong, S. Han. Airborne near infrared three-dimensional ghost imaging lidar via sparsity constraint. Remote Sens., 10, 732(2018).
[7] B. Sun, M. P. Edgar, R. Bowman, L. E. Vittert, S. Welsh, A. Bowman, M. J. Padgett. 3D computational imaging with single-pixel detectors. Science, 340, 844-847(2013).
[8] M. Sun, M. P. Edgar, G. M. Gibson, B. Sun, N. Radwell, R. Lamb, M. J. Padgett. Single-pixel three-dimensional imaging with time-based depth resolution. Nat. Commun., 7, 12010(2016).
[9] L. Bian, J. Suo, G. Situ, Z. Li, J. Fan, F. Chen, Q. Dai. Multispectral imaging using a single bucket detector. Sci. Rep., 6, 24752(2016).
[10] F. Magalhães, F. M. Araújo, M. Correia, M. Abolbashari, F. Farahi. High-resolution hyperspectral single-pixel imaging system based on compressive sensing. Opt. Eng., 51, 071406(2012).
[11] N. Radwell, K. J. Mitchell, G. M. Gibson, M. P. Edgar, R. Bowman, M. J. Padgett. Single-pixel infrared and visible microscope. Optica, 1, 285-289(2014).
[12] G. M. Gibson, S. D. Johnson, M. J. Padgett. Single-pixel imaging 12 years on: a review. Opt. Express, 28, 28190-28208(2020).
[13] Z. Zhang, X. Wang, G. Zheng, J. Zhong. Fast Fourier single-pixel imaging via binary illumination. Sci. Rep., 7, 12029(2017).
[14] M. Sun, L. Meng, M. P. Edgar, M. J. Padgett, N. Radwell. A Russian dolls ordering of the Hadamard basis for compressive single-pixel imaging. Sci. Rep., 7, 3464(2017).
[15] Z.-H. Xu, W. Chen, J. Penuelas, M. Padgett, M.-J. Sun. 1000 fps computational ghost imaging using led-based structured illumination. Opt. Express, 26, 2427-2434(2018).
[16] O. Katz, Y. Bromberg, Y. Silberberg. Compressive ghost imaging. Appl. Phys. Lett., 95, 131110(2009).
[17] M. Lyu, W. Wang, H. Wang, H. Wang, G. Li, N. Chen, G. Situ. Deep-learning-based ghost imaging. Sci. Rep., 7, 17865(2017).
[18] Y. LeCun, Y. Bengio, G. Hinton. Deep learning. Nature, 521, 436-444(2015).
[19] G. Barbastathis, A. Ozcan, G. Situ. On the use of deep learning for computational imaging. Optica, 6, 921-943(2019).
[20] F. H. Catherine, M.-S. Roderick, J. P. Miles, P. E. Matthew. Deep learning for real-time single-pixel video. Sci. Rep., 8, 2369(2010).
[21] F. Wang, H. Wang, H. Wang, G. Li, G. Situ. Learning from simulation: an end-to-end deep-learning approach for computational ghost imaging. Opt. Express, 27, 25560-25572(2019).
[22] B. Neyshabur, S. Bhojanapalli, D. McAllester, N. Srebro. Exploring generalization in deep learning. Advances in Neural Information Processing Systems (NIPS), 1-10(2017).
[23] R. Shang, K. Hoffer-Hawlik, F. Wang, G. Situ, G. P. Luke. Two-step training deep learning framework for computational imaging without physics priors. Opt. Express, 29, 15239-15254(2021).
[24] A. Goy, G. Rughoobur, S. Li, K. Arthur, A. I. Akinwande, G. Barbastathis. High-resolution limited-angle phase tomography of dense layered objects using deep neural networks. Proc. Natl. Acad. Sci. USA, 116, 19848-19856(2019).
[25] A. Goy, K. Arthur, S. Li, G. Barbastathis. Low photon count phase retrieval using deep learning. Phys. Rev. Lett., 121, 243902(2018).
[26] F. Wang, Y. Bian, H. Wang, M. Lyu, G. Pedrini, W. Osten, G. Barbastathis, G. Situ. Phase imaging with an untrained neural network. Light Sci. Appl., 9, 77(2020).
[27] R. Iten, T. Metger, H. Wilming, L. del Rio, R. Renner. Discovering physical concepts with neural networks. Phys. Rev. Lett., 124, 010508(2020).
[28] C.-H. Lee, Z. Liu, L. Wu, P. Luo. Maskgan: towards diverse and interactive facial image manipulation. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 5548-5557(2020).
[29] F. Ferri, D. Magatti, L. A. Lugiato, A. Gatti. Differential ghost imaging. Phys. Rev. Lett., 104, 253603(2010).
[30] W. Gong, S. Han. A method to improve the visibility of ghost images obtained by thermal light. Phys. Lett. A, 374, 1005-1008(2010).
[31] L. Bian, J. Suo, Q. Dai, F. Chen. Experimental comparison of single-pixel imaging algorithms. J. Opt. Soc. Am. A, 35, 78-87(2018).
[32] S. Boyd, L. Vandenberghe. Convex Optimization(2004).
[33] F. Zhuang, Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong, Q. He. A comprehensive survey on transfer learning. Proc. IEEE, 109, 43-76(2020).
[34] X. Zhang, F. Wang, G. Situ. BlindNet: an untrained learning approach toward computational imaging with model uncertainty. J. Phys. D, 55, 034001(2022).
[35] K. M. Czajkowski, A. Pastuszczak, R. Kotyński. Real-time single-pixel video imaging with Fourier domain regularization. Opt. Express, 26, 20009-20022(2018).
[36] A. Pastuszczak, R. Stojek, P. Wróbel, R. Kotyński. Differential real-time single-pixel imaging with fourier domain regularization: applications to VIS-IR imaging and polarization imaging. Opt. Express, 29, 26685-26700(2021).
[37] D. Ulyanov, A. Vedaldi, V. Lempitsky. Deep image prior. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 9446-9454(2018).
[38] C. Li. An Efficient Algorithm for Total Variation Regularization with Applications to the Single Pixel Camera and Compressive Sensing(2010).
[39] A. Coates, A. Ng, H. Lee. An analysis of single-layer networks in unsupervised feature learning. 14th International Conference on Artificial Intelligence and Statistics, 215-223(2011).
[40] Z.-P. Li, X. Huang, Y. Cao, B. Wang, Y.-H. Li, W. Jin, C. Yu, J. Zhang, Q. Zhang, C.-Z. Peng, F. Xu, J.-W. Pan. Single-photon computational 3D imaging at 45km. Photon. Res., 8, 1532-1540(2020).

Set citation alerts for the article
Please enter your email address