• Chinese Optics Letters
  • Vol. 19, Issue 10, 101101 (2021)
Hao Zhang1 and Deyang Duan1、2、*
Author Affiliations
  • 1School of Physics and Physical Engineering, Qufu Normal University, Qufu 273165, China
  • 2Shandong Provincial Key Laboratory of Laser Polarization and Information Technology, Research Institute of Laser, Qufu Normal University, Qufu 273165, China
  • show less
    DOI: 10.3788/COL202119.101101 Cite this Article Set citation alerts
    Hao Zhang, Deyang Duan. Computational ghost imaging with compressed sensing based on a convolutional neural network[J]. Chinese Optics Letters, 2021, 19(10): 101101 Copy Citation Text show less

    Abstract

    Computational ghost imaging (CGI) has recently been intensively studied as an indirect imaging technique. However, the image quality of CGI cannot meet the requirements of practical applications. Here, we propose a novel CGI scheme to significantly improve the imaging quality. In our scenario, the conventional CGI data processing algorithm is optimized to a new compressed sensing (CS) algorithm based on a convolutional neural network (CNN). CS is used to process the data collected by a conventional CGI device. Then, the processed data are trained by a CNN to reconstruct the image. The experimental results show that our scheme can produce higher quality images with the same sampling than conventional CGI. Moreover, detailed comparisons between the images reconstructed using the deep learning approach and with conventional CS show that our method outperforms the conventional approach and achieves a ghost image with higher image quality.

    1. Introduction

    Ghost imaging is an indirect imaging technique based on quantum properties (e.g.,  quantum entanglement or intensity correlation) of the light field[13]. Compared to conventional optical imaging techniques, ghost imaging requires two light beams: a reference light beam, which never illuminates the object and is directly measured by a detector with a spatial resolution (e.g.,  a charge-coupled device) and an object light beam, which, after illuminating the object, is measured by a bucket detector with no spatial resolution. By correlating the photocurrents from the two detectors, the ghost image is retrieved. Previous works show that ghost imaging has potential applications in remote sensing[4,5], industrial imaging[6,7], medical imaging[810], and super-resolution imaging[11,12]. However, conventional ghost imaging requires two optical paths, which severely limits its application. Fortunately, Shapiro creatively introduced the concept of computational ghost imaging (CGI) in 2008[13]. In the CGI setup, the idle light is obtained by calculation, so the reference light path is omitted in the experimental apparatus[14]. Compared with conventional ghost imaging, CGI is more suitable for application in remote sensing, radar, and other fields.

    After more than 10 years, CGI theory and experiments have matured. However, CGI is still in the laboratory stage. One of the critical problems is that the image quality cannot meet practical applications. Generally, to produce a clear image, conventional CGI, including conventional ghost imaging, takes approximately tens of thousands of sets of data, which obviously cannot meet the requirements of practical application, especially those of moving target imaging. How to improve the image quality of ghost imaging is one of the key factors for realizing its application. Compressed sensing (CS)[1518] and deep learning (DL)[1922] greatly improve the image quality, but there is still a gap compared with the quality of classical optical imaging.

    In this article, we propose a novel CGI scheme with CS based on a convolutional neural network (CNN) to improve the image quality. The setup is based on a conventional CGI experimental apparatus. First, the data collected by the CGI device are compressed by the conventional CS algorithm; then, the processed data is trained to reconstruct the ghost image. This scheme combines the advantages of CS with a low sampling rate and a CNN for accurate image reconstruction. Theoretical and experimental results show that this scheme is significantly better than conventional CS and a conventional DL algorithm with a CNN under the same amount of data.

    2. Theory

    We use a conventional CGI experimental device in our work. The setup is shown in Fig. 1. In the setup, a quasi-monochromatic laser illuminates an object T(ρ), and the reflected light carrying the object information is modulated by a spatial light modulator. A bucket detector collects the modulated light Edi(ρ,t). Correspondingly, the calculated light Eci(ρ,t) can be obtained by diffraction theory. The object image can be reconstructed by correlating the signal output by the bucket detector and calculated signal[2325], i.e.,  G(ρ,ρ)=1ni=1n(|Edi(ρ,t)|2|Eci(ρ,t)|2|Edi(ρ,t)|2|Eci(ρ,t)|2),where · stands for an ensemble average. The subscript i=1,2,,n denotes the ith measurement, and n denotes the total number of measurements. For simplicity, the object function T(ρ) is contained in Edi(ρ,t).

    Setup of the CGI system with CS-CNN. SLM, spatial light modulator; BD, bucket detector.

    Figure 1.Setup of the CGI system with CS-CNN. SLM, spatial light modulator; BD, bucket detector.

    The flow chart of the CS-CNN is shown in Fig. 2. In the following, we briefly introduce the process of this algorithm. The algorithm mainly consists of three parts: (i) a conventional CS program to compress the data collected by the CGI device; (ii) a conventional CGI process program; and (iii) a 10-layer CNN constructed for the training data.

    Network structure of the proposed CS-CNN.

    Figure 2.Network structure of the proposed CS-CNN.

    In the conventional CGI device, a set of data (n) is measured by a bucket detector. Correspondingly, according to the diffraction theory of light, the distribution of the idle light field in the object plane can be obtained. Thus, we obtain n200×200 data points. Each data point is divided into 20×20 blocks without overlapping. According to CS theory[15,16], the random Gaussian matrix is used to process the data. The rows of 20×20 data blocks are arranged into a column vector to obtain a 400-dimensional column vector. In this article, the measurement rate is MR=0.25, and thus the size of the measurement matrix is 100×400. Finally, a 100-dimensional measurement vector is obtained. The above process can be expressed as y=ϕx,where ϕRM×N(MN) is the measurement basis matrix, xRN represents the vectorized image block, and yRM is the measurement vector. N/M represents the measurement rate. Following the above steps, we can further compress the data to 50 dimensions.

    A new set of data is obtained by processing the above data with a conventional CGI program. Then, a 10-layer CNN is constructed to train the data. Layers 1–4 of the network are stacked autoencoders, and layers 5–10 are convolution layers. The measurement matrix is replaced by a stacked autoencoder, and the input layer is 20×20 data blocks. All of the rows are arranged into a 400×1 column vector. If the number of neurons in the first layer is C, the measurement rate is MR=C/400. The first layer of the network is connected to the column vector x converted from the input image block, and the number of neurons C is set according to different measurement rates. The activation function is a rectified linear unit (ReLU) function, which outputs the C-dimensional column vector y, i.e.,  y=T(W1x+b1),where T represents the ReLU activation function, W1 represents the weight parameter vector of neurons in the first layer, and b1 represents the bias of neurons in the first layer.

    The second layer of the network is fully connected to the first layer, which has 400 neurons. Take the output y of the first layer as the input, output x, and the activation function is the ReLU function. In the same way, the third layer is fully connected to the second layer with 100 neurons. The fourth layer is fully connected to the third layer with 400 neurons. The initial reconstructed image block vector is rearranged into 20×20 image blocks according to the original row and column to obtain the preliminary reconstructed image block.

    Finally, the CNN is used to reconstruct the image block accurately. The output data of the fourth layer are taken as the input of the fifth layer. In the fifth layer, sixty-four 11×11 convolution kernels are used to generate sixty-four 10×10 feature maps. The sixth layer of the network is connected to the fifth layer (a convolution layer), and thirty-two 1×1 convolution kernels are used to generate thirty-two 20×20 characteristic graphs. The seventh layer of the network is connected to the sixth layer (a convolution layer), and a 7×7 convolution kernel is used to generate a 20×20 feature map. The eighth layer of the network is connected to the seventh layer (a convolution layer), and sixty-four 11×11 convolution cores are used to generate sixty-four 20×20 feature maps. The ninth layer of the network is connected to the eighth layer (a convolution layer), and thirty-two 1×1 convolution kernels are used to generate thirty-two 20×20 characteristic graphs. The activation function of the above process is an ReLU function. The tenth layer of the network is connected to the ninth layer (a convolution layer). A 7×7 convolution kernel is used. The number of zeros in the tenth layer (a convolution layer) is three, and the output of the activation function is not used to generate the reconstructed image block of size 20×20.

    In the DL framework Caffe, the 10-layer network is trained in an unsupervised way, and the loss function is L({W})=1Ti=1TF(xi,{W})xi2.

    The number of input neurons in the first layer is zero, and the number of output neurons in the fourth layer is zero. In the 5th to 10th layers of the network, the initial weight distribution is subject to a Gaussian distribution with a mean of zero and a variance of 0.01. In layers 1–10 of the network, the initial offset values are set to zero. After the deep neural network, the reconstructed image blocks are obtained, then the image blocks are rearranged according to the original row, and the row values are rearranged according to the index.

    3. Results

    The experimental setup is schematically shown in Fig. 1. A standard monochromatic laser (30 mW, Changchun New Industries Optoelectronics Technology Co., Ltd., MGL-III-532) with wavelength λ=532nm illuminates an object (Rubik’s Cube). The light reflected by the object focuses on a two-dimensional amplitude-only ferroelectric liquid crystal spatial light modulator (Meadowlark Optics A512-450-850) with 512×512 addressable 15μm×15μm pixels through the lens. A bucket detector collects the modulated light. Correspondingly, the reference signal is obtained by MATLAB software. The ghost image is reconstructed by the CS-CNN. In this experiment, the sampling rate is MR=0.25, and the number of training sets is 1000.

    Figure 3 shows a set of experimental results. Figure 3(a1) is the object. Figures 3(a2)3(a5) represent reconstructed ghost images with different numbers of frames. The results show that the image quality is significantly improved by increasing the number of frames. High-quality ghost images comparable to classical optical imaging can be produced with little data. To quantitatively analyze the quality of the reconstructed image at different frames, the peak signal to noise ratio (PSNR) and structural similarity index (SSIM) are used as our evaluation indexes. As can be seen from Fig. 3(b), despite the number of samples being very small, the reconstructions are still in reasonable quality.

    Ghost images reconstructed by CGI with CS-CNN. (a1) Classical image. The numbers of frames in the reconstructed ghost images are (a2) 30, (a3) 50, (a4) 70, and (a5) 90. (b) PSNR and SSIM curves of the reconstructed images with different frame numbers.

    Figure 3.Ghost images reconstructed by CGI with CS-CNN. (a1) Classical image. The numbers of frames in the reconstructed ghost images are (a2) 30, (a3) 50, (a4) 70, and (a5) 90. (b) PSNR and SSIM curves of the reconstructed images with different frame numbers.

    We compare the conventional CS, DL, and CS-CNN CGI algorithms based on the same experimental data in Fig. 4. CGI can not effectively reconstruct the image when the number of frames is less than 100. Consequently, there is no experimental result of CGI in Fig. 4. The conventional CS algorithm and CS-CNN algorithm have the same sampling rate, i.e., MR=0.25. The DL algorithm and CS-CNN algorithm set the same dataset, i.e., 1000. When the number of samples is very low, Fig. 4 shows that with the same number of frames the image quality obtained by this scheme is the best. The quantitative results (Fig. 5) show that the PSNR of CGI with CS-CNN is on average 28.4% higher than that of CGI with DL under the same reconstructed frame number, and SSIM increases by 93.8% on average[26].

    Detailed comparison between the ghost images reconstructed using the conventional CS algorithm, DL algorithm, and CS-CNN algorithm. The number of frames is (a) 30, (b) 50, (c) 70, and (d) 90.

    Figure 4.Detailed comparison between the ghost images reconstructed using the conventional CS algorithm, DL algorithm, and CS-CNN algorithm. The number of frames is (a) 30, (b) 50, (c) 70, and (d) 90.

    PSNR and SSIM curves of reconstructed images of CS, DL, and CS-CNN with different frame numbers.

    Figure 5.PSNR and SSIM curves of reconstructed images of CS, DL, and CS-CNN with different frame numbers.

    4. Summary

    In summary, we have proposed a novel method to improve the image quality of CGI. This method combines the advantages of the CS algorithm and CNN algorithm. We analyzed the performance of the conventional CGI, CS, and DL algorithms under the same conditions and observed that our CS-CNN scheme outperforms the other methods, especially when the sampling rate is small. CS based on a CNN is the better CGI method to date. This method provides a promising solution to these challenges that prohibit the use of CGI in practical applications.

    References

    [1] T. B. Pittman, Y. H. Shih, D. V. Strekalov, A. V. Sergienko. Optical imaging by means of two-photon quantum entanglement. Phys. Rev. A, 52, R3429(1995).

    [2] J. Cheng, S.-S. Han. Incoherent coincidence imaging and its applicability in X-ray diffraction. Phys. Rev. Lett., 92, 093903(2004).

    [3] X. H. Chen, Q. Liu, K. H. Luo, L. A. Wu. Lensless ghost imaging with true thermal light. Opt. Lett., 34, 695(2009).

    [4] B. I. Erkmen. Computational ghost imaging for remote sensing. J. Opt. Soc. A, 29, 782(2012).

    [5] D. Y. Duan, Z. X. Man, Y. J. Xia. Nondegenerate wavelength computational ghost imaging with thermal light. Opt. Express, 27, 25187(2019).

    [6] J. H. Gu, S. Sun, Y. K. Xu, H. Z. Lin, W. T. Liu. Feedback ghost imaging by gradually distinguishing and concentrating onto the edge area. Chin. Opt. Lett., 19, 041102(2021).

    [7] G. Wang, H. B. Zheng, Z. G. Tang, Y. C. He, Y. Zhou, H. Chen, J. B. Liu, Y. Yuan, F. L. Li, Z. Xu. Naked-eye ghost imaging via photoelectric feedback. Chin. Opt. Lett., 18, 091101(2020).

    [8] D. Pelliccia, A. Rack, M. Scheel, V. Cantelli, D. M. Paganin. Experimental X-ray ghost imaging. Phys. Rev. Lett., 117, 113902(2016).

    [9] H. Yu, R. Lu, S. Han, H. Xie, G. Du, T. Xiao, D. Zhu. Fourier-transform ghost imaging with hard X rays. Phys. Rev. Lett., 117, 113901(2016).

    [10] A. Zhang, Y. He, L. Wu, L. Chen, B. Wang. Tabletop X-ray ghost imaging with ultra-low radiation. Optica, 5, 374(2018).

    [11] W. Li, Z. Tong, K. Xiao, Z. Liu, Q. Gao, J. Sun, S. Liu, S. Han, Z. Wang. Single-frame wide-field nanoscopy based on ghost imaging via sparsity constraints. Optica, 6, 1515(2019).

    [12] W. Gong, S. Han. High-resolution far-field ghost imaging via sparsity constraint. Sci. Rep., 5, 9280(2015).

    [13] J. H. Shapiro. Computational ghost imaging. Phys. Rev. A, 78, 061802(R)(2008).

    [14] Y. Bromberg, O. Katz, Y. Silberberg. Ghost imaging with a single detector. Phys. Rev. A, 79, 053840(2009).

    [15] O. Katza, Y. Bromberg, Y. Silberberg. Compressive ghost imaging. Appl. Phys. Lett., 95, 131110(2009).

    [16] V. Katkovnik, J. Astola. Compressive sensing computational ghost imaging. J. Opt. Soc. Am. A, 29, 1556(2012).

    [17] P. W. Wang, C. L. Wang, C. P. Yu, S. Yue, W. L. Gong, S. S. Han. Color ghost imaging via sparsity constraint and non-local self-similarity. Chin. Opt. Lett., 19, 021102(2021).

    [18] Z. Chen, J. Shi, G. Zeng. Object authentication based on compressive ghost imaging. Appl. Opt., 55, 8644(2016).

    [19] M. Lyu, W. Wang, H. Wang, W. Wang, G. Li, N. Chen, G. Situ. Deep-learning-based ghost imaging. Sci. Rep., 7, 17865(2017).

    [20] Y. He, G. Wang, G. Dong, S. Zhu, H. Chen, A. Zhang, Z. Xu. Ghost imaging based on deep learning. Sci. Rep., 8, 6469(2018).

    [21] T. Shimobaba, Y. Endo, T. Nishitsuji, T. Takahashi, Y. Nagahama, T. Hasegawa, M. Sano, R. Hirayama, T. Kakue, A. Shiraki, T. Ito. Computational ghost imaging using deep learning. Opt. Commun., 413, 147(2018).

    [22] G. Barbastathis, A. Ozcan, G. Situ. On the use of deep learning for computational imaging. Optica, 6, 921(2019).

    [23] X. L. Yin, Y. J. Xia, D. Y. Duan. Theoretical and experimental study of the color of ghost imaging. Opt. Express, 26, 18944(2018).

    [24] W. J. Jiang, X. Y. Li, X. L. Peng, B. Q. Sun. Imaging high-speed moving targets with a single-pixel detector. Opt. Express, 28, 7889(2020).

    [25] D. F. Shi, C. Y. Fan, P. F. Zhang, H. Shen, J. H. Zhang, C. H. Qiao, Y. J. Wang. Two-wavelength ghost imaging through atmospheric turbulence. Opt. Express, 21, 2050(2013).

    [26] Y. H. Liu, S. Y. Liu, F. X. Fu. Optimization of compressed sensing reconstruction algorithms based on convolutional neural network. Comput. Sci., 47, 143(2020).

    Hao Zhang, Deyang Duan. Computational ghost imaging with compressed sensing based on a convolutional neural network[J]. Chinese Optics Letters, 2021, 19(10): 101101
    Download Citation