• Photonics Research
  • Vol. 9, Issue 2, B30 (2021)
Chengshuai Yang1, Yunhua Yao1、6、*, Chengzhi Jin1, Dalong Qi1, Fengyan Cao1, Yilin He1, Jiali Yao1, Pengpeng Ding1, Liang Gao2, Tianqing Jia1, Jinyang Liang3, Zhenrong Sun1, and Shian Zhang1、4、5、7、*
Author Affiliations
  • 1State Key Laboratory of Precision Spectroscopy, School of Physics and Electronic Science, East China Normal University, Shanghai 200062, China
  • 2Department of Electrical and Computer Engineering, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
  • 3Institut National de la Recherche Scientifique, Centre Énergie Matériaux Télécommunications, Laboratory of Applied Computational Imaging, Varennes, Québec J3X1S2, Canada
  • 4Collaborative Innovation Center of Extreme Optics, Shanxi University, Taiyuan 030006, China
  • 5Collaborative Innovation Center of Light Manipulations and Applications, Shandong Normal University, Jinan 250358, China
  • 6e-mail: yhyao@lps.ecnu.edu.cn
  • 7e-mail: sazhang@phy.ecnu.edu.cn
  • show less
    DOI: 10.1364/PRJ.410018 Cite this Article Set citation alerts
    Chengshuai Yang, Yunhua Yao, Chengzhi Jin, Dalong Qi, Fengyan Cao, Yilin He, Jiali Yao, Pengpeng Ding, Liang Gao, Tianqing Jia, Jinyang Liang, Zhenrong Sun, Shian Zhang. High-fidelity image reconstruction for compressed ultrafast photography via an augmented-Lagrangian and deep-learning hybrid algorithm[J]. Photonics Research, 2021, 9(2): B30 Copy Citation Text show less

    Abstract

    Compressed ultrafast photography (CUP) is the fastest single-shot passive ultrafast optical imaging technique, which has shown to be a powerful tool in recording self-luminous or non-repeatable ultrafast phenomena. However, the low fidelity of image reconstruction based on the conventional augmented-Lagrangian (AL) and two-step iterative shrinkage/thresholding (TwIST) algorithms greatly prevents practical applications of CUP, especially for those ultrafast phenomena that need high spatial resolution. Here, we develop a novel AL and deep-learning (DL) hybrid (i.e., AL+DL) algorithm to realize high-fidelity image reconstruction for CUP. The AL+DL algorithm not only optimizes the sparse domain and relevant iteration parameters via learning the dataset but also simplifies the mathematical architecture, so it greatly improves the image reconstruction accuracy. Our theoretical simulation and experimental results validate the superior performance of the AL+DL algorithm in image fidelity over conventional AL and TwIST algorithms, where the peak signal-to-noise ratio and structural similarity index can be increased at least by 4 dB (9 dB) and 0.1 (0.05) for a complex (simple) dynamic scene, respectively. This study can promote the applications of CUP in related fields, and it will also enable a new strategy for recovering high-dimensional signals from low-dimensional detection.

    1. INTRODUCTION

    Ultrafast imaging has played an indispensable role in photochemistry [1,2], biomedicine [35], microfluidics [6], shock waves [7], and plasma physics [8]. Recently, various ultrafast imaging techniques have been developed, including compressed ultrafast photography (CUP) [911]. Unlike some active ultrafast imaging techniques that need specific illumination light [1214] or a pump–probe technique that requires multiple measurements [1517], CUP is a single-shot and passive ultrafast imaging technique. Its temporal resolution and number of frames can reach tens of femtoseconds and several hundred, respectively. Therefore, CUP has great advantages for measuring some self-luminous or non-repeatable ultrafast phenomena, which is attributed mainly to the novel model of CUP, which combines compressed sensing (CS) theory and time–space conversion technology. So far, CUP has been successfully applied to measure light reflection and refraction [9], femtosecond temporal focusing [10], photonic Mach cones [18], dissipative solitons [19], phase-sensitive transparent objects [20], three-dimensional (3D) objects [21], ultrashort laser spatiotemporal evolution [22], and photoluminescence processes [9]. However, due to the high data compression ratio, the fidelity of reconstructed images for CUP is relatively low by the conventional two-step iterative shrinkage/thresholding (TwIST) algorithm, which limits its practicality. To improve image fidelity, a variety of methods have been proposed, such as a space- and intensity-constrained image reconstruction algorithm [23], augmented-Lagrangian (AL)-based image reconstruction algorithm [24], plug-and-play alternating direction method of multipliers algorithm [25], optimizing the codes for CUP [26], lossless CUP [18], and multi-encoding CUP [27]. These proposed schemes can improve image fidelity to a certain extent, but there are still great challenges in measuring the complex dynamic scenes.

    In image reconstruction of CUP, all selections of the sparse domain, determination of relevant iteration parameters, and denoising after iteration calculation greatly limit image fidelity. To completely solve these problems, we developed a novel image reconstruction method based on an AL and deep-learning (DL) hybrid (i.e., AL+DL) algorithm. This idea is borrowed mainly from some early algorithms, such as the AL algorithm [24,28,29], learning iteration parameters [3033], learning sparse domain [3437], and U-net architecture [38], but there are still many differences compared to each of the early algorithms. First, the AL+DL algorithm utilizes multiple learning transformations to seek the best sparse domain. Typically, the sparse domain in conventional TwIST and AL algorithms is determined before image reconstruction [24,39], so it is usually not optimal for one dynamic scene. In contrast, the sparse domain in the AL+DL algorithm can be optimized in multiple transformations, which is more pertinent. Second, the AL+DL algorithm takes full advantage of gradient descent (GD), DL, and AL algorithms, which simplifies the mathematical architecture to deal with the 3D tensor problem, and these advantages can reduce the cost of each iteration and decrease the number of iterations. Third, the AL+DL algorithm optimizes the relevant iteration parameters by learning the dataset, which is different from previous AL and TwIST algorithms, where these parameters are artificially predetermined. Finally, the AL+DL algorithm uses a U-net architecture containing attention layers to help denoise and retain the spatial details of the images after iteration calculation. Importantly, our theoretical simulation and experimental results show the AL+DL algorithm can obtain much higher image fidelity than conventional AL and TwIST algorithms for CUP, which strongly supports our theory.

    2. PRINCIPLE

    In CUP, a 3D dynamic scene I(x,y,t) is encoded by operator C, sheared by operator S, and integrated by operator T, and finally a two-dimensional (2D) image E(x,y) is obtained. For convenience, hereafter, I(x,y,t) is abbreviated to I, and E(x,y) is abbreviated to E. Mathematically, this process can be described as E=TSCI.

    For simplicity, we define O=TSC. Thus, Eq. (1) can be further written as E=OI.

    To recover 3D I from 2D E, we need to solve the inverse problem of Eq. (2). The number of elements in I is much larger than that in E, so the inverse problem of Eq. (2) is undetermined. The CUP strategy is to introduce a CS theory [9]. The CS theory makes full use of the sparsity of I in a certain domain to recover the original information. This sparsity in one domain means that only a few elements are nonzero, while most of the elements are zero. Consider a case in which I has n elements and E has m elements in the original domain, and I has s nonzero elements in a sparse domain, i.e., the sparsity where ns and n>m>s. Due to the fact that m is generally larger than s, this makes it possible to solve the inverse problem of Eq. (2). In a practical solution, the CS algorithm minimizes I in a sparse domain on condition of Eq. (1), which is shown as {minIΦ(I)s.t.EOI=0,where Φ(I) is the expression of I in the sparse domain. According to CS theory [40,41], the original dynamic scene I can be completely recovered when m>fsμ2,where f is a constant correlated with the number of elements n, μ is the mutual coherence between the sparse basis of the sparse domain Φ and operator O, determined by operators C, S, and T. From Eq. (4), one can see that both increasing m and reducing s and μ are feasible schemes to improve the quality of image reconstruction. However, increasing m, i.e., increasing the sampling rate, will reduce the spatial resolution or requires many streak cameras, which is impractical in the actual CUP system. Thus, reducing s and μ is the best choice. Optimizing O can reduce only μ, while optimizing Φ can reduce both s and μ; therefore, here we employ the method of optimizing Φ. To optimize Φ, we impose a low-rank property on the entire dynamic scene I (tensor) with many different transformations [42], which is different from traditional methods with only one transformation. Thus, problem (3) can be further written as {minI(x,y,t)pqψpIs.t.EOI=0,where ψp represents one transformation in the sparse domain, and q denotes the total number of transformations. In transforming problem (5) from a constraint into an unconstraint, there exist two frameworks: the penalty function method and the AL method. The performance of the AL method is better than that of the penalty function method, which has been proved in previous works [24,29], and therefore, here the AL method is adopted. Thus, problem (5) can be transformed into minI{pqψpIγ(EOI)+ζ2EOIF2},where γ and ζ are the Lagrangian multiplier and penalty parameter, respectively, which are associated with I. For convenience, problem (6) is further written as minI{pqψpI+ζ2EγζOIF2}.

    To solve problem (7), an auxiliary variable J is introduced into problem (7), and is written as {minI{pqψpJ+ζ2EγζOIF2}s.t.I=J.

    By adopting the AL method, the constrained problem (8) can be transformed into minI,J{pq[ψpJλp(IJ)+δp2IJF2]+ζ2EγζOIF2},where λ={λ1,λ2,,λq} and δ={δ1,δ2,,δq} are the Lagrangian multipliers and penalty parameters, respectively, which are associated with J in different transformations. By transformation, problem (9) can be further written as minI,J{pq(ψpJ+δp2IλpδpJF2)+ζ2EγζOIF2}.

    Problem (10) can be solved by an alternating direction method of multipliers (ADMM) based on an iteration of solving the I-subproblem and J-subproblem alternatively. However, in the J-subproblem, the sparse domains in different transformations lead to different solutions at the beginning of the iteration. Therefore, some independent auxiliary variables W={W1,W2,,Wq} are introduced for each transformation, and thus problem (10) can be written as minI,W{pq(ψpWp+δp2IλpδpWpF2)+ζ2EγζOIF2}.

    To solve problem (11), the ADMM is also adopted to solve the I-subproblem and W-subproblem alternatively. In the kth iteration, the I-subproblem can be written as Ik=argminI{pqδpk2IλpkδpkWpk1F2+ζk2EγkζkOIF2},and the W-subproblem can be written as Wpk=argminWpψpWp+δpk2IkλpkδpkWpF2.

    The I-subproblem in Eq. (12) is a quadratic regularized least-squares problem, and its direct solution is given in a closed form as {Ik=HARD[ζkOTEγkOT+(pfδpkWpk1λpk)],HARD=(ζkOTO+pfδpkIiden)1,where Iiden is an identity operator. It is expensive to compute HARD due to large data, so here the GD algorithm is used to solve the I-subproblem. The main shortcoming of the GD algorithm is that the number of iterations is very large because it is difficult to choose the step size. Some methods have been developed to obtain a better step size through much computation, such as the Barzilai and Borwein (BB) method [43,44]. Here, a learning method is used to seek the optimized step size. In our method, the number of iterations is much less than that of the BB method. Based on the GD algorithm, the solution to Eq. (12) can be expressed as Ik=Ik-1αk[pqδpk(Ik1λpkδpkWpk1)ζkOT(EγkζkOIk1)],where α is the step size in the GD algorithm, and here the GD algorithm is utilized to calculate Eq. (15). From Eqs. (15) and (13), one can see that the solution to problem (6) depends mainly on Wp instead of I. For convenience, the solution to Wp can be further written as Wpk=Sp(Ikλpkδpk).

    Data flow chart of the AL+DL algorithm. (a) Solver Sp(·) in a sparse domain; (b) general framework by connecting each iteration in a sequence order. Here, each Si(·) is calculated in parallel to producing Wi, and GD algorithm is employed to calculate I.

    Figure 1.Data flow chart of the AL+DL algorithm. (a) Solver Sp(·) in a sparse domain; (b) general framework by connecting each iteration in a sequence order. Here, each Si(·) is calculated in parallel to producing Wi, and GD algorithm is employed to calculate I.

    (a) U-net architecture in the AL+DL algorithm; (b) self-attention model.

    Figure 2.(a) U-net architecture in the AL+DL algorithm; (b) self-attention model.

    3. THEORETICAL SIMULATIONS

    Reconstructed results of (a) boatman, (b) ocean animal, and (c) finger by the AL+DL (second row), AL (third row), and TwIST (fourth row) algorithms, together with the ground truth (first row) for comparison. The last column is the enlarged image in the corresponding red squares.

    Figure 3.Reconstructed results of (a) boatman, (b) ocean animal, and (c) finger by the AL+DL (second row), AL (third row), and TwIST (fourth row) algorithms, together with the ground truth (first row) for comparison. The last column is the enlarged image in the corresponding red squares.

    Here, only three representative pictures are selected, and an interesting area in each dynamic scene is enlarged for observation. Spatial details in the boatman, ocean animal, and finger can be clearly observed by the AL+DL algorithm, while these details are submerged by the AL and TwIST algorithms, which is disadvantageous for high-spatial-resolution imaging of a dynamic scene. To intuitively compare the improved efficiency in image fidelity by the AL+DL algorithm, we calculate PSNR and SSIM, and the calculated results are given in Table 1. Compared to the AL and TwIST algorithms, both PSNR and SSIM by the AL+DL algorithm are significantly improved. Here, PSNR (SSIM) is increased by at least 4.35 dB (0.136) for the boatman, 5.47 dB (0.114) for the ocean animal, and 9.78 dB (0.051) for the finger. Based on these calculated results, a rule can be found, which is, the simpler the spatial structure of the dynamic scene, the higher the improvement efficiency of PSNR, while the improvement efficiency of SSIM shows the opposite behavior. This phenomenon should be related to the sparsity of the dynamic scene; the simpler dynamic scene usually has higher sparsity, and vice versa. PSNR is based on a logarithmic function, which is not very well matched to perceived visual quality, but SSIM is based on visible structures in the image. Thus, PSNR has high improvement efficiency for a simple dynamic scene (i.e., finger), while SSIM has high improvement efficiency for a complex dynamic scene (i.e., boatman). In addition, the AL+DL algorithm can reconstruct a dynamic scene in only a few seconds, which is much shorter than the AL and TwIST algorithms, which need tens of seconds; the computing efficiency is improved by an order of magnitude, which is very beneficial in practical applications of CUP.

    Average PSNR (in dB) and SSIM by Different Image Reconstruction Algorithms in Different Dynamic Scenes

    SceneAL+DLALTwIST
    PSNRSSIMPSNRSSIMPSNRSSIM
    Boatman28.500.83624.150.70022.470.589
    Ocean animal30.470.91625.000.80224.720.781
    Finger42.000.98332.220.93228.560.894

    4. EXPERIMENTAL RESULTS

    System configuration of CUP. DMD, digital micromirror device; CMOS, complementary metal–oxide-semiconductor.

    Figure 4.System configuration of CUP. DMD, digital micromirror device; CMOS, complementary metal–oxide-semiconductor.

    Measuring temporal evolution of a spatially modulated picosecond laser spot. (a) Experimental design. (b)–(d) Reconstructed results by the AL+DL, AL, and TwIST algorithms, respectively. (e) Measured static image by external CCD. (f)–(h) Extracted images from (b)–(d), respectively, at the time of 14 ps; curves on the right are the integration results of the corresponding images along the horizontal direction.

    Figure 5.Measuring temporal evolution of a spatially modulated picosecond laser spot. (a) Experimental design. (b)–(d) Reconstructed results by the AL+DL, AL, and TwIST algorithms, respectively. (e) Measured static image by external CCD. (f)–(h) Extracted images from (b)–(d), respectively, at the time of 14 ps; curves on the right are the integration results of the corresponding images along the horizontal direction.

    Here, the static image is achieved by external CCD measurement without encoding operator C and shearing operator S, as shown in Fig. 5(e). Meanwhile, the intensities of Figs. 5(e)–5(h) are also integrated along the horizontal direction, and the calculated results are given on the right of the relative images. The AL+DL algorithm can retain very high image fidelity, but the AL and TwIST algorithms cause a certain degree of image distortion. The fundamental reason should be the mismatch of the sparse domain in image reconstruction. More importantly, like the static image, the blocked part in the laser spot (see light blue squares) can be clearly distinguished by the AL+DL algorithm, where an obvious valley in the intensity curve is observed, but not by either the AL or TwIST algorithm, especially the TwIST algorithm.

    Measuring wavefront movement by obliquely illuminating a collimated femtosecond laser pulse on a transverse fan pattern. (a) Experimental design. (b)–(d) Reconstructed results by the AL+DL, AL, and TwIST algorithms, respectively. (e) Measured static image by external CCD. (f)–(h) Integrated images from (b)–(d), respectively. (i)–(l) Results of Fourier transform from (e)–(h), respectively.

    Figure 6.Measuring wavefront movement by obliquely illuminating a collimated femtosecond laser pulse on a transverse fan pattern. (a) Experimental design. (b)–(d) Reconstructed results by the AL+DL, AL, and TwIST algorithms, respectively. (e) Measured static image by external CCD. (f)–(h) Integrated images from (b)–(d), respectively. (i)–(l) Results of Fourier transform from (e)–(h), respectively.

    5. DISCUSSION

    The AL+DL algorithm is a data-driven method, which can optimize the sparse domain and relevant iteration parameters by learning instead of hand-crafted determination. For CS, the sparse domain is the core part that determines the sparsity and affects mainly the coherence. Thus, the sparse domain almost determines the image reconstruction quality. In general, the learning method can seek better sparse domain and iteration parameters, and therefore the AL+DL algorithm can get higher image fidelity than conventional AL and TwIST algorithms. Because of learning the sparse domain and iteration parameters, the AL+DL algorithm has high robustness and allows the encoding operator C to be different in training and testing processes, while the pure neural network algorithms cannot, such as deep fully connected networks [58], ReconNet [59], DR2-Net [60], λ-net [38], and DeepCubeNet [61]. Also, the AL+DL algorithm embeds a GD algorithm into tensor computation, which involves massive data. In calculation, the GD algorithm does not easily find the appropriate step size, so it needs to perform many iterations, i.e., the convergence speed is low. To decrease the number of iterations, data scientists prefer Newton’s method or a conjugate gradient algorithm by increasing the cost of each iteration [62]. However, some mathematicians seek a better step size in the GD algorithm to decrease the number of iterations by increasing the cost of each iteration, such as the BB method. Here, we utilize the GD algorithm to calculate the large data by a data-driven method based on the learning model, which can find the optimal step size to decrease the number of iterations without increasing the cost of each iteration and make the gradient show better orthogonality. It is noted that the AL+DL algorithm needs just 15 iterations, while the corresponding traditional algorithm based on the BB method needs more than 100 iterations.

    As shown in Figs. 3, 5, and 6, compared to the AL and TwIST algorithms, the AL+DL algorithm shows great advantages in image reconstruction accuracy, but it also inherits the shortcoming of the data-driven method, i.e., the dependence on a learning dataset. In image reconstruction, these images in the dataset should have some similarities to those in the dynamic scene. An inappropriate training dataset may lead to results worse than those obtained by the AL and TwIST algorithms. In some special dynamic scenes, it may be difficult to find a similar dataset for training. In this case, it is feasible to increase the sampling rate m, such as lossless-CUP or multi-encoding CUP. Moreover, it is also a good idea to optimize the codes, which is similar to optimizing the sparse domain, which can reduce coherence. However, the AL+DL algorithm cannot be adopted directly to optimize the codes, because here the codes are considered as constant. Optimizing the codes demands that the mathematical architecture regard the codes as a variable; thus, the whole AL+DL architecture needs to be redesigned. In the future, we will strive to seek some new algorithms to simultaneously optimize the codes, sparse domain, and iteration parameters by learning the dataset.

    6. CONCLUSION

    In summary, we have developed a new AL+DL algorithm to realize high-fidelity image reconstruction for CUP. In our method, there are four key points: (1) optimizing the sparse domain in multiple transformation; (2) optimizing the relevant calculation parameters in the iteration process; (3) employing the GD algorithm to improve computing efficiency; (4) embedding the U-net architecture to help denoise. Key points (1), (2), and (4) are implemented by the DL method, and improving key point (3) also needs the DL method. However, the whole framework is determined by the AL method, which combines these four key points. Thus, the AL+DL algorithm not only utilizes the training neural networks, but also has some potential mathematical interpretations. More importantly, these results from theoretical simulations and experimental measurements show that the AL+DL algorithm is superior to conventional AL and TwIST algorithms in image fidelity and computing efficiency. Additionally, the AL+DL algorithm is a simple mathematical architecture, so it is easy to extend to other high-dimensional tensor fields. In future studies, we will continue to search for better image reconstruction algorithms for CUP to achieve super-high image fidelity.

    References

    [1] P. R. Poulin, K. A. Nelson. Irreversible organic crystalline chemistry monitored in real time. Science, 313, 1756-1760(2006).

    [2] P. Hockett, C. Z. Bisgaard, O. J. Clarkin, A. Stolow. Time-resolved imaging of purely valence-electron dynamics during a chemical reaction. Nat. Phys., 7, 612-615(2011).

    [3] R. Horstmeyer, H. Ruan, C. Yang. Guidestar-assisted wavefront-shaping methods for focusing light into biological tissue. Nat. Photonics, 9, 563-571(2015).

    [4] J. W. Borst, A. J. Visser. Fluorescence lifetime imaging microscopy in life sciences. Meas. Sci. Technol., 21, 102002(2010).

    [5] H. R. Petty. Spatiotemporal chemical dynamics in living cells: from information trafficking to cell physiology. Biosystems, 83, 217-224(2006).

    [6] T. M. Squires, S. R. Quake. Microfluidics: fluid physics at the nanoliter scale. Rev. Mod. Phys., 77, 977-1026(2005).

    [7] N. Šiaulys, L. Gallais, A. Melninkaitis. Direct holographic imaging of ultrafast laser damage process in thin films. Opt. Lett., 39, 2164-2167(2014).

    [8] R. L. Kodama, P. A. Norreys, K. Mima, A. E. Dangor, R. G. Evans, H. Fujita, Y. Kitagawa, K. Krushelnick, T. Miyakoshi, N. Miyanaga. Fast heating of ultrahigh-density plasma as a step towards laser fusion ignition. Nature, 412, 798-802(2001).

    [9] L. Gao, J. Liang, C. Li, L. V. Wang. Single-shot compressed ultrafast photography at one hundred billion frames per second. Nature, 516, 74-77(2014).

    [10] J. Liang, L. Zhu, L. V. Wang. Single-shot real-time femtosecond imaging of temporal focusing. Light Sci. Appl., 7, 42(2018).

    [11] D. Qi, S. Zhang, C. Yang, Y. He, F. Cao, J. Yao, P. Ding, L. Gao, T. Jia, J. Liang. Single-shot compressed ultrafast photography: a review. Adv. Photon., 2, 014003(2020).

    [12] K. Nakagawa, A. Iwasaki, Y. Oishi, R. Horisaki, A. Tsukamoto, A. Nakamura, K. Hirosawa, H. Liao, T. Ushida, K. Goda. Sequentially timed all-optical mapping photography (STAMP). Nat. Photonics, 8, 695-700(2014).

    [13] T. Suzuki, R. Hida, Y. Yamaguchi, K. Nakagawa, T. Saiki, F. Kannari. Single-shot 25-frame burst imaging of ultrafast phase transition of Ge2Sb2Te5 with a sub-picosecond resolution. Appl. Phys. Express, 10, 092502(2017).

    [14] Y. Lu, T. T. Wong, F. Chen, L. Wang. Compressed ultrafast spectral-temporal photography. Phys. Rev. Lett., 122, 193904(2019).

    [15] A. Velten, T. Willwacher, O. Gupta, A. Veeraraghavan, M. G. Bawendi, R. Raskar. Recovering three-dimensional shape around a corner using ultrafast time-of-flight imaging. Nat. Commun., 3, 745(2012).

    [16] A. H. Zewail. Four-dimensional electron microscopy. Science, 328, 187-193(2010).

    [17] A. Barty, S. Boutet, M. J. Bogan, S. Hau-Riege, S. Marchesini, K. Sokolowski-Tinten, N. Stojanovic, H. Ehrke, A. Cavalleri, S. Düsterer. Ultrafast single-shot diffraction imaging of nanoscale dynamics. Nat. Photonics, 2, 415-419(2008).

    [18] J. Liang, C. Ma, L. Zhu, Y. Chen, L. Gao, L. V. Wang. Single-shot real-time video recording of a photonic Mach cone induced by a scattered light pulse. Sci. Adv., 3, e1601814(2017).

    [19] J. C. Jing, X. Wei, L. V. Wang. Spatio-temporal-spectral imaging of non-repeatable dissipative soliton dynamics. Nat. Commun., 11, 2059(2020).

    [20] T. Kim, J. Liang, L. Zhu, L. V. Wang. Picosecond-resolution phase-sensitive imaging of transparent objects in a single shot. Sci. Adv., 6, e6200(2020).

    [21] J. Liang, L. Gao, P. Hai, C. Li, L. V. Wang. Encrypted three-dimensional dynamic imaging using snapshot time-of-flight compressed ultrafast photography. Sci. Rep., 5, 15504(2015).

    [22] F. Cao, C. Yang, D. Qi, J. Yao, Y. He, X. Wang, W. Wen, J. Tian, T. Jia, Z. Sun. Single-shot spatiotemporal intensity measurement of picosecond laser pulses with compressed ultrafast photography. Opt. Laser Eng., 116, 89-93(2019).

    [23] L. Zhu, Y. Chen, J. Liang, Q. Xu, L. Gao, C. Ma, L. V. Wang. Space-and intensity-constrained reconstruction for compressed ultrafast photography. Optica, 3, 694-697(2016).

    [24] C. Yang, D. Qi, F. Cao, Y. He, X. Wang, W. Wen, J. Tian, T. Jia, Z. Sun, S. Zhang. Improving the image reconstruction quality of compressed ultrafast photography via an augmented Lagrangian algorithm. J. Opt., 21, 035703(2019).

    [25] Y. Lai, Y. Xue, C. Y. Côté, X. Liu, A. Laramée, N. Jaouen, F. Légaré, L. Tian, J. Liang. Single-shot ultraviolet compressed ultrafast photography. Laser Photon. Rev., 14, 2000122(2020).

    [26] C. Yang, D. Qi, X. Wang, F. Cao, Y. He, W. Wen, T. Jia, J. Tian, Z. Sun, L. Gao. Optimizing codes for compressed ultrafast photography by the genetic algorithm. Optica, 5, 147-151(2018).

    [27] C. Yang, D. Qi, J. Liang, X. Wang, F. Cao, Y. He, X. Ouyang, B. Zhu, W. Wen, T. Jia. Compressed ultrafast photography by multi-encoding imaging. Laser Phys. Lett., 15, 116202(2018).

    [28] C. Li. An efficient algorithm for total variation regularization with applications to the single pixel camera and compressive sensing(2010).

    [29] M. V. Afonso, J. M. Bioucas-Dias, M. A. Figueiredo. An augmented Lagrangian approach to the constrained optimization formulation of imaging inverse problems. IEEE Trans. Image Process., 20, 681-695(2010).

    [30] Y. Yang, J. Sun, H. Li, Z. Xu. ADMM-CSNet: a deep learning approach for image compressive sensing. IEEE Trans. Pattern Anal., 42, 521-538(2018).

    [31] J. Zhang, B. Ghanem. ISTA-Net: interpretable optimization-inspired deep network for image compressive sensing. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1828-1837(2018).

    [32] J. Ma, X. Liu, Z. Shou, X. Yuan. Deep tensor ADMM-net for snapshot compressive imaging. Proceedings of the IEEE International Conference on Computer Vision, 10223-10232(2019).

    [33] K. Monakhova, J. Yurtsever, G. Kuo, N. Antipa, K. Yanny, L. Waller. Learned reconstructions for practical mask-based lensless imaging. Opt. Express, 27, 28075-28090(2019).

    [34] Q. Xie, Q. Zhao, D. Meng, Z. Xu. Kronecker-basis-representation based tensor sparsity and its applications to tensor recovery. IEEE Trans. Pattern Anal. Mach. Intell., 40, 1888-1902(2017).

    [35] Y. Wang, J. Peng, Q. Zhao, Y. Leung, X. Zhao, D. Meng. Hyperspectral image restoration via total variation regularized low-rank tensor decomposition. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sensing, 11, 1227-1243(2017).

    [36] L. Wang, C. Sun, Y. Fu, M. H. Kim, H. Huang. Hyperspectral image reconstruction using a deep spatial-spectral prior. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 8032-8041(2019).

    [37] Z. Wu, Y. Sun, A. Matlock, J. Liu, L. Tian, U. S. Kamilov. SIMBA: scalable inversion in optical tomography using deep denoising priors. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sensing, 14, 1163-1175(2020).

    [38] X. Miao, X. Yuan, Y. Pu, V. Athitsos. Lambda-net: reconstruct hyperspectral images from a snapshot measurement. IEEE/CVF International Conference on Computer Vision (ICCV), 4058-4068(2019).

    [39] J. M. Bioucas-Dias, M. A. Figueiredo. A new TwIST: two-step iterative shrinkage/thresholding algorithms for image restoration. IEEE Trans. Image Process., 16, 2992-3004(2007).

    [40] E. J. Candes, T. Tao. Near-optimal signal recovery from random projections: universal encoding strategies?. IEEE Trans. Inform. Theory, 52, 5406-5425(2006).

    [41] E. J. Candes, J. K. Romberg, T. Tao. Stable signal recovery from incomplete and inaccurate measurements. Commun. Pure Appl. Math., 59, 1207-1223(2006).

    [42] X. Liu, X. Wang. Fourth-order tensors with multidimensional discrete transforms(2017).

    [43] J. Barzilai, J. M. Borwein. Two-point step size gradient methods. IMA J. Numer. Anal., 8, 141-148(1988).

    [44] M. Raydan. Convergence properties of the Barzilai and Borwein gradient method(1991).

    [45] B. Lim, S. Son, H. Kim, S. Nah, K. Mu Lee. Enhanced deep residual networks for single image super-resolution. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 136-144(2017).

    [46] L. Yue, X. Miao, P. Wang, B. Zhang, X. Zhen, X. Cao. Attentional alignment networks. 29th British Machine Vision Conference, 1-14(2018).

    [47] S. Min, X. Chen, Z. Zha, F. Wu, Y. Zhang. A two-stream mutual attention network for semi-supervised biomedical segmentation with noisy labels. Proceedings of the AAAI Conference on Artificial Intelligence, 4578-4585(2019).

    [48] Y. Li, Z. Xiao, X. Zhen, X. Cao. Attentional information fusion networks for cross-scene power line detection. IEEE Geosci. Remote Sens. Lett., 16, 1635-1639(2019).

    [49] Y. Huang, X. Cao, X. Zhen, J. Han. Attentive temporal pyramid network for dynamic scene classification. Proceedings of the AAAI Conference on Artificial Intelligence, 8497-8504(2019).

    [50] D. P. Kingma, J. Ba. Adam: a method for stochastic optimization(2014).

    [51] S. H. Chan, R. Khoshabeh, K. B. Gibson, P. E. Gill, T. Q. Nguyen. An augmented Lagrangian method for video restoration. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 941-944(2011).

    [52] https://drive.google.com/file/d/1PGav_1Yju2lfXlXQsxnvkMPu_j_EWm0B/view. https://drive.google.com/file/d/1PGav_1Yju2lfXlXQsxnvkMPu_j_EWm0B/view

    [53] https://www.pexels.com/video/blue-jellyfish-856882/. https://www.pexels.com/video/blue-jellyfish-856882/

    [54] https://www.pexels.com/video/a-person-using-a-smart-cellular-phone-3143531/. https://www.pexels.com/video/a-person-using-a-smart-cellular-phone-3143531/

    [55] H. Yu, S. Winkler. Image complexity and spatial information. Fifth International Workshop on Quality of Multimedia Experience (QoMEX), 12-17(2013).

    [56] C. Yang, D. Qi, F. Cao, Y. He, J. Yao, P. Ding, X. Ouyang, Y. Yu, T. Jia, S. Xu. Single-shot receive-only ultrafast electro-optical deflection imaging. Phys. Rev. Appl., 13, 024001(2020).

    [57] C. Yang, F. Cao, D. Qi, Y. He, P. Ding, J. Yao, T. Jia, Z. Sun, S. Zhang. Hyperspectrally compressed ultrafast photography. Phys. Rev. Lett., 124, 023902(2020).

    [58] M. Iliadis, L. Spinoulas, A. K. Katsaggelos. Deep fully-connected networks for video compressive sensing. Digit. Signal Process., 72, 9-18(2018).

    [59] K. Kulkarni, S. Lohit, P. Turaga, R. Kerviche, A. Ashok. ReconNet: non-iterative reconstruction of images from compressively sensed measurements. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 449-458(2016).

    [60] H. Yao, F. Dai, S. Zhang, Y. Zhang, Q. Tian, C. Xu. DR2-Net: deep residual reconstruction network for image compressive sensing. Neurocomputing, 359, 483-493(2019).

    [61] D. Gedalin, Y. Oiknine, A. Stern. DeepCubeNet: reconstruction of spectrally compressive sensed hyperspectral images with deep neural networks. Opt. Express, 27, 35811-35822(2019).

    [62] J. Nocedal, S. Wright. Numerical Optimization(2006).

    Chengshuai Yang, Yunhua Yao, Chengzhi Jin, Dalong Qi, Fengyan Cao, Yilin He, Jiali Yao, Pengpeng Ding, Liang Gao, Tianqing Jia, Jinyang Liang, Zhenrong Sun, Shian Zhang. High-fidelity image reconstruction for compressed ultrafast photography via an augmented-Lagrangian and deep-learning hybrid algorithm[J]. Photonics Research, 2021, 9(2): B30
    Download Citation