
- Photonics Research
- Vol. 13, Issue 2, 511 (2025)
Abstract
1. INTRODUCTION
Hyperspectral imaging is employed to capture multi-band spectral images by examining the reflection or radiation data from an object or scene across successive wavelengths of light. With its capacity for high spatial and spectral resolution, hyperspectral imaging has great potential in the applications of biology and medicine [1–3], agriculture and forestry [4–6], oceans and astronomy [7,8], military and defense [9], and art and cultural relics [10]. Traditional hyperspectral imaging systems acquire spectral data through diverse techniques, such as whiskbroom, pushbroom, and wavelength scanning [11–13]. Although yielding accurate spectral components, these systems sacrifice imaging time or space, resulting in the inability to capture dynamic scenes in real time. Thus, researchers have conceived a range of snapshot hyperspectral imaging (SHI) systems to achieve real-time wide-field imaging spectrometers.
SHI consists of an optical hardware encoder and a software decoder. Based on their encoding strategies, contemporary SHI systems can be classified into amplitude-encoding [14–17] and phase-encoding [18–22] categories. The amplitude-encoding SHI systems, typical examples like the coded aperture snapshot spectral imaging (CASSI) system and its variants, consist of front optics, a pseudorandom binary coded aperture, relay lenses, dispersive elements (e.g., grating or prism), and a focal plane array detector [14]. Among them, the pseudorandom binary pattern has a 50%-transmission ratio and is placed at the effective field stop of an imaging system [14,23,24]. While successfully retrieving spectral images through compressed measurements, they fall short in optical throughput and a bulky system [12,25]. In contrast, the phase-encoding methods manipulate the phase of incident light through a custom-designed ultrathin diffractive lens, which yields a coded diffractive image with a spectrum separation [18,21,26]. The phase modulation element in SHI typically involves a diffractive optical element (DOE) [20,27–30], metasurfaces [31,32], and other nanomaterials [33–35]. Among these components, DOE-based SHI stands out due to its simple-to-manufacture, compact, and ultrathin structure, and remarkable dispersion capabilities.
The hardware encoder of DOE-based SHI introduces specific phase delays for different wavelengths by customizing the height map of the DOE. Different patterns can be used to design the height map to achieve phase modulation, such as Fresnel, cubic, multi-focal, diffractive achromat, hybrid diffractive-refractive, and square cubic [36,37]. Related research works have progressively addressed problems such as point spread function (PSF) inhomogeneity at different spectral bands, chromatic aberrations, and mismatches between design and fabrication [18,36,38–40]. Still, non-negligible gaps exist between the ideal physical design and the actual practice of DOE. For instance, limitations in stabilized lithography restrict the quantization levels of DOE height maps to eight levels [39]. Furthermore, when the incident wavelength deviates from the design wavelength, the diffraction efficiency of DOE is significantly reduced [40,41]. In terms of the manufacturing of DOE, the existing deployed DOE configuration in deep optics includes a single DOE [28,39] (conducting imaging and phase modulation) and a DOE (implementing phase modulation) coupled with a simple lens (dedicated to imaging) [42]. As a result, the single DOE configuration is preferred to make the system compact.
Sign up for Photonics Research TOC. Get the latest issue of Photonics Research delivered right to you!Sign up now
The software decoder in the DOE-based SHI is used to solve ill-posed inverse problems to retrieve high-fidelity spectral data cubes from the captured single measurement and various reconstruction methods include analytical-modeling-based methods [18,43] and deep-learning-based methods [12,44–46]. Analytical-modeling-based algorithms adopt handcrafted priors, e.g., total variation and non-local self-similarity, to filter the solution to the desired signal space, which relies on a long iteration time and empirical parameter tuning to get optimum results [43]. The deep-learning-based methods include U-net [47] and its variants [48], RNN [49], Transformer [50], LSTM [17], and Mamba [51]. They have been proposed to achieve end-to-end high-fidelity spectral image reconstruction. However, the neglect of the low-level limitations associated with lithography techniques and the diffraction efficiencies of DOE presents practical challenges in real data reconstruction. These challenges include alignment errors and stray light from the transition area of the ring structure.
Liquid-crystal-on-silicon spatial light modulators (LCoS-SLMs) can dynamically simulate the phase modulation of DOE to generate specific phase delay or optical range difference for different wavelengths by controlling the state of liquid crystal pixels on its surface. The LCoS-SLM supports quantization levels up to 256, allowing floating-point gray-level design schemes [52]. These features can mitigate the low-accuracy issues arising from the height limitation of less than 16th levels in the manufactured DOE. Moreover, the LCoS-SLM can dynamically load multiple DOE simulation patterns with different design wavelengths at a frame rate of 180 Hz. This feature enables higher diffraction efficiency and improves spectral recovery accuracy, meanwhile allowing for portable alteration of the focal length for imaging. Furthermore, the repeatable refresh ability of LCoS-SLM dramatically improves the efficiency and reduces the cost required for manufacturing DOEs, facilitating efficient real-time model debugging in the field. With all these resilient merits, LCoS-SLMs have been employed for phase modulation in achromatic imaging [52], super-resolution imaging [53], ultrafast imaging [54], extended depth of field enhancement [55], and computational holographic imaging [56,57], but little research has been conducted on DOE-based SHI.
To bridge this gap, we propose a new lensless efficient SHI (LESHI) aided by an LCoS-SLM. LESHI utilizes an LCoS-SLM to replace a single fabricated DOE as the hardware encoder, simultaneously realizing imaging and phase modulation. For the software decoding process, we developed a learning algorithm based on the ResU-net architecture, taking into account the sensor’s response function and the diffraction efficiency of the DOE. Using the developed algorithm, high-resolution 31-channel spectral images can be reconstructed from the captured three-channel red-green-blue (RGB) image. To improve diffraction efficiency with a single simulated DOE and explore the switch ability of LCoS-SLM, we propose a distributed diffractive optics (DDO) model by dynamically controlling the light phase. Thus, multiple different phase modulation patterns can be loaded onto the LCoS-SLM, resulting in high reconstruction accuracy and high diffraction efficiency throughout the full visible spectral range (400–700 nm). Furthermore, the LESHI system can realize the modification of the focal length and the field of view without adding other optical components, demonstrating the feature of tunability. The entire imaging system adopts an end-to-end approach to modeling, training, and optimizing, ensuring a high level of integration and coordination to achieve optimal performance. In a nutshell, LESHI not only solves the errors between high-level DOE design and fabrication, as well as the optical alignment difficulties during assembly, but also leverages multiple simulated DOEs for imaging in different spectral bands, thus improving the diffraction efficiency and spectral reconstruction accuracy in the entire visible spectrum. At the same time, it enables convenient modification of focus lengths and real-time on-site debugging, greatly diminishing the production cost and time of DOE. Extensive comprehensive simulations and real-world hardware experiments validate the superior performance of the system.
2. RESULTS
A. Operating Principle of LESHI
The schematic of the LESHI system is shown in Fig. 1. A light source (CIE standard illuminant D65, Datacolor Tru-Vue light booth) is used to illuminate the object. The reflected light of the sample passes through the polarizer (GCL-050003), is reflected by a beam splitter (GCC-M402103), and impinges on the LCoS-SLM (FSLM-2K39-P02, 8-bit grayscale level of 256 steps, 180-Hz refresh rate) loaded with optimized DOE patterns. Since the liquid crystal layer has different refractive indices for different wavelengths of the spectrum [52,53], it can produce different phase delays for the entire spectrum like DOE, splitting the continuous hyperspectral data cube. Thus, when a light wave passes through the liquid crystal layer of the LCoS-SLM, the modulation of each pixel causes the phase of the light wave to change. Finally, the phase-modulated light reflected from the LCoS-SLM transmits the beam splitter and is recorded by a color CMOS camera (ME2P-1230-23U3C, which contains a Bayer filter).
Figure 1.Schematic of the lensless efficient snapshot hyperspectral imaging (LESHI) system. LCoS-SLM, liquid crystal on silicon-based spatial light modulator. LESHI comprises hardware-based diffractive imaging and software-based hyperspectral reconstruction algorithms. The diffractive imaging component includes an LCoS-SLM, a polarizer, a beam splitter, and a color CMOS camera. The hyperspectral reconstruction algorithm employs a ResU-net to decode the spectral information.
The working principle of LESHI is illustrated in Fig. 2(a). In the forward propagation of the model, LESHI sequentially performs the compression of the spectral dataset into a three-channel RGB snapshot, the image reconstruction of the 31-channel spectral cube from the snapshot, and the calculation of the loss function between the reconstruction results and the ground truth. In backward propagation of the model, the model optimizes its variables (e.g., the values of each pixel of the phase modulation pattern and parameters in the neural networks) by minimizing the loss function using the gradient descent methods. Notably, we take the diffraction efficiencies into account in the model, which is missed in existing learning methods [18,22,27–30]. Besides, a rotationally symmetric design [28] was used to reduce the computational complexity of the phase delay pattern.
Figure 2.Working principle of LESHI. (a) Pipeline of LESHI.
Figure 2(b) shows the imaging process of the LESHI system with a representative PSF (details in Fig. 7 of Appendix A). A spectral dataset in the visible band with 31 channels and 10-nm spectral resolution convolves the PSF and yields the snapshot. The forward mode of LESHI is expressed as
To improve the diffraction efficiency and account for noise effects on the quality of the reconstructed images, the ideal PSF without diffraction efficiency can be transformed into the first-order degenerate PSF (D-PSF). The D-PSF provides a more accurate representation of the imaging system and is expressed as
After the data acquisition, the captured images are used as input to a customized ResU-net [Fig. 2(d), details in Appendix D], which can retrieve 31-channel spectral images. The loss function
B. Validation of the LESHI Model
To verify the LESHI model, we conducted a comprehensive simulation using the ICVL dataset [58], and it consists of 201 spectral scenes, randomly distributed for training (160 scenes), validation (21 scenes), and testing (20 scenes). To match the hyperparameters of the model, each scene with the size of
We conducted a simulation to generate the PSFs by setting the parameters of diffraction imaging and camera spectral response functions. Figure 3(a) shows the ground truth in the test set. We systematically simulated the phase modulation patterns. Specifically, LESHI employs an end-to-end optimization approach to generate a phase modulation pattern loaded onto the LCoS-SLM. The resulting pattern, as shown in Fig. 3(b), is a grayscale pattern with 8-bit and 256-level precision. By adjusting the gray level of each pixel on the liquid crystal, the phase delay magnitudes can be modified across various spectra. The inconsistent diffracted spot sizes in different channels due to the varying degree of phase modulation of the spectra by the system result in a white haze covering the captured RGB images. Figure 3(c) displays the simulated captured images by the color CMOS camera, providing a visual representation of this white haze phenomenon. The customized ResU-net network takes the snapshot as input and reconstructs 31-channel hyperspectral images. Figure 3(d) visually represents the reconstructed spectral image in RGB. Figure 3(e) shows the reconstructed 31-channel spectral images using a single LCoS-SLM loaded with a single simulated DOE, colored with the RGB values of the corresponding wavelengths. In addition, we validated the diffraction efficiency effectiveness of the DDO model.
Figure 3.Validation of LESHI model. (a) Ground truth from the ICVL dataset. (b) The trained simulated DOE pattern loaded on the LCoS-SLM. (c) RGB image generated by the LESHI model with a single DOE pattern. (d) Reconstructed result of (c). (e) Reconstructed hyperspectral images using LESHI model with a single DOE pattern. (f) Ground truth and reconstructed values of the spectral radiance curves for local area “1” marked in (a). (g) Same as (f) but for local area “2”. (h) Diffraction efficiency as a function of wavelength, using single DOE pattern (LCoS-S) and multiple DOE patterns (LCoS-D) in the LESHI model. The table shows the relative diffraction efficiency gain (RDEG) of LCoS-D compared to LCoS-S at three different bands (400–500 nm, 500–600 nm, 600–700 nm).
To verify the accuracy of the models for spectral reconstruction, we compared the average spectral radiance of the reconstructed and true spectral images. Two
C. Quantification of the System’s Performance of LESHI
Upon the LESHI model, we built the LESHI system. To characterize the spatial resolution of the LESHI system, the resolution test chart of ISO12233 (3nh, SIQ) was used. The distance between the resolution test chart and the LCoS-SLM was 1.2 m, the screen ratio of the test chart was 4:3, and the focal length of LESHI was set to 50 mm. Moreover, we mitigate the effect of multiple orders by adding a polarizer in front of the LCoS-SLM and increasing its phase quantization level to the highest level (256-level). Figure 4(a) shows the reconstructed resolution test chart, which preserves lots of low- and high-frequency information about the chart. Figures 4(b) and 4(c) plot the reconstructed intensity profiles of the two groups of lines at different locations [marked by light orange and teal boxes in Fig. 4(a)] on the resolution target against the ground truth intensity profiles. With the Rayleigh resolution criterion, the effective spatial resolution of the LESHI system was characterized as 15.74 μm.
Figure 4.Characterization of the LESHI system performance. (a) Reconstructed image of ISO12233 test chart. (b) Spatial line profiles of two regions on the test chart, highlighted in light orange and teal boxes at the location of label 1 in (a). (c) Spatial line profiles of two regions on the test chart, highlighted in light blue and teal boxes at the location of label 2 in (a). (d) Measurement of the LEHSI system. (e) Reconstruction result of (c) in RGB format. (f) Root mean square error (RMSE) and maximum error of reconstructed image and measurement by the CS-2000 spectrometer at six local regions [marked by white boxes in (c)]. (g) Reconstruction radiance curves of six local regions [marked by white boxes in (c)] as a function of wavelength. Ground truth is obtained by the CS-2000 spectrometer. (h) Seven representative reconstructed spectral channels of (d).
The spectral resolution of the LESHI system was evaluated by comparing the spectral values obtained from the spectrometer capturing a ColorChecker Digital SG with the reconstruction results. The measurement of the color calibrator using the LESHI system is shown in Fig. 4(d). Figure 4(e) shows the reconstructed 31-channel spectral composite image in RGB form. Figure 4(f) shows the root mean square error (RMSE, left
D. Demonstration of Distributed Diffractive Optical Model
To demonstrate the feasibility of applying the DDO model to LESHI, a Thorlabs’ Lab Snacks box is used as the test sample. First, we loaded the three different designed DOE patterns sequentially onto the LCoS-SLM and captured the corresponding RGB images. Second, we extracted the R, G, and B channels from the three captured images. Third, the selected R, G, and B channels with the highest diffraction efficiencies were combined. Finally, the newly synthesized RGB image was fed into the reconstruction network to retrieve 31-channel spectral images. Figure 5(a) shows the measured RGB image using a single DOE pattern (LCoS-S) and the seven representative reconstructed channels (center wavelengths at 410 nm, 450 nm, 490 nm, 530 nm, 570 nm, 630 nm, and 680 nm). Figure 5(b) is the same as Fig. 5(a) except using multiple simulated DOE patterns (LCoS-D). The comparison results show that the DDO-model-based reconstruction results are better than those of the single DOE pattern. All reconstructed spectral images generated by LCoS-S and LCoS-D are shown in
Figure 5.Demonstration of distributed diffractive optics (DDO) imaging. (a) Captured and reconstructed images based on a single simulation of DOE. (b) Captured and reconstruction images based on multiple simulated DOEs (DDO model). (c) Reconstructed values and ground truth of spectral radiance based on LCoS-S and LCoS-D models at the location of label 1 in (a). (d) Reconstructed values and ground truth of spectral radiance based on LCoS-S and LCoS-D models at the location of label 2 in (a). (e) Images and simulated diffraction efficiency (DE) of the R, G, and B channels captured by the model based on LCoS-S and LCoS-D.
To quantitively analyze the reconstruction result, we measured the spectral radiance of two local areas [
E. Application of Range Sensing via a Tunable Focal Length
The tunable and convenient focal length of the LESHI system enables it to meet the different needs of the imaging field of view and the range of the captured scene. The focal length of the LESHI system can be modified by loading DOE patterns with different focal lengths. First, we trained the patterns of DOEs with focal lengths ranging from 50 mm to 100 mm, with a step of 2 mm, and a total of 25 images. Six representative DOE patterns (with focal lengths of 50 mm, 60 mm, 70 mm, 80 mm, 90 mm, and 100 mm) are shown in Fig. 6(a). Second, each of the well-trained patterns was loaded on the LCoS-SLM, and the CMOS camera was moved to the corresponding position according to the focal length of the used pattern. Using the captured RGB images [Fig. 6(b)] under different focal lengths as the input of the well-trained neural network, the corresponding reconstructed spectral images can be retrieved with high-fidelity image quality [Fig. 6(c)]. The results show that the field of view of the scene shrinks as the focal length increases, which can be explained by the Lagrange-Helmholtz invariant (i.e., a bigger focal length gives a smaller aperture angle in image space and thus a smaller object height). Figure 6(d) shows one representative reconstructed spectral image under these focal lengths. The reconstructed spectral images under the focal length range (50–100 mm) are shown in
Figure 6.Application results for focal length modification. (a) Phase modulation patterns loaded onto LCoS-SLM with different focal lengths by end-to-end training. (b) Corresponding captured RGB images of (a). (c) Results of spectral image recovery by applying the LESHI system at different focal lengths. (d) Six representative reconstructed spectral channels corresponding to (c).
3. DISCUSSION AND CONCLUSIONS
We have developed the LESHI system based on diffractive optics via the LCoS-SLM. LESHI employs a learning-based DOE pattern loaded onto the LCoS-SLM to perform phase modulation and imaging, instead of a physically fabricated DOE. Using the customized ResU-net algorithm, we have retrieved the 31-channel spectral cube with an image resolution of
Compared to diffractive hyperspectral imaging via a fabricated DOE, the LESHI system has significant advantages in terms of spectral reconstruction accuracy, system flexibility, diffraction efficiency, and cost of fabrication. The limitation of stabilized lithography technology restricts the number of quantization levels supported by fabricated DOE to only eight. This reduction in quantization levels results in a decrease in the resolution of spectral phase modulation by DOE. Consequently, the accuracy of the reconstruction capability of the entire system is weakened. The LCoS-SLM technology offers a phase modulation level of 256 gray levels, allowing for a floating-point gray level design. This feature enables higher phase resolution, which is beneficial for optimizing and replacing fabricated DOE. The high diffraction efficiency of the fabricated DOE is challenging to maintain across the entire 400–700 nm band due to limitations in material and design wavelength. The LESHI system employs the DDO model to dynamically load multiple phase modulation patterns for different spectral bands. This implementation enhances the diffraction efficiency of imaging. Besides, the high cost of DOE fabrication significantly restricts its potential applications. By dynamically loading patterns, LCoS-SLM can save the time and cost of fabricating DOEs, improving the efficiency of real-time system debugging. In addition, the micrometer-scale level presents practical challenges when attempting to achieve pixel-level alignment for DOE. This can result in calibration errors between the idealized camera model and the actual experiment. In contrast, the pattern loaded on the LCoS-SLM has pixel-level translation, rotation, and grayscale flipping, which mitigates the difficulty of optical alignment in practical assembly.
The principle of LESHI could be extended to other DOE-based imaging modalities. The LCoS-SLM can simulate DOE based on various patterns using high-level encoding and reloadable features, thereby improving the performance and efficiency of existing fabricated-DOE-based systems such as full-spectrum computational imaging [18], high-dynamic-range imaging [30], depth-spectral imaging [27], and achromatic extended depth of field and super-resolution imaging [36]. Besides, with an ultrashort chirped pulse as a light source, LESHI could be directly applied to ultrafast imaging [59] because the reconstructed spectral frames of LESHI can be linked to time information benefiting from the chirped pulse (i.e., the wavelength changes during the duration of the pulse).
While the proposed distributed LESHI system improves the spectral imaging performance of the scene, the current model is based on the training of one dataset, which limits its ability to generalize to scenes in wide applications. In the future, the system will be comprehensively optimized by adding the required scene object information to the model training to improve the generalization ability of the model. In addition, deep unfolding networks [60] and plug-and-play mechanisms [61] will be considered to improve the flexibility of the network structure in handling different sizes of spectral cubes. Finally, the entire network model can be miniaturized by optimizing the network parameters, and the trained model can be loaded using FPGA hardware instead of GPU to improve the reconstruction speed of the spectrum.
Acknowledgment
Acknowledgment. The data table of LCoS-SLM spectral phase delay at different center wavelengths was provided by Xi’an CAS Microstar Optoelectronic Technology Co., Ltd.
APPENDIX A: DERIVATION OF THE LESHI MODEL
The traditional diffractive optical imaging model [
The PSF elucidates a mathematical model [
Assume that the complex amplitude of the point source P with wavelength
When the wavefield
The phase delay
The Fresnel diffraction occurring from
The ideal power density
Figure
According to the principle of incoherent optical imaging, the diffractive imaging process is modeled as the convolution of the original image
Figure 7.LESHI-based point spread function for 31 channels at 400–700 nm. Due to the phase delay of LCoS-SLM for different spectra, the system has different point spread functions for different bands.
Figure 8.Spectral response and modulation simulation curves of camera and LCoS-SLM. (a) Sensor spectral response curves. (b) Phase modulation curves of LCoS-SLM with different center wavelengths. (c) Diffraction efficiency of LCoS-SLM with different center wavelengths.
APPENDIX B: DEFINITION OF DIFFRACTION EFFICIENCY
Diffraction efficiency is a crucial metric for assessing the imaging capability of diffractive optical elements. It plays a significant role in determining the spectral range of spectral imaging. By measuring the diffraction efficiency, one can evaluate the effectiveness of these elements in diffraction light and produce high-quality images. This metric provides valuable insights into the performance and potential applications of diffractive optical elements in various fields, such as microscopy, spectroscopy, and remote sensing. The diffraction efficiency of a single-layer diffractive element can be expressed as follows [
The LCoS-SLM system, which is typically used for phase modulation at a single wavelength, is being utilized in this case to modulate the spectrum across the full visible spectral band, which ranges from 400 to 700 nm. Therefore, we simulate the phase modulation values of LCoS-SLM in the full spectral bands. Figure
APPENDIX C: DISTRIBUTED DIFFRACTIVE OPTICAL IMAGING
The distributed diffractive optics model employs spatio-temporal multiplexing to perform distributed imaging of the same scene at different spectrum bands by sequentially loading multiple DOEs. This system utilizes LCoS-SLM to load multiple simulated DOE patterns to realize the DDO model in batches. As shown in Fig.
DDO uses an LCoS-SLM to dynamically load the grayscale patterns of three simulated DOEs at different central wavelengths and image the same field of view. Then, the images of different simulated DOEs are extracted from the corresponding high-diffraction-efficiency bands to synthesize the final captured image. Therefore, the final image acquired on the sensor by using the distributed diffraction imaging model consists of three different parts, specifically as follows:
APPENDIX D: SPECTRAL RECONSTRUCTION NETWORK
RGB images encoded by hyperspectral cubes require an image parser to reconstruct the hyperspectral image. LESHI uses ResU-net as a computational decoder for spectral reconstruction. As shown in Fig.
APPENDIX E: INVESTIGATION OF DOE PATTERN WITH DIFFERENT LEVELS
To verify the effect of level of DOE on the accuracy of spectral reconstruction, we simulated a set of hyperspectral imaging models with different-level patterns onto LCoS-SLM. Figure
Figure 9.The effect of different levels of the simulated DOE for spectral reconstruction. Comparing the reconstruction performance for 4, 16, 64, and 256 levels, it can be concluded that the reconstruction performance gradually improves with the growth of levels.
APPENDIX F: COMPARISON OF FABRICATED DOE, SINGLE DOE PATTERN, AND MULTIPLE DOE PATTERNS IN SNAPSHOT HYPERSPECTRAL IMAGING
The proposed LESHI was verified by comparing the reconstruction results of three different hyperspectral imaging systems: fabricated DOE, LCoS-S, and LCoS-D. The fabricated DOE system is based on a fabricated DOE, the LCoS-S system uses a single phase-modulated pattern loaded on LCoS-SLM, and the LCoS-D system utilizes a DDO model loaded with multiple patterns. The quantization level of the height map of the fabricated DOE can range from zero to eight levels, depending on the actual lithography conditions. By contrast, the phase modulation pattern for both LCoS-S and LCoS-D can range from 0 to 256 levels. Figure
Figure 10.Comparison of spectral reconstruction simulations for different models. (a) Comparing the four reconstruction data results and visual effects, the diffractive optical imaging model based on LCoS-SLM can effectively improve the reconstruction performance and avoid the degradation of the reconstruction results caused by the quantized DOE. (b) Spectral radiance curves for different models. The spectral curves show that the reconstructed spectral curves of LCoS-D are closer to the ground truth values.
Figure 11.Performance comparison of hyperspectral reconstruction using fabricated DOE and simulated DOE loaded onto LCoS-SLM. (a) Comparison of PSNR for hyperspectral image reconstruction with different models. (b) Comparison of SSIM metrics for hyperspectral image reconstruction with different models. (c) Comparison of RMSE metrics for hyperspectral image reconstruction with different models. (d) Comparison of ERGAS metrics for hyperspectral image reconstruction with different models.
We conducted simulations to evaluate the root mean square error (RMSE) and the normalized absolute spectral error (ERGAS) metrics of reconstruction. Four models were used: full-precision DOE (DOE-DO), quantization-aware DOE (DOE-QDO), LCoS-S, and LCoS-D. These models were chosen based on the comparative methods outlined in Ref. [
APPENDIX G: COMPARISON OF LESHI WITH TYPICAL HYPERSPECTRAL IMAGING MODALITIES
To assess the performance of the proposed models in LESHI, we conducted simulations and compared them with representative SHI systems, namely, Fresnel lens [
PSNR, SSIM, RMSE, and ERGAS Simulation Results of Spectral Image Reconstruction Using Different Models
Encoding | PSNR ↑ | SSIM ↑ | RMSE ↓ | ERGAS ↓ |
---|---|---|---|---|
CASSI [ | 30.65 | 0.897 | 0.0356 | 20.72 |
Fresnel [ | 27.42 | 0.868 | 0.0557 | 30.16 |
DOE [ | 31.69 | 0.935 | 0.0322 | 19.93 |
LCoS-S | 33.90 | 0.960 | 0.0285 | 14.88 |
LCoS-D | 35.42 | 0.9768 | 0.0209 | 12.85 |
References
[1] G. Lu, B. Fei. Medical hyperspectral imaging: a review. J. Biomed. Opt., 19, 010901(2014).
[7] V. Marx. When microbiologists plunge into the ocean. Nat. Methods, 17, 133-136(2020).
[12] L. Huang, R. Luo, X. Liu. Spectral imaging with deep learning. Light Sci. Appl., 11, 61(2022).
[18] Y. Peng, Q. Fu, W. Heidrich. The diffractive achromat full spectrum computational imaging with diffractive optics. SIGGRAPH ASIA 2016 Virtual Reality meets Physical Reality: Modelling and Simulating Virtual Humans and Environments, 1-2(2016).
[26] J. W. Goodman. Introduction to Fourier Optics(2005).
[27] S. H. Baek, H. Ikoma, M. H. Kim. Single-shot hyperspectral-depth imaging with learned diffractive optics. IEEE/CVF International Conference on Computer Vision, 2651-2660(2021).
[30] C. A. Metzler, H. Ikoma, G. Wetzstein. Deep optics for single-shot high-dynamic-range imaging. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1375-1385(2020).
[34] A. Arbabi, A. Faraon. Advances in optical metalenses. Nat. Photonics, 17, 16-25(2023).
[35] W. J. Padilla, R. D. Averitt. Imaging with metamaterials. Nat. Rev. Phys., 4, 85-100(2022).
[39] L. Li, L. Wang, W. Song. Quantization-aware deep optics for diffractive snapshot hyperspectral imaging. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 19780-19789(2022).
[47] O. Ronneberger, P. Fischer, T. Brox. U-Net: convolutional networks for biomedical image segmentation. 18th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), 234-241(2015).
[57] D. Wang, N. N. Li, Y. L. Li. Large viewing angle holographic 3D display system based on maximum diffraction modulation. Light Adv. Manuf., 4, 195-205(2023).
[58] B. Arad, O. Ben-Shahar. Sparse recovery of hyperspectral signal from natural RGB images. 14th European Conference on Computer Vision (ECCV), 19-34(2016).
[59] J. Liang, F. Légaré, F. Calegari. Ultrafast imaging. Ultrafast Sci., 4, 0059(2024).
[60] K. Zhang, L. V. Gool, R. Timofte. Deep unfolding network for image super-resolution. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 3217-3226(2020).

Set citation alerts for the article
Please enter your email address