
- Photonics Research
- Vol. 13, Issue 2, 488 (2025)
Abstract
1. INTRODUCTION
Large-volume photoacoustic microscopy has garnered significant attention within the biomedical photonics research community. This is due to its capability to visualize the intricate optical absorption properties of biological tissue with both high contrast and resolution [1–3]. In previous work, we introduced a flexible depth-of-focus (DOF), depth-invariant resolution photoacoustic microscopy (FDIR-PAM), employing non-diffraction Airy beams to facilitate large-volume imaging while upholding exceptional resolution [4]. Despite the evident superiority of FDIR-PAM, two inherent limitations persist, constraining its widespread adoption as a favored imaging modality.
The initial challenge confronting FDIR-PAM imaging pertains to the trade-off between spatial resolution and imaging speed [5]. Advancing clinical translation necessitates the imperative of rapid screening and repeatable monitoring in optical biopsy for clinical disease evaluation. While the FDIR-PAM imaging technique showcases notable advantages in delineating intricate histopathological biomarkers, the pace of imaging across expansive fields of interest remains sluggish due to the reliance on high-magnification objective lenses and traditional full-sampling point-by-point mechanical scanning [5]. This is particularly evident without compromising spatial resolution and the demand for large-volume imaging. Rapid optical biopsy and repeatable monitoring serve as pivotal measures for guiding PAM-assisted ablation and assessing treatment efficacy [6–8]. Conforming to Nyquist sampling theory, ensuring high spatial resolution in PAM images hinges on maintaining a full sampling scan step size typically not exceeding half of the anticipated spatial resolution [5]. Besides sophisticated and costly hardware methods [9–12], there is a growing preference for straightforward and economical software solutions that streamline system complexity. In modern high-speed microscopy imaging, a favored approach for enhancing imaging speed involves reducing the number of sampled points through sparse sampling [13]. Consequently, sparse sampling presents an indispensable compromise to improve imaging speed. The task of single-image super-resolution endeavors to reconstruct high-resolution images from sparsely sampled data, offering a significant improvement in imaging speed while preserving detection accuracy. In recent years, with the ascendancy of deep learning and deep image prior methods [14,15], image super-resolution reconstruction techniques tailored for under-sampled rapid PAM imaging have attained the pinnacle of performance without increasing system complexity. A secondary challenge that has plagued FDIR-PAM since it was proposed is the artifact-removed reconstruction task. In the realm of microscopic imaging, non-diffraction beams such as Airy beams have shown promise in effectively extending the DoF. However, despite the theoretical superiority of these non-diffracted beam imaging systems, practical implementation remains constrained. The primary hurdle arises from the side-lobes of non-diffraction beams, which induce a reduction in image contrast [16,17], particularly evident under low signal-to-noise ratios or in the presence of speckle and scattering, thus posing a severe challenge to traditional deconvolution algorithms [18]. Consequently, PAM systems employing non-diffraction beams often fail to yield superior contrast compared to Gaussian beams at the focal plane. In recent years, the advent of deep learning techniques has heralded new avenues to address this issue. Several studies have employed improved generative adversarial networks (GANs) to eliminate artifacts stemming from non-diffraction beam side-lobes in microscope images [19,20], with experimental findings demonstrating the superiority of deep-learning-based methods over traditional deconvolution approaches. While sparse reconstruction and artifact suppression tasks have been extensively investigated, they have conventionally been treated as distinct endeavors. In the realm of practical PAM for optical biopsy applications, images captured through non-diffraction Airy beam excitation and sparse scanning typically exhibit diminished spatial resolution and conspicuous side-lobe artifacts. Consequently, the synergistic execution of super-resolution reconstruction and side-lobe artifact removal emerges as paramount for achieving rapid, large-volume, and high-resolution PAM simultaneously.
In this study, we developed a hybrid hardware-software method that integrates Airy beam excitation and sparse sampling techniques to achieve rapid, large-volume imaging. This approach, termed multitask learning-powered large-volume, rapid photoacoustic microscopy with Airy beam excitation and sparse sampling (ML-LR-PAM), leverages the system’s known point spread function to simultaneously achieve super-resolution image reconstruction and remove Airy beam artifacts. These tasks are achieved through multitask learning, employing scaled dot-product attention mechanism (SDAM) Wasserstein-based CycleGAN (SW-CycleGAN). Experimental results confirm the efficacy of this method, facilitating the widespread adoption of Airy beam-sparse sampling combined with SW-CycleGAN in biomedical characterization. This modality offers rapid, large-volume imaging with high resolution and artifact-free performance. While existing techniques such as sparse sampling reconstruction and side-lobe suppression have demonstrated excellent results [5,21], our multitask learning approach provides a unique solution to the challenges of high-speed, extended DoF PAM, overcoming the trade-off between spatial resolution and imaging speed. Its key advantage lies in the integration of multiple tasks into a unified model, making it a powerful tool for large-volume, rapid imaging applications in fields like biological monitoring and functional imaging. This approach has the potential to complement and enhance existing methods, particularly in scenarios where both high speed and large DoF are essential. It is especially valuable for clinical applications, such as disease monitoring and treatment evaluation, where both spatial and temporal resolutions are critical for improving diagnostic accuracy and supporting real-time decision-making.
Sign up for Photonics Research TOC. Get the latest issue of Photonics Research delivered right to you!Sign up now
2. METHODS
A. Experimental Setup
Figure 1(a) illustrates the configuration of the experimental setup. The setup architecture has been extensively described in our previous work, with recent minor adjustments made to certain optical path hardware details [4]. Here, we provide a concise outline. The Airy beam FDIR-PAM entails the utilization of a nanosecond pulsed laser (Poplar-532-10B, Huaray; wavelength 532 nm). A spatial light modulator (SLM, UPOLabs, HDSLM80R) operating in diffractive mode is employed to generate Airy beam illumination through the encoding of a flexible cubic phase pattern for precise phase modulation and laser intensity control. Subsequently, the modulated beam undergoes focusing and Fourier transformation by passing through the objective lens (
Figure 1.ML-LR-PAM system. (a) ML-LR-PAM microscopy system with Airy beam excitation. (b) Phase pattern, Airy beam obtained from the focal plane of the objective lens, and Airy beam profile in deep propagation direction. (c) Design of two-dimensional sparse-sampling raster scanning and multitask SW-CycleGAN-based super-resolution reconstruction combined with Airy beam artifact removal. (d) Airy beam PSF at
In the process of shaping the long focal length beam, the non-diffracting Airy beam is generated through phase modulation of a Gaussian beam using a phase mask and Fourier transformation with a Fourier lens. This technique enhances imaging resolution and DoF, providing significant advantages for biomedical imaging applications, particularly in capturing detailed structural and functional information. A high-resolution CMOS camera (DS-CFM300-H) was positioned at the rear focal plane of the objective lens to observe the evolution of the beam field distribution and the depth propagation direction of the non-diffused optical profile, as depicted in Fig. 1(b). The represented maps illustrate the phase pattern mask of the Airy beam, the sectional diagram captured at the focal plane of the objective lens, and the depth propagation direction of the Airy beam. To ensure image quality of high spatial resolution, full-sampling scanning adheres to the Nyquist sampling theorem, dictating that the scanning step size in PAM should be less than half of the lateral resolution of the system. Typically, the region of interest spans approximately
B. SW-CycleGAN-Based Multitask Network
The architecture of the SW-CycleGAN utilized in this study is illustrated in Fig. 2 and is derived from an enhanced CycleGAN model. Notably, CycleGAN offers a remarkable advantage due to its minimal requirement for paired training samples [22,23], substantially broadening its applicability. Specifically, CycleGAN streamlines data preparation by introducing two distinct sets of images, such as one containing an Airy beam image and another containing a Gaussian beam image without necessitating direct pairwise mapping. This approach opens up the potential for multiple image conversion tasks. A fundamental mechanism of CycleGAN is the cyclic consistency loss, which ensures the preservation of core content during image transformation. Consequently, the converted image not only maintains visual similarity to the original image but also achieves the desired shift in style or texture while preserving the image’s shape and structural features. This aspect is crucial for ensuring the quality of conversion, particularly when applied to microscopy image processing. These combined advantages render CycleGAN an ideal choice for Airy beam side-lobe removal, especially in scenarios where extensive data acquisition is unfeasible. It not only simplifies the complexity of data collection but also guarantees the quality and efficiency of the conversion process by preserving the fundamental structure and features of the image. However, CycleGAN is designed for the broader field of image processing and is not specifically tailored for the image super-resolution and Airy beam artifact removal required in this study. Therefore, an improvement was introduced based on CycleGAN to achieve the specified objectives, leading to the development of the improved network known as SW-CycleGAN. Specifically, a scaled dot-product attention mechanism (SDAM) [24] was integrated into the network, and WGAN’s W-loss [25] was utilized to optimize against GAN loss, as depicted in Fig. 2(b).
Figure 2.SW-CycleGAN principle and deep neural network architectures. (a) Unsupervised training is performed with blood vessel and microsphere simulation data. Forward and reverse generator transformations
Self-attention mechanisms play a pivotal role in enabling models to discern and leverage intricate interrelations among different regions within an image. In the realm of vascular image processing, this capability empowers the model to concentrate on specific nuances and intricacies of the blood vessel, particularly those areas demanding meticulous attention during artifact removal and reconstruction. By enhancing the model’s comprehension of vascular architecture, especially amidst the presence of side-lobe artifacts, self-attention facilitates the model in discriminating between artifacts and authentic vascular structures, thereby facilitating more precise artifact removal and reconstruction. SDAM, a core calculation method of a typical self-attention mechanism in deep learning, holds significant utility in image processing domains. Its algorithmic flow, delineated in Fig. 2(b), involves the linear transformation of the input matrix X, represented as a series of “words”, into matrices Query (Q), Key (K), and Value (V). Herein, Q and K play pivotal roles in establishing semantic associations among the “words”, with the ensuing distance matrix reflecting these semantic linkages, computed through the core self-attention algorithm. The resultant output matrix, embodying global correlations, arises from the matrix multiplication of the distance matrix and matrix V. Wasserstein distance, also known as “bulldozer distance”, has showcased its efficacy across numerous studies since the inception of W-GAN [26,27]. Compared to conventional GAN loss, Wasserstein distance mitigates mode collapse during training by providing more continuous and useful gradients, which helps the generator learn a more accurate image transformation. When used in CycleGAN, Wasserstein distance enhances stability, promotes diversity in the generated images, and improves their overall quality. Additionally, it enables smoother and more realistic image transformations between domains. Hence, Wasserstein distance supplants the traditional GAN loss within CycleGAN to bolster the performance of the network proposed in this study.
C. Experimental Verification Based on Simulation Data
To comprehensively assess the performance of the SW-CycleGAN network proposed herein, we introduce benchmark methods for comparison, namely, the sparse image recovery networks FD U-net and Res U-net, alongside the renowned Richardson-Lucy (RL) algorithm. Previous experiments have proven that, when dealing with sparse image restoration and Airy beam artifact removal, prioritizing artifact removal can improve image quality more effectively. Accordingly, in this study, we employ two control methods, namely, RL + FD Unet and RL + Res U-net, which first address Airy beam side-lobe artifact removal follow by super-resolution reconstruction. Simultaneously leveraging three Airy beam images with varying scale factor parameters (
Figure 3.Comparison of SW-CycleGAN and other methods from mouse brain microvascular dataset [5]. (a) The first column (LR) represents
Drawing on the unique effects of CycleGAN, SW-CycleGAN, trained for the two tasks of image super-resolution reconstruction and Airy beam side-lobe artifact removal, can also perform artifact removal independently. With this in mind, in order to further verify the artifact removal effect of SW-CycleGAN, a comparative experiment was conducted. First, the mouse blood vessel image with Gaussian beam imaging was generated into an Airy-beam-based PAM image (
Figure 4.Comparison verification of the PAM vascular structure maps of Airy beam side-lobes artifact removal by SW-CycleGAN and RL methods [5]. (a) Comparison of the results of SW-CycleGAN and RL methods on simulated vascular data. (b), (c) Quantitative analysis profiles in the simulation blood vessel maps corresponding to the blue and green dotted lines in (a) (see
D. Experimental Verification Based on System Acquisition Data
To validate the effectiveness of the SW-CycleGAN in mitigating Airy beam side-lobes and enhancing image quality, particularly in comparison to conventional separate approaches, simulated vascular data were initially used. However, due to variations in system configurations, experimental settings, and inherent uncertainties, there may be discrepancies between simulated and real-world data. The performance of the proposed method has been further validated in various complex environments, including leaf skeletons, mouse cerebral vasculature, and zebrafish adult pigmented stripes [Figs. 5(a)–5(c)]. All data were acquired using an Airy-beam-based photoacoustic microscopy (PAM) system with a step size of
Figure 5.The results from real data collected using ML-RL-PAM in various complex environments, including leaf skeletons, mouse cerebral vasculature, and zebrafish adult pigmented stripes, are shown. (a)–(c) depict Airy-beam-based PAM imaging of leaf skeletons, mouse cerebral vasculature, and zebrafish adult pigmented stripes, respectively. The first two columns compare imaging results using Gaussian and Airy beams. The third to sixth columns show results for LR Airy, Airy plus RL + FD Unet, Airy plus SW-CycleGAN, and HR, Airy plus SW-CycleGAN. LR denotes low resolution, and HR denotes high resolution. (d) presents the quantitative analysis corresponding to the white dashed lines L1–L3 in (a)–(c) (see
To further validate these results and highlight the impact of the improved resolution and signal-to-background ratio (SBR), we conducted a quantitative analysis [Fig. 5(d)] on the regions marked by the white dashed lines of L1–L3. The presence of a smaller peak, higher SBR, and improved resolution, as indicated by the blue and green lines in Fig. 5(d), demonstrates the superiority of the Airy plus SW-CycleGAN and HR, Airy plus SW-CycleGAN methods. These approaches effectively remove side-lobe artifacts and restore sparse images. Overall, these results clearly show that the SW-CycleGAN method significantly enhances the SBR and image resolution.
Finally, to further validate the generalization of the proposed method, its performance was evaluated and compared with high-quality co-registered Gaussian beam images obtained from the physiopathological tumor microenvironment of mouse skin cancers (Fig. 6), which depict vascular structures and blood flow information [7]. Airy beam PSF images with a scale factor value (
Figure 6.Verification of the generalization of the proposed method using the physiopathological tumor microenvironment of mouse skin cancers. (a), (b) show the characterization of vascular structures and blood flow information in the melanoma and basal cell carcinoma models, respectively. The second through fifth columns display the corresponding results for the yellow boxes highlighted in the first column, including LR Airy (
3. DISCUSSION AND CONCLUSIONS
In this study, we developed a novel approach called multitask learning-powered large-volume, rapid photoacoustic microscopy with Airy beams (ML-LR-PAM), which utilizes SW-CycleGAN to drive Airy beam-sparse sampling photoacoustic microscopy (PAM). This method significantly advances the field of fast, high-volume imaging by combining hardware and deep learning techniques to achieve artifact removal and image super-resolution reconstruction. The Fourier transform of the Airy wave solution is given by a Gaussian beam with an additional cubic phase term [28]. To generate various Airy beam excitation, the pupil phase can be manipulated by imposing a cubic phase with different scale factors onto a Gaussian laser mode. This can be achieved, for example, using a spatial light modulator (SLM). Notably, SW-CycleGAN has been optimized by incorporating self-attention mechanisms and Wasserstein distances into the original CycleGAN network for effective image style migration. To validate the performance of SW-CycleGAN in removing Airy beam artifacts and improving image quality, extensive comparisons were made with traditional RL methods using simulated data sets. Remarkably, SW-CycleGAN outperformed the traditional methods even when physical parameters were unknown. Furthermore, the efficacy of SW-CycleGAN was verified on experimental images obtained from an Airy beam-sparse sampling PAM system. We anticipate that ML-LR-PAM will serve as a convenient and powerful tool for rapid, large-volume imaging and analysis, providing accurate image quality. Compared to traditional hardware methods, the multitask learning approach offers significant advantages in mitigating side-lobe effects. Conventional hardware solutions, such as spatial filtering, rely on fixed adjustments that may not adapt well to variations in imaging conditions, leading to suboptimal image quality. In contrast, the multitask learning method dynamically optimizes imaging parameters in real-time, effectively addressing side-lobe issues based on the specific imaging scenario. Additionally, this approach achieves super-resolution enhancement through software alone, eliminating the need for costly hardware upgrades and resolving the trade-off between image resolution and imaging speed. Overall, the multitask learning method provides greater flexibility, adaptability, and cost-effectiveness, making it suitable for dynamic environments and accessible to laboratories with limited resources. The most compelling application of the proposed method is in the clinical monitoring of brain diseases and atherosclerotic plaques. These conditions involve complex morphological changes that affect imaging sensitivity, particularly in deeper tissue layers. By enhancing image resolution and minimizing side-lobe artifacts, this advanced imaging technique enables high-quality, deep tissue imaging, improving diagnostic accuracy and monitoring disease progression. It allows for rapid, repeatable assessments, which are crucial for evaluating treatment efficacy and making timely clinical decisions, ultimately improving patient outcomes and supporting longitudinal studies. In the future, to achieve greater DoF and ultra-high resolution in pathological microscopic imaging with high numerical apertures, strategies such as Airy-beam-based multi-focus image fusion can be employed. This approach extends the DoF while maintaining focus across uneven tissue surfaces. Additionally, deep learning techniques like the SW-CycleGAN algorithm can enhance image quality and resolution, which is crucial for both clinical and high-throughput research applications. By using non-diffracting Airy beams, the DoF is extended without the need for axial refocusing, enabling faster and deeper imaging. The ML-LR-PAM technique enhances image resolution and reduces artifacts through efficient sparse data acquisition combined with the SW-CycleGAN algorithm. This approach overcomes the mechanical limitations and response time delays associated with traditional electrically tunable lenses (ETLs), allowing for high-speed imaging even at MHz laser repetition rates. Overall, the method improves imaging efficiency, depth capability, and image quality, offering greater versatility and compatibility across different imaging systems. Moreover, the superiority of ML-LR-PAM opens up possibilities for its integration with other imaging modalities, such as fluorescence microscopy, where real-time, large-volume feedback and high image quality are crucial.
In conclusion, ML-LR-PAM has been developed based on a combination of software and hardware solutions. The integration of Airy beam excitation and sparse sampling, coupled with SW-CycleGAN-based deep learning, facilitates rapid and high-volume imaging. By simultaneously addressing out-of-focus and under-sampled image resolution issues through this combined approach, we demonstrate improvements in image quality, including enhanced resolution and suppression of side-lobe artifacts. This new technique holds promise for applications in large-field volumetric imaging, rapid optical biopsy, as well as repeatable monitoring to guide PAM-assisted ablation procedures and assess treatment outcomes.
References
[23] H. Hakimnejad, Z. Azimifar, M. S. Nazemi. Unsupervised photoacoustic tomography image reconstruction from limited-view unpaired data using an improved CycleGAN. 28th International Computer Conference, Computer Society of Iran (CSICC), 1-6(2023).
[24] A. Vaswani, N. Shazeer, N. Parmar. Attention is all you need. Proceedings of the 31st International Conference on Neural Information Processing Systems, 6000-6010(2017).
[25] M. Arjovsky, S. Chintala, L. Bottou. Wasserstein generative adversarial networks. International Conference on Machine Learning (ICML), 214-223(2017).

Set citation alerts for the article
Please enter your email address