• Photonics Research
  • Vol. 9, Issue 4, B159 (2021)
Che Liu1、2, Wen Ming Yu1、2, Qian Ma1、2, Lianlin Li3, and Tie Jun Cui1、2、*
Author Affiliations
  • 1Institute of Electromagnetic Space, Southeast University, Nanjing 210096, China
  • 2State Key Laboratory of Millimeter Wave, Southeast University, Nanjing 210096, China
  • 3School of Electronic Engineering and Computer Sciences, Peking University, Beijing 100871, China
  • show less
    DOI: 10.1364/PRJ.416287 Cite this Article Set citation alerts
    Che Liu, Wen Ming Yu, Qian Ma, Lianlin Li, Tie Jun Cui. Intelligent coding metasurface holograms by physics-assisted unsupervised generative adversarial network[J]. Photonics Research, 2021, 9(4): B159 Copy Citation Text show less

    Abstract

    Intelligent coding metasurface is a kind of information-carrying metasurface that can manipulate electromagnetic waves and associate digital information simultaneously in a smart way. One of its widely explored applications is to develop advanced schemes of dynamic holographic imaging. By now, the controlling coding sequences of the metasurface are usually designed by performing iterative approaches, including the Gerchberg–Saxton (GS) algorithm and stochastic optimization algorithm, which set a large barrier on the deployment of the intelligent coding metasurface in many practical scenarios with strong demands on high efficiency and capability. Here, we propose an efficient non-iterative algorithm for designing intelligent coding metasurface holograms in the context of unsupervised conditional generative adversarial networks (cGANs), which is referred to as physics-driven variational auto-encoder (VAE) cGAN (VAE-cGAN). Sharply different from the conventional cGAN with a harsh requirement on a large amount of manual-marked training data, the proposed VAE-cGAN behaves in a physics-driving way and thus can fundamentally remove the difficulties in the conventional cGAN. Specifically, the physical operation mechanism between the electric-field distribution and metasurface is introduced to model the VAE decoding module of the developed VAE-cGAN. Selected simulation and experimental results have been provided to demonstrate the state-of-the-art reliability and high efficiency of our VAE-cGAN. It could be faithfully expected that smart holograms could be developed by deploying our VAE-cGAN on neural network chips, finding more valuable applications in communication, microscopy, and so on.

    1. INTRODUCTION

    Electromagnetic (EM) metasurfaces use periodic or quasi-periodic macroscopic basic units to simulate the atoms or molecules on the microscale of traditional material science. These macroscopic basic units could interact with external electric fields via resonance effects and express unique EM properties [16] to realize various novel functional devices such as cloaks [710], concentrators [11], illusion optics devices [12,13], special lenses [14,15], and diffuse reflections [16]. Recently, a series of special EM metasurfaces called coding, digital, and programmable metasurfaces [17] (‘coding metasurfaces’ for short in the remainder of this article) have gained more attention due to the unique methods for manipulating the EM waves. By encoding the phase responses of the metasurfaces as digital numbers ‘0’ or ‘1’, the EM property of each meta-unit in the coding metasurfaces could be switched in real time when controlled by a field-programmable gate array (FPGA), which allows people to design metasurfaces in digital space rather than the analog domain. Further, the digital representation of meta-units could link the EM space and digital world and has been widely used in manipulating amplitude [18,19], polarization [2022], and orbital angular momentum [2325], yielding the concept of information metasurfaces [2629] and many other novel applications [3033].

    With the development of communication technology, people need to deal with more and more information in daily life, which adds the burden of the whole society and promotes the development of artificial intelligence (AI) to help people deal with various information handling tasks such as speech recognition [3436], image recognition [3739], automatic translation [4042], and robot control [4345]. The growth of AI has also brought changes to the design of EM metasurfaces, especially coding metasurfaces. The digital representation of the coding metasurfaces makes it more convenient to put AI technology into the design of the meta-unit states [4648] and has already sprung out various interesting applications such as the smart system [49], high-resolution imager [47,50], and recognizer [51].

    Despite the wide usage of AI in coding metasurfaces, the design of meta-unit states for holographic imaging by coding metasurfaces still rests on iteration optimization algorithms like the Gerchberg–Saxton (GS) algorithm [52,53] or greedy algorithm [54]. Recently, a valuable trail has been made for using supervised deep neural networks (DNNs) to recover holograms from the generated speckles distorted by a thin diffuser [55]. However, the demand for thousands of training samples acquired by real measurements remarkably increases the cost in the usage of this supervised DNN. Here, we propose a new method based on AI or deep learning that could rapidly generate the coding pattern (metasurface hologram) of the 1 bit coding metasurface when a target holographic image is given. The deep learning structure composed of unsupervised variational auto-encoder (VAE) and conditional generative adversarial networks (cGANs) is presented in our method. The use of VAE [56], as an unsupervised algorithm, could remarkedly reduce the time consumption for the preparation of training data, and the EM propagation model described by the rigorous dyadic Green’s function (DGF) [53] is used to make the VAE structure possible while keeping the physical interpretability. We merge the Wasserstein distance [5759] [by the form of Wasserstein GAN (WGAN) [60,61]] and mean square error (MSE) in the design of loss function to make the distributions of generated holographic images and target holographic images closer. The structure of cGAN [6264] is also used to avoid the confusion of the generative network. We design a 40×40 one bit coding metasurface working at the frequency of 35 GHz to validate our intelligent method, and the simulation and experimental results both validate the efficiency and reliability of the proposed approach.

    2. THEORY AND METHOD

    A. EM Propagation Model

    (a) Sketch of our hologram system that consists of a 1 bit coding metasurface loading the hologram and a feed antenna. The distance between the holographic imaging plane and coding metasurface is 30 cm. (b) The meta-unit of the 1 bit coding metasurface. (c) The side-looking photograph of our hologram system. The interval between meta-units on the coding metasurface is 3.8 mm, and the feed antenna radiates the EM waves with frequencies from 34 GHz to 36 GHz.

    Figure 1.(a) Sketch of our hologram system that consists of a 1 bit coding metasurface loading the hologram and a feed antenna. The distance between the holographic imaging plane and coding metasurface is 30 cm. (b) The meta-unit of the 1 bit coding metasurface. (c) The side-looking photograph of our hologram system. The interval between meta-units on the coding metasurface is 3.8 mm, and the feed antenna radiates the EM waves with frequencies from 34 GHz to 36 GHz.

    We accept rigorous DGF [53] as the basic computational kernel of the forward propagation from the source currents of meta-units to the near-field EM distribution: G(r,r)=(I+k2)g(r,r),where I is a 3×3 dyadic identity matrix, r and r are source and field points, respectively, and g(r,r)=ejkR4πRis the free-space Green’s function, in which R=|rr|. The forward propagation formula could be represented as E(r)=jωμVdrG(r,r)·J(r),where E(r) and J(r) represent the electric field at the field point and the current at the source point, respectively. Owing to the discrete array form of the coding meta-units, the source currents can also be expressed in a discrete form. For convenience, we use a current element to represent a coding meta-unit and discrete the EM field into M points at the same time. Then, Eq. (3) could reduce to E(rm)=jωμn=1NG(rm,rn)·J(rn),m=1,,M,where N is the number of coding meta-units. Further, we organize the scalar components of E(rm) and J(rn) in Eq. (4) for all m and n into column vectors E and J, respectively. As we can see from Eq. (4), the relationship between E(rm) and J(rn) is linear, and, hence, they could be connected with a complex-value coefficient matrix. Then, Eq. (4) could ultimately reduce to E=W·J,where W is the coefficient matrix that links the source J and field E. We only care about the vertical polarization EM waves and the relative amplitude of the near-field EM distribution. Hence, the current vector J could be described as J=JrJφ,where Jr is a complex-value vector and represents the current part that is directly caused by the incident EM waves, which is proportional to the incident electric-field values; means element-wise multiplication; and Jφ represents the current part that is controlled by each coding meta-unit, which is a real-value vector whose elements are ‘1’ or ‘1’ to represent the phase of 0 or π, corresponding to code ‘0’ or ‘1’. Then, Eq. (5) could be rewritten as E=W·JrJφ=W·diag(Jr)·Jφ=Wr·Jφ,Wr=W·diag(Jr).Thus, the forward propagation process could be represented as a form of matrix multiplication from a real-value vector Jφ to the near-field electric-field distribution E. The design objective of VAE-cGAN is to get the current vector Jφ to generate a target near-field electric-field distribution E, whose amplitude distribution represents a target holographic image.

    B. Network Structure

    Schematic diagram of the proposed VAE-cWGAN. The generator together with the EM propagation process makes up the VAE structure. Two kinds of distance criteria (MSE and Wasserstein distance) are used to improve the imaging quality of the generator.

    Figure 2.Schematic diagram of the proposed VAE-cWGAN. The generator together with the EM propagation process makes up the VAE structure. Two kinds of distance criteria (MSE and Wasserstein distance) are used to improve the imaging quality of the generator.

    The discriminator adopts the structure of cGAN [63] and WGAN [61], which is responsible for calculating the approximate Wasserstein distance between the generated holographic image and the target holographic image. It will be trained in the adversarial process towards the generator and improve the imaging quality of the generated holographic image.

    C. Backpropagation Process

    The mainstream deep learning platform such as TensorFlow and Pytorch could not directly deal with the complex values, which will cause difficulty in the backpropagation process from loss function to Jφ. Thus, we need to derive the backpropagation partial derivative equation from the generated holographic image (the amplitude of E) to Jφ to get rid of the calculation of complex values, which is expressed as (using numerator layout) |E|Jφ=|E|1Re[E*EJφ]=|E|1Re[E*Wr],where the superscript notation −1 means taking the reciprocal element-wisely and * means taking the conjugate matrix. Then, the backpropagation process from loss function to Jφ can be calculated by real values, LossJφ=Loss|E|·|E|Jφ.

    D. Design of Loss Function

    We merge the MSE loss and Wasserstein distance as our final loss function. The MSE loss and Wasserstein distance are both criterions to indicate the distance between two distributions. Although using the Wasserstein distance evaluation function alone could also reach an equally good result, we still add the MSE evaluation to the loss function of the generator because the MSE evaluation could act as the ‘lubricant’ for the training process to accelerate and stabilize the convergence of the generator and discriminator. The discriminator cannot tell fake samples apart from real ones at the beginning of the training process; thus, it should be trained to learn a K-Lipschitz continuous function [60,61] so as to compute the Wasserstein distance. Therefore, at the beginning of training process, the generator is not able to get any effective guides from the discriminator for the updates of parameters, which would raise the risk of ‘mode collapse’. Luckily, the MSE evaluation function could help to guide the updates of the generator before the discriminator is well trained so as to prevent the generator from the ‘mode collapse’. This is the reason why we add it into the final loss function of the generator.

    The MSE loss is described as MSE=1Nsum[(|E||Et|)2],MSE|E|=2N[(|E||Et|)]T,where |Et| represents the target holographic image and superscript 2 means element-wise square.

    (a) Three discrete-sequence sets sampled from distributions P1, P2, and P3, respectively. (b) The target holographic image is input to the trained generator. (c) The generated holographic image [corresponding to Fig. 3(b)] output by a generator trained only using the MSE loss. (d) The generated holographic image output by a generator trained using MSE loss and Wasserstein distance simultaneously.

    Figure 3.(a) Three discrete-sequence sets sampled from distributions P1, P2, and P3, respectively. (b) The target holographic image is input to the trained generator. (c) The generated holographic image [corresponding to Fig. 3(b)] output by a generator trained only using the MSE loss. (d) The generated holographic image output by a generator trained using MSE loss and Wasserstein distance simultaneously.

    The mathematical expectation of p-norm distance between the discrete sequences sampled from distributions P1 and P2 is the same as that between P1 and P3. However, it is obvious that P2 is more visually similar to P1 than P3. Thus, the p-norm criterion may fail to indicate the distance between two sparse distributions, just as demonstrated in Figs. 3(b) and 3(c), in which the target holographic image [Fig. 3(b)] can be expressed as a sparse matrix, making it difficult to find the direction of optimization and fall into local minimum with the generator trained using only MSE loss, and eventually output a wrong holographic image [Fig. 3(c)]. Luckily, the Wasserstein distance could help denote this distribution difference, and the generator trained by the Wasserstein distance could generate the correct holographic image [Fig. 3(d)].

    For the loss function of the discriminator, we use the loss function of WGAN [61] to simulate the calculation of Wasserstein distance and introduce the concept of cGAN [63] to match the target holographic image with the generated holographic image: Loss_D=Ex˜Pg|Pr[D(x˜)]ExPr|Pr[D(x)]+λEx^Px^[(x^D(x^)21)2],where Pg and Pr represent the distributions of the generated holographic images and target holographic images, respectively. Meanwhile, by adding the MSE loss in Eq. (10), our loss function for the generator could be expressed as Loss_G=Ex˜Pg|Pr[D(x˜)]+MSE.Figure 3(d) shows the generated holographic image output by the joint optimization of MSE loss and Wasserstein distance, which matches well with the target holographic image [Fig. 3(b)] and demonstrates the necessity of the mixed distance criterion.

    3. RESULTS AND DISCUSSION

    Generated holographic images at each training time corresponding to the valid target holographic images. One time of training is when training generator has three iterations after the training discriminator has one iteration.

    Figure 4.Generated holographic images at each training time corresponding to the valid target holographic images. One time of training is when training generator has three iterations after the training discriminator has one iteration.

    For the experimental process, we input a testing target holographic image into the trained generator and get its output current vector Jφ. Then, the generated Jφ is binarized to ‘1’ or ‘1’, which corresponds to the reflection phase of 0 or π (the coding meta-unit state of 0 or 1), respectively. The simulated holographic images are calculated by Eq. (7) with binarized current vectors Jφ. The coding metasurface [Fig. 1(b)] is set using an FPGA with the coding meta-unit states generated from Jφ. Last, we measure the holographic images radiated by the digital coding metasurface holograms in a standard microwave chamber.

    Testing results of our proposed intelligent metasurface hologram system. The target holographic images are randomly chosen from the testing MNIST dataset or images of handwritten letters. The simulation holographic images are calculated by Eq. (7) with the binarized current vectors Jφ output by the generator. The experimental holographic images are radiated by our 1 bit coding metasurface configured with the corresponding metasurface holograms generated by the binarized current vectors Jφ from the generator.

    Figure 5.Testing results of our proposed intelligent metasurface hologram system. The target holographic images are randomly chosen from the testing MNIST dataset or images of handwritten letters. The simulation holographic images are calculated by Eq. (7) with the binarized current vectors Jφ output by the generator. The experimental holographic images are radiated by our 1 bit coding metasurface configured with the corresponding metasurface holograms generated by the binarized current vectors Jφ from the generator.

    Comparison results between the VAE-cGAN and GS algorithms. The top half presents the simulated results of the generated holographic images radiated by metasurface holograms designed by our VAE-cGAN and GS algorithms, respectively. Here, GS1, GS2, and GS3 are the generated holographic images obtained by running the GS algorithm three times in sequence. The values of MSE and PSNR evaluations are marked below the corresponding holographic images. The bottom half of this figure illustrates the statistical frequency histograms of the holographic image quality in terms of MSE and PSNR.

    Figure 6.Comparison results between the VAE-cGAN and GS algorithms. The top half presents the simulated results of the generated holographic images radiated by metasurface holograms designed by our VAE-cGAN and GS algorithms, respectively. Here, GS1, GS2, and GS3 are the generated holographic images obtained by running the GS algorithm three times in sequence. The values of MSE and PSNR evaluations are marked below the corresponding holographic images. The bottom half of this figure illustrates the statistical frequency histograms of the holographic image quality in terms of MSE and PSNR.

    The mean MSE and PSNR values of our VAE-cGAN evaluated with the whole testing dataset are 0.0382 and 14.33, respectively. Compared with 0.0407 (mean-MSE) and 14.05 (mean-PSNR) of the GS algorithm, our VAE-cGAN shows a better capability in searching the global optimum than the GS algorithm.

    4. CONCLUSION

    We propose a new intelligent non-iterative approach (VAE-cGAN) based on deep learning methods for metasurface holograms. The usage of an unsupervised VAE structure makes our system easily trained from scratch, and the introduction of Wasserstein distance criterion improves the imaging quality of holographic images. After the VAE-cGAN is well trained, we just need to use the generator part to generate the coding patterns (metasurface holograms) corresponding to the target holographic images. The non-iterative structure of the generator enables the realization of holographic imaging with high quality and high efficiency, which are validated by both simulation and experimental results.

    When deploying our trained generator on a neural network chip, our system could become a real-time holographical imager to rapidly generate the desired holographic images. It could be expected that our intelligent metasurface hologram system could become an efficient tool for microwave or even optical holograms [69], and more valuable applications could be explored in wireless communications [32], smart EM environment, health monitoring [51,70], and so on.

    References

    [1] V. G. Veselago. Electrodynamics of substances with simultaneously negative values of sigma and μ. Sov. Phys. Usp., 10, 509-514(1968).

    [2] R. Zhao, L. Huang, Y. Wang. Recent advances in multi-dimensional metasurfaces holographic technologies. PhotoniX, 1, 20(2020).

    [3] X. Ding, Z. Wang, G. Hu, J. Liu, K. Zhang, H. Li, B. Ratni, S. N. Burokur, Q. Wu, J. Tan, C.-W. Qiu. Metasurface holographic image projection based on mathematical properties of Fourier transform. PhotoniX, 1, 16(2020).

    [4] Q. Ma, Q. R. Hong, X. X. Gao, H. B. Jing, C. Liu, G. D. Bai, Q. Cheng, T. J. Cui. Smart sensing metasurface with self-defined functions in dual polarizations. Nanophotonics, 9, 3271-3278(2020).

    [5] X. Ni, N. K. Emani, A. V. Kildishev, A. Boltasseva, V. M. Shalaev. Broadband light bending with plasmonic nanoantennas. Science, 335, 427(2012).

    [6] N. Yu, P. Genevet, M. A. Kats, F. Aieta, J.-P. Tetienne, F. Capasso, Z. Gaburro. Light propagation with phase discontinuities: generalized laws of reflection and refraction. Science, 334, 333-337(2011).

    [7] D. Schurig, J. J. Mock, B. J. Justice, S. A. Cummer, J. B. Pendry, A. F. Starr, D. R. Smith. Metamaterial electromagnetic cloak at microwave frequencies. Science, 314, 977-980(2006).

    [8] J. Li, J. B. Pendry. Hiding under the carpet: a new strategy for cloaking. Phys. Rev. Lett., 101, 203901(2008).

    [9] R. Liu, C. Ji, J. J. Mock, J. Y. Chin, T. J. Cui, D. R. Smith. Broadband ground-plane cloak. Science, 323, 366-369(2009).

    [10] Q. Ma, Z. L. Mei, S. K. Zhu, T. Y. Jin, T. J. Cui. Experiments on active cloaking and illusion for Laplace equation. Phys. Rev. Lett., 111, 173901(2013).

    [11] W. X. Jiang, T. J. Cui, Q. Cheng, J. Y. Chin, X. M. Yang, R. Liu, D. R. Smith. Design of arbitrarily shaped concentrators based on conformally optical transformation of nonuniform rational B-spline surfaces. Appl. Phys. Lett., 92, 264101(2008).

    [12] Y. Lai, J. Ng, H. Chen, D. Han, J. Xiao, Z.-Q. Zhang, C. T. Chan. Illusion optics: the optical transformation of an object into another object. Phys. Rev. Lett., 102, 253902(2009).

    [13] L. Chen, Q. Ma, Q. F. Nie, Q. R. Hong, H. Y. Cui, Y. Ruan, T. J. Cui. Dual-polarization programmable metasurface modulator for near-field information encoding and transmission. Photon. Res., 9, 116-124(2021).

    [14] N. Kundtz, D. R. Smith. Extreme-angle broadband metamaterial lens. Nat. Mater., 9, 129-132(2010).

    [15] W. X. Jiang, C.-W. Qiu, T. C. Han, Q. Cheng, H. F. Ma, S. Zhang, T. J. Cui. Broadband all-dielectric magnifying lens for far-field high-resolution imaging. Adv. Mater., 25, 6963-6968(2013).

    [16] X. M. Yang, X. Y. Zhou, Q. Cheng, H. F. Ma, T. J. Cui. Diffuse reflections by randomly gradient index metamaterials. Opt. Lett., 35, 808-810(2010).

    [17] T. J. Cui, M. Q. Qi, X. Wan, J. Zhao, Q. Cheng. Coding metamaterials, digital metamaterials and programmable metamaterials. Light Sci. Appl., 3, e218(2014).

    [18] J. Li, Y. Zhang, J. Li, X. Yan, L. Liang, Z. Zhang, J. Huang, J. Li, Y. Yang, J. Yao. Amplitude modulation of anomalously reflected terahertz beams using all-optical active Pancharatnam–Berry coding metasurfaces. Nanoscale, 11, 5746-5753(2019).

    [19] R. Y. Wu, L. Zhang, L. Bao, L. W. Wu, Q. Ma, G. D. Bai, H. T. Wu, T. J. Cui. Digital metasurface with phase code and reflection-transmission amplitude code for flexible full-space electromagnetic manipulations. Adv. Opt. Mater., 7, 1801429(2019).

    [20] Q. Ma, Q. R. Hong, G. D. Bai, H. B. Jing, R. Y. Wu, L. Bao, Q. Cheng, T. J. Cui. Editing arbitrarily linear polarizations using programmable metasurface. Phys. Rev. Appl., 13, 021003(2020).

    [21] Q. Ma, C. B. Shi, G. D. Bai, T. Y. Chen, A. Noor, T. J. Cui. Beam-editing coding metasurfaces based on polarization bit and orbital-angular-momentum-mode bit. Adv. Opt. Mater., 5, 1700548(2017).

    [22] Q. Ma, L. Chen, H. B. Jing, Q. R. Hong, H. Y. Cui, Y. Liu, L. Li, T. J. Cui. Controllable and programmable nonreciprocity based on detachable digital coding metasurface. Adv. Opt. Mater., 7, 1901285(2019).

    [23] G. Ding, K. Chen, X. Luo, J. Zhao, T. Jiang, Y. Feng. Dual-helicity decoupled coding metasurface for independent spin-to-orbital angular momentum conversion. Phys. Rev. Appl., 11, 044043(2019).

    [24] J. Han, L. Li, H. Yi, Y. Shi. 1-bit digital orbital angular momentum vortex beam generator based on a coding reflective metasurface. Opt. Mater. Express, 8, 3470-3478(2018).

    [25] Q. Zheng, Y. Li, Y. Han, M. Feng, Y. Pang, J. Wang, H. Ma, S. Qu, J. Zhang. Efficient orbital angular momentum vortex beam generation by generalized coding metasurface. Appl. Phys. A, 125, 136(2019).

    [26] T. J. Cui, L. Li, S. Liu, Q. Ma, Q. Cheng. Information metamaterial systems. iScience, 23, 101403(2020).

    [27] L. Li, T. J. Cui. Information metamaterials - from effective media to real-time information processing systems. Nanophotonics, 8, 703-724(2019).

    [28] T. J. Cui, S. Liu, L. Zhang. Information metamaterials and metasurfaces. J. Mater. Chem. C, 5, 3644-3668(2017).

    [29] Q. Ma, T. J. Cui. Information metamaterials: bridging the physical world and digital world. PhotoniX, 1, 1(2020).

    [30] L. Zhang, X. Q. Chen, S. Liu, Q. Zhang, J. Zhao, J. Y. Dai, G. D. Bai, X. Wan, Q. Cheng, G. Castaldi, V. Galdi, T. J. Cui. Space-time-coding digital metasurfaces. Nat. Commun., 9, 4334(2018).

    [31] T. J. Cui, S. Liu, G. D. Bai, Q. Ma. Direct transmission of digital message via programmable coding metasurface. Research, 2019, 2584509(2019).

    [32] H. Zhao, Y. Shuang, M. Wei, T. J. Cui, P. D. Hougne, L. Li. Metasurface-assisted massive backscatter wireless communication with commodity Wi-Fi signals. Nat. Commun., 11, 3926(2020).

    [33] Y. Shuang, H. Zhao, W. Ji, T. J. Cui, L. Li. Programmable high-order OAM-carrying beams for direct-modulation wireless communications. IEEE J. Emerg. Sel. Top. Circuits Syst., 10, 29-37(2020).

    [34] G. Hinton, L. Deng, D. Yu, G. E. Dahl, A. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen, T. N. Sainath, B. Kingsbury. Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process. Mag., 29, 82-97(2012).

    [35] G. Pironkov, S. U. N. Wood, S. Dupont. Hybrid-task learning for robust automatic speech recognition. Comput. Speech Lang., 64, 101103(2020).

    [36] A. Graves, A. R. Mohamed, G. Hinton. Speech recognition with deep recurrent neural networks. IEEE International Conference on Acoustics, Speech and Signal Processing, 6645-6649(2013).

    [37] K. Lu. Intelligent recognition system for high precision image significant features in large data background. Cyber Security Intelligence and Analytics, 1056-1062(2020).

    [38] T. Tong, X. Mu, L. Zhang, Z. Yi, P. Hu. MBVCNN: joint convolutional neural networks method for image recognition. AIP Conf. Proc., 1839, 020091(2017).

    [39] H. Sun, Q. Zhang, H. Wang, A. Li. Research of images recognition method based on RBF neural network. 7th International Conference on System of Systems Engineering, 24-26(2012).

    [40] K. Cho, B. Van Merrienboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, Y. J. C. Bengio. Learning phrase representations using RNN encoder-decoder for statistical machine translation(2014).

    [41] C. Escolano, M. R. Costa-Jussa, J. A. R. Fonollosa. From bilingual to multilingual neural-based machine translation by incremental training. 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, 236-242(2019).

    [42] S. Kwon, B. H. Go, J. H. Lee. A text-based visual context modulation neural model for multimodal machine translation. Pattern Recogn. Lett., 136, 212-218(2020).

    [43] C. He, Y. Wan, Y. Gu, F. L. Lewis. Integral reinforcement learning-based multi-robot minimum time-energy path planning subject to collision avoidance and unknown environmental disturbances. IEEE Control Syst. Lett., 5, 983-988(2021).

    [44] B. Sangiovanni, G. P. Incremona, M. Piastra, A. Ferrara. Self-configuring robot path planning with obstacle avoidance via deep reinforcement learning. IEEE Control Syst. Lett., 5, 397-402(2021).

    [45] J. Yoo, D. Jang, H. J. Kim, K. H. Johansson. Hybrid reinforcement learning control for a micro quadrotor flight. IEEE Control Syst. Lett., 5, 505-510(2021).

    [46] Q. Zhang, X. Wan, S. Liu, J. Y. Yin, L. Zhang, T. J. Cui. Shaping electromagnetic waves using software-automatically-designed metasurfaces. Sci. Rep., 7, 3588(2017).

    [47] L. L. Li, H. X. Ruan, C. Liu, Y. Li, Y. Shuang, A. Alu, C. W. Qiu, T. J. Cui. Machine-learning reprogrammable metasurface imager. Nat. Commun., 10, 1082(2019).

    [48] T. Qiu, X. Shi, J. Wang, Y. Li, S. Qu, Q. Cheng, T. Cui, S. Sui. Deep learning: a rapid and efficient route to automatic metasurface design. Adv. Sci., 6, 1900128(2019).

    [49] Q. Ma, G. D. Bai, H. B. Jing, C. Yang, L. Li, T. J. Cui. Smart metasurface with self-adaptively reprogrammable functions. Light Sci. Appl., 8, 98(2019).

    [50] H. Li, H. Zhao, M. Wei, H. Ruan, Y. Shuang, T. J. Cui, P. del Hougne, L. Li. Intelligent electromagnetic sensing with learnable data acquisition and processing. Patterns, 1, 100006(2020).

    [51] L. Li, Y. Shuang, Q. Ma, H. Li, H. Zhao, M. Wei, C. Liu, C. Hao, C.-W. Qiu, T. J. Cui. Intelligent metasurface imager and recognizer. Light Sci. Appl., 8, 97(2019).

    [52] L. Li, T. J. Cui, W. Ji, S. Liu, J. Ding, X. Wan, Y. B. Li, M. Jiang, C.-W. Qiu, S. Zhang. Electromagnetic reprogrammable coding-metasurface holograms. Nat. Commun., 8, 197(2017).

    [53] J. Wu, Z. Wang, L. Zhang, Q. Cheng, S. Liu, S. Zhang, J. Song, C. T. Jun. Anisotropic metasurface holography in 3D space with high resolution and efficiency. IEEE Trans. Antennas Propag., 69, 302-316(2020).

    [54] C. Liu, Q. Ma, L. Li, T. J. Cui. Work in progress: intelligent metasurface holograms. 1st ACM International Workshop on Nanoscale Computing, Communication, and Applications, 45-48(2020).

    [55] Z. Y. Zhou, J. Xia, J. Wu, C. L. Chang, X. Ye, S. G. Li, B. T. Du, H. Zhang, G. D. Tong. Learning-based phase imaging using a low-bit-depth pattern. Photon. Res., 8, 1624-1633(2020).

    [56] R. Lopez, J. Regier, M. I. Jordan, N. Yosef. Information constraints on auto-encoding variational Bayes. 32nd Conference on Neural Information Processing Systems (NeurIPS), 6114-6125(2018).

    [57] I. Deshpande, Z. Zhang, A. Schwing. Generative modeling using the sliced Wasserstein distance. IEEE Conference on Computer Vision and Pattern Recognition, 3483-3491(2018).

    [58] M. Arjovsky, L. Bottou. Towards principled methods for training generative adversarial networks. 5th International Conference on Learning Representations (ICLR), 1-17(2017).

    [59] V. M. Panaretos, Y. Zemely. Statistical aspects of Wasserstein distances(2019).

    [60] M. Arjovsky, S. Chintala, L. Bottou. Wasserstein generative adversarial networks. 34th International Conference on Machine Learning (ICML), 214-223(2017).

    [61] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, A. Courville. Improved training of wasserstein GANs. 31st Annual Conference on Neural Information Processing Systems (NIPS), 5768-5778(2017).

    [62] M. Mirza. Conditional generative adversarial nets(2014).

    [63] P. Isola, J.-Y. Zhu, T. Zhou, A. A. Efros. Image-to-image translation with conditional adversarial networks. 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 5967-5976(2017).

    [64] T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz, B. Catanzaro. High-resolution image synthesis and semantic manipulation with conditional GANs. 31st Meeting of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 8798-8807(2018).

    [65] K. He, X. Zhang, S. Ren, J. Sun. Identity mappings in deep residual networks. 21st ACM Conference on Computer and Communications Security (CCS), 630-645(2016).

    [66] S. Ioffe, C. Szegedy. Batch normalization: accelerating deep network training by reducing internal covariate shift. 32nd International Conference on Machine Learning (ICML), 448-456(2015).

    [67] D. Ulyanov, A. Vedaldi. Instance normalization: the missing ingredient for fast stylization(2017).

    [68] Y. LeCun, C. Cortes, C. J. C. Burges. The MNIST database of handwritten digits(2012).

    [69] X. Zou, G. Zheng, Q. Yuan, W. Zang, R. Chen, T. Li, L. Li, S. Wang, Z. Wang, S. Zhu. Imaging based on metalenses. PhotoniX, 1, 2(2020).

    [70] A. Ahad, M. Tahir, K. A. Yau. 5G-based smart healthcare network: architecture, taxonomy, challenges and future research directions. IEEE Access, 7, 100747(2019).

    Che Liu, Wen Ming Yu, Qian Ma, Lianlin Li, Tie Jun Cui. Intelligent coding metasurface holograms by physics-assisted unsupervised generative adversarial network[J]. Photonics Research, 2021, 9(4): B159
    Download Citation