• Photonics Research
  • Vol. 9, Issue 5, B201 (2021)
Yahui Zhang1、2, Joshua Robertson1, Shuiying Xiang2、*, Matěj Hejda1, Julián Bueno1, and Antonio Hurtado1、3
Author Affiliations
  • 1Institute of Photonics, SUPA Department of Physics, University of Strathclyde, Glasgow G1 1RD, UK
  • 2State Key Laboratory of Integrated Service Networks, Xidian University, Xi’an 710071, China
  • 3e-mail: antonio.hurtado@strath.ac.uk
  • show less
    DOI: 10.1364/PRJ.412141 Cite this Article Set citation alerts
    Yahui Zhang, Joshua Robertson, Shuiying Xiang, Matěj Hejda, Julián Bueno, Antonio Hurtado. All-optical neuromorphic binary convolution with a spiking VCSEL neuron for image gradient magnitudes[J]. Photonics Research, 2021, 9(5): B201 Copy Citation Text show less

    Abstract

    All-optical binary convolution with a photonic spiking vertical-cavity surface-emitting laser (VCSEL) neuron is proposed and demonstrated experimentally for the first time, to the best of our knowledge. Optical inputs, extracted from digital images and temporally encoded using rectangular pulses, are injected in the VCSEL neuron, which delivers the convolution result in the number of fast (<100 ps long) spikes fired. Experimental and numerical results show that binary convolution is achieved successfully with a single spiking VCSEL neuron and that all-optical binary convolution can be used to calculate image gradient magnitudes to detect edge features and separate vertical and horizontal components in source images. We also show that this all-optical spiking binary convolution system is robust to noise and can operate with high-resolution images. Additionally, the proposed system offers important advantages such as ultrafast speed, high-energy efficiency, and simple hardware implementation, highlighting the potentials of spiking photonic VCSEL neurons for high-speed neuromorphic image processing systems and future photonic spiking convolutional neural networks.

    1. INTRODUCTION

    Convolutional neural networks (CNNs) have seen tremendous success in many applications, such as speech and image recognition [1,2], computer vision [3], and document analysis [4]. However, CNN-based systems are computationally expensive due to their complicated architectures and the large number of parameters they rely on. CNNs therefore typically require the implementation of multicore central processing units and graphics processing units to compensate for the rather high computational expense [5,6]. This makes CNN architectures often unsuitable for smaller devices like phones and smart cameras, where power and speed have strict limitations. To address these drawbacks, the optimization and discovery of new high-speed and low power consumption platforms for CNNs are urgently required. For the optimization of CNNs, binary CNNs, which are simple, efficient, and accurate approximations of complete CNNs, can be introduced [79]. In binary CNNs, the weights given to the inputs of each convolutional layer are approximated with binary values [7]. Therefore, binary CNNs boast 58× faster convolutional operations and 32× less memory requirements than traditional CNNs [7]. Several optimized binary versions of CNNs have been proposed for training processes and image classification tasks [7,10,11]. However, beyond the optimization of CNNs, a new platform offering high speed and low power consumption remains highly desirable.

    Photonics is considered a highly promising candidate for future neural network implementations given the unique advantages it provides such as high speed, wide bandwidth, and low power consumption [1221]. Photonics-based CNNs have therefore been proposed in order to increase the speed of convolutional operations [1821]. A photonic CNN accelerator was proposed based on silicon photonic micro-ring weighting banks [18]. The full system design offers more than three orders of magnitude improvement in execution time, and its optical core potentially offers more than five orders of magnitude improvement compared to state-of-the-art electronic counterparts [18]. Xu et al. also proposed high-accuracy optical convolution unit architecture based on acousto-optical modulator arrays, where the optical convolution unit was shown to perform well on inferences of typical CNN tasks [20]. However, the size of the system is based on the size of the kernel utilized in these emerging works on photonic CNNs.

    In this work, we propose an all-optical binary convolution system using a single vertical-cavity surface-emitting laser (VCSEL) operating as a spiking optical neuron, hence, dramatically reducing hardware requirements. In our approach, temporal encoding is used instead of spatial encoding, thus crucially helping to reduce (optical) hardware complexity. In our all-optical binary convolution technique, results are represented by the number of fast (<100  ps long) spiking responses delivered by the optical spiking VCSEL neuron. This has unique advantages in terms of robustness to noise and high precision. Additionally, VCSELs have unique inherent advantages, such as high-energy efficiency, high-speed modulation capability, low bias currents, easy packaging, and highly integrable structures [22,23]. In particular, VCSELs have demonstrated the ability to generate fast spiking dynamics analogous to those of biological neurons known for their robustness to input noise [2429]. The controlled activation, inhibition, and communication of these neuronal dynamics has been demonstrated, and recently a single VCSEL device was used to perform spiking pattern recognition and rate coding [2431]. Thus, photonic spiking VCSELs make suitable candidates for a new future photonic platform for ultrafast energy efficient spiking CNNs.

    In this work, we use a VCSEL-based photonic approach for binary convolution to demonstrate image gradient magnitude calculation. This delivers an essential portion of the image edge detection functionality used by computer vision and image recognition systems. Here, a single VCSEL system is developed to solely perform a convolution operation; hence, no VCSEL-based CNN architecture, capable of providing learning and classification capabilities, is discussed in this work. The rest of the paper is organized as follows. Section 2 is devoted to the experimental setup of this work for the demonstration of all-optical binary convolution with a spiking VCSEL neuron and the theoretical model used to predict the response of the system. In Section 3, convolutional results are analyzed before the full calculation of image gradient magnitudes is performed both experimentally and theoretically. Finally, Section 4 summarizes the conclusions of this work.

    2. EXPERIMENTAL SETUP AND THEORETICAL MODEL

    We present here the experimental arrangement and theoretical model of the all-optical binary convolution system based on a photonic spiking VCSEL neuron. In this work, we set a source digital image and a kernel as the two inputs of the convolution system. The value of any one pixel in the source image or kernel is limited to 0 or 1.

    A. Experimental Setup

    Experimental setup of the binary convolution system based on a single VCSEL. TL, tunable laser; OI, optical isolator; VOA, variable optical attenuator; PC1, PC2, and PC3, polarization controllers; AWG, arbitrary waveform generator; Mod1, Mod2, Mach–Zehnder modulators; OC1, OC2, optical couplers; CIRC, circulator; Bias & T Controller, bias and temperature controller; PD, photodetector; PM, power meter; SCOPE, oscilloscope; OSA, optical spectrum analyzer.

    Figure 1.Experimental setup of the binary convolution system based on a single VCSEL. TL, tunable laser; OI, optical isolator; VOA, variable optical attenuator; PC1, PC2, and PC3, polarization controllers; AWG, arbitrary waveform generator; Mod1, Mod2, Mach–Zehnder modulators; OC1, OC2, optical couplers; CIRC, circulator; Bias & T Controller, bias and temperature controller; PD, photodetector; PM, power meter; SCOPE, oscilloscope; OSA, optical spectrum analyzer.

    (a) Optical spectrum of free-running VCSEL used in the experiment. (b) Optical spectrum of the VCSEL subject to constant optical injection. Two polarization modes of VCSELs are referred to as λy (parallel) and λx (orthogonal).

    Figure 2.(a) Optical spectrum of free-running VCSEL used in the experiment. (b) Optical spectrum of the VCSEL subject to constant optical injection. Two polarization modes of VCSELs are referred to as λy (parallel) and λx (orthogonal).

    B. Theoretical Model

    We use an extension of the well-known spin-flip model (SFM) to model the operation of the VCSEL acting as a spiking optical neuron. In our formulation, we add extra terms to the model’s equations to account for the source image and kernel inputs. The rate equations can be described as follows [26,27]: dEx,ydt=(k±γa)Ex,yi(kα±γp)Ex,y+k(1+iα)(NEx,y±inEy,x)+kinj[Einjx1(t)+Einjx2(t)]eiΔωxt+Fx,y,dNdt=γN[N(1+|Ex|2+|Ey|2)μ+in(EyEx*ExEy*)],dndt=γsnγN[n(|Ex|2+|Ey|2)+iN(EyEx*ExEy*)],where the subscripts x, y represent the XP and YP modes of the VCSEL, respectively. Ex,y is the slowly varying complex amplitude of the field in the XP and YP modes. N is the total carrier inversion between conduction and valence bands. n is the difference between carrier inversions with opposite spins. k denotes the field decay rate. γa and γp are the linear dichroism and the birefringence rate, respectively. α is the linewidth enhancement factor. γN is the decay rate of N. γs is the spin-flip rate. μ represents the normalized pump current. kinj is the injected strength, and Einjx1 and Einjx2 indicate, respectively, the source image and kernel inputs. Δωx is defined as Δωx=ωinjxω0, where ωinjx is the angular frequency of the externally injected light in the XP mode, and ω0=(ωx+ωy)/2 is the center frequency between the XP and YP modes with ωx=ω0+αγaγp and ωy=ω0αγa+γp. The frequency detuning between the externally injected signal and the XP mode is set as Δfx=finjxfx. Hence, in Eq. (1), Δωx=2πΔfx+αγaγp. Fx,y are the spontaneous emission noise terms, which can be written as Fx=βspγN2(N+nξ1+Nnξ2),Fy=iβspγN2(N+nξ1Nnξ2),where βsp is the strength of the spontaneous emission, and ξ1 and ξ2 are independent complex Gaussian white noise terms of zero mean and a unit variance. We numerically solve Eqs. (1)–(4) using the fourth-order Runge–Kutta method. The parameter values configured for the 1300 nm VCSEL are as follows [27]: k=185  ns1, γa=2  ns1, γp=128  ns1, α=2, γN=0.5  ns1, γs=110  ns1, βsp=106, and kinj=125  ns1. With these parameters, the YP mode is the main lasing mode, and the XP mode is the subsidiary mode, as in Fig. 2(a).

    3. EXPERIMENTAL AND NUMERICAL RESULTS

    In this section, we firstly provide an experimental proof-of-concept demonstration of all-optical binary convolution with a spiking VCSEL neuron. We then calculate the image gradient magnitudes from a basic “Square” source image and a complex “Horse head” source image by means of all-optical binary convolution. Simulation results on the binary convolution and the calculation of image gradient magnitudes are also presented using a “Horse” source image from the latest version of the Berkeley Segmentation Data Set [32]. Finally, the robustness of our binary convolution system is also tested numerically by adding noise to the source image and kernel inputs.

    A. Experimental Results

    Example of a single step during a 2D binary convolution operation. During this step, a Hadamard (element-wise) product is calculated for a submatrix of the image and the kernel, and all of the values in the multiplication result are summed up to obtain a single value.

    Figure 3.Example of a single step during a 2D binary convolution operation. During this step, a Hadamard (element-wise) product is calculated for a submatrix of the image and the kernel, and all of the values in the multiplication result are summed up to obtain a single value.

    Experimental convolution operation. (a) Inputs of Channel 1 (image in Fig. 3). (b) Inputs of Channel 2 (kernel in Fig. 3). (c) Inputs of VCSEL. (d) Outputs of VCSEL (the results of convolution).

    Figure 4.Experimental convolution operation. (a) Inputs of Channel 1 (image in Fig. 3). (b) Inputs of Channel 2 (kernel in Fig. 3). (c) Inputs of VCSEL. (d) Outputs of VCSEL (the results of convolution).

    Temporal map of 100 superimposed consecutive convolutional results measured experimentally at the output of spiking VCSEL neuron.

    Figure 5.Temporal map of 100 superimposed consecutive convolutional results measured experimentally at the output of spiking VCSEL neuron.

    B. Calculation of Image Gradient Magnitudes

    In this section, the image gradient magnitude, critical to image edge detection, is calculated using our approach based on a single spiking VCSEL neuron and optical binary convolution. The image gradient magnitude G(x) of a given pixel x is calculated using the following equations [33]: G(x)=GX(x)2+GY(x)2,GX(x)=[B(x)BX+][B(x)BX],GY(x)=[B(x)BY+][B(x)BY].

    Four binary convolutions, i.e., B(x)BX,Y±, are used in GX(x) and GY(x). B(x)=p=0N1s(ip,ix)·2p is the N bit local binary pattern descriptor of pixel x. ix is the central pixel intensity, and ip is the intensity of the pth neighbor of x in the source pattern. The comparison operator is defined as s(ip,ix)={1if  |ipix|>Tx0otherwise,where Tx=14ix+20 and N=5×51.

    (a) Gray color: range of the local binary pattern descriptor of pixels. (b) A 24×24 pixels “Square” source image. The red highlight indicates a given pixel in the image. (c) The four convolutions (BX+, BX−, BY+, and BY−) of the 5×5 binary pattern. Bits that fall outside the highlighted areas for a given string are set to zero.

    Figure 6.(a) Gray color: range of the local binary pattern descriptor of pixels. (b) A 24×24 pixels “Square” source image. The red highlight indicates a given pixel in the image. (c) The four convolutions (BX+, BX, BY+, and BY) of the 5×5 binary pattern. Bits that fall outside the highlighted areas for a given string are set to zero.

    For the red-highlighted pixel x in Fig. 6(b), “1” in B(x) corresponds to a white pixel, and “0” corresponds to a black pixel in the source image.

    In Eqs. (7) and (8), BX+, BX, BY+, and BY are the four kernels that are adopted as in Ref. [33]. Figure 6(c) shows the areas of the four different kernels. Pixels that fall outside of the highlighted areas in Fig. 6(c) for a given string are set to zero. For example, BX+=[1000001000110000100010000].

    Four convolutional results with four highlighted area kernels for one pixel, which has red box in Fig. 6.

    Figure 7.Four convolutional results with four highlighted area kernels for one pixel, which has red box in Fig. 6.

    Gradient maps of the “Square” source image. Visualizations of (a) G, (b) GX, and (c) GY maps of the “Square” source image based on the optical binary convolution performed by the VCSEL neuron.

    Figure 8.Gradient maps of the “Square” source image. Visualizations of (a) G, (b) GX, and (c) GY maps of the “Square” source image based on the optical binary convolution performed by the VCSEL neuron.

    “Horse head” image and the gradient maps of the “Horse head” image. (a) Source “Horse” image. The blue box indicates the “Horse Head” image used for analysis in (b). Visualizations of the (c) G, (d) GX, and (e) GY maps of the “Horse head” image obtained from the optical binary convolution performed with the VCSEL neuron.

    Figure 9.“Horse head” image and the gradient maps of the “Horse head” image. (a) Source “Horse” image. The blue box indicates the “Horse Head” image used for analysis in (b). Visualizations of the (c) G, (d) GX, and (e) GY maps of the “Horse head” image obtained from the optical binary convolution performed with the VCSEL neuron.

    C. Numerical Results

    In this section, binary convolution based on a single VCSEL neuron is performed numerically. The robustness of the system to perform all-optical binary convolution under noisy inputs and for larger kernels is investigated. Finally, the calculation of image gradient magnitudes with our photonic approach using a single VCSEL neuron is presented numerically using the “Horse” image from the latest version of the Berkeley Segmentation Data Set [32].

    (a1)–(a3) Inputs of Channel 1 (image in Fig. 3). (b1)–(b3) Inputs of Channel 2 (kernel in Fig. 3). (c1)–(c3) VCSEL neuron’s output. (a1)–(c1) Convolutional operation in the VCSEL neuron without noise. (a2)–(c2) Convolutional operation in the VCSEL neuron with added input noise of SNR=20 dB. (a3)–(c3) Convolution operation with a 5×5 pixels kernel.

    Figure 10.(a1)–(a3) Inputs of Channel 1 (image in Fig. 3). (b1)–(b3) Inputs of Channel 2 (kernel in Fig. 3). (c1)–(c3) VCSEL neuron’s output. (a1)–(c1) Convolutional operation in the VCSEL neuron without noise. (a2)–(c2) Convolutional operation in the VCSEL neuron with added input noise of SNR=20  dB. (a3)–(c3) Convolution operation with a 5×5 pixels kernel.

    “Horse” image and gradient maps of the “Horse” image. (a) “Horse” image. Visualizations of (b) G, (c) GX, and (d) GY maps of “Horse” image based on the numerical optical binary convolution in VCSEL.

    Figure 11.“Horse” image and gradient maps of the “Horse” image. (a) “Horse” image. Visualizations of (b) G, (c) GX, and (d) GY maps of “Horse” image based on the numerical optical binary convolution in VCSEL.

    Optical binary convolution can be used in systems where simplified convolutional operations, with binary inputs (still able to provide high-performance accuracy), provide other key advantages in terms of increased operation speed, lowered energy consumption, and reduced hardware requirements. This is the case of the system reported in this work, using an extremely hardware friendly implementation of a single VCSEL to perform high-speed and low energy (<pJ/spike) image edge-feature detection. Besides, in our approach, the results of the binary convolution are output in an optical spiking representation, providing unique advantages in terms of robustness to noise and high precision of convolutional results. This spiking representation therefore enables our platform to successfully perform with noisy optical and electronic signals. Whilst other recent works have recently reported complex systems for optical convolution operation using temporally modulated inputs and weights for image processing tasks showing excellent performance [36], our technique benefits from an extremely simple architecture using just one off-the-shelf, inexpensive 1300 nm VCSEL to perform the binary convolution operation for image edge-feature detection. Our approach combining a VCSEL-based spiking photonic neuron with time multiplexing is able to deliver the operation of a full neuronal layer, where each 1.5-ns-long time slot operates in fact as a virtual neuron (or node) processing specific image pixel information. This offers great promise for future implementation of interconnected VCSEL-based neuronal network architectures for image processing tasks of increased complexity (e.g.,  image classification) and using neuron-like spiking signals to operate.

    The utilization of binary convolution in the calculation of image gradient maps has been reported to outperform the alternative Canny implementation [35] of image gradient maps convolution, in the Intel i7 mobile processor [33]. In that report, the frequency of the binary gradient-based edge detector was 4.7 Hz, while the Canny convolution approach was found to operate at 0.5 Hz. This indicates that binary convolution can be performed at speeds faster than alternative convolution approaches. Additionally, mobile processors operate with powers of several watts (for example, the Intel i7 has a power of 15 W) [37], whilst VCSELs, such as the one used in this work, provide low power performance typically at milliwatt (mW) and sub-mW power levels. Hence, the energy consumption for the calculation of image gradient maps obtained with our VCSEL-based optical binary convolution system can be significantly more energy efficient, as well as yield faster operation speeds, than the performance achieved with traditional or binary convolution methods in digital processors.

    4. CONCLUSION

    In this work, we proposed and investigated experimentally and numerically an all-optical binary convolution system using a VCSEL operating as a photonic spiking neuron. The inputs (image and kernel) are encoded temporally using fast rectangular pulses (1.5-ns-long) and optically injected into the VCSEL neuron. The latter’s optical output directly provides the results of the convolution in the number of (sub-ns long) spikes fired. In addition to performing all-optical binary convolution, we demonstrated experimentally and numerically the ability of the proposed system to calculate the image gradient magnitudes from digital source images. This feature was successfully used to identify key edge features from a source image as well as its separate horizontal and vertical components. Furthermore, we investigated numerically the robustness of the proposed VCSEL-based convolutional system to input noise. This simple system, using a single commercially available VCSEL operating at the key telecom wavelength of 1300 nm, offers a novel photonic solution to binary convolution with the advantage of being highly energy efficient and hardware friendly. This opens exciting prospects for a new photonic spiking platform for future optical binary spiking CNNs. Furthermore, the high-speed, low cost, and neuronal functionalities of these photonic spiking systems hold promise for numerous processing tasks expanding into fields such as computer vision and artificial intelligence.

    Acknowledgment

    Acknowledgment. We thank Prof. T. Ackemann and Prof. A. Kemp (University of Strathclyde) for lending some of the equipment used in this work. All data underpinning this publication are openly available from the University of Strathclyde KnowledgeBase at https://doi.org/10.15129/51af27bc-8cc2-46a7-9088-1aad87e4340c.

    References

    [1] O. Abdel-Hamid, A. R. Mohamed, H. Jiang, G. Penn. Applying convolutional neural networks concepts to hybrid NN-HMM model for speech recognition. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 4277-4280(2012).

    [2] J. Fu, H. Zheng, T. Mei. Look closer to see better: recurrent attention convolutional neural network for fine-grained image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 4438-4446(2017).

    [3] K. Gopalakrishnan, S. K. Khaitan, A. Choudhary, A. Agrawal. Deep convolutional neural networks with transfer learning for computer vision-based data-driven pavement distress detection. Constr. Build. Mater., 157, 322-330(2017).

    [4] P. Y. Simard, D. Steinkraus, J. C. Platt. Best practices for convolutional neural networks applied to visual document analysis. International Conference on Document Analysis and Recognition (ICDAR), 958-963(2003).

    [5] C. Farabet, C. Couprie, L. Najman, Y. LeCun. Learning hierarchical features for scene labeling. IEEE Trans. Pattern Anal. Mach. Intell., 35, 1915-1929(2013).

    [6] L. Cavigelli, M. Magno, L. Benini. Accelerating real-time embedded scene labeling with convolutional networks. Proceedings of the 52nd Annual Design Automation Conference, 108(2015).

    [7] M. Rastegari, V. Ordonez, J. Redmon, A. Farhadi. Xnor-net: ImageNet classification using binary convolutional neural networks. European Conference on Computer Vision (ECCV), 525-542(2016).

    [8] F. Juefei-Xu, V. N. Boddeti, M. Savvides. Local binary convolutional neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 19-28(2017).

    [9] X. Lin, C. Zhao, W. Pan. Towards accurate binary convolutional neural network. Advances in Neural Information Processing Systems (NIPS), 345-353(2017).

    [10] M. Courbariaux, Y. Bengio, J.-P. David. BinaryConnect: training deep neural networks with binary weights during propagations. Advances in Neural Information Processing Systems (NIPS), 3105-3113(2015).

    [11] M. Courbariaux, I. Hubara, D. Soudry, R. El-Yaniv, Y. Bengio. Binarized neural networks: training deep neural networks with weights and activations constrained to +1 or -1(2016).

    [12] M. Turconi, B. Garbin, M. Feyereisen, M. Giudici, S. Barland. Control of excitable pulses in an injection-locked semiconductor laser. Phys. Rev. E, 88, 022923(2013).

    [13] P. R. Prucnal, B. J. Shastri, T. F. de Lima, M. A. Nahmias, A. N. Tait. Recent progress in semiconductor excitable lasers for photonic spike processing. Adv. Opt. Photon., 8, 228-299(2016).

    [14] S. Xiang, Y. Zhang, J. Gong, X. Guo, L. Lin, Y. Hao. STDP-based unsupervised spike pattern learning in a photonic spiking neural network with VCSELs and VCSOAs. IEEE J. Sel. Top. Quantum Electron., 25, 1700109(2019).

    [15] Y. Zhang, S. Xiang, X. Guo, A. Wen, Y. Hao. All-optical inhibitory dynamics in photonic neuron based on polarization mode competition in a VCSEL with an embedded saturable absorber. Opt. Lett., 44, 1548-1551(2019).

    [16] J. Feldmann, N. Youngblood, C. D. Wright, H. Bhaskaran, W. H. P. Pernice. All-optical spiking neurosynaptic networks with self-learning capabilities. Nature, 569, 208-214(2019).

    [17] J. Robertson, E. Wade, Y. Kopp, J. Bueno, A. Hurtado. Toward neuromorphic photonic networks of ultrafast spiking laser neurons. IEEE J. Sel. Top. Quantum Electron., 26, 7700715(2019).

    [18] A. Mehrabian, Y. Al-Kabani, V. J. Sorger, T. El-Ghazawi. PCNNA: a photonic convolutional neural network accelerator. IEEE International System-on-Chip Conference (SOCC), 169-173(2018).

    [19] H. Bagherian, S. Skirlo, Y. Shen, H. Meng, V. Ceperic, M. Soljacic. On-chip optical convolutional neural networks(2018).

    [20] S. Xu, J. Wang, R. Wang, J. Chen, W. Zou. High-accuracy optical convolution unit architecture for convolutional neural networks by cascaded acousto-optical modulator arrays. Opt. Express, 27, 19778-19787(2019).

    [21] S. Xu, J. Wang, W. Zou. High-energy-efficiency integrated photonic convolutional neural networks(2019).

    [22] K. Iga, H. E. Li. Vertical-Cavity Surface-Emitting Laser Devices(2003).

    [23] R. Michalzik. VCSELs: Fundamentals, Technology and Applications of Vertical-Cavity Surface-Emitting Lasers, 166(2013).

    [24] A. Hurtado, J. Javaloyes. Controllable spiking patterns in long-wavelength vertical cavity surface emitting lasers for neuromorphic photonics systems. Appl. Phys. Lett., 107, 241103(2015).

    [25] B. Garbin, J. Javaloyes, G. Tissoni, S. Barland. Topological solitons as addressable phase bits in a driven laser. Nat. Commun., 6, 5915(2015).

    [26] S. Y. Xiang, A. J. Wen, W. Pan. Emulation of spiking response and spiking frequency property in VCSEL-based photonic neuron. IEEE Photon. J., 8, 1504109(2016).

    [27] T. Deng, J. Robertson, A. Hurtado. Controlled propagation of spiking dynamics in vertical-cavity surface-emitting lasers: towards neuromorphic photonic networks. IEEE J. Sel. Top. Quantum Electron., 23, 1800408(2017).

    [28] J. Robertson, T. Deng, J. Javaloyes, A. Hurtado. Controlled inhibition of spiking dynamics in VCSELs for neuromorphic photonics: theory and experiments. Opt. Lett., 42, 1560-1563(2017).

    [29] A. Dolcemascolo, B. Garbin, B. Peyce, R. Veltz, S. Barland. Resonator neuron and triggering multipulse excitability in laser with injected signal. Phys. Rev. E, 98, 062211(2018).

    [30] J. Robertson, M. Hejda, J. Bueno, A. Hurtado. Ultrafast optical integration and pattern classification for neuromorphic photonics based on spiking VCSEL neurons. Sci. Rep., 10, 1(2020).

    [31] M. Hejda, J. Robertson, J. Bueno, A. Hurtado. Spike-based information encoding in VCSELs for neuromorphic photonic systems. J. Phys., 2, 044001(2020).

    [32] P. Arbelaez, M. Maire, C. Fowlkes, J. Malik. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell., 33, 898-916(2011).

    [33] P. L. St-Charles, G. A. Bilodeau, R. Bergevin. Fast image gradients using binary feature convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1-9(2016).

    [34] E. Nadernejad. Edge detection techniques: evaluations and comparisons. Appl. Math. Sci., 2, 1507-1520(2008).

    [35] J. Canny. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell., 8, 679-698(1986).

    [36] R. Hamerly, L. Bernstein, A. Sludds, M. Soljačić, D. Englund. Large-scale optical neural networks based on photoelectric multiplication. Phys. Rev. X, 9, 021032(2019).

    [37] V. Spiliopoulos, G. Keramidas, S. Kaxiras, K. Efstathiou. Power-performance adaptation in Inter core i7. Proceedings of 2nd Workshop Computer Architecture and Operating System Co-design, 1-10(2011).

    Yahui Zhang, Joshua Robertson, Shuiying Xiang, Matěj Hejda, Julián Bueno, Antonio Hurtado. All-optical neuromorphic binary convolution with a spiking VCSEL neuron for image gradient magnitudes[J]. Photonics Research, 2021, 9(5): B201
    Download Citation