• Photonics Research
  • Vol. 9, Issue 8, 1493 (2021)
Qiuquan Yan1, Qinghui Deng2, Jun Zhang1, Ying Zhu2, Ke Yin3, Teng Li2、4, Dan Wu1、5, and Tian Jiang2、4、*
Author Affiliations
  • 1College of Computer, National University of Defense Technology, Changsha 410073, China
  • 2College of Advanced Interdisciplinary Studies, National University of Defense Technology, Changsha 410073, China
  • 3National Innovation Institute of Defense Technology, Academy of Military Sciences PLA China, Beijing 100071, China
  • 4Beijing Institute for Advanced Study, National University of Defense Technology, Beijing 100020, China
  • 5Hefei Interdisciplinary Center, National University of Defense Technology, Hefei 230037, China
  • show less
    DOI: 10.1364/PRJ.428117 Cite this Article Set citation alerts
    Qiuquan Yan, Qinghui Deng, Jun Zhang, Ying Zhu, Ke Yin, Teng Li, Dan Wu, Tian Jiang. Low-latency deep-reinforcement learning algorithm for ultrafast fiber lasers[J]. Photonics Research, 2021, 9(8): 1493 Copy Citation Text show less

    Abstract

    The application of machine learning to the field of ultrafast photonics is becoming more and more extensive. In this paper, for the automatic mode-locked operation in a saturable absorber-based ultrafast fiber laser (UFL), a deep-reinforcement learning algorithm with low latency is proposed and implemented. The algorithm contains two actor neural networks providing strategies to modify the intracavity lasing polarization state and two critic neural networks evaluating the effect of the actor networks. With this algorithm, a stable fundamental mode-locked (FML) state of the UFL is demonstrated. To guarantee its effectiveness and robustness, two experiments are put forward. As for effectiveness, one experiment verifies the performance of the trained network model by applying it to recover the mode-locked state with environmental vibrations, which mimics the condition that the UFL loses the mode-locked state quickly. As for robustness, the other experiment, at first, builds a database with UFL at different temperatures. It then trains the model and tests its performance. The results show that the average mode-locked recovery time of the trained network model is 1.948 s. As far as we know, it is 62.8% of the fastest average mode-locked recovery time in the existing work. At different temperatures, the trained network model can also recover the mode-locked state of the UFL in a short time. Remote algorithm training and automatic mode-locked control are proved in this work, laying the foundation for long-distance maintenance and centralized control of UFLs.
    lossa=1nj=1nQpredict_j=1nj=1ncritic(sj,aj),

    View in Article

    lossc=1ni=1n[Qpredict_i(ri+γQtarget_i)]2,

    View in Article

    {ωa_t=τωa_t+(1τ)ωa_newθc_t=τθc_t+(1τ)θc_new,

    View in Article

    costtime={20if  n(t_p)nt_lockstd(t_p)+abs(avg_t_pcuravg_t_plock)else,

    View in Article

    costfreq={20if  n(f_p)nf_lockstd(f_p)+abs(min_f_pcur/avg_f_plock)else,

    View in Article

    r=20(costtime+costfreq),

    View in Article

    Qiuquan Yan, Qinghui Deng, Jun Zhang, Ying Zhu, Ke Yin, Teng Li, Dan Wu, Tian Jiang. Low-latency deep-reinforcement learning algorithm for ultrafast fiber lasers[J]. Photonics Research, 2021, 9(8): 1493
    Download Citation