[2] LIN C,WU G W,MOHAMMAD S,et al. Clustering and splitting charging algorithms for large scaled wireless rechargeable sensor networks[J]. The Journal of Systems & Software, 2016,113(1):381-394.
[3] HE L,KONG L,GU Y,et al. Evaluating the on-demand mobile charging in Wireless Sensor Networks[J]. IEEE Transactions on Mobile Computing, 2015,14(9):1861-1875.
[6] LI F Y,QIN J H,ZHENG W X. Distributed Q-Learning based online optimization algorithm for unit commitment and dispatch in smart grid[J]. IEEE Transactions on Cybernetics, 2019,50(9):4146-4156.
[7] YUAN Y L, YU Z L, GU Z H, et al. A novel multi-step Q-learning method to improve data efficiency for deep reinforcement learning[J]. Knowledge-Based Systems, 2019,175(1):107-117.
[9] NISIOTI E,THOMOS N. Fast Q-learning for improved finite length performance of irregular repetition slotted ALOHA[J]. IEEE Transactions on Cognitive Communications and Networking, 2020,6(2):844-857.
[10] XIAO L, XU D J, XIE C X, et al. Cloud storage defense against advanced persistent threats: a prospect theoretic study[J]. IEEE Journal on Selected Areas in Communications, 2017,35(3):534-544.
[11] DABBAGHJAMANESH M, MOEINI A, KAVOUSI-FARD A. Reinforcement learning-based load forecasting of electric vehicle charging station using Q-learning technique[J]. IEEE Transactions on Industrial Informatics, 2021,17(6):4229-4237.
[12] NAJEEB N, DETWEILER C. Extending Wireless Rechargeable Sensor Network life without full knowledge[J]. Sensors(Basel,Switzerland), 2017,17(7):200-227.
[16] TONY A, HIRYANTO L. A review on energy harvesting and storage for rechargeable wireless sensor networks[J]. IOP Conference Series:Materials Science and Engineering, 2019,42(8):386-474.