• Laser & Optoelectronics Progress
  • Vol. 57, Issue 24, 241003 (2020)
Deyong Gao1、2, Zibing Kang1、*, Song Wang1、2, and Yangping Wang1、3
Author Affiliations
  • 1School of Electronic & Information Engineering, Lanzhou Jiaotong University, Lanzhou, Gansu 730070, China;
  • 2Gansu Provincial Engineering Research Center for Artificial Intelligence and Graphic & Image Processing, Lanzhou, Gansu 730070, China;
  • 3Gansu Provincial Key Laboratory of System Dynamics and Reliability of Rail Transport Equipment, Lanzhou, Gansu 730070, China
  • show less
    DOI: 10.3788/LOP57.241003 Cite this Article Set citation alerts
    Deyong Gao, Zibing Kang, Song Wang, Yangping Wang. Human-Body Action Recognition Based on Dense Trajectories and Video Saliency[J]. Laser & Optoelectronics Progress, 2020, 57(24): 241003 Copy Citation Text show less
    Saliency-based action recognition algorithm framework
    Fig. 1. Saliency-based action recognition algorithm framework
    Dense trajectory algorithm framework
    Fig. 2. Dense trajectory algorithm framework
    Sample frames from UCF Sports and YouTube. (a) UCF Sports; (b) YouTube
    Fig. 3. Sample frames from UCF Sports and YouTube. (a) UCF Sports; (b) YouTube
    Estimation of saliency detection parameters
    Fig. 4. Estimation of saliency detection parameters
    Comparison of the DT and our method. (a) UCF Sports; (b) YouTube
    Fig. 5. Comparison of the DT and our method. (a) UCF Sports; (b) YouTube
    Accuracy comparison of each class by DT and our method. (a) UCF Sports; (b) YouTube
    Fig. 6. Accuracy comparison of each class by DT and our method. (a) UCF Sports; (b) YouTube
    Experimental environmentDetail information
    OSUbuntu14.04
    CPUIntel(R) i7-8700 @3.20 GHz
    GPUNvidia GeForce GTX 1060 3 GB
    RAM16 GB
    CompilerMatlab2016
    Table 1. Experimental environment
    DatasetsDTS-Traj
    UCF Sports88.290.3
    YouTube84.189.6
    Table 2. Comparison of mean accuracy by DT and our method unit: %
    UCF SportsYouTube
    MethodMean accuracyMethodMean accuracy
    Wang et al[6]89.10Wang et al[6]85.40
    Yi et al[13]90.08Yang et al[22]88.00
    Somasundaram et al[14]87.30Peng et al[23]87.60
    Li et al[15]93.40Guo et al[24]89.50
    Cho et al[21]89.70Duan et al[25]90.00
    Our method90.30Our method89.60
    Table 3. Results comparison of our method and the state-of-the-art method unit: %
    Deyong Gao, Zibing Kang, Song Wang, Yangping Wang. Human-Body Action Recognition Based on Dense Trajectories and Video Saliency[J]. Laser & Optoelectronics Progress, 2020, 57(24): 241003
    Download Citation