• Laser & Optoelectronics Progress
  • Vol. 59, Issue 8, 0815001 (2022)
Guorong Xie1,2, Yi Qu2,*, and Rongqi Jiang1,2
Author Affiliations
  • 1Postgraduate Brigade, Engineering University of PAP, Xi’an , Shaanxi 710086, China
  • 2School of Information Engineering, Engineering University of PAP, Xi’an , Shaanxi 710086, China
  • show less
    DOI: 10.3788/LOP202259.0815001 Cite this Article Set citation alerts
    Guorong Xie, Yi Qu, Rongqi Jiang. Tracking Algorithms Based on Antiocclusion Object Models[J]. Laser & Optoelectronics Progress, 2022, 59(8): 0815001 Copy Citation Text show less
    Kalman filter tracker flow diagram
    Fig. 1. Kalman filter tracker flow diagram
    Particle filter tracker flow diagram
    Fig. 2. Particle filter tracker flow diagram
    Feature descriptorRepresentation methodsScene adaptability

    Manual

    visual

    feature

    IntensityMOSSE8, CSK9Single stable scene
    GraySingle stable scene
    Color

    CN10, ASMS11

    DAT12, CSCT13

    Partial occlusion, fast move, scale invariation
    CN, LabInapplicable to light changes
    GradientKCF14, BACF15, LCT16, DSST17Translation, rotation, illumination invariation
    HOG, SIFTInapplicable to deformation and motion blur
    TextureTLD18Illumination, scale, fast motion invariation
    LBPInapplicable to deformation, time-consuming
    Optical FlowFlowTrack19Suitable for short-term occlusion and scale change
    Not suitable for severe occlusion, camera shake, and out-of-view
    Deep featureDeep SRDCF20, CFNet21Strong robustness to complex scenes including occlusion
    Poor real-time performance and interpretability
    Table 1. Scene adaptability of tracking features and representative methods
    Limitations of traditional methodsRelated improvements measuresMethods
    Time-consumingKey point optical flow calculation instead of global filed22-23
    Limitation of small displacementConstruct an optical flow pyramid for downsampling24
    Inaccurate expressionDeep optical flow estimation25-26
    Uncertainty under occlusionSymmetry between optical map and occlusion area26
    Table 2. Development of optical flow estimation methods
    Feature typeHigh-level feature of VGG-NetLow-level feature of VGG-Net
    Extraction layerConv4-4, Conv5-4Conv1-2, Conv2-2
    DescriptionSemantic nature informationContour and texture
    AdvantagesRobust to appearance change, occlusionEssential positioning
    DisadvantagesLow spatial resolutionPoor anti-interference ability
    Table 3. Expression traits of features extracted from VGG-Net different layers
    TypeAlgorithmFeature extractedObject model characteristicsFitted scenes

    Color and

    gradient

    feature

    SAMF29HOG, CN, GreyScale adaptive with multiple featuresPartial occlusion, deformation, rotations, interference
    MOCA30MC-HOGHOG extracted from CN channels
    Staple31Color, HOGGlobal-local features, 80 fps of speed
    SAT32RGB, SIFTEncoding structure, keypoints spatial layout
    DPCF33Color, HOGCollaborative and deformable model
    MvCFT35HOG, CN, GreyFusion model of multi-perspective feature

    Appearance feature of

    different

    layers

    HCF37Conv3-4, Conv4-4,Conv5-4Holistic model with weighted multi-feature filter confidence mapOcclusion, clutters, deformation, semantics-distractors
    C-COT38Conv1, Conv5Continuous spatial interpolation, preciser
    ECO39Conv1, Conv5PCA decomposed convolution, fast
    SANet44CNN and RNN featureSkip concatenation, self-structure encoding

    Appearance and motion

    feature

    DFCNet41Optical flow, CNN feataureAdaptive keyframe scheduling mechanismClutters, fast movements, deformation, heavy occlusion
    DCTN43Optical flow, CNN feataurePyramidal feature hierarchy
    STSGS42Conv3-4, Conv4-4, Conv5-4Quantum mechanics based saliency detection and motion flow map generation
    FPRNet45Optical flow, CNN and RNN featureMulti-scale spatiotemporal representations by flow pyramid recurrent framework
    Table 4. Object model traits and fitted scenes of representative algorithms based on fusion features
    MethodsTypeSystem requirementsCharacteristics and applicability of tracker methods
    StatusNoise
    3447-50KFLinearGaussianVulnerable to interference

    Applicable to severe

    occlusion temporarily

    51EKFNon-linearGaussianEasy to accumulate error
    52-53UKFNon-linearNon-gaussianHigher speed and accuracy
    Table 5. Characteristics and scenario applicability of three kinds of KF tracking methods
    CharacteristicsKalman filterParticle filter
    Probability modelState and observation modelsWeighted particle swarm estimate
    RequirmentsGaussian and linear systemNon-linear and non-gaussian system
    Anti-occlusionPrediction mechanism with fast convergencePrediction and multi-modal maintenance
    Applicable scenesShort-term occlusionPartial occlusion, background interference
    DisadvantageApplicable scene limitationsParticles degenerate, time-consuming
    Table 6. Scenes applicability and traits of status estimation model between Kalman filter and particle filter
    Anti-occlusion schemeMethodsContribution and characteristicsOcclusion application
    Occlusion detection54Normalization factor for occlusion detectionLong-term, fully, partial
    55Internality of weight value and distribution region
    56Third-order cumulants of reconstruction error
    Trajectory prediction57Bhattacharyya coefficient as judgment criterionShot-term

    Modified

    model to

    maintain

    particle

    diversity

    Robust observation58Multiple likelihood models of HSV and HOG

    Partial, serious,

    long-term

    59Color distribution model
    60Deep feature and color histogram in adaptive mode
    Redistribution61Particles generation are independentlySerious
    62Redistribution based on region growth
    Memory Mechanism63Memory-based state estimation schemeSerious, long-term
    64Adaptive update strategy with well model saved
    Mark flag65Binary occlusion flag state representation for particlesPartial
    Deep learning66Observation model built by RBF neural networkLong-term
    Table 7. Anti-occlusion scheme, innovation and occlusion adaptation of algothrims based on PF state estimation
    CharacteristicConvolutional neural networksRecurrent Neural Network
    LSTMGRU
    NetworkCompositionConvolutional layer, pooling layer, downsampling layerForgot gateUpdate gate, reset gate
    Structural features

    Local receptive field

    Weight sharing

    Spatiotemporal downsampling

    Better effectHigher efficiency
    Memory and feedback functions by current and historical states connection
    TrackingInformationSpatial semantic informationTemporal information
    DisadvantageSensitive to similar interference, lacking relevant information, may lose spatial detailsHigh time consumption
    Table 8. Comparison of structure and tracking applications characteristics between CNN and RNN
    Tracker

    Object

    models

    Success ratePrecisionFPSPC(CPU,RAM,Nvidia GPU)Test dataset
    AllOccAllOcc
    CSKGrey0.3980.545152Intel i5-760 2.80 GHz CPU, 16 G RAMOTB50
    KCFG0.5140.5130.7400.749172Intel i5-760 2.80 GHz CPU, 16 G RAM
    CNC0.4440.4280.6350.629105Intel i5-760 2.80 GHz CPU, 16 G RAM
    SAMFG+C0.5670.6130.7740.8397Intel i5-760 2.80 GHz CPU, 16 G RAM
    DSSTG0.5550.5340.54924Intel-i5-4590 CPU, 8 G RAM
    MvCFTMAF0.5320.55525.5Intel-i5-4590 CPU, 8 G RAM
    MOCAG+C0.5690.8240.87716.5Intel Xeon2 2.50 GHz CPU, 256 G RAM
    StapleG+C0.6940.51380Intel Core i7-4790K 4.0 GHz CPU
    FPRNetDMF, DAF0.6130.8541Intel Core i7-4790K 4.0 GHz CPU
    RPTPF, Pb0.5760.8124.15Intel-i7-3770 3.4 GHz CPU, 16 G RAM
    SPWTDMF, Kb0.5300.79262.3Intel i7-6700 CPU
    RTTPb, RF0.5880.8273‒42.80 GHz CPU, 16 G RAM
    Re3DAF, T-C0.4220.390150Intel Xeon 2.20 GHz CPU, Nvidia Titan X
    KCFG0.4760.6930.625172Intel i7 3.7 GHz CPU, 12 GB RAMOTB100
    C-COTDAF0.6730.9020.3Intel i7 3.7 GHz, GTX TITAN Z GPU
    SANetDAF, RF0.6920.9281Intel i7 3.7 GHz, GTX TITAN Z GPU
    SAMFG+C0.5350.5290.74316.8Intel Xeon 2.6 GHz CPU, 256 G RAM
    SAMF-CAST-C0.5750.5500.79313Intel Xeon 2.6 GHz CPU, 256 G RAM
    StapleG+C0.5790.5430.78459.8Intel Xeon 2.6 GHz CPU, 256 G RAM
    Staple-CAST-C0.5790.5580.81035.2Intel Xeon 2.6 GHz CPU, 256 G RAM
    DSSTG0.4750.6950.61524Intel i7 3.7 GHz CPU, 12 GB RAM
    LGCFLGPb0.5850.7820.7198Intel i7 3.7 GHz CPU, 12 GB RAM
    RPTKb0.7150.93620GeForce GTX 1080Ti GPU
    Table 9. Performance comparison of representative trackers with variety models on OTB50 and OTB100