• Laser & Optoelectronics Progress
  • Vol. 57, Issue 12, 121011 (2020)
Yang Wang and Libo Liu*
Author Affiliations
  • School of Information Engineering, Ningxia University, Yinchuan, Ningxia 750021, China
  • show less
    DOI: 10.3788/LOP57.121011 Cite this Article Set citation alerts
    Yang Wang, Libo Liu. Bilinear Residual Attention Networks for Fine-Grained Image Classification[J]. Laser & Optoelectronics Progress, 2020, 57(12): 121011 Copy Citation Text show less
    Architecture of B-CNN model
    Fig. 1. Architecture of B-CNN model
    Schematic of bilinear feature combination
    Fig. 2. Schematic of bilinear feature combination
    Structure of residual unit
    Fig. 3. Structure of residual unit
    Channel attention module
    Fig. 4. Channel attention module
    Spatial attention module
    Fig. 5. Spatial attention module
    Residual attention structure of BRAN model
    Fig. 6. Residual attention structure of BRAN model
    Example of training data augmentation
    Fig. 7. Example of training data augmentation
    Visualization of different feature maps. (a) Original images; (b) B-CNN; (c) channel attention maps; (d) spatial attention maps
    Fig. 8. Visualization of different feature maps. (a) Original images; (b) B-CNN; (c) channel attention maps; (d) spatial attention maps
    DatasetClassTrainTestTotal
    CUB-200-20112005994579411788
    Stanford Dogs12012000858020580
    Stanford Cars1968144804116185
    Table 1. Detailed statistics of three fine-grained image datasets
    ApproachBackboneAccuracy /%
    B-CNN(baseline)VGG-M+VGG-D84.1
    B-CNN(resnet×2)ResNet-34×285.0
    BRAN(cha. attention)ResNet-34×2 + channel attention86.2
    BRAN(spa. attention)ResNet-34×2 + spatial attention85.5
    BRAN(cha.& spa. attention)ResNet-34×2 + cha. & spa. attention87.2
    Table 2. Ablation experiment and analysis of proposed method on CUB-200-2011 dataset
    ApproachBackboneAccuracy /%
    Birds[9]Dogs[10]Cars[11]
    Two-level attention[22]VGG1977.9--
    NAC[23]VGG1981.0168.61-
    B-CNN[6]VGG-M+VGG-D84.1-91.3
    ST-CNN[24]Inception-v2×384.1--
    DVAN[25]VGG-16×379.081.587.1
    RA-CNN[26]VGG-19×385.387.392.5
    MA-CNN[19]VGG-19×386.5-92.8
    MAMC[27]ResNet-10186.585.293.0
    BRANResNet-34×287.289.292.5
    Table 3. Comparison with weakly-supervised methods in terms of classification accuracy
    Yang Wang, Libo Liu. Bilinear Residual Attention Networks for Fine-Grained Image Classification[J]. Laser & Optoelectronics Progress, 2020, 57(12): 121011
    Download Citation