• Infrared and Laser Engineering
  • Vol. 50, Issue 5, 20200364 (2021)
Yuanhong Mao, Zhong Ma*, and Zhanzhuang He
Author Affiliations
  • Xi’an Microelectronics Technology Institute, Xi’an 710065, China
  • show less
    DOI: 10.3788/IRLA20200364 Cite this Article
    Yuanhong Mao, Zhong Ma, Zhanzhuang He. Infrared-visible image patches matching via convolutional neural networks[J]. Infrared and Laser Engineering, 2021, 50(5): 20200364 Copy Citation Text show less

    Abstract

    Infrared-visible image patches matching is widely used in many applications, such as vision-based navigation and target recognition. As infrared and visible sensors have different imaging principles, it is a challenge for the infrared-visible image patches matching. The deep learning has achieved state-of-the-art performance in patch-based image matching. However, it mainly focuses on visible image patches matching, which is rarely involved in the infrared-visible image patches. An infrared-visible image patch matching network (InViNet) based on convolutional neural networks (CNNs) was proposed. It consisted of two parts: feature extraction and feature matching. It focused more on images content themselves contrast, rather than imaging differences in infrared-visible images. In feature extraction, the contrastive loss and the triplet loss function could maximize the inter-class feature distance and reduce the intra-class distance. In this way, infrared-visible image features for matching were more distinguishable. Besides, the multi-scale spatial feature could provide region and shape information of infrared-visible images. The integration of low-level features and high-level features in InViNet could enhance the feature representation and facilitate subsequent image patches matching. With the improvements above, the accuracy of InViNet increased by 9.8%, compared with the state-of-the-art image matching networks.
    $l({x_1},{x_2}) = \left\{ \begin{array}{l} {\mathop{d}\nolimits} (f({x_1}),f({x_2})),\;\;\;\;\;\;\;\;\;{p_1} = {p_2}\\ \max (0,{\mathop{\rm margin}\nolimits} - {\mathop{d}\nolimits} (f({x_1}),f({x_2}))),\;\begin{array}{*{20}{c}} {{p_1} \ne {p_2}} \end{array} \end{array} \right.$(1)

    View in Article

    ${\rm{max}}({d}(f({x_a}),f({x_p})) - {d} (f({x_a}),f({x_n})) + {\rm{margin}},0)$(2)

    View in Article

    ${d(}f({x_a}),f({x_p})) + {\rm{margin}} \leqslant {d}(f({x_a}),f({x_n}))$(3)

    View in Article

    Yuanhong Mao, Zhong Ma, Zhanzhuang He. Infrared-visible image patches matching via convolutional neural networks[J]. Infrared and Laser Engineering, 2021, 50(5): 20200364
    Download Citation