• Journal of Infrared and Millimeter Waves
  • Vol. 40, Issue 1, 122 (2021)
Miao LI1、*, Zai-Ping LIN1, Jian-Peng FAN1, Wei-Dong SHENG1, Jun LI1, Wei AN1, and Xin-Lei LI2
Author Affiliations
  • 1College of electronic science and technology,National University of Defense Technology,Changsha 410073,China
  • 2The Xian Chinese Space Tracking Control Center,Xian,Shanxi 710000,China
  • show less
    DOI: 10.11972/j.issn.1001-9014.2021.01.017 Cite this Article
    Miao LI, Zai-Ping LIN, Jian-Peng FAN, Wei-Dong SHENG, Jun LI, Wei AN, Xin-Lei LI. Point target detection based on deep spatial-temporal convolution neural network[J]. Journal of Infrared and Millimeter Waves, 2021, 40(1): 122 Copy Citation Text show less

    Abstract

    Point target detection in Infrared Search and Track (IRST) is a challenging task. due to less information. Traditional methods based on hand-crafted features are hard to finish detection intelligently. A novel deep spatial-temporal convolution neural network is proposed to suppress background and detect point targets. The proposed method is realized based on fully convolution network. So input of arbitrary size can be put into the network and correspondingly-sized output can be obtained. In order to meet the requirement of real time for practical application, the factorized technique is adopted. 3D convolution is decomposed into 2D convolution and 1D convolution, and it leads to significantly less computation. Multi-weighted loss function is designed according to the relation between prediction error and detection performance for point target. Number-balance weight and intensity-balance weight are introduced to deal with the imbalanced sample distribution and imbalanced error distribution. The experimental results show that the proposed method can effectively suppress background clutters, and detect point targets with less runtime.

    Introduction

    Infrared Search and Track(IRST)systems were developed to automatically search,capture and track small incoming targets from infrared sequences. They have been wildly applied to many important fields,including unmanned aerial vehicle(UAV)defense,territory surveillance,space situation awareness(SSA),precise guidance,and so on1-5. Obviously,small target detection is the key process in IRST,and the detection performance decides whether the IRST is erected success or not. Unfortunately,point target detection is still a great challenge for many negative effects.(1)Although the targets having a total spatial extent of less than 80 pixels(9×9)are defined as small target by SPIE,the targets in IRST may occupy fewer pixels(e.g.,3×3 pixels)6.(2)Very few or even no obvious textural features can be extracted by hand due to limited spatial resolution7.(3)The intensity of target signal is very weak because of distant observation.(4)The sensor noise and background clutter may be salient. In fact,the scattered clouds appearing at the edge of the big clouds are very similar as point targets. For the above reasons,the weak point targets are usually submerged by noises and clutters,and they are hard to be detected simultaneously with low false alarm ratio,high probability of detection and strong robustness.

    Many methods have been presented in the past decades,such as mean subtraction filter 8,median subtraction filter 9,TopHat filter 10,Max-Mean/Max-Median filter 11,spatial-temporal accumulative difference 12,matched filter 13,and local contrast measure(LCM)7]. However,these methods cannot detect point targets in complex background. Because they were designed based on hand-crafted features,the performances heavily depend on the completeness and accuracy of the features. Unmatched hand-crafted features will result in the performance degradation,and consequently the missed detection ratio and false-detecting ratio will increase rapidly.

    Generally,those assumptions often adopted to support hand-crafted filters include:(1)point target can be modeled as 2D Gauss model314.(2)The target has a signature of discontinuity with its neighboring regions and concentrates in a relatively small region.(3)The background is consistent with its neighboring regions. However,in practical application,single hand-crafted filter cannot well deal with different scenes,especially when there are heavy clutters in the background. In order to overcome the defects of single hand-crafted filter,the combination of multiple filters based on different features is often used in practical system. But the combination and order of different filters are still an intractable difficulty. Besides,large number parameters of multiple filters must be tuned very carefully. Unfortunately,this work is very hard for human,and the algorithm optimization heavily depend on the experience of designer.

    In this paper,a deep spatial-temporal convolution neural network based on deep learning theory is proposed to detect point targets intelligently. To detect targets from any image size,the fully convolution is adopted. Thus,the proposed network and be trained by small-size images,and tested by large-size images without any modification. In order to meet the requirement of real time,the 3D convolution in the proposed method is factorized into 2D spatial convolution and 1D temporal convolution,fewer parameters are needed and the computing burden is greatly decreased. Additional,the specified loss function is introduced to take the number imbalance and error imbalance into account simultaneously. Simulation results demonstrate that the proposed approach can robustly and effectively suppress background clutters and detect weak point targets in infrared sequences.

    The rest of the paper is organized as follows:Section 1 reviews the related work about point target detection and deep learning based methods. Section 2 shows the deep spatial-temporal convolution neural network. The network architecture,factorized 3D convolution,fully convolution and novel loss function are introduced in detail. The overall performance of the proposed method and comparison results with other methods are presented in Section 3. Finally,conclusions are drawn in Section 4.

    1 Related Work

    Benefiting from the enhancement of computer,deep learning techniques have been recently used in object detection,visual recognition,and time series classification. Especially,the deep convolutional neural network(CNN)achieved impressive results at 2015 ImageNet contest. After that,many deep learning based methods are proposed in the field of target location and identification,such as R-CNN 15,Fast R-CNN16,Faster R-CNN 17], Mask-RCNN 18,and YOLO 19-20 and. It has been proved by many researchers that deep learning based methods can automatically excavate more deep and obscure features from the raw images directly from a mass of training images,which are more beneficial to discriminate different objects than hand-crafted features.

    Although many deep learning based methods are studied in recent years,the targets involved in the above methods are large targets,called as area targets,such as human face,vehicle,and animal. The features of area targets are distinctly different from point targets in IRST. In fact,the area target may extend to hundreds of pixels with abundant texture information,geometry information,and color information,which can provide plenty of details for processing. However,the point target is extremely not obvious in shape,size and color characteristics,because they are generated from point source at long distance,and only gray information is obtained by infrared sensor. As a result,the existing detection methods based on deep learning for area targets are not suitable for point target detection in IRST.

    Some methods inspired by deep neural networks to detect point targets have been proposed recently. In brief,these methods can be divided into three categories.

    (1)Some methods simply convert the detection problem into pattern recognition problem. Ming LIU used traditional 2D CNN networks to judge whether there are infrared small targets in infrared patches 21. The input is fixed-sized in these methods. The network cannot be self-adaptive for arbitrary size,because of the fully connected layers. Essentially,these approaches can be regarded as fixed patch-wised recognition methods.

    (2)Many researchers try to combine traditional pre-detection and deep learning based recognition. At May 2018,DDP-CNN was proposed based on data-driven proposal and deep learning-based classification by Ryu 1,including region proposal step and classification step. The region proposal step is realized based hand-crafted features(mean subtract filter)to find suspected target regions,and the classification step is finished by deep learning(consisting of two convolution layers,one max pooling and two fully connected layers). So,the DDP-CNN can be regarded as partial intelligent recognition method. Besides,the pooling layers lead to low positioning accuracy.

    (3)2D CNN networks based on spatial features are used to detect small targets by some researchers. Lin’s method is designed to detect infrared targets in oversampling images 22. The oversampling sensors are very different from that of usual sensors. As a result,the target size in oversampling system may be increased several times as much,and they are considered as extended target. Larger size leads to more texture features. Besides,only spatial features are not enough for the point target detection. Thus,Lin’s method does not well deal with point target detection.

    This work will focus on detecting point targets with high performance and less runtime by deep spatial-temporal convolution network.

    2 Proposed method

    2.1 Network architecture

    For some reasons,exiting CNN networks cannot be directly used to detect point targets. Firstly,the traditional CNN networks are used to detect large area targets,and accurate spatial coordinates are thrown away due to pooling layers and fully connected layers. However,the targets in IRST are point targets,which must be located precisely by the pixel or sub-pixel. Secondly,for point target detection,each pixel may present a small target,and infrared image should be processed pixel by pixel. Thus,batch detection method is wanted. Thirdly,the input of traditional networks usually is fixed-size because of fully connected layers,which limits the flexibility in practical application. Thus,the special network must be designed based on the characters of point target.

    Fortunately,we have found some characters of point target,and they can make it possible to overcome the above shortcomings. For example,the main features of point target can be obtained by statistical analysis based on its small neighboring region. Thus,smaller receptive filed(RF)of network is enough for point targets in comparison with area targets. As a result,one hand,fewer stacked layers in network are needed,which can reduce the complexity and computation. On the other hand,pooling layers and fully connected layers,used to enlarge receptive filed by image compression and feature integration,can be given up.

    For the above reasons,the proposed network is hierarchically constructed by stacking different convolutional layers. The network architecture is shown in Fig. 1.

    The proposed network architecture.

    Figure 1.The proposed network architecture.

    The proposed point target detection network takes video clip as input,and produces a residual image with the same size of the input. The residual image represents the estimated point target intensity after background suppression. The final target index can be obtained by threshold segmentation easily.

    In Fig. 1,“C-n” indicates the index of convolution layer. The number of feature maps is denoted by the number before @,and the size of kernel is represented by the number after @. For example,the kernel with 1×3×3 represents the depth,height and width are 1 pixel,3 pixels and 3 pixels,respectively.

    This method consists of three parts. Firstly,the bottom of this network(C-1 to C-4)is a stack of 3D convolutional layers,which are focus on low-order spatiotemporal features. To improve their efficiency,the factorized 3D convolution is adopted,as introduced in Section 2.2. In second part,the 3D spatiotemporal feature maps generated from video clip are compressed into one 2D hybrid feature map. This operation is carried out by the convolution over the whole video clip in time dimension. In third part,the high-order hybrid features are intensively learned by more convolution kernels. Finally,feature fusion across different channels is achieved by 1×1 convolution in channel dimension,which makes sure that the output size is same as that of input. The 1×1 convolution is equivalent to cross-channel parametric pooling layer,and allows complex interactions of cross channel information.

    The process is modeled with a fully convolutional network. Convolutional layer is architecture with shared parameters,so all pixels can be processed by the same operation. The feature maps,input and output of each convolutional layer,can be modeled as feature results with size d×c×h×w,where dch and w are depth,the number of channels,height,width,respectively. For the first convolutional layer,the input is the infrared video clip,and the size is h×w,the length of video clip is dd is set to 5 in this paper),the number of channel is 1(because the output of infrared sensor is gray image). The output feature map indicates a particular feature representation extracted based on all locations on the input,which can be obtained via convolving the input feature map with a trainable kernel and adding a trainable bias parameter. In this work,the input feature map is denoted as X. The weights and bias of convolution kernel are represented by W and b. Thus,the output feature map can be computed by 23

    fsX;W,b=W*sX+b ,

    where *s denotes the convolution operation with stride ss is 1 in this work). Feature representation ability can be enhanced by point-wise nonlinearity operation following with each convolutional layer,and ReLU is adopted in this method. Unlike traditional methods,non-linear down-sampling operation(e.g.,max pooling)is thrown away,because pixel-wise prediction is very important for point target detection.

    2.2 Factorized 3D convolution

    Point target can be detected by 2D convolutional neural networks under smooth background. However,these methods do not provide robust detection in complex background,especially when the clutters are strong as shown in Section 3.2. The reason is that only spatial information is not enough to discriminate true or false targets. Thus,both spatial and temporal information must be fully utilized.

    2D convolution is performed only spatially,and temporal information of the input is lost. 3D convolution is done spatiotemporally,and both spatial information and temporal information of the input are preserved. Thus,3D convolution is well-suited for spatiotemporal feature learning,and it is adopted in the proposed method.

    Traditional 3D convolution can be regarded as that 2D spatial convolution and 1D depth projections are performed simultaneously. Thus,the cost of computational complexity is exceptionally high,even higher than the peak of common computers. Although many studies have proven that deep 3D convolutional neural networks can obtain spatiotemporal features even better than human level accuracy,it is beyond the applicable level. To solve the problem of real time and limit memory space in applicable application,the factorized 3D is adopted,which unravels spatial and temporal convolutions apart2425. It means that 2D spatial convolution layer and 1D temporal convolution layer are sequentially carried out as shown in Fig. 2. In this figure,k,i,j represents the pixel at k frame with the coordinationi,j. Acceptable accuracy with significantly less computation can be achieved by stacking them together.

    The sketches of 3D covolution and factorized 3D convolution: (a) 3D covolution; (b) factorized 3D convolution.

    Figure 2.The sketches of 3D covolution and factorized 3D convolution: (a) 3D covolution; (b) factorized 3D convolution.

    As shown in Fig. 2,the factorized 3D convolution is equivalent to 3D convolution. The factorization can be modeled by

    Conv3DConv2DConv1D ,

    where represents the Keonecker product. Conv3D is the 3D convolution. Conv2D and Conv1D denote the 2D convolution(spatial convolution)and 1D convolution(temporal convolution),respectively.

    The computation comparison of 3D convolution and factorized 3D convolution is listed as Table 1. The number of parameters of traditional 3D convolution with 1024 3D kernels is about 28.67K,and the requirement of computation is about 0.06GFlops. It can be approximated by 32 2D kernels and 32 1D kernels. As a result,only 9.38K parameters and 0.03GFlops are required by factorized 3D convolution. So,the factorization can reduce the computation burden by 50%.

    ItemFlops(G)Parameters(K)
    3D Conv.0.0628.67
    Factorized 3D Conv.0.039.38

    Table 1. Computation comparison of different convolutions.

    2.3 Fully convolution

    Pooling layer and fully connected layer take important role in traditional deep-learning methods for area target detection.

    Typical pooling operations include average pooling and max pooling. They can be considered as non-linear down-sampling. For example,the size of feature map is reduced to a quarter of original after 2×2 pooling layer. Obviously,the compressed feature maps are coarse and reduced-resolution,and lots of detailed information has been lost 27. Even so,such coarse features(including color,shape,texture,and so on)are enough to locate and recognize area targets. For point target,single pixel may represent one important target,so any information lost can cause unexpected consequence. Thus,the whole feature maps should be fully and intensively analyzed,and pooling layer should not appear in this network.

    Furthermore,the fully connected layer brings multiple local features from different regions together,but it limits the input size. It means that the size(height and width)of train image must be same as the size of test image.

    To overcome those shortages,the proposed network only consists of many convolution layers,called fully convolution method. The 2D convolution layers with 1×H×W kernel are used to extract spatial features,while the 1D convolution layers with D×1×1 kernel are adopted for feature extraction in time domain. Different convolution layers are connected in series,and it makes the features from both spatial convolution and temporal convolution to be deeply integrated. By choosing slip step=1,the intensity generated from point target is estimated pixel by pixel. So,pixel-wise processing is carried out by the proposed fully convolution network.

    Additionally,it has been demonstrated that small receptive fields of 3×3 convolution kernels can better learn complex features with deeper architectures26. Consequently,the spatial receptive field is set to 3×3 in the proposed network,while the temporal depth of the 3D convolution kernels are adjusted as needed.

    2.4 Multi-weighted loss function

    The multi-weighted L1 norm loss function is proposed in this work,and can be expressed as

    =1Ni=1Nj=1Mwi,jNumBalwi,jIntBalyi,j-yi,j' ,

    where N is the number of training samples,M is the number of pixels in training sample,yi,jis the ground truth for the jth pixel in ith training sample. yi,j'is the output of the proposed network. wi,jNumBalis the number-balance weighting parameter,and wi,jIntBal is the intensity-balance weighting parameter. The different weighting parameters are jointly used to trade off the false alarms and missing alarms. The loss function is minimized during the training,and it indicates that the predicted target intensity gradually reaches the truth.

    The sample imbalance encountered in training of point target detection may bring extreme error. The imbalance between target samples and background samples can overwhelm training and lead to degenerate model,especially when background samples are far more than target samples. In fact,the background samples generally belong to majority class,while the target samples are in the minority. In practical application,the background images can be easy and often obtained,however the true targets are relatively rare. If the sample imbalance cannot be solved,the training is inefficient. For example,the extreme sample imbalance may lead to true target can be completely ignored. In order to alleviate the bias in performance caused by imbalanced sample distribution,the number-balance weight wi,jNumBal is assigned to each sample to weaken the relative impact of background samples,while strengthen the impact of target samples. wi,jNumBal can be calculated by

    wi,jNumBalNiNBackground
    Ni=NBackgroundNTrueifithsampleisbackgroundifithsampleistrue target

    where NBackground and NTrue are the number of background samples and the number of target samples,respectively. The training sample includes the background samples and the target samples,i.e. N=NTrue+NBackground.

    For point target,the detection result can be obtained by threshold segmentation based on the output after background suppression. Although the error sum may be same for different predicted results,the detection results are obviously different as shown in Fig. 3. In this figure,each rectangle represents one pixel,and indicated by i,v,where i is the index of pixel and v is the predicted value. The ground truth is shown in Fig. 3(a),and all pixels belong to background. Two possible error distributions are shown in Fig. 3(b)and(c). Though the error sums of both(b)and(c)are 0.9,the result of(c)may bring more false alarms,because the 5th pixel can extremely likely over the threshold during segmentation. Thus,different weighting parameters should be adaptively given to different error distributions,called intensity-balance weight.

    The example of different error distributions: (a) the ground truth; (b) 1th predicted result with uniform error; (c) 2th predicted result with concentrated error.

    Figure 3.The example of different error distributions: (a) the ground truth; (b) 1th predicted result with uniform error; (c) 2th predicted result with concentrated error.

    In this work,the intensity-balance weight is represented by wi,jIntBal,and can be calculated by

    wi,jIntBal=1,ifyi,j-yi,j'dthyi,j-yi,j'dth,ifyi,j-yi,j'>dth

    The calculation of wi,jIntBal can be shown as Fig. 4.

    The function of intensity weighting parameter.

    Figure 4.The function of intensity weighting parameter.

    As described in Fig. 4,when the original error is greater than dth,the weighting parameter of the pixel is assigned with larger value such that the network is trained with less false alarms. dth is set based on the prior segmentation threshold.

    3 Simulation results

    3.1 Experiment scheme and evaluation metrics

    In this section,three experiments are performed to evaluate the performance of the proposed method. A large amount of infrared samples are generated based on point target model and real background images shown as Fig. 5,and the sequences are 5 frames long.

    The example of samples: (a) the target samples; (b) the background samples.

    Figure 5.The example of samples: (a) the target samples; (b) the background samples.

    In this paper,the weak target is regarded as point target,because of long-range observation. The point targets are generated by 2D Gaussian function as following 314.

    Tx,y=Ipexp-12x-xc2δx2+y-yc2δy2 ,

    where Tx,y is target projection on image planar,x and y represent the spatial coordinates,xc,yc represents the target center position,Ipdenotes the peak intensity,δx2 and δy2 are the variance in row and column direction.

    Furthermore,the observation of optical image embedded with dim point target can be obtained as following 314.

    Fx,y=Tx,y+Bx,y+Cx,y ,

    where Fx,y represents the output image,Bx,y denotes cloud background,Cx,y is noise.

    For the background samples,the ground truth is a fully zero-value image. If there are target pixels and background pixels at the same time in the training sample,the ground truth only contains the gray values of target pixels,and the others are set to zeros. The preprocessing before entering the network is necessary for all training and testing samples. In this work,the preprocessing is carried out by de-averaging and normalization. It should be noted that the same average is used for all samples,because the sequences are obtained by same sensor. As a result,the absolute intensity of point targets can be better preserves to support distinction. Besides,dth=0.2 in these experiments.

    In the first experiment,the point targets are detected by the proposed method,Lin’s method22,Max-Mean filter 11,TopHat filter10 and Spatial-Temporal Accumulative Difference method(STDA)12,respectively. The proposed method and Lin’s method belong to deep learning based solution,and the others are traditional methods based on hand-crafted filters. Besides,the STDA is a classical method based on the spatial-temporal fusion. The experiment is performed to validate the detection performance of the proposed method compared with existing methods. It illustrates the proposed method can significantly improve point target detection performance in heavy clutters.

    In the second experiment,the availability of the proposed method with different input size is proved. It demonstrates the advantage inheriting from fully convolution.

    In the third experiment,the detection performance of the proposed method is evaluated under different conditions including original signal-to-clutter ratio(SCR)and jitter of sensor.

    In order to measure the performance,the following evaluation metrics are introduced.

    To measure the ability of removing background,the background suppression factor(BSF)is introduced30. It can be computed by

    BSF=σinσout,

    where σin represents the standard deviation of original image,and σout denotes the one of output image,respectively.

    The quality of image can also be indicated by SCR. It is defined as

    SCR=μtσb,

    where μt is the target intensity without background,and σb represents the standard deviation of local background region 22.

    The comprehensive detection result is evaluated by Receiver Operating Characteristic(ROC). The ROC curve can describe the detection result by a function with the probability of detection(pd)and the probability of false alarms(pf222930. pd is defined by

    pd=NdNt,

    where Nd denotes the number of detection reports from true point targets,and Nt denotes the number of true targets. pf is defined by

    Pf=NfNi,

    where Nf represents the number of detection reports from false alarms,and Ni is the number of pixels of all testing images.

    The simulation environment in this work is shown in the Table 2.

    ItemParameter
    CPUIntel i7,2.8GHz×12
    GPUNvidia-1080Ti
    RAM64GB
    SystemUbuntu 18.04
    Disk2TB
    SoftwarePytorch 1.1
    LanguagePython 3.6

    Table 2. The simulation environment.

    3.2 Comparison with other methods

    In this simulation,10000 training samples are obtained based on 10 real background sequences. Meanwhile,10000 testing samples are obtained from another 10 real background sequences. The point taregets in samples are randomly added based on point target model. Mean SCR of original image is about 6. The jitter of sensor is simulated as random Gaussian distribution with σ=0.2. The size of infrared image is 25×25 pixels,

    Fig. 6 and Fig. 7 show the results of different methods for background images. In these figures,(a)is the original images,and(b)shows the result of the proposed method. Subsequently,(c)-(f)are the results of Lin’s method,Max-Mean,TopHat and STDA,respectively. The background in Fig. 6(a)is smooth,thus these methods can obtain good results. There are many clutter signals in Fig. 7(a),and they cannot be suppressed by traditional spatial methods(see Fig. 7(c)-(e),where the clutter signals after processing are even over 2). As shown in Fig. 7(f),benefiting from the fusion of spatial information and temporal information,the STDA achieves better result than traditional spatial methods. However,the max value of the output of the proposed method is less than 0.05,and the clutter signals can be rejected more easily.

    The original image and results of different methods for 1th Background.: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    Figure 6.The original image and results of different methods for 1th Background.: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    The original image and results of different methods for 2th Background.: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    Figure 7.The original image and results of different methods for 2th Background.: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    Fig. 8 and Fig. 9 show the results for target samples,in which the targets are marked by white circles. The target in Fig. 8 is stronger than background in the original image,and the target can be detected by the above methods. However,the target in Fig. 9 is far weaker than many clutters in the original image(the target is about 1,and some clutters are about 3). The weak point target cannot be found in the results of Max-Mean and TopHat. Those traditional spatial methods fail to catch the point target. The Lin’s method and STDA can well extract target intensity,but many clutters are still kept. So,the performance of Lin’ method and STDA greatly degrade under complex background. However,not only the weak target can be detected,but also the clutters can be well suppressed by the proposed method as shown in Fig. 9(b). The result of our method has fewer clutters compared to the other methods,which is important to keep lower false-alarm rates under the same probability of detection.

    The original image and results of different methods for Target 1: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    Figure 8.The original image and results of different methods for Target 1: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    The original image and results of different methods for Target 2: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    Figure 9.The original image and results of different methods for Target 2: (a) the original input; (b) the result of our method; (c) the result of Lin’s method; (d) the result of Max-Mean; (e) the result of TopHat; (f) the result of STDA.

    The original image and standard deviation in the time domain of Target 2 are shown in Fig. 10. In fact,Target 2 is very weak and hard to be detected just based on spatial feature or temporal feature. The proposed method can extract spatial-temporal feature to suppress background,and the fusion of spatial-temporal feature is automatically achieved.

    The display of Target 2: (a) the original gray image; (b) the standard deviation in the time domain.

    Figure 10.The display of Target 2: (a) the original gray image; (b) the standard deviation in the time domain.

    In order to intuitively show the background suppression performance,the comparisons of SCR and BSF for two point targets are listed in Table 3 and Table 4. The SCR is computed based on the output images of different methods. It is proved that the proposed method can enhance the ability of background suppression,which is very import for detection by threshold segmentation. More specifically,the mean SCRs of Max-Mean,TopHat,Lin’s method,STDA and our method are 12.1859,11.5509,14.3741,18.2125 and 19.7507,respectively. The BSFs of those methods are 1.2498,1.0821,2.6755,3.4656 and 4.2671,respectively. It is clear that the proposed method can obtain the best background suppression.

    ProposedLin’sMax-MeanTopHatSTDA
    Target 120.406116.825320.930820.039419..8378
    Target 219.095311.92283.44103.062316.5872
    Mean19.750714.374112.185911.550918.2125

    Table 3. Background suppression comparison by SCR in output.

    ProposedLin’sMax-MeanTopHatSTAD
    Target 11.35271.02931.24771.12101.3375
    Target 27.18154.32181.25201.04315.5936
    Mean4.26712.67551.24981.08213.4656

    Table 4. Background suppression comparison by BSF.

    The ROC curves of different methods are depicted in Fig. 11. The methods based on spatial-temporal fusion(including the proposed method and STDA method)demonstrate better performance than the spatial methods(including Max-Mean,Tophat and Lin’s method). The proposed method shows the best detection performance with low false alarm probability and high detection probability.

    The ROC curves of different methods.

    Figure 11.The ROC curves of different methods.

    For IRST application,the high real time is required. The comparison of average runtime is listed in Table 5. More time is needed by the proposed method than Lin’s method,because infrared sequences are put into the network to extract the spatial-temporal features in this work. However,the proposed method is still faster than Max-Mean,TopHat and STDA,and can meet the requirement of real time.

    ProposedLin’sMax-MeanTopHatSTDA

    Average

    runtime(s)/sample

    5.84×10-43.38×10-43.30×10-31.33×10-24.51×10-3

    Table 5. Average runtime comparison.

    3.3 Evaluation with different input size

    In the above experiment,the input size is set to 25×25pixels for convenience. But,it is not fixed. Benefiting from fully convolution,each pixel can be processed using the same parameters at the same time. As a result,though the proposed network is trained with the input size25×25,input of arbitrary size can be put into the network and correspondingly-sized output can be obtained at once. On the contrary,the patch-wise method is fixed.

    The image of Fig. 9(a)is extended to 35×35 pixels and 45×45 pixels as shown in Fig. 12(a)and Fig. 12(c),respectively. The position of target is changed to(25,23)and(34,28)from(16,18). Meanwhile,the results are shown in Fig. 12(b)and Fig. 12(d),respectively. Comparison shows that results in three cases are in agreement with each other. So,under sufficient resources,the proposed network is applicable to images with different sizes. The important advantages of the proposed method are shift-invariant and position invariant.

    The result of different input size: (a) the input image with 35×35 pixels; (b) the result of image with35×35 pixels; (c) the input image with 45×45 pixels; (b) the result of image with45×45 pixels.

    Figure 12.The result of different input size: (a) the input image with 35×35 pixels; (b) the result of image with35×35 pixels; (c) the input image with 45×45 pixels; (b) the result of image with45×45 pixels.

    3.4 Comparison under different conditions

    In order to illustrate the detection performance of the proposed method under different conditions,the infrared sequences with different jitters and original SCRs are tested.

    In Fig. 13,the targets are added based on mean original SCR 6,the ROC curves under different jitters are shown. The standard deviations of inter-frame jitters are set to 0.1,0.2,and 0.5,respectively. Along with the increase of jitter,the probability of detection decreases gradually. The probability of detection changes from 98% to 77%,when the false alarm ratio is 10-4. Thus,it is very important for the proposed method that keeping sensor motionless.

    The ROC curves with different jitters.

    Figure 13.The ROC curves with different jitters.

    The detection performances of the proposed method for point targets with different original SCRs are analyzed,and jitter is fixed to 0.2. Fig. 14 shows ROC curves under different SCRs,which are set to 4,6 and 8,respectively. Though some targets may be missed,the proposed method can achieve fairly high probability of detection. For example,the probability of detection can reach about 90% at the false alarm ratio is 10-4,when the mean original SCR is 6. It should be noted that the SCRs of some targets in this test sets are lower than mean SCR.

    The ROC curves with different mean original SCRs.

    Figure 14.The ROC curves with different mean original SCRs.

    4 Conclusions

    In IRST,point detection is still a great challenge for some reasons. Traditional methods can’t robustly and intelligently detect point targets in complex background. In this work,a deep spatial-temporal convolution neural network is proposed to address this problem. The network is built based on fully convolution without pooling layer and fully connected layer,factorized 3D convolution and multi-weighted loss function are adopted to enhance the performance. The proposed method is compared to other four methods,including traditional methods(e.g.,Max-Mean filter TopHat filter and Spatial-Temporal Accumulative Difference method)and deep learning based method(e.g.,Lin’s method). The detection performance is evaluated by different metrics,such as signal-to-clutter ratio,background suppression factor. Meanwhile,ROC curves are drew to confirm the robustness of the proposed approach. Additionally,the comparison under different conditions is carried out for the proposed method,and the affections of original SCR and sensor’s jitter are demonstrated in detail. Consequently,the deep spatial-temporal convolution neural network can effectively detect point targets using less runtime.

    References

    [1] R Junhwan, K Sungho. Small infrared target detection by data-driven proposal and deep learning-based classification, 10621J-10624J(2018).

    [2] X Sui, Q Chen, L Bai. Detection algorithm of targets for infrared search system based on area infrared focal plane array under complicated background. Optik, 123, 235-239(2012).

    [3] J Zhao, Z Tang, J Yang et al. Infrared small target detection using sparse representation. Journal of systems engineering and electronics, 22, 897-904(2011).

    [4] M Li, Z Lin, Y Long et al. Joint detection and tracking of size-varying infrared targets based on block-wise sparse decomposition. Infrared Physics and Technology, 76, 131-138(2016).

    [5] J Gao, Z Lin, W An. Infrared small target detection using a temporal variance and spatial patch contrast filter. IEEE Access, 7, 32217-32226(2019).

    [6] W Zhang, M Cong, L Wang. Algorithms for optical weak small targets detection and tracking: Review, 643-647(2003).

    [7] C Chen, H Li, Y Wei, T Xia et al. A local contrast method for small infrared target detection. IEEE transactions on geoscience and remote sensing, 52, 574-581(2014).

    [8] R Warren. Detection of distant airborne targets in cluttered backgrounds in infrared image sequences(2002).

    [9] J Barnett. Statistical analysis of median subtraction filtering with application to point target detection in infrared backgrounds, 10-18(1989).

    [10] V Tom, T Peli, M Leung, J Bondaryk. Morphology-based algorithm for point target detection in infrared backgrounds, 2-11(1993).

    [11] S D Deshpande, H E Meng, R Venkateswarlu et al. Max-mean and max-median filters for detection of small targets, 74-83(1999).

    [12] Q Yu, S Huang, W Zhao et al. A fusion detection algorithm of small infrared target based on spatial-temporal accumulative difference. Journal of projectiles, rockets, missiles and guidance, 34, 181-189(2014).

    [13] W Schmidt. Modified matched filter for cloud clutter suppression. IEEE transactions on pattern analysis and machine intelligence, 12, 594-600(1990).

    [14] C Gao, D Meng, Y Yang et al. Infrared patch-image model for small target detection in a single image. IEEE transactions on image processing, 22, 5996-5009(2013).

    [15] R Girshick, J Donahue, T Darrell. Region-based convolutional networks for accurate object detection and segmentation. IEEE Transactions on pattern analysis and machine intelligence, 38, 142-158(2016).

    [16] R Girshick. Fast R-CNN, 1440-1448(2015).

    [17] S Ren, K He, R Girshick et al. Faster RCNN: Towards real-time object detection with region proposal networks, 91-99(2015).

    [18] K He, G Gkioxari, P Dollar. Mask R-CNN, 2980-2988(2017).

    [19] J Redmon, S Divvala, R Gishick. You only look once: Unified, real-time object detection, 779-788(2016).

    [20] R Joseph, A Farhadi. YOLO3: An incremental improvement. ArXiv preprint, 1-6(2018).

    [21] M Liu, H Du, Y Zhao et al. Image small target detection based on deep learning with SNR controlled sample generation, 211-220(2017).

    [22] L Lin, S Wang, Z Tang. Using deep learning to detect small targets in infrared oversampling images. Journal of Systems Engineering and Electronics, 5, 947-952(2018).

    [23] G Ian, B Yoshua, C Aaron. Deep learning(2017).

    [24] M Wang, B Liu, F Hassan. Factorized convolutional neural networks, 1-10(2017).

    [25] L Sun, K Jia, D Yeung et al. Human action recognition using factorized spatio-temporal convolutional networks, 4597-4605(2015).

    [26] K Simonyan, A Zisserman. Very deep convolutional networks for large-scale image recognition. ArXiv preprint, 1409-1556(2014).

    [27] J Long, S Evan, D Trevor. Fully convolutional networks for semantic segmentation, 1-10(2015).

    [28] S Kim. Double layered-background removal filter for detecting small infrared targets in heterogenous backgrounds. Journal of infrared, millimeter and terahertz waves, 32, 79-101(2011).

    [29] C Guo, M Deyu, Y Yi et al. Infrared patch-image model for small target detection in a single image. IEEE transactions on image processing, 22, 4996-5009(2013).

    [30] Y He, M Li, J Zhang et al. Small infrared target detection based on low-rank and sparse representation. Infrared physics & technology, 68, 98-109(2015).

    Miao LI, Zai-Ping LIN, Jian-Peng FAN, Wei-Dong SHENG, Jun LI, Wei AN, Xin-Lei LI. Point target detection based on deep spatial-temporal convolution neural network[J]. Journal of Infrared and Millimeter Waves, 2021, 40(1): 122
    Download Citation