[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Separation of Temperature-Induced Response for Bridge Long-Term Monitoring Data Using Local Outlier Correction and Savitzky–Golay Convolution Smoothing
Previous Article in Journal
COVID-Net USPro: An Explainable Few-Shot Deep Prototypical Network for COVID-19 Screening Using Point-of-Care Ultrasound
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast and Robust Infrared Small Target Detection Using Weighted Local Difference Variance Measure

Key Laboratory of Signal Detection and Processing, Department of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(5), 2630; https://doi.org/10.3390/s23052630
Submission received: 9 January 2023 / Revised: 22 February 2023 / Accepted: 24 February 2023 / Published: 27 February 2023
(This article belongs to the Section Sensor Networks)
Figure 1
<p>The proposed algorithm framework.</p> ">
Figure 2
<p>(<b>a</b>) The double-layer window. (<b>b</b>) The tri-layer window. (<b>c</b>) The new tri-layer filtering window. (<b>d</b>) Gaussian distribution characteristics of IR small target. (<b>e</b>) Situations that the algorithm needs to handle.</p> ">
Figure 3
<p>SMs of different algorithms. (<b>a1</b>–<b>a9</b>) Nine representative IR images. (<b>b1</b>–<b>b9</b>) LCM. (<b>c1</b>–<b>c9</b>) MPCM. (<b>d1</b>–<b>d9</b>) RLCM. (<b>e1</b>–<b>e9</b>) TLLCM. (<b>f1</b>–<b>f9</b>) VAR_DIFF. (<b>g1</b>–<b>g9</b>) ADMD. (<b>h1</b>–<b>h9</b>) WSLCM. (<b>i1</b>–<b>i9</b>) Proposed methods. (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>a</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>a</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math>), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>b</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>b</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>c</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>c</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>d</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>d</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>e</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>e</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>f</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>f</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>g</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>g</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>h</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>h</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), and (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mrow> <mi mathvariant="normal">i</mi> <mn mathvariant="bold">1</mn> </mrow> </mrow> <mo stretchy="true">¯</mo> </mover> <mtext>–</mtext> <mover accent="true"> <mrow> <mrow> <mi mathvariant="normal-bold">i</mi> <mn mathvariant="bold">9</mn> </mrow> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ) are the 3D gray distribution maps of (<b>a1</b>–<b>a9</b>), (<b>b1</b>–<b>b9</b>), (<b>c1</b>–<b>c9</b>), (<b>d1</b>–<b>d9</b>), (<b>e1</b>–<b>e9</b>), (<b>f1</b>–<b>f9</b>), (<b>g1</b>–<b>g9</b>), (<b>h1</b>–<b>h9</b>), and (<b>i1</b>–<b>i9</b>), respectively.</p> ">
Figure 3 Cont.
<p>SMs of different algorithms. (<b>a1</b>–<b>a9</b>) Nine representative IR images. (<b>b1</b>–<b>b9</b>) LCM. (<b>c1</b>–<b>c9</b>) MPCM. (<b>d1</b>–<b>d9</b>) RLCM. (<b>e1</b>–<b>e9</b>) TLLCM. (<b>f1</b>–<b>f9</b>) VAR_DIFF. (<b>g1</b>–<b>g9</b>) ADMD. (<b>h1</b>–<b>h9</b>) WSLCM. (<b>i1</b>–<b>i9</b>) Proposed methods. (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>a</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>a</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math>), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>b</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>b</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>c</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>c</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>d</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>d</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>e</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>e</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>f</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>f</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>g</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>g</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>h</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>h</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), and (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mrow> <mi mathvariant="normal">i</mi> <mn mathvariant="bold">1</mn> </mrow> </mrow> <mo stretchy="true">¯</mo> </mover> <mtext>–</mtext> <mover accent="true"> <mrow> <mrow> <mi mathvariant="normal-bold">i</mi> <mn mathvariant="bold">9</mn> </mrow> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ) are the 3D gray distribution maps of (<b>a1</b>–<b>a9</b>), (<b>b1</b>–<b>b9</b>), (<b>c1</b>–<b>c9</b>), (<b>d1</b>–<b>d9</b>), (<b>e1</b>–<b>e9</b>), (<b>f1</b>–<b>f9</b>), (<b>g1</b>–<b>g9</b>), (<b>h1</b>–<b>h9</b>), and (<b>i1</b>–<b>i9</b>), respectively.</p> ">
Figure 3 Cont.
<p>SMs of different algorithms. (<b>a1</b>–<b>a9</b>) Nine representative IR images. (<b>b1</b>–<b>b9</b>) LCM. (<b>c1</b>–<b>c9</b>) MPCM. (<b>d1</b>–<b>d9</b>) RLCM. (<b>e1</b>–<b>e9</b>) TLLCM. (<b>f1</b>–<b>f9</b>) VAR_DIFF. (<b>g1</b>–<b>g9</b>) ADMD. (<b>h1</b>–<b>h9</b>) WSLCM. (<b>i1</b>–<b>i9</b>) Proposed methods. (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>a</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>a</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math>), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>b</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>b</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>c</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>c</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>d</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>d</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>e</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>e</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>f</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>f</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>g</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>g</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>h</mi> </mstyle> <mn mathvariant="bold">1</mn> </mrow> <mo stretchy="true">¯</mo> </mover> <mo>–</mo> <mover accent="true"> <mrow> <mstyle mathvariant="bold" mathsize="normal"> <mi>h</mi> </mstyle> <mn mathvariant="bold">9</mn> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ), and (<math display="inline"><semantics> <mrow> <mover accent="true"> <mrow> <mrow> <mi mathvariant="normal">i</mi> <mn mathvariant="bold">1</mn> </mrow> </mrow> <mo stretchy="true">¯</mo> </mover> <mtext>–</mtext> <mover accent="true"> <mrow> <mrow> <mi mathvariant="normal-bold">i</mi> <mn mathvariant="bold">9</mn> </mrow> </mrow> <mo stretchy="true">¯</mo> </mover> </mrow> </semantics></math> ) are the 3D gray distribution maps of (<b>a1</b>–<b>a9</b>), (<b>b1</b>–<b>b9</b>), (<b>c1</b>–<b>c9</b>), (<b>d1</b>–<b>d9</b>), (<b>e1</b>–<b>e9</b>), (<b>f1</b>–<b>f9</b>), (<b>g1</b>–<b>g9</b>), (<b>h1</b>–<b>h9</b>), and (<b>i1</b>–<b>i9</b>), respectively.</p> ">
Figure 4
<p>The ROC curves of different detection methods. (<b>a</b>−<b>i</b>) The experimental results of datasets 1−9.</p> ">
Figure 5
<p>(<b>a</b>) Representative images of raw IR dataset 3. (<b>b</b>) Gaussian white noise. (<b>c</b>) Poisson noise. (<b>d</b>) Rayleigh noise. (<b>e</b>) Multiplicative noise. (<b>f</b>) Uniform noise. (<b>a1</b>–<b>f1</b>) The 3D gray distribution maps of (<b>a</b>–<b>f</b>).</p> ">
Figure 6
<p>The number of missed images in different algorithms under the influence of noise.</p> ">
Figure 7
<p>The number of false-alarm images in different algorithms under the influence of noise.</p> ">
Versions Notes

Abstract

:
Infrared (IR) small-target-detection performance restricts the development of infrared search and track (IRST) systems. Existing detection methods easily lead to missed detection and false alarms under complex backgrounds and interference, and only focus on the target position while ignoring the target shape features, which cannot further identify the category of IR targets. To address these issues and guarantee a certain runtime, a weighted local difference variance measure (WLDVM) algorithm is proposed. First, Gaussian filtering is used to preprocess the image by using the idea of a matched filter to purposefully enhance the target and suppress noise. Then, the target area is divided into a new tri-layer filtering window according to the distribution characteristics of the target area, and a window intensity level (WIL) is proposed to represent the complexity level of each layer of windows. Secondly, a local difference variance measure (LDVM) is proposed, which can eliminate the high-brightness background through the difference-form, and further use the local variance to make the target area appear brighter. The background estimation is then adopted to calculate the weighting function to determine the shape of the real small target. Finally, a simple adaptive threshold is used after obtaining the WLDVM saliency map (SM) to capture the true target. Experiments on nine groups of IR small-target datasets with complex backgrounds illustrate that the proposed method can effectively solve the above problems, and its detection performance is better than seven classic and widely used methods.

1. Introduction

Infrared (IR) imaging technology has been widely used in civilian fields such as car navigation, diseased-cell diagnosis, industrial-flaw detection, physiological performance of animal life processes, and plant monitoring [1]. It is worth noting that the infrared search and track (IRST) system based on IR imaging technology has the advantages of passive surveillance, all-weather use, and high spatial resolution, and uses the difference in thermal radiation between the target and the background to achieve long-distance target detection [2,3]. It has very important application value in military fields such as precision guidance, early-warning systems, space-based surveillance, and geological analysis [4,5]. IR small-target detection plays a vital role in these applications. To find the target as early as possible, long-distance detection and tracking are required, so the target has few pixel and texture features and lacks shape and structure information [6]. Furthermore, targets are usually immersed in complex backgrounds, and targets can be affected by a low signal-to-clutter ratio (SCR) [7]. Therefore, IR small-target detection is still a difficult and challenging task.
IR small-target detection methods in complex scenes can be divided into sequence detection methods and single frame detection methods [8,9]. Compared with the sequence detection method, the single frame detection method has a small amount of calculation and strong scene adaptability. Since real-time target detection becomes urgent in the military application of an IRST system, research based on single frame detection method is very necessary [10,11].
Existing single-frame detection methods can be divided into four categories. The first category is based on filtering methods, which are divided into algorithms based on spatial filtering and algorithms based on frequency-domain filtering. The algorithm based on spatial filtering is simple in design, fast in calculation speed, and has better performance in a uniform background, but it is easy to cause false detection in a complex background and has poor robustness [12,13]. Although algorithms based on frequency-domain filtering can suppress complex backgrounds, they have high computational complexity [14,15]. The second category is based on low-rank sparse restoration methods, which have high detection performance under strong noise background conditions, but have high computational complexity when dealing with large-scale images [16,17,18,19]. The third category is methods based on deep learning, which can improve the detection accuracy of small targets to a certain extent, but lack many datasets in various forms, which is challenging [20,21,22,23,24]. The fourth category is methods based on the human visual system. This system is relatively real-time and it is not easy to lose target features during the detection process and, but it is easy to cause false positives in complex scenes [25,26,27,28,29,30,31,32,33,34,35]. Given the importance of real-time detection and detection rate, this paper was inspired by the human visual system, a brief overview of detection methods based on the human visual system follows.
IR small-target detection algorithms based on the local-contrast method of the human visual system have attracted much attention. These algorithms focus on the differences between the target and the background surrounding it. For instance, Chen et al. [5] proposed a local contrast measure (LCM) that uses nested windows with eight orientations to suppress background edges; Han et al. [25] proposed an improved LCM (ILCM) that uses the target area average to suppress pixel-sized noise with high brightness (PNHB); Han et al. [26] proposed the relative LCM (RLCM) computed by combining ratio differences, and then generalizing it to the sub-block level [27]; Wei et al. [28] used the multi-scale patch-based contrast measure (MPCM) algorithm to fuse the corresponding two directions into a whole to capture the target; Han et al. [29] adopted a multi-scale three-layer local contrast measure (TLLCM), used Gaussian filtering to enhance the target area, and took the average value of several largest pixels in the surrounding area; Moradi et al. [30] proposed absolute directional mean difference (ADMD), which uses an orientation method to suppress the structural background; and Zhang et al. [20] proposed a multi-scale strengthened directional difference (MSDD) algorithm, which combines the local directional-intensity measure and the local directional-fluctuation measure to effectively suppress the angular clutter. Furthermore, in existing studies, many researchers are keen to employ weighting functions on top of basic local-contrast algorithms to improve detection performance. For example, Qin et al. [10] used the variance of the central unit as the weight function; Deng et al. [31] improved the local entropy as the weight function; Nasiri et al. [32] used the center and surrounding variance difference (VAR_DIFF) as the weighting function; Liu et al. [33] proposed a weighted LCM, which defines a weighting function based on the strong clutter edge features; Lv et al. [34] proposed the regional intensity level (RIL) algorithm to assess the complexity level of each unit, taking the RIL difference between the central unit and its surrounding background as a weighting function; and Han et al. [35] proposed weighted strengthened LCM (WSLCM) and proposed an improved RIL (IRIL) that replaces the maximum with the average of several maximum grayscale calculations.
The weighted LCM using more local information can reduce the false-alarm rate to a certain extent. However, there are still some problems. First, current algorithms usually directly compute the contrastive information between the target area and surrounding areas, but when the target scale is small, the edge information cannot be captured for effective enhancement. Second, some weighting algorithms increase the time of image processing during detection. Third, the existing methods do not sufficiently consider the shape of the true target, and the detection process is easily disturbed by noise.
To better enhance targets of different scales in different complex scenes, ensure a certain detection time, better preserve target shape characteristics, and reduce false-alarm rates, a detection framework based on weighted local difference variance measure (WLDVM) is proposed. First, the image is preprocessed by Gaussian filtering, and then according to the distribution characteristics in the target area, the target area is divided into a new tri-layer filtering window and the window intensity level (WIL) value of each layer of windows is calculated. Second, the local difference variance measure (LDVM) and weighting function are calculated by ratio and difference operations using the obtained position and WIL value of each layer window. Finally, a simple threshold is used to segment the fused result WLDVM to capture the true target. The contributions of this paper are as follows:
  • The new tri-layer filtering window is proposed. The target area is divided according to its distribution characteristics and size, which can adapt to the detection of targets of different scales and save detection time.
  • WIL is proposed. Each layer of window uses the mean of the two largest subblock averages instead of the single largest subblock average to better capture the target and suppress edge noise.
  • LDVM is proposed. Through the idea of local fluctuation, the target area is further enhanced, and the high-brightness background is eliminated.
  • A detection framework based on WLDVM is proposed. The experimental results using multiple sets of IR datasets show that the proposed algorithm has the best detection performance and consumes less time.

2. Proposed Algorithm

Figure 1 shows the proposed WLDVM algorithm framework. First, the image is preprocessed by Gaussian filtering, and the WIL values of each layer are calculated through the new tri-layer filtering window. Then, according to the WIL value and location of each layer, the idea of local fluctuation and background estimation is introduced to calculate LDVM and weighting function. The true small target is the most prominent in the final weighted result, which can be easily captured with a simple threshold segmentation.

2.1. Gaussian Filtering Pre-Processing

Small targets in IR image usually have a low SCR and are susceptible to noise interference because of the effects of long-distance and atmospheric transmission. These factors make detection more difficult, requiring noise suppression and target enhancement. The best filter for improving the target should have the same distribution as the target, according to the matched filter theory [36] and given that small IR targets have Gaussian-like properties and that Gaussian filters are excellent at suppressing high-frequency IR image components including scattered noises, Gaussian noises, and PNHB [29,37]. In this study, noise is reduced and small targets are enhanced using Gaussian filtering. The result of the Gaussian filtering operation is expressed as
G I ( x , y ) = l = 1 1 k = 1 1 G ( l , k ) I ( x + l , y + k ) ,   G = 1 16 [ 1     2     1 2     4     2 1     2     1 ] .
where  G  is the Gaussian template and  I  is the original IR image.

2.2. Construction of the New Tri-Layer Filtering Window

Traditional LCM and its improved algorithms adopt a double-layer filtering window, the central unit captures the target area, and the surrounding units capture the background area around the target, see Figure 2a. But when the scale of the true target area is smaller than the central unit scale, the detected target will be enlarged. Therefore, Nasiri et al. [32] made an improvement and proposed a three-layer nested window to divide the central unit into two parts, namely the core layer and the reserve layer. The core layer captures the main energy of the target area, and the reserve layer separates the target from its surrounding units, see Figure 2b. Usually, PNHB in complex backgrounds is difficult to suppress because its core layer differs significantly from surrounding layers.
It is well known that the real target area has a compact two-dimensional Gaussian shape distribution whose intensity weakens towards the surroundings, as shown in Figure 2d, while PNHB does not possess such a distribution. In this paper, according to the target area distribution characteristics in Figure 2d, a new tri-layer filtering window is proposed to capture the target area, namely inner layer ( T 0  yellow area), middle layer ( T 1  green area), and outer layer ( T 2  blue area), see Figure 2c. According to SPIE, the total spatial extent of the small target is usually less than 80 pixels [5]. Therefore, the inner layer is set to 1 × 1; through four directions and the middle layer and the outer layer are each divided into four subblocks. The subblock of the middle layer is a symmetrical trapezoid with a height of 2, an upper base of 1, and a lower base of 3. The subblock of the outer layer is a symmetrical trapezoid with a height of 2, an upper base of 5, and a lower base of 7. The proposed new tri-layer filtering window can adapt to the detection of targets of different scales, and its total space is small, which will make the algorithm run faster.

2.3. Calculation of the Window Intensity Level (WIL)

Apply the new tri-layer filtering window from top to bottom and left to right on the Gaussian filtered image and follow the steps below to calculate the WIL value for each layer of each pixel.
1.
For the inner layer:
WIL T 0 = G I T 0
where  G I T 0  is the pixel in cell  T 0  of the Gaussian filtered image  G I .
2.
For the middle and outer layers:
First, the average value of each subblock in the layer is calculated as the key parameter for the next calculation:
M T i j = 1 N T i j   k = 1 N T i j G I T i j k i = 1 , 2 ; j = 1 , 2 , 3 , 4
where  N T i j  is the total number of pixels in cell  T i j , and  G I T i j k  is the gray value of the  k th pixel in cell  T i j .
WIL T i  is the mean of the  m  largest  M T i j  values in  T i  area, that is,
WIL T i = 1 m l = 1 m M T i l i = 1 , 2 ; m = 2
where  M T i l  is the  l th largest  M T i j  value in the  T i  area. The distribution trend of the cloud layer is a gradual process; the interior of the cloud layer changes slowly, and the gray value of the edge fluctuates greatly, as shown in Figure 2e. With this type of edge it is easy to cause the occlusion of the weak target, and the gray value of the inner layer of the small target at the cloud edge is at least not much different from the average gray value of a sub-block of other layers. To effectively enhance this type of small target to avoid missed detection,  m  in Equation (4) needs to be greater than 1. When  m  is greater than 2, edge clutter will be enhanced to cause false alarms, so 2 is the most suitable value for  m .

2.4. Local Difference Variance Measure (LDVM)

The local contrast in the form of differences can eliminate the high-brightness background. The difference of WIL is defined by the difference between layers as
D o W I L = { WIL T q WIL T p ,   p > q 0 , others
where  WIL T q  indicates that the maximum value in  WIL T i  is in the  T q  layer and  WIL T p  indicates that the minimum value in  WIL T i  is in the  T p  layer. Clutter can be further suppressed by non-negative constraints.
Through the above calculation, there are cases where pixels are suppressed at the edges inside the target area. To prevent these pixels from being suppressed by further calculations, this paper enhances areas with large local fluctuations by computing the mean filtering of the square of the image minus the square of its mean filtering. The LDVM of each pixel is defined as
LDVM ( x , y ) = M 2 L ( x , y ) ( M L ( x , y ) ) 2
where  M 2 L  and  M L  are defined as
M 2 L ( x , y ) = l = 2 2 k = 2 2 M F ( l , k ) ( D o W I L ( x + l , y + k ) ) 2
M L ( x , y ) = l = 2 2 k = 2 2 M F ( l , k ) D o W I L ( x + l , y + k )
where  M F  is a 5 × 5 normalized mean filtering template. Obviously, the gray value of the local area of the pixels at the edges inside the target area fluctuates greatly, so these pixels are effectively enhanced.

2.5. Weighting Function

The local contrast in the form of ratio can enhance the true target. The ratio of WIL is defined by the difference between layers as
R o W I L ( x , y ) = { WIL T q / WIL T p ,   p > q 0 , others .
Mean filtering can reduce the sharp change of image gray value to achieve the purpose of smoothing the image. In this paper, the background estimation is performed by mean filtering as
B E ( x , y ) = l = 2 2 k = 2 2 M F ( l , k ) G I ( x + l , y + k )
Although  R o W I L  as an enhancement factor can effectively enhance the target area, there is still a lot of background clutter. In this paper, background estimation is used to calculate the weight function of each pixel in the form of ratio difference combination to suppress part of the background clutter, which is defined as
W ( x , y ) = m a x { 0 , R o W I L ( x , y ) G I ( x , y ) B E ( x , y ) }
In general, the weight of the true target is very large, and its surrounding local background is completely suppressed, so the weighting function fully considers the shape of the target.

2.6. Weighted Local Difference Variance Measure (WLDVM)

The LDVM and weighting function are fused to obtain the WLDVM of the current pixel, that is
WLDVM ( x , y ) = W ( x , y ) LDVM ( x , y ) .
The calculation of LDVM can better eliminate the high-brightness background and make the whole target appear high-brightness. The operation result of the weighting function can fully consider the shape of the target. The WLDVM algorithm preserves the shape of the original target, the target is effectively enhanced, and the background is effectively suppressed. In most cases the target size is unknown and multi-scale detection is required. In this paper, multi-scale detection is not required, and efficient detection can be performed, which greatly saves detection time.

2.7. Threshold Operation

The saliency map (SM) of each IR image can be obtained by computing WLDVM and the different results produced by the pixels of different situations are analyzed.
  • For a pixel in the real target area, since the target area often presents a compact two-dimensional Gaussian shape, its  D o W I L  will be large and  R o W I L > 1 , and its  LDVM  and weight will be very large. Hence, the resulting value of  WLDVM  will be large.
  • For a pixel in the pure background area, since the pure background area is often continuous and evenly distributed, its  D o W I L 0  and  R o W I L 1 , then its  LDVM 0  and  W 0 . Therefore,  WLDVM 0 .
  • For a pixel at the edge of the background, its  D o W I L  may be greater than 0 but less than that of the true target, so its  LDVM  is much less than that of the true target; in addition,  R o W I L  may be greater than 1, but its enhancement effect is not much different from the local background estimation, so the corresponding  W  will be less than the true target’s  W . Hence, its  WLDVM  is much less than that of the true target.
  • For a pixel in the PNHB area, its  D o W I L  will be less than that of the true target, and thus its  LDVM  will be less than the true target’s  LDVM ; in addition, its  W  will be less than the true target’s  W . Hence, its  WLDVM  is much less than that of the true target.
As can be seen from this discussion, the true target area will be the most salient in SM, so a simple threshold operation is used to extract it, the threshold is defined as
T h = λ m a x S M + ( 1 λ ) m e a n S M
where  m a x S M  is the maximum gray value of SM, and  m e a n S M  is the average gray value of SM.  λ  is an experimental constant between 0 and 1. In the experimental part, the value of  λ  is analyzed in detail, and the experiment shows that  λ  can take any value between 0.5 and 0.6.

3. Experimental Results

To demonstrate the detection performance of the proposed algorithm, nine groups of IR datasets were used, including three sets of real IR sequences (datasets 1, 3, and 4), five sets of simulated IR sequences (datasets 2, 5, 6, 7, and 8), and one non-sequential dataset (dataset 9). Datasets are shown in [5,38,39,40,41,42,43]. The targets in dataset 1 are all immersed in very complex dense cloud cover and most of the targets have very low contrast. The targets in dataset 2 are all immersed in a dimly lit background. The target in dataset 3 moves from the cloud layer to the cloudless area, some targets have low contrast, and the background contains a lot of noise. The aircraft target in dataset 4 is large in scale and immersed in a cloudless area with a few thin clouds in the background. The target in dataset 5 moves from a cloudless dark background area into thin cloud cover. The target in dataset 6 is immersed in a complex air and sea background, which contains many PNHBs. Targets in dataset 7 move from a background containing buildings. The targets in dataset 8 are immersed in complex and changing land backgrounds. Dataset 9 consists of representative images of different sequences, with both targets and backgrounds differing between images. Additional details are shown in Table 1.
First, we analyzed the effect of  λ  value on detection performance. Table 2 shows the number of false-alarm images  N F A  and the number of missed images  N M D  corresponding to different datasets under different values of  λ , where  λ  increases from 0 to 1 with a step size of 0.1. The experiments showed that when the value of  λ  was between 0.5 and 0.6, small targets in different complex scenes could be effectively captured, there were no missed detections and false alarms in any dataset, and high classification accuracy could be obtained.
Then, seven LCM-based algorithms were selected from multiple perspectives for comparison with the proposed algorithm, including LCM [5], MPCM [28], RLCM [26], TLLCM [29], VAR_DIFF [32], ADMD [30], and WSLCM [35]. Among them, VAR_DIFF and TLLCM are local-contrast algorithms based on tri-layer windows, and the rest are local-contrast algorithms based on double-layer windows; RLCM, TLLCM, and WSLCM are local-contrast methods using ratio difference joint operations; and VA_DIFF and WSLCM are local-contrast algorithms that use the weighting function.
To analyze different methods intuitively, Figure 3 shows the SMs of different algorithms. Each dataset’s original image sample may be found in the first column. The target size was variable, the backdrop was intricate, and there were various levels of noise present. As shown in the second column of the figure, LCM enhanced the target and made the target area larger, while enhancing the noise, and the background suppression effect was not good. As shown in the third column of the figure, MPCM enhanced the target but did not preserve the target shape very well and had a certain suppression effect on the background and noise, but when the background was more complex, the detection effect was not good. As shown in the fourth column of the figure, RLCM enhanced the target and made the target area larger and had a general effect on background and noise suppression. As shown in the fifth column, TLLCM had a mediocre level of noise and background suppression efficiency, but the detection effect was poor when the background was complicated. As shown in the sixth, seventh, and eighth columns of the figure, VAR_DIFF, ADMD, and WSLCM had better background suppression effects, but when the background was complex, the noise suppression effect was average, and the detection performance was unstable. As shown in the ninth column of the figure, the proposed method effectively improved the target SCR and better preserved the target outline, could better suppress the background and noise, and the detection performance was the best.
To illustrate the detection performance of these algorithms, the indicator’s signal-to-clutter ratio gain (SCRG) and background suppression factor (BSF) before thresholding are used simultaneously, and defined as
S C R G = S C R o u t S C R i n ,   B S F = σ i n σ o u t   ,   S C R = | m t m b | σ b .
where  S C R i n  is the SCR value of the original image,  S C R o u t  is the SCR value of the SM,  δ i n  is the standard deviation of the non-target area in the original image and  δ o u t  is the standard deviation of the non-target area in the SM, and  m t  is the mean of the target area,  m b  and  σ b  are the mean and standard deviation of the local background area around the target, respectively. It can be seen in Table 3 that VAR_DIFF had one set with the highest SCRG value, WSLCM had two sets with the highest SCRG value, and the proposed algorithm had six sets with the highest SCRG value and the highest average SCRG value. The results show that the proposed method achieved more significant target enhancement before thresholding than other methods. VAR_DIFF had one set with the highest BSF value, WSLCM had five sets with the highest BSF value, and the proposed algorithm had three sets with the highest BSF value and the highest mean. It shows that the background-suppression ability of the proposed algorithm is equivalent to that of the WSLCM algorithm, and better than that of other algorithms.
Figure 4 depicts the receiver operating characteristic (ROC) curves for different algorithms to evaluate the target-enhancement ability and background-suppression ability after thresholding, where the false-positive rate (FPR) and the true-positive rate (TPR) are the horizontal and vertical coordinates of the ROC curve [44], respectively, and are defined as
FPR = N f a l s e N p i x e l   ,   TPR = N d e t e c t e d N t u r e   .
where  N f a l s e  is the number of detected false targets,  N p i x e l  is the total number of pixels in the whole image,  N d e t e c t e d  is the number of detected true targets, and  N t u r e  is the total number of true targets.
In the ROC curve, the more the curve shifts to the upper left corner, the better the detection performance will be. Under the same FPR, the larger the TPR, the better the performance of the algorithm. As can be seen from Figure 4, when  FPR = 10 5 :
  • LCM had a TPR greater than 0.9 and less than 1 in dataset 4, and performed poorly in other datasets;
  • MPCM had a TPR greater than 0.9 and less than 1 in dataset 5, and performed poorly in other datasets;
  • RLCM achieved the highest TPR in dataset 2, and performed poorly in other datasets;
  • TLLCM achieved the highest TPR in dataset 2, TPR greater than 0.9 and less than 1 in dataset 8, while performing poorly in other datasets;
  • VAR_DIFF achieved the highest TPR in datasets 2 and 4, TPR greater than 0.8 and less than 1 in datasets 1, 7, 8, and 9, and performed poorly in other datasets;
  • ADMD achieved the highest TPR in dataset 7, while performing poorly in other datasets;
  • WSLCM achieved the highest TPR in datasets 2, 4, 6, 7, and 8, and the TPR was greater than 0.8 and less than 1 in datasets 1, 3, 5, and 9;
  • The proposed algorithm achieved the highest TPR in all nine datasets.
Obviously, the proposed algorithms achieved satisfactory results, but the existing algorithms were affected by varying degrees of background clutter, resulting in algorithm instability. Compared with existing algorithms, the proposed algorithm was more stable, could effectively handle different scenarios, and had the best detection performance.
Table 4 reports the full specification of the implementation environment. The mean runtime was used to demonstrate the computational complexity of different detection algorithms. As can be seen in Table 5, the VAR_DIFF algorithm was faster than other existing algorithms, and the proposed algorithm was second only to the VAR_DIFF algorithm. Although our method was not the most time efficient, it was still relatively fast.
It can be seen from all the above experimental results that none of the existing algorithms could preserve the shape characteristics of the target well. Among them, LCM diffused the target, RLCM diffused the smaller scale target; and LCM, MPCM, RLCM, and TLLCM had poor background suppression. Although VAR_DIFF, ADMD, and WSLCM had strong background suppression capabilities, the detection rate of ADMD was average, and VAR_DIFF and WSLCM had missed detection in scenes with low SCR targets. Although only WSLCM had a relatively low false-alarm rate among existing algorithms, the detection of the WSLCM algorithm is particularly time-consuming; however, the proposed algorithm can preserve the target shape well, has strong background suppression ability, high detection rate, low false-alarm rate, and faster detection speed. In general, the proposed algorithm can effectively preserve the shape features of targets of different scales and types, and can adapt to detection in different scenarios, which further guarantees the speed of the algorithm based on effective detection. Therefore, the proposed algorithm performs better overall.
Furthermore, to evaluate the robustness of the proposed algorithm against noise, different types of noise were added to dataset 3 which already contained different degrees of noise for performance comparison. Figure 5 shows representative images of the original IR dataset 3 and images with different types of noise added. Five types of noise were added in the experiment, including Gaussian white noise with a variance of 0.001, Poisson noise, Rayleigh noise with a variance of 15, multiplicative noise with a variance of 3, and uniform noise with a minimum of −14 and a maximum of 14. Table 6 shows the adaptive threshold calculation formulas corresponding to different algorithms and the range of experimental constants. VAR_DIFF and ADMD do not give specific threshold formulas, so the other 5 algorithms were selected for comparison, and the middle value of the applicable range of the constant was used as the constant value in the experiment, the specific information is shown in Table 6. Figure 6 and Figure 7, respectively, show the number of missed images and the number of false-alarm images after different detection algorithms pass the corresponding threshold operation under different noise-type datasets. The experimental results show that the proposed algorithm did not miss detection under the influence of different types of noise. Although the proposed algorithm had false alarms under the influence of Poisson noise and Rayleigh noise, other algorithms had missed detection and false positives under the influence of different noises. Overall, compared with other methods, the proposed algorithm could successfully suppress most of the noise and had strong robustness against noise.

4. Conclusions

This paper proposes an IR small-target detection algorithm based on WLDVM. The proposed algorithm performs preprocessing operations through the idea of matched filtering, which can reduce noise and enhance small targets to a certain extent. The distribution characteristics of the target area are fully utilized to divide the window area, which can adapt to the detection of small targets of different scales. LDVM can more effectively highlight the target area and eliminate the bright background, thereby effectively improving the detection rate and reducing the false-alarm rate. The weighting function can improve the adaptability to complex backgrounds and can preserve the shape features of targets of different scales. The fused results can further reduce the missed-detection rate and false-positive rate in complex scenes, thus achieving strong robust detection. Experiments show that the algorithm has good anti-noise ability and is robust to objects of different scales and categories under complex backgrounds. Compared with other methods, the proposed method has obvious advantages in quantitative results such as BSF, SCRG, and the mean runtime, and can better preserve the shape features of targets visually. In future work, we will further study the application of this method in the recognition of IR target categories such as tanks, warships, and aircraft.

Author Contributions

Conceptualization, Y.Z. (Ying Zheng) and Y.Z. (Yuye Zhang); methodology, Y.Z. (Ying Zheng); software, Y.Z. (Ying Zheng); validation, Y.Z. (Ying Zheng), Y.Z. (Yuye Zhang), R.D., C.M. and X.L.; formal analysis, Y.Z. (Yuye Zhang); investigation, R.D.; resources, Y.Z. (Yuye Zhang).; data curation, C.M.; writing—original draft preparation, Y.Z. (Yuye Zhang); writing—review and editing, Y.Z. (Ying Zheng); visualization, Y.Z. (Ying Zheng); supervision, X.L.; project administration, X.L.; funding acquisition, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural Science Foundation of Xinjiang Province under grant number 2020D01C026, and by National Natural Science Foundation of China under grant numbers U1911401 and 61433012.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kastberger, G.; Stachl, R. Infrared imaging technology and biological applications. Behav. Res. Methods Instrum. Comput. 2003, 35, 429–439. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Jong, A. IRST and its perspective. Proc. SPIE Int. Soc. Opt. Eng. 1995, 2552, 206–213. [Google Scholar]
  3. Fang, B.; Chen, T. Study on the Key Techniques of the Imaging Infrared Guidance for AAM. Infrared Technol. 2003, 25, 45–48. [Google Scholar]
  4. Malanowski, M.; Kulpa, K. Detection of moving targets with continuous-wave noise radar: Theory and measurements. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3502–3509. [Google Scholar] [CrossRef]
  5. Chen, C.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  6. Yao, S.; Chang, Y.; Qin, X. A Coarse-to-Fine Method for Infrared Small Target Detection. IEEE Geosci. Remote Sens. 2019, 16, 256–260. [Google Scholar] [CrossRef]
  7. Zhang, P.; Wang, X.; Wang, X.; Fei, C.; Guo, Z. Infrared small target detection based on spatial-temporal enhancement using quaternion discrete cosine transform. IEEE Access 2019, 7, 54712–54723. [Google Scholar] [CrossRef]
  8. Nie, J.; Qu, S.; Wei, Y.; Zhang, L.; Deng, L. An Infrared Small Target Detection Method Based on Multiscale Local Homogeneity Measure. Infrared Phys. Technol. 2018, 90, 186–194. [Google Scholar] [CrossRef]
  9. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A. Infrared patch-image model for small target detection in a single image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  10. Qin, Y.; Li, B. Effective infrared small target detection utilizing a novel local contrast method. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1890–1894. [Google Scholar] [CrossRef]
  11. Shao, X.; Fan, H.; Lu, G.; Xu, J. An improved infrared dim and small target detection algorithm based on the contrast mechanism of human visual system. Infrared Phys. Technol. 2012, 55, 403–408. [Google Scholar] [CrossRef]
  12. Rivest, J.; Fortin, R. Detection of dim targets in digital infrared imagery by morphological image processing. Opt. Eng. 1996, 35, 1886–1893. [Google Scholar] [CrossRef]
  13. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  14. Hou, W.; Lei, Z.; Yu, Q.; Liu, X. Small target detection using main directional suppression high pass filter. Optik 2014, 125, 3017–3022. [Google Scholar] [CrossRef]
  15. Xin, Y.; Zhou, J.; Chen, Y. Dual multi-scale filter with SSS and GW for infrared small target detection. Infrared Phys. Technol. 2017, 81, 97–108. [Google Scholar] [CrossRef]
  16. Dai, Y.; Wu, Y. Reweighted infrared patch-tensor model with both nonlocal and local priors for single-frame small target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef] [Green Version]
  17. He, Y.; Li, M.; Zhang, J.; An, Q. Small infrared target detection based on low-rank and sparse representation. Infrared Phys. Technol. 2015, 68, 98–109. [Google Scholar] [CrossRef]
  18. Zheng, C.; Li, H. Small infrared target detection based on low-rank and sparse matrix decomposition. In Applied Mechanics and Materials; Trans Tech Publications Ltd.: Zurich, Switzerland, 2013; Volume 239, pp. 214–218. [Google Scholar]
  19. Dai, Y.; Wu, Y.; Song, Y. Infrared small target and background separation via column-wise weighted robust principal component analysis. Infrared Phys. Technol. 2016, 77, 421–430. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Zheng, Y.; Li, X. Multi-Scale Strengthened Directional Difference Algorithm Based on the Human Vision System. Sensors 2022, 22, 10009. [Google Scholar] [CrossRef]
  21. Wang, H.; Zhou, L.; Wang, L. Miss detection vs. false alarm: Adversarial learning for small object segmentation in infrared images. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2019, Seoul, Republic of Korea, 27–28 October 2019; pp. 8509–8518. [Google Scholar]
  22. Xu, X.; Sun, Y.; Ding, L.; Yang, F. A Novel Infrared Small Target Detection Algorithm Based on Deep Learning. In Proceedings of the 2020 4th International Conference on Advances in Image Processing, Chengdu China, 13–15 November 2020; pp. 8–14. [Google Scholar]
  23. Lin, L.; Wang, S.; Tang, Z. Using deep learning to detect small targets in infrared oversampling images. J. Syst. Eng. Electron. 2018, 29, 947–952. [Google Scholar]
  24. Ryu, J.; Kim, S. Small infrared target detection by data-driven proposal and deep learning-based classification. In Proceedings of the Infrared Technology and Applications XLIV, Orlando, FL, USA, 16–19 April 2018; SPIE: Bellingham, WA, USA, 2018; Volume 10624, pp. 134–143. [Google Scholar]
  25. Han, J.; Ma, Y.; Zhou, B.; Fan, F.; Liang, K.; Fang, Y. A robust infrared small target detection algorithm based on human visual system. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2168–2172. [Google Scholar]
  26. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared small target detection utilizing the multiscale relative local contrast measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  27. Han, J.; Yu, Y.; Liang, K.; Zhang, H. Infrared small-target detection under complex background based on subblock-level ratio-difference joint local contrast measure. Opt. Eng. 2018, 57, 103105. [Google Scholar] [CrossRef]
  28. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  29. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. A Local Contrast Method for Infrared Small-Target Detection Utilizing a Tri-Layer Window. IEEE Geosci. Remote Sens. 2020, 17, 1822–1826. [Google Scholar] [CrossRef]
  30. Moradi, S.; Moallem, P.; Sabahi, M.F. Fast and robust small infrared target detection using absolute directional mean difference algorithm. Signal Process. 2020, 117, 107727. [Google Scholar] [CrossRef]
  31. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Entropy-based window selection for detecting dim and small infrared targets. Pattern Recognit. 2017, 61, 66–77. [Google Scholar] [CrossRef]
  32. Nasiri, M.; Chehresa, S. Infrared small target enhancement based on variance difference. Infrared Phys. Technol. 2017, 82, 107–119. [Google Scholar] [CrossRef]
  33. Liu, J.; He, Z.; Chen, Z.; Shao, L. Tiny and dim infrared target detection based on weighted local contrast. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1780–1784. [Google Scholar] [CrossRef]
  34. Lv, P.; Sun, S.; Lin, C.; Liu, G. A method for weak target detection based on human visual contrast mechanism. IEEE Geosci. Remote Sens. Lett. 2019, 16, 261–265. [Google Scholar] [CrossRef]
  35. Han, J.; Moradi, S.; Faramarzi, I.; Zhang, H.; Zhao, Q.; Zhang, X.; Li, N. Infrared small target detection based on the weighted strengthened local contrast measure. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1670–1674. [Google Scholar] [CrossRef]
  36. Moradi, S.; Moallem, P.; Sabahi, M. Scale-space point spread function based framework to boost infrared target detection algorithms. Infrared Phys. Technol. 2016, 77, 27–34. [Google Scholar] [CrossRef]
  37. Guan, X.; Peng, Z.; Huang, S.; Chen, Y. Gaussian scale-space enhanced local contrast measure for small infrared target detection. IEEE Geosci. Remote Sens. Lett. 2019, 17, 327–331. [Google Scholar] [CrossRef]
  38. Du, P.; Hamdulla, A. Infrared moving small-target detection using spatial–temporal local difference measure. IEEE Geosci. Remote Sens. 2019, 17, 1817–1821. [Google Scholar] [CrossRef]
  39. Wu, W.; Hamdulla, A. Space-time-based Detection of Infrared Small Moving Target. J. Appl. Sci. Eng. 2020, 23, 497–507. [Google Scholar]
  40. Rawat, S.; Verma, K.; Kumar, Y. Review on recent development in infrared small target detection algorithms. Procedia Comput. Sci. 2020, 167, 2496–2505. [Google Scholar] [CrossRef]
  41. Shahraki, H.; Aalaei, S.; Moradi, S. Infrared small target detection based on the dynamic particle swarm optimization. Infrared Phys. Technol. 2021, 117, 103837. [Google Scholar] [CrossRef]
  42. Hui, B.; Song, Z.; Fan, H.; Zhong, P.; Hu, W.; Zhang, X.; Zhang, Y. A dataset for infrared image dim-small aircraft target detection and tracking under ground/air background. Sci. Data Bank. 2019, 5, 12. [Google Scholar]
  43. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric contextual modulation for infrared small target detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 950–959. [Google Scholar]
  44. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Infrared small-target detection using multiscale gray difference weighted image entropy. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 60–72. [Google Scholar] [CrossRef]
Figure 1. The proposed algorithm framework.
Figure 1. The proposed algorithm framework.
Sensors 23 02630 g001
Figure 2. (a) The double-layer window. (b) The tri-layer window. (c) The new tri-layer filtering window. (d) Gaussian distribution characteristics of IR small target. (e) Situations that the algorithm needs to handle.
Figure 2. (a) The double-layer window. (b) The tri-layer window. (c) The new tri-layer filtering window. (d) Gaussian distribution characteristics of IR small target. (e) Situations that the algorithm needs to handle.
Sensors 23 02630 g002
Figure 3. SMs of different algorithms. (a1a9) Nine representative IR images. (b1b9) LCM. (c1c9) MPCM. (d1d9) RLCM. (e1e9) TLLCM. (f1f9) VAR_DIFF. (g1g9) ADMD. (h1h9) WSLCM. (i1i9) Proposed methods. ( a 1 ¯ a 9 ¯ ), ( b 1 ¯ b 9 ¯  ), ( c 1 ¯ c 9 ¯  ), ( d 1 ¯ d 9 ¯  ), ( e 1 ¯ e 9 ¯  ), ( f 1 ¯ f 9 ¯  ), ( g 1 ¯ g 9 ¯  ), ( h 1 ¯ h 9 ¯  ), and ( i 1 ¯ i 9 ¯  ) are the 3D gray distribution maps of (a1a9), (b1b9), (c1c9), (d1d9), (e1e9), (f1f9), (g1g9), (h1h9), and (i1i9), respectively.
Figure 3. SMs of different algorithms. (a1a9) Nine representative IR images. (b1b9) LCM. (c1c9) MPCM. (d1d9) RLCM. (e1e9) TLLCM. (f1f9) VAR_DIFF. (g1g9) ADMD. (h1h9) WSLCM. (i1i9) Proposed methods. ( a 1 ¯ a 9 ¯ ), ( b 1 ¯ b 9 ¯  ), ( c 1 ¯ c 9 ¯  ), ( d 1 ¯ d 9 ¯  ), ( e 1 ¯ e 9 ¯  ), ( f 1 ¯ f 9 ¯  ), ( g 1 ¯ g 9 ¯  ), ( h 1 ¯ h 9 ¯  ), and ( i 1 ¯ i 9 ¯  ) are the 3D gray distribution maps of (a1a9), (b1b9), (c1c9), (d1d9), (e1e9), (f1f9), (g1g9), (h1h9), and (i1i9), respectively.
Sensors 23 02630 g003aSensors 23 02630 g003bSensors 23 02630 g003c
Figure 4. The ROC curves of different detection methods. (ai) The experimental results of datasets 1−9.
Figure 4. The ROC curves of different detection methods. (ai) The experimental results of datasets 1−9.
Sensors 23 02630 g004
Figure 5. (a) Representative images of raw IR dataset 3. (b) Gaussian white noise. (c) Poisson noise. (d) Rayleigh noise. (e) Multiplicative noise. (f) Uniform noise. (a1f1) The 3D gray distribution maps of (af).
Figure 5. (a) Representative images of raw IR dataset 3. (b) Gaussian white noise. (c) Poisson noise. (d) Rayleigh noise. (e) Multiplicative noise. (f) Uniform noise. (a1f1) The 3D gray distribution maps of (af).
Sensors 23 02630 g005
Figure 6. The number of missed images in different algorithms under the influence of noise.
Figure 6. The number of missed images in different algorithms under the influence of noise.
Sensors 23 02630 g006
Figure 7. The number of false-alarm images in different algorithms under the influence of noise.
Figure 7. The number of false-alarm images in different algorithms under the influence of noise.
Sensors 23 02630 g007
Table 1. Details of the nine datasets.
Table 1. Details of the nine datasets.
Number of ImagesImage SizeTarget SizeTarget NumberDataset TypeTarget DetailBackground Detail
Dataset 1 [38]170   250 × 250   6 × 6 1Real sequencePoint target,
incomplete occlusion
Complex clouds,
complex background
Dataset 2 [39]429   250 × 250   6 × 6 1Simulated sequencePoint target,
low contrast
Heavy noise,
dim background
Dataset 3 [5]30   256 × 200   4 × 6   to   5 × 8 1Real sequenceFast-moving,
low contrast
Heavy noise,
dense clouds
Dataset 4 [40]39   256 × 200   4 × 14   to   7 × 15 1Real sequenceAircraft target,
fast-moving
Changing background,
thin clouds
Dataset 5 [41]50   302 × 202   4 × 8 1Simulated sequencePoint target,
incomplete occlusion
Remaining almost the same
Thin clouds
Dataset 6 [41]50   238 × 158   4 × 8 1Simulated sequencePoint target,
fast-moving
Heavy noise,
complex background
Dataset 7 [41]50   256 × 239   4 × 8 1Simulated sequencePoint target,
low contrast
Multiple buildings, heavy noise
Dataset 8 [42]100   256 × 256   5 × 5   1Simulated sequencePoint target,
continuously moving
Heavy noise,
land background
Dataset 9 [43]152Variety   3 × 3   to   11 × 11 1Non-sequentialVarietyVariety
Table 2. The number of false-alarm images  N F A  and the number of missed images  N M D  of different datasets under different  λ  values.
Table 2. The number of false-alarm images  N F A  and the number of missed images  N M D  of different datasets under different  λ  values.
N F A Dataset λ = 0 λ = 0.1 λ = 0.2 λ = 0.3 λ = 0.4 λ = 0.5 λ = 0.6 λ = 0.7 λ = 0.8 λ = 0.9 λ = 1
1687431000000
21300000000000
3303110000000
4390000000000
5502000000000
65011300000000
75030000000000
810010057144000000
912520621000000
N M D 10000000037118170
20000000040294429
30000000382430
40000000052739
50000000192450
60000000002750
70000000032950
8000000065492100
90000000113283152
Table 3. SCRG and BSF of different detection algorithms.
Table 3. SCRG and BSF of different detection algorithms.
SCRGDatasetLCMMPCMRLCMTLLCMVAR_DIFFADMDWSLCMProposed
10.60012.02290.72732.243928.339625.926426.827744.0452
20.32562.21930.65741.624614.109311.727025.354927.8967
30.63221.06710.79031.08080.918216.608530.5386105.4981
40.20550.29580.24430.80306.96730.34183.852518.2806
50.85490.63880.79180.85554.539923.174521.284451.1140
60.79831.13330.91341.28985.326821.881052.813138.8965
70.31711.69630.40420.76711.45413.299115.97809.5035
80.40213.41640.68871.56802.513610.582730.056539.5189
90.38861.82860.71661.968324.205312.271621.752323.6554
Mean0.50271.59090.65931.35579.819313.979225.384239.8232
BSF10.31851.67830.83500.6472115.033012.6716372.70781144.7466
20.98053.66172.08151.7074212.732117.2603758.07123894.0449
32.17486.40612.38681.801138.571523.490537.883435.9936
42.221411.26765.72763.3599100.814328.13953704.5536572.1345
50.58512.73401.33400.858130.986821.4587125.5700151.6036
61.03184.04021.39121.126922.3591105.0573165.307089.8998
70.43191.86720.69750.52158.714414.0430144.658218.4970
81.47646.75782.53752.380429.950522.9731241.081438.1197
91.41038.27443.55252.5970974.0334117.42752401.43762206.5812
Mean1.18125.18752.28261.6666170.355040.2802883.4745905.7357
Table 4. The implementation environment.
Table 4. The implementation environment.
Operating SystemWindows (Windows 10 21H1, x64)
MATLAB versionMATLAB R2020a
CPUIntel Core i7-10875H @ 2.30 GHz
Memory16.0 GB
Table 5. Comparison of the mean runtime of different algorithms.
Table 5. Comparison of the mean runtime of different algorithms.
MethodLCMMPCMRLCMTLLCMVAR_DIFFADMDWSLCMProposed
Time (s)0.02740.03121.13840.32160.00680.01671.47800.0153
Table 6. Threshold calculation formulas of different algorithms.
Table 6. Threshold calculation formulas of different algorithms.
Formula for Threshold CalculationRange of Experimental ConstantConstant Value Used in the Experiment
LCM [5] T h = μ S M + k σ S M
    μ S M :   Mean of SM
    σ S M :   SM standard deviation
    k : Experimental constant
  k [ 3 , 5 ]   k = 4
MPCM [28]   k [ 3 , 14 ]   k = 8.5
RLCM [26]   k [ 2 , 9 ]   k = 5.5
TLLCM [29] T h = λ m a x S M + ( 1 λ ) m e a n S M
    m a x S M : Maximum of SM
    m e a n S M : Mean of SM
    λ : Experimental constant between 0 and 1
  λ [ 0.7 , 0.9 ]   λ = 0.8
WSLCM [35]   λ [ 0.6 , 0.9 ]   λ = 0.75
Proposed   λ [ 0.5 , 0.6 ]   λ = 0.55
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, Y.; Zhang, Y.; Ding, R.; Ma, C.; Li, X. Fast and Robust Infrared Small Target Detection Using Weighted Local Difference Variance Measure. Sensors 2023, 23, 2630. https://doi.org/10.3390/s23052630

AMA Style

Zheng Y, Zhang Y, Ding R, Ma C, Li X. Fast and Robust Infrared Small Target Detection Using Weighted Local Difference Variance Measure. Sensors. 2023; 23(5):2630. https://doi.org/10.3390/s23052630

Chicago/Turabian Style

Zheng, Ying, Yuye Zhang, Ruichen Ding, Chunming Ma, and Xiuhong Li. 2023. "Fast and Robust Infrared Small Target Detection Using Weighted Local Difference Variance Measure" Sensors 23, no. 5: 2630. https://doi.org/10.3390/s23052630

APA Style

Zheng, Y., Zhang, Y., Ding, R., Ma, C., & Li, X. (2023). Fast and Robust Infrared Small Target Detection Using Weighted Local Difference Variance Measure. Sensors, 23(5), 2630. https://doi.org/10.3390/s23052630

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop