[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Development of a Microfluidic-Based Optical Sensing Device for Label-Free Detection of Circulating Tumor Cells (CTCs) Through Their Lactic Acid Metabolism
Next Article in Special Issue
High-Speed Incoming Infrared Target Detection by Fusion of Spatial and Temporal Detectors
Previous Article in Journal
Location Detection and Tracking of Moving Targets by a 2D IR-UWB Radar System
Previous Article in Special Issue
Plasmonic Structure Integrated Single-Photon Detector Configurations to Improve Absorptance and Polarization Contrast
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Detection Based on the Generation of a Background Image by Using a Far-Infrared Light Camera

Division of Electronics and Electrical Engineering, Dongguk University, 26 Pil-dong 3-ga, Jung-gu, Seoul 100-715, Korea
*
Author to whom correspondence should be addressed.
Sensors 2015, 15(3), 6763-6788; https://doi.org/10.3390/s150306763
Submission received: 15 January 2015 / Revised: 17 February 2015 / Accepted: 9 March 2015 / Published: 19 March 2015
(This article belongs to the Special Issue Frontiers in Infrared Photodetection)
Figure 1
<p>Overall procedure of the proposed method.</p> ">
Figure 2
<p>Flow chart of generating a background image.</p> ">
Figure 3
<p>Examples of obtaining the background image from an open database: (<b>a</b>) preliminary background image obtained by temporal averaging; (<b>b</b>) extracted human areas by the binarization, labeling, size filtering and a morphological operation of <a href="#sensors-15-06763-f002" class="html-fig">Figure 2</a>; and (<b>c</b>) the generated final background image.</p> ">
Figure 4
<p>The first example for obtaining a background image from our database: (<b>a</b>) preliminary background image obtained by temporal averaging; (<b>b</b>) extracted human areas; and (<b>c</b>) the generated final background image.</p> ">
Figure 5
<p>The second example for obtaining a background image from our database: (<b>a</b>) preliminary background image obtained by temporal averaging; (<b>b</b>) extracted human areas; and (<b>c</b>) the generated final background image.</p> ">
Figure 6
<p>Example of the fusion of two difference images: (<b>a</b>) input image; (<b>b</b>) background image; (<b>c</b>) pixel difference image; (<b>d</b>) edge difference image; and (<b>e</b>) fusion of the pixel and edge difference images.</p> ">
Figure 7
<p>Division of the candidate region within an input image based on the horizontal histogram: (<b>a</b>) input image; (<b>b</b>) detected candidate region and its horizontal histogram; and (<b>c</b>) the division result of the candidate region.</p> ">
Figure 7 Cont.
<p>Division of the candidate region within an input image based on the horizontal histogram: (<b>a</b>) input image; (<b>b</b>) detected candidate region and its horizontal histogram; and (<b>c</b>) the division result of the candidate region.</p> ">
Figure 8
<p>Division of the candidate region within an input image based on the vertical histogram: (<b>a</b>) input image; (<b>b</b>) detected candidate region and its vertical histogram; and (<b>c</b>) the division result of the candidate region.</p> ">
Figure 9
<p>Example of different sizes of human areas resulting from camera viewing direction and perspective projection: (<b>a</b>) input image, including three detected areas of humans; and (<b>b</b>) information of the width, height and size of the three detected human areas, respectively.</p> ">
Figure 10
<p>Result of obtaining the final region of the human area: (<b>a</b>) result after the process based on the separation of histogram information; (<b>b</b>) result after the process based on camera viewing direction and perspective projection; and (<b>c</b>) result of the final detected region of the human area.</p> ">
Figure 10 Cont.
<p>Result of obtaining the final region of the human area: (<b>a</b>) result after the process based on the separation of histogram information; (<b>b</b>) result after the process based on camera viewing direction and perspective projection; and (<b>c</b>) result of the final detected region of the human area.</p> ">
Figure 11
<p>Comparisons of generated background images with the OTCBVS benchmark dataset. The left-upper [<a href="#B28-sensors-15-06763" class="html-bibr">28</a>], right-upper [<a href="#B26-sensors-15-06763" class="html-bibr">26</a>], left-lower [<a href="#B24-sensors-15-06763" class="html-bibr">24</a>,<a href="#B25-sensors-15-06763" class="html-bibr">25</a>,<a href="#B33-sensors-15-06763" class="html-bibr">33</a>], and right-lower figures are generated by previous methods and the proposed one, respectively.</p> ">
Figure 12
<p>Comparisons of generated background images with our database (second database). The left, middle and right figures of (<b>a</b>,<b>b</b>) are by simple temporal averaging operation [<a href="#B28-sensors-15-06763" class="html-bibr">28</a>], averaging the frames in two difference sequences [<a href="#B26-sensors-15-06763" class="html-bibr">26</a>] and the proposed method, respectively: (<b>a</b>) with Sequence 4 of [<a href="#B28-sensors-15-06763" class="html-bibr">28</a>] (left figure) and the proposed method (right figure) and with Sequences 4 and 1 of [<a href="#B26-sensors-15-06763" class="html-bibr">26</a>] (middle figure); (<b>b</b>) with Sequence 5 of [<a href="#B28-sensors-15-06763" class="html-bibr">28</a>] (left figure) and the proposed method (right figure) and with Sequences 5 and 2 of [<a href="#B26-sensors-15-06763" class="html-bibr">26</a>] (middle figure).</p> ">
Figure 13
<p>Detection results with the OTCBVS benchmark dataset (<b>a</b>–<b>f</b>) and our database (<b>g</b>–<b>j</b>). Results of images in: (<b>a</b>) Sequence 1; (<b>b</b>) Sequence 3; (<b>c</b>) Sequence 4; (<b>d</b>) Sequence 5; (<b>e</b>) Sequence 6; (<b>f</b>) Sequence 7; (<b>g</b>) Sequence 2; (<b>h</b>) Sequence 3; (<b>i</b>) Sequence 4; and (<b>j</b>) Sequence 6.</p> ">
Figure 13 Cont.
<p>Detection results with the OTCBVS benchmark dataset (<b>a</b>–<b>f</b>) and our database (<b>g</b>–<b>j</b>). Results of images in: (<b>a</b>) Sequence 1; (<b>b</b>) Sequence 3; (<b>c</b>) Sequence 4; (<b>d</b>) Sequence 5; (<b>e</b>) Sequence 6; (<b>f</b>) Sequence 7; (<b>g</b>) Sequence 2; (<b>h</b>) Sequence 3; (<b>i</b>) Sequence 4; and (<b>j</b>) Sequence 6.</p> ">
Figure 14
<p>Overlapping area of ground truth and detected boxes.</p> ">
Figure 15
<p>Detection error cases with the OTCBVS benchmark dataset: (<b>a</b>) original image; (<b>b</b>) result of the proposed method.</p> ">
Figure 16
<p>Detection error cases with our database: (<b>a</b>) original image; (<b>b</b>) result of the proposed method.</p> ">
Versions Notes

Abstract

:
The need for computer vision-based human detection has increased in fields, such as security, intelligent surveillance and monitoring systems. However, performance enhancement of human detection based on visible light cameras is limited, because of factors, such as nonuniform illumination, shadows and low external light in the evening and night. Consequently, human detection based on thermal (far-infrared light) cameras has been considered as an alternative. However, its performance is influenced by the factors, such as low image resolution, low contrast and the large noises of thermal images. It is also affected by the high temperature of backgrounds during the day. To solve these problems, we propose a new method for detecting human areas in thermal camera images. Compared to previous works, the proposed research is novel in the following four aspects. One background image is generated by median and average filtering. Additional filtering procedures based on maximum gray level, size filtering and region erasing are applied to remove the human areas from the background image. Secondly, candidate human regions in the input image are located by combining the pixel and edge difference images between the input and background images. The thresholds for the difference images are adaptively determined based on the brightness of the generated background image. Noise components are removed by component labeling, a morphological operation and size filtering. Third, detected areas that may have more than two human regions are merged or separated based on the information in the horizontal and vertical histograms of the detected area. This procedure is adaptively operated based on the brightness of the generated background image. Fourth, a further procedure for the separation and removal of the candidate human regions is performed based on the size and ratio of the height to width information of the candidate regions considering the camera viewing direction and perspective projection. Experimental results with two types of databases confirm that the proposed method outperforms other methods.

1. Introduction

The need for computer vision-based human detection has increased in fields, such as security, intelligent surveillance and monitoring systems [1,2,3,4,5]. However, performance enhancement of human detection based on visible light cameras is limited because of factors, such as nonuniform illumination, shadows and low external light in the evening and night. Consequently, thermal cameras have been widely used for tasks, such as security and outdoor surveillance. In general, the infra-red (IR) spectrum can be classified into four sub-bands, such as near-IR (NIR), whose wavelength ranges from 0.75 to 1.4 μm, short-wave IR (SWIR), whose wavelength ranges from 1.4 to 3 μm, medium-wave IR (MWIR), whose wavelength ranges from 3 to 8 μm, and long-wave IR (LWIR), whose wavelength ranges from 8 to 15 μm [6]. Because significant heat energy is reported to be emitted in MWIR and LWIR sub-bands, both of these sub-bands are usually used for sensing the human body, including the face, without an additional light source, and they are referred to as the thermal sub-band [6]. We call the image acquired in this sub-band the thermal image in this paper.
The gray level of an object in a thermal image depends on the temperature of the object. If the object is hotter than the surrounding environment, it is represented as a brighter region. Typically, the gray level of human areas in a thermal image appears differently from the background, allowing thermal images to be applied to the detection of human areas in an image. However, under different conditions, the characteristics of the human areas and background can change. This can influence the accuracy of detecting the precise locations and shapes of the human areas in the thermal image [7,8,9,10,11,12,13,14,15].
Recently, significant work has focused on detecting human areas in thermal camera images, because the performance of the detection influences the accuracy of the human tracking and behavioral recognition. Previous research into human detection in thermal camera images can be classified into two categories: without background generation [2,15,16,17,18,19,20] and with background generation [21,22,23,24,25,26,27,28,29]. For the former category, Martin et al. proposed a detector and tracker of objects based on motion and scale-invariant feature transform (SIFT) methods [2]. In addition, the former used the methods for feature extraction based on the histogram of the oriented gradient (HOG) [7,8,9,15,16,17] and geometric characteristics [15] with a classifier based on support vector machine (SVM) [7,15,17,18], the adaptive boosting (Adaboost) method [19] and the soft-label boosting algorithm [20]. The advantage of these methods is that they can detect an object without a background image. However, they have the disadvantage that they require a pre-defined template or classifier for human areas; this must be obtained through training. Moreover, their performance is influenced by environmental factors, such as rain, snow and the amount of sunlight, and they require an excessive processing time to detect the human areas in the entire image by scanning.
Therefore, techniques for detecting human areas using background generation can be employed as an alternative. These are based on a Gaussian background-subtraction approach [21,22,24,25], texture change [23], expectation minimization (EM) [26,27] and image averaging [28].
To detect an object, methods, such as a contour saliency map (CSM) [21,24,25], CSM template matching [22], shape and appearance-based detection [26,27], spatiotemporal texture vectors [23] and a boosting framework [28], can be applied. In some research [26,27], shape descriptors, such as compactness and leanness, are obtained from the skeleton of the object, and they are used as the input values of SVM for shape-based detection. In addition, the features obtained by principal component analysis (PCA) with the pixel values from a window of fixed size [26] and varying size [27] are used for appearance-based detection. Using both shape and appearance-based detection is one of the main novelties of their research. In previous research [29], a particle filter framework and histogram based on the intensity-distance projection space were used. The advantage of these methods is that they are applicable to multiple environmental conditions and can detect objects of varying scale. However, the disadvantage is that performance degradation can occur if the intensity of the object is similar to the background. Further, if there exist motionless humans (at the same positions) in all of the image frames, these approaches can produce an erroneous background, which has the areas of the image occupied by motionless humans, although the accurate background should not include the areas of the image occupied by humans. Moreover, they do not consider the parameters of the detector, which are adaptively determined by the background information. The pixel brightness of the background is usually affected by environmental temperature. That is, the pixel brightness of the background is higher in the daytime than that that at dawn or night. If the pixel brightness is too high or too low, the difference of the pixel brightness between the background and human usually decreases. Therefore, the thresholds used for obtaining the binarization image from background subtraction should be reduced in these cases. This scheme means that the parameters of the detector of the human by background subtraction are adaptively determined by the background information (the pixel brightness of the background). Other research proposed the usage of stereo thermal cameras [10] and a dual system of visible light and thermal cameras [11]. However, these systems are expensive and oversized, owing to the use of two cameras.
Table 1. Comparison of previous and the proposed methods. HOG, histogram of the oriented gradient; EM, expectation minimization; CSM, contour saliency map.
Table 1. Comparison of previous and the proposed methods. HOG, histogram of the oriented gradient; EM, expectation minimization; CSM, contour saliency map.
CategoryWithout Background Generation [7,8,9,15,16,17,18,19,20]With Background Generation
Not Adjusting the Parameters for Detection Based on Background Information [21,22,23,24,25,26,27,28,29]Adjusting the Parameters for Detection Based on Background Information (Proposed Method)
Examples
-
Motion + SIFT-based [2], HOG-based [7,8,9,15,16,17], geometric characteristics-based [15,17,18], Adaboost-based [19] and soft-label boosting-based [20] detection of object.
-
Gaussian model-based [21,22,23,25], texture change-based [23], EM-based [26,27] and image averaging-based [28] background modeling and subtraction.
-
The correct background image can be generated by image averaging, various filtering and erasing of the human area with adaptive determination of thresholds and parameters for the human detector.
-
CSM-based [21,24,25], CSM template matching-based [22], shape and appearance-based [26,27], spatiotemporal texture vectors-based [23], boosting framework-based [28] and particle filter and histogram-based [29] detection of object.
Advantages
-
Can detect object without a background image.
-
Can be applicable to various environmental conditions and can detect objects of various scales.
-
Robust detection of the human area can be obtained by adaptively determining the thresholds and parameters for detection considering background information.
-
Does not require a training procedure to obtain the classifier of human detection.
Disadvantages
-
Requires a pre-defined template or classifier for the human area that must be obtained through training.
-
Performance degradation can occur if the intensity of the object is similar to the background.
-
Additional procedure is required for obtaining correct background image.
-
Performance is influenced by various environmental factors, such as rain, snow and the amount of sunlight.
-
The approach can produce erroneous background, including the image areas occupied by humans, if the image frames include the image areas occupied by motionless humans.
-
Requires significant processing time to detect the human area in the entire image by scanning.
-
Parameters for the detector of the image areas occupied by humans are not adaptively determined based on background information.
To address these problems, we present a new method of detecting human areas in a thermal image in varying environmental conditions. The proposed approach is based on the method of background subtraction. One background image is generated using median and average filtering. Additional filtering procedures based on maximum gray level, size filtering and region erasing are applied to remove human areas from the background image. Candidate human regions in the input image are located by combining the pixel and edge difference images between the input and background images. The thresholds for the difference images are adaptively determined based on the brightness of the generated background image. Noise components are removed by component labeling, a morphological operation and size filtering. The contiguous foreground area may be occupied by more than a single human. In addition, the size of this area can be too small to be accepted as the foreground area, which is caused by the incorrect detection of the foreground area. Therefore, these detected areas are merged or separated by the further procedure based on the information in the horizontal and vertical histograms of the detected area. This procedure is adaptively operated based on the brightness of the generated background image. Finally, a further procedure for the separation and removal of the candidate human regions is performed based on the information of the size and the ratio of the height to width of the candidate regions considering the camera viewing direction and perspective projection.
Table 1 presents a summarized comparison of the previous research of human detection in thermal camera images and the proposed method.
This paper is organized as follows. We provide an overview of the proposed algorithm in Section 2. We present experimental results and analysis in Section 3. Finally, conclusions are presented in Section 4.

2. Proposed Method

2.1. Proposed Method

An overview of the proposed method is presented in Figure 1.
Our approach for detecting humans in a thermal image can be divided into three steps: generating a background image, obtaining a difference image with the background and input image and detecting humans in the difference image. As explained in Section 1, we call the image acquired in the sub-bands of MWIR and LWIR the thermal image [6] in this paper.
To begin, a background image is generated. In this step, a correct background image is obtained by various filtering and erasing of human areas (see the details in Section 2.2). Then, two (pixel and edge) difference images are obtained from the background and input images. These two binary difference images are combined using pixel-wise conjunction. The thresholds for the difference images are adaptively determined based on the brightness information of the generated background image (see the details in Section 2.3).
The third step addresses human detection in the combined difference image. After applying size filtering and a morphology operation based on the size of the candidate areas, noise is removed. The remaining areas are separated by a vertical and horizontal histogram of the detected regions using the intensity of the background. Detected regions that may have more than two human areas are merged. Therefore, further procedures are performed to separate the candidate regions and to remove noise regions. These are based on the information of the size and ratio (of the height to width) of the candidate regions considering the camera-viewing angle and perspective projection (see the details in Section 2.4). Finally, we obtain the correct human areas.
Figure 1. Overall procedure of the proposed method.
Figure 1. Overall procedure of the proposed method.
Sensors 15 06763 g001

2.2. Generating a Background Image

Research into detecting humans has focused on creating background models to detect candidate regions of human areas [21,22,23,24,25,26,27,28,29]. Generating a robust background image is essential to extract candidate regions of human areas based on background subtraction. However, the methods of creating a background image from a sequence containing motionless humans in all frames have a problem in that an incorrect background is generated with a human included. This influences the performance of detecting a human based on the background-subtraction method. To solve this problem, two images from different sequences are averaged to obtain a background image [26]. This, however, has the problem that the brightness of the generated background image is changed compared to the input image, causing a human area detection error. For example, if the first image is from the sequence captured in the daytime and the second one is from the sequence captured at night, the pixel values, even in the same position, are different from each other in these two images. That is because the pixel value in the thermal image is usually higher in the daytime than that at night. Therefore, if a temporal-averaged image is obtained from these two images as a background image, the pixel value in each position of the background image is different from that of the first or second image. Therefore, although a current input image is obtained in the daytime or at night, the pixel value in each position of the current input image is different from that of the background image, even at the same position without humans. Consequently, these pixel differences produce a lot of erroneous regions as the human area by background subtraction.
Therefore, we propose an approach for generating a correct background image under these conditions. A flow chart of the proposed method is presented in Figure 2.
In the proposed research, one background image is generated with training before testing. In detail, one background image is generated using training images, whereas the performance of our method is measured using the testing images that are not used for generating the background image. Although the training images include both background and humans, an accurate background image (excluding the image areas occupied by humans) is obtained by our method in Figure 2. The sequence of training images is processed by a median filter of 3 × 3 pixels to reduce noise; pixels with the same positions as the training images are temporally averaged as illustrated in Figure 3a. In previous research [28], Calafut et al. obtained the background image using the simple operation of temporal averaging. However, an incorrect background is generated when areas of motionless humans are included in the image sequence. Therefore, we propose a method for obtaining the correct background image by including an additional procedure to remove human areas in the background image as follows. Normally, a human area has the characteristics that its gray level is higher than the background area in thermal image. In detail, the pixel of higher temperature usually has a larger value than that of lower temperature in the thermal image. Therefore, because humans actively emit heat, the corresponding pixel values usually have a larger one in the thermal image compared to those of the background.
Figure 2. Flow chart of generating a background image.
Figure 2. Flow chart of generating a background image.
Sensors 15 06763 g002
Figure 3. Examples of obtaining the background image from an open database: (a) preliminary background image obtained by temporal averaging; (b) extracted human areas by the binarization, labeling, size filtering and a morphological operation of Figure 2; and (c) the generated final background image.
Figure 3. Examples of obtaining the background image from an open database: (a) preliminary background image obtained by temporal averaging; (b) extracted human areas by the binarization, labeling, size filtering and a morphological operation of Figure 2; and (c) the generated final background image.
Sensors 15 06763 g003
Based on this, to force the human area to be more distinctive compared to the background region, we apply a 3 × 3 local max filter (that selects a pixel if it has the maximum value in the window) to the background image. Then, the processed image is converted to a binarized version based on the threshold method of Equation (3) [30]:
μ k =   i = 1 M j = 1 N I k ( i , j ) M × N
σ k =   i = 1 M j = 1 N ( I k ( i , j ) μ k ) 2 M × N 1
B k ( i , j ) = {   0   i f   I k ( i , j ) > μ k + T h · σ k 1   o t h e r w i s e
where Ik(i, j) is the gray value of the pixel at the position (i, j). M and N are the width and height of the image, respectively. μk and σk are the average value and standard deviation of the image pixel, respectively. Bk(i, j) is a binary image, and k is the number of the input image in the sequence. Th is the optimal parameter, which is experimentally determined.
Then, the candidates of human areas are detected as shown in Figure 3b. As shown in Figure 3b, most of the detected candidate areas represent the humans, and these areas should not be included in the generated background image. Therefore, our method removes them by the procedure of erasing the areas. A detailed explanation of the erasing algorithm follows. By horizontal scanning, the left- and right-most positions of each candidate area per each row are located. Then, the pixels whose X positions are smaller and larger than the left- and right-most positions are determined as the pixels of the nonhuman region, respectively. Finally, the candidate area per each row is erased by linear interpolation with these pixels. This procedure is iterated within the entire image.
For example, we assume that the left-upper-most position of the image is (0, 0) at the (x, y) coordinate, and one candidate area is composed of seven pixels, such as (50, 1), (51, 1), (52, 1), (49, 2), (50, 2), (51, 2) and (52, 2), respectively. Then, there is no detected left- and right-most position by the horizontal scanning of the 1st row. By the horizontal scanning of the 2nd row, the left- and right-most positions are detected as (50, 1) and (52, 1), respectively. Then, (48, 1) (the pixel whose X position is smaller than the left-most position) and (54, 1) (the pixel whose X position is larger than the right-most position) can be determined as the pixels of the nonhuman region, respectively. If the gray levels at the positions (48, 1) and (54, 1) are 10 and 70, respectively, the candidate area of the 2nd row is erased by linear interpolation with 10 and 70. That is, the gray values of 10, 20, 30, 40, 50, 60 and 70 are newly assigned to the pixels of (48, 1), (49, 1), (50, 1), (51, 1), (52, 1), (53, 1) and (54, 1), respectively. This procedure is iterated from the 3rd row to the last row. Through this procedure, the discontinuous boundary between the human candidate area and its neighboring background can disappear, and the consequent candidate area can be removed.
From these methods, a correct background image is generated, even though motionless humans exist in all of the images at the same location. In the proposed research, we used two kinds of databases, an open database and our own database. Detailed descriptions of these databases are provided in Section 3.1. An example of the final background image with the open database is displayed in Figure 3c. We can observe that the background image does not include the human areas.
In Figure 4 and Figure 5, we present additional examples for obtaining the background images using the proposed method with our database. We can confirm that the proposed method can generate correct backgrounds that do not include human areas.
Figure 4. The first example for obtaining a background image from our database: (a) preliminary background image obtained by temporal averaging; (b) extracted human areas; and (c) the generated final background image.
Figure 4. The first example for obtaining a background image from our database: (a) preliminary background image obtained by temporal averaging; (b) extracted human areas; and (c) the generated final background image.
Sensors 15 06763 g004
Figure 5. The second example for obtaining a background image from our database: (a) preliminary background image obtained by temporal averaging; (b) extracted human areas; and (c) the generated final background image.
Figure 5. The second example for obtaining a background image from our database: (a) preliminary background image obtained by temporal averaging; (b) extracted human areas; and (c) the generated final background image.
Sensors 15 06763 g005

2.3. Generating a Difference Image with the Background and Input Image

To identify the regions-of-interest (ROI) that contain humans, we use a background-subtraction technique, as illustrated in Figure 1. First, we apply median and average filters of 3 × 3 pixels to an input image to reduce noise. A pixel difference image is created with the pixels between the background and input images using optimal thresholds based on the intensity of the background image. The operator is presented in Equations (4) and (5).
i = 1 M j = 1 N B k ( i , j ) > T   a n d   i = 1 M j = 1 N { I k ( i , j ) B k ( i , j ) } > U
where Bk(i, j) and Ik(i, j) are the pixel gray levels of a generated background and an input image at the position (i, j), respectively. M and N are the width and height of the image, respectively. k is number of the input image in the sequence. T and U are the optimal thresholds, which were experimentally determined. If Equation (4) with pixel Ik(i, j) is satisfied, a binarized image (Dk(i, j)) is obtained using Equation (5).
D k ( i , j ) = {   1   i f   | I k ( i , j ) B k ( i , j ) | > X 0   o t h e r w i s e  
where X is an optimal threshold that is adaptively determined based on the brightness of the background image. In detail, with too large or small of a value of the brightness of the background image, the proposed system assigns a smaller value to X, because the pixel difference between the input and background image usually decreases in these cases. This adaptive scheme enables the detection of a human area by background subtraction to be robust to various environmental conditions of the background.
As indicated in Figure 6c, the difference image is created by the pixel difference between the input and background image. We can determine the rough regions of the human candidates using these procedures. However, using the pixel difference only is ineffective for detecting the shape of the objects. If only the pixel difference is used, it is difficult to detect regions of human areas whose intensities are similar to those of the background image. Furthermore, some regions of human areas can be removed or separated as small parts (as indicated in the two small areas within the red-dotted circle in the middle portion of Figure 6c) that may be removed by the subsequent filtering procedures, even though these regions contain humans.
To overcome these problems, we also create an edge difference image with edges between the background and input image. Using a standard Sobel mask of 3 × 3 pixels, the edge of the background and input image can be extracted. Based on the same methods of Equations (4) and (5), an edge image is obtained by background subtraction with the input and background edge images considering the brightness of the background image (Figure 6d).
The pixel and edge (binary) difference images are then combined (Figure 6e) using pixel-wise conjunction. This means that if either pixel of the pixel or edge difference image represents a human area, it is determined to be a human region.
Figure 6. Example of the fusion of two difference images: (a) input image; (b) background image; (c) pixel difference image; (d) edge difference image; and (e) fusion of the pixel and edge difference images.
Figure 6. Example of the fusion of two difference images: (a) input image; (b) background image; (c) pixel difference image; (d) edge difference image; and (e) fusion of the pixel and edge difference images.
Sensors 15 06763 g006

2.4. Human Detection

With the combined image of pixel and edge differences (as explained in Section 2.3 and Figure 6e), human detection is performed as shown in Figure 1. Many noises of a small size exist in the combined image. In order to eliminate noises and figure out the candidate region of the human, component labeling, morphological operation (erosion and dilation) and size filtering are applied to the image. Then, we can get the regions of human candidates. However, it is often the case that more than two people can be detected as one candidate region. Therefore, the proposed method determines whether the candidate region is separated or not based on the histogram information of the region (see the details in Section 2.4.1).

2.4.1. Division of Candidate Region Based on Histogram Information

Horizontal and vertical histograms of each candidate region are obtained to determine whether the candidate region should be separated. In Figure 7, we present an example where a candidate region is divided into two parts using the horizontal histogram based on the size, the ratio of height to width and the intensity of the background. In detail, if the size of a detected region is greater than a threshold or the ratio of the height to width is not satisfied with the condition, the region is separated into two parts based on the histogram information. This procedure is executed only if the brightness of the background is less than the threshold. This is because, in this case, the pixel and edge differences between the human area and background image become large, and consequently, the credibility of the detected candidate region is high. The horizontal histogram is obtained by Equation (6):
F x = y = 0 I y P ( I ( x , y ) )
where I(x, y) is the pixel intensity at a location (x, y) within the candidate region and P(·) is equal to one if I(x, y) is true, otherwise zero. Iy is the height of the candidate region. In Figure 7b, Bx is the horizontal index of the candidate region within the image. As indicated in Figure 7b, if the minimum value of Fx, which represents the position of separation, is lower than the threshold, the candidate region is separated into two parts based on the position, as illustrated in Figure 7c.
Figure 7. Division of the candidate region within an input image based on the horizontal histogram: (a) input image; (b) detected candidate region and its horizontal histogram; and (c) the division result of the candidate region.
Figure 7. Division of the candidate region within an input image based on the horizontal histogram: (a) input image; (b) detected candidate region and its horizontal histogram; and (c) the division result of the candidate region.
Sensors 15 06763 g007aSensors 15 06763 g007b
Then, the proposed method determines whether the candidate region should be separated based on the vertical histogram information of the region. In Figure 8, a candidate region is divided into two parts by the vertical histogram based on the size, the ratio of the height to width and the intensity of the background. In detail, if either the size or the ratio of the height to width of the detected box is larger than the threshold considering the prior knowledge of the image area occupied by the human, the detected box is separated into two parts based on the histogram information. This procedure is executed only if the brightness of the background is less than the threshold. This is because, in this case, the pixel and edge differences between the human area and background image become large, and consequently, the credibility of the detected candidate region is high. The vertical histogram is obtained by Equation (7):
F y = x = 0 I x P ( I ( x , y ) )
where I(x, y) is the pixel intensity at location (x, y) within the candidate region and P(·) is equal to one if I(x, y) is true, otherwise zero. Ix is the width of the candidate region. In Figure 8b, By is the vertical index of the candidate region within the image. As indicated in Figure 8b, if the minimum value of Fy, which represents the position of separation, is lower than the threshold, the candidate region is separated into two parts based on the position, as illustrated in Figure 8c.
Figure 8. Division of the candidate region within an input image based on the vertical histogram: (a) input image; (b) detected candidate region and its vertical histogram; and (c) the division result of the candidate region.
Figure 8. Division of the candidate region within an input image based on the vertical histogram: (a) input image; (b) detected candidate region and its vertical histogram; and (c) the division result of the candidate region.
Sensors 15 06763 g008

2.4.2. Division of the Candidate Region Based on Camera Viewing Direction with Perspective Projection

To separate the candidate regions more accurately, additional procedures are performed. In Figure 9, the sizes of the candidates are considerably different from each other. This is caused by the camera viewing direction and perspective projection. Because the position of the thermal camera is near the bottom of the image and the camera captures the scene in a slanted direction, the Z distance between the camera and object in the bottom area of the image is closer than that in the upper area. Therefore, if the position of an object is located in the upper area of the image, the size of the object inevitably becomes smaller, owing to the greater Z distance based on the principle of perspective projection. Therefore, the proposed system determines whether the detected candidate is separated. In detail, a larger candidate region is allowed when it is detected in the bottom area of the image, whereas a smaller candidate region is permitted if it is detected in the upper area of the image. Candidates whose size exceeds a threshold are separated into multiple parts.
Figure 9. Example of different sizes of human areas resulting from camera viewing direction and perspective projection: (a) input image, including three detected areas of humans; and (b) information of the width, height and size of the three detected human areas, respectively.
Figure 9. Example of different sizes of human areas resulting from camera viewing direction and perspective projection: (a) input image, including three detected areas of humans; and (b) information of the width, height and size of the three detected human areas, respectively.
Sensors 15 06763 g009
In our research, we do not actually perform the method of camera perspective projection. However, we consider only the concept of camera perspective projection (the size of the object in the camera image becomes larger with a smaller Z distance, whereas it becomes smaller with a larger Z distance) for determining whether the detected candidate should be separated or not. Therefore, actual camera calibration is not performed in our method.
Then, the correct areas of the human are detected by using size filtering based on the size and the ratio of the height to width of the region. From that, we can obtain the final human areas, excluding other objects, such as vehicles, as shown in Figure 10c.
Figure 10. Result of obtaining the final region of the human area: (a) result after the process based on the separation of histogram information; (b) result after the process based on camera viewing direction and perspective projection; and (c) result of the final detected region of the human area.
Figure 10. Result of obtaining the final region of the human area: (a) result after the process based on the separation of histogram information; (b) result after the process based on camera viewing direction and perspective projection; and (c) result of the final detected region of the human area.
Sensors 15 06763 g010aSensors 15 06763 g010b

3. Experimental Results

3.1. Dataset Description

In this research, we used two thermal databases for the experiments. As explained in Section 1, we call the image acquired in the sub-bands of MWIR and LWIR the thermal image [6] in this paper. The first database was object tracking and classification beyond visible spectrum (OTCBVS) benchmark dataset [31]. This has been widely used as an open database for the performance evaluation of object detection in thermal imaging. It contains ten categorized sequences of thermal images obtained at different times and in different weather conditions. The dataset covers various environmental conditions, such as morning, afternoon and rainy and sunny days. Each sequence contains from 18 to 73 frames, captured within one minute, such that the environmental factors, such as precipitation and temperature, remain unchanged [27]. There are 284 images. They are from 30-Hz video from an IR camera. The size of each image is 360 × 240 pixels. They were captured from the same location. Motionless humans are presented in Sequence 8. We tested all of the database images in the ten sequences.
To validate the applicability of the proposed method irrespective of the kind of database, we also created a second database with our thermal camera. For convenience, we call this “our database”. This dataset has seven categorized sequences of thermal images that were captured with different behaviors, such as walking, running, standing and sitting. Each sequence contains from 64 to 144 frames. The total number of images is 768. They were captured using an ICI 7320 thermal camera [32]. The size of each image is 320 × 240 pixels of 14 bits. Humans in the images are approximately 20 to 68 pixels in width and 34 to 103 pixels in height. Motionless humans are presented in Sequences 4, 5, 6 and 7.
We present the results of generating the background image in Section 3.2 and detection results with comparisons with other methods in Section 3.3.

3.2. Results of Generating Background

To demonstrate the performance of the proposed method in generating a background, the background image from the proposed method is compared with those of other methods, as presented in Figure 11. In the previous method [28], they obtained a background image using a simple averaging operation. In [26], they obtained a background image by averaging the frames in two different sequences.
Figure 11. Comparisons of generated background images with the OTCBVS benchmark dataset. The left-upper [28], right-upper [26], left-lower [24,25,33], and right-lower figures are generated by previous methods and the proposed one, respectively.
Figure 11. Comparisons of generated background images with the OTCBVS benchmark dataset. The left-upper [28], right-upper [26], left-lower [24,25,33], and right-lower figures are generated by previous methods and the proposed one, respectively.
Sensors 15 06763 g011
Because there exist motionless people in all of the images of Figure 11, the human areas are not completely removed from background images by the previous methods [26,28], as illustrated in the left and middle images of Figure 11, respectively. The brightness of the entire image is also changed by averaging the frames in two different sequences in [26]. This brightness change of some parts, such as trees and vehicles, causes the incorrect detection of human areas in an input image by background subtraction, because there is a difference between the brightness of the parts of background and the input images. However, all of these human areas are removed using the proposed method while maintaining the brightness of the entire image, including trees and vehicles, as displayed in the right image of Figure 11. From that, we can confirm that the correct background can be obtained using the proposed method.
Most previous research used the simple temporal averaging [26,28,34,35,36] and temporal median filtering [37] for background generation, which cannot solve the problems of motionless humans. We additionally compared other methods (not using simple temporal averaging and temporal median filtering, however; explicitly trying to solve the problems of motionless humans) [24,25,33] to our method for background generation. In the research by [24,25,33], they generated the statistical background model by calculating weighted means and variances of the sampled values. As shown in Figure 11, the human areas are not completely removed from background images by the previous methods [24,25,33], whereas all of these human areas are removed using the proposed method, which shows that our method of background generation outperforms the previous method [24,25,33].
In Figure 12, the background images from our database (second database) generated by the proposed method in various environmental conditions are presented. The human areas are not completely removed from the background images by the previous methods [26,28], as indicated in the left and middle images of Figure 12, respectively. Comparing the images from [26,28], the distinctiveness of the human areas is reduced significantly more by [26,28]. The brightness of the entire image is changed by averaging the frames in the two different sequences in [26]. This brightness change of the entire image causes the incorrect detection of human areas in the input image by background subtraction, because there is a difference between the brightness of the background and the input images. However, human areas are not contained in the background images from the proposed method, and we can confirm that a correct background can be obtained using the proposed method.
Figure 12. Comparisons of generated background images with our database (second database). The left, middle and right figures of (a,b) are by simple temporal averaging operation [28], averaging the frames in two difference sequences [26] and the proposed method, respectively: (a) with Sequence 4 of [28] (left figure) and the proposed method (right figure) and with Sequences 4 and 1 of [26] (middle figure); (b) with Sequence 5 of [28] (left figure) and the proposed method (right figure) and with Sequences 5 and 2 of [26] (middle figure).
Figure 12. Comparisons of generated background images with our database (second database). The left, middle and right figures of (a,b) are by simple temporal averaging operation [28], averaging the frames in two difference sequences [26] and the proposed method, respectively: (a) with Sequence 4 of [28] (left figure) and the proposed method (right figure) and with Sequences 4 and 1 of [26] (middle figure); (b) with Sequence 5 of [28] (left figure) and the proposed method (right figure) and with Sequences 5 and 2 of [26] (middle figure).
Sensors 15 06763 g012

3.3. Detection Results

In this subsection, we present the detection results of the next experiment using the proposed method. In Figure 13, the detection results of various frames are presented. The images were captured at different times and in different weather conditions. Each detected region of a human is surrounded by a white box. In all cases, the human areas were detected successfully, in spite of the overlapping of humans (Figure 13a,d–f), darker human areas than the background (Figure 13b), vehicles, which have a similar intensity as humans (Figure 13c,d), a human using an umbrella (Figure 13a) and various kinds of human behavior, such as walking, running, sitting and standing (Figure 13g–j).
Figure 13. Detection results with the OTCBVS benchmark dataset (af) and our database (gj). Results of images in: (a) Sequence 1; (b) Sequence 3; (c) Sequence 4; (d) Sequence 5; (e) Sequence 6; (f) Sequence 7; (g) Sequence 2; (h) Sequence 3; (i) Sequence 4; and (j) Sequence 6.
Figure 13. Detection results with the OTCBVS benchmark dataset (af) and our database (gj). Results of images in: (a) Sequence 1; (b) Sequence 3; (c) Sequence 4; (d) Sequence 5; (e) Sequence 6; (f) Sequence 7; (g) Sequence 2; (h) Sequence 3; (i) Sequence 4; and (j) Sequence 6.
Sensors 15 06763 g013aSensors 15 06763 g013b
In the next experiments, we quantitatively compared the detection accuracies using the proposed method to those of other methods. For this, we manually depicted bounding boxes on the human areas in the images as ground truth regions. The detection results were evaluated using the Pascal criteria [38] to judge true/false positives by measuring the overlap of the bounding box and a ground truth box. If the area of overlap a o   between the detected bounding box Bd and ground truth box Bgt of Figure 14 exceeded 0.5 (50%) using Equation (8) [19,38,39], we counted the result as a correct detection [39].
a o =   a r e a ( B d B g t ) a r e a ( B d B g t )
where Bd Bgt denotes the intersection of the detected and ground truth bounding boxes. Bd Bgt is their union [39].
Figure 14. Overlapping area of ground truth and detected boxes.
Figure 14. Overlapping area of ground truth and detected boxes.
Sensors 15 06763 g014
Based on Equation (8), we can count the number of true positive (TP) and false positive (FP) detection cases. Positive and negative samples mean the human area and background area, respectively. That is, the TP case is that the human area is correctly detected as a human, whereas the FP case is that the background area is incorrectly detected as a human. Based on this, we measured the accuracies of the human detection in terms of positive predictive value (PPV) (precision) and sensitivity (recall), as indicated in Equations (9) and (10) [15,40]. In these equations, #TP, #FP and #human areas in all of the images represent the number of TP cases, FP cases and human areas in all of the images, respectively. As indicated in Equations (9) and (10), a higher value of PPV and sensitivity means a higher accuracy of human detection.
PPV =   # TP # TP + # FP
Sensitivity =   # TP # human areas in all the images
As indicated in Table 2, the detection accuracy of the proposed method with the OTCBVS benchmark dataset is compared with other methods [15,22,26]. Experimental results confirm that the proposed method outperformed the other methods [15,22,26] in terms of both PPV and sensitivity.
Table 2. Comparison of the detection results for the proposed method and other methods with the OTCBVS benchmark dataset. PPV, positive predictive value.
Table 2. Comparison of the detection results for the proposed method and other methods with the OTCBVS benchmark dataset. PPV, positive predictive value.
Sequence No.12345678910Total
#Frames31282318231822247324284
#People911001011091019794999597984
#TP[15]789570109918864829177845
[22]8894101107909392759595930
[26]91991001091019794999594979
Proposed method91100991091019594999597980
#FP[15]23131062209041
[22]00110001036
[26]00200001036
Proposed method00130000004
PPV[15]0.980.970.840.920.940.980.9410.9110.95
[22]110.990.991110.9910.970.9936
[26]110.9811110.9910.970.9939
Proposed method110.990.971111110.9959
Sensitivity[15]0.860.950.6910.830.910.680.830.960.790.86
[22]0.970.9410.980.890.960.980.7610.980.9459
[26]10.990.991111110.970.9949
Proposed method110.98110.9811110.9959
In Table 3, the detection accuracy of the proposed method with our database (second database) is presented. In Table 3, the PPV and sensitivity are 98.05% and 97.35%, respectively. From Table 2 and Table 3, we can conclude that the proposed method can be applied to thermal images irrespective of the kind of database.
Table 3. Comparison of the detection results for the proposed method and other methods with “our database” (second database).
Table 3. Comparison of the detection results for the proposed method and other methods with “our database” (second database).
Sequence No.1234567Total
#Frames137144648512712784768
#People2033271162382924671681,811
#TP[22]1742851052192893501671,589
[26]203319982382924121681,730
Proposed method2033141142352924371681,763
#FP[22]47212113202115
[26]117060165292
Proposed method01356011035
PPV[22]0.78730.93140.83330.99550.98970.94590.98820.9325
[26]0.99510.949410.975410.96260.76360.9495
Proposed method10.96020.95800.975110.975410.9805
Sensitivity[22]0.85710.87160.90520.92020.98970.74950.9940.8774
[26]10.97550.8448110.882210.9553
Proposed method10.96020.98280.987410.935810.9735
In Figure 15, we show the detection error cases by the proposed method with the open database. There are five people in the upper-left area of an image. However, four people are detected by the proposed algorithm. This is because of erroneous separation caused by the close positions of five people with occlusion.
Figure 15. Detection error cases with the OTCBVS benchmark dataset: (a) original image; (b) result of the proposed method.
Figure 15. Detection error cases with the OTCBVS benchmark dataset: (a) original image; (b) result of the proposed method.
Sensors 15 06763 g015
In Figure 16, we show the detection error cases by the proposed method with our database. In Figure 16, one candidate region including two people is incorrectly detected in the upper area of an image. This is because the two people are overlapped. As shown in Figure 15 and Figure 16, most of the errors are caused by the occlusion of people, and we would research the method for solving this problem based on more accurate segmentation of human areas as future work.
Figure 16. Detection error cases with our database: (a) original image; (b) result of the proposed method.
Figure 16. Detection error cases with our database: (a) original image; (b) result of the proposed method.
Sensors 15 06763 g016

4. Conclusions

In this study, we presented a new approach to detect humans in thermal images based on the generation of a background image in various environmental conditions. A correct background image was generated by image averaging and detection and erasing methods of human areas. A difference image was obtained by combining pixel and edge difference images. Human candidate regions, detected in the difference image, were divided and redefined based on histogram information, perspective projection and intensity of the background image. Based on the redefined candidate region, correct human areas were detected. The optimal thresholds for generating difference images and defining the candidate region were adaptively determined based on the brightness of the generated background. The experimental results confirmed that the detection accuracies of the proposed method were higher than other methods.
In future work, we plan to apply the proposed method to images captured by a visible light camera. Additionally, we will research the method for solving the problem of the inaccurate detection of human areas caused by occlusion. Furthermore, we will expand the proposed research into human activity recognition.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2012R1A1A2038666).

Author Contributions

Eun Som Jeon and Kang Ryoung Park designed the overall system and made the human detection algorithm. In addition, they wrote and revised the paper. Jong-Suk Choi, Ji Hoon Lee and Kwang Yong Shin helped to make the algorithm for background generation. Yeong Gon Kim and Toan Thanh Le helped to collect our database and to measure the accuracies of human detection.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arandjelović, O. Contextually Learnt Detection of Unusual Motion-Based Behaviour in Crowded Public Spaces. In Proceedings of the 26th Annual International Symposium on Computer and Information Science, London, UK, 26–28 September 2011; pp. 403–410.
  2. Martin, R.; Arandjelović, O. Multiple-Object Tracking in Cluttered and Crowded Public Spaces. Lect. Notes Comput. Sci. 2010, 6455, 89–98. [Google Scholar]
  3. Khatoon, R.; Saqlain, S.M.; Bibi, S. A Robust and Enhanced Approach for Human Detection in Crowd. In Proceedings of the International Multitopic Conference, Islamabad, Pakistan, 13–15 December 2012; pp. 215–221.
  4. Rajaei, A.; Shayegh, H.; Charkari, N.M. Human Detection in Semi-Dense Scenes Using HOG descriptor and Mixture of SVMs. In Proceedings of the International Conference on Computer and Knowledge Engineering, Mashhad, Iran, 31 October–1 November 2013; pp. 229–234.
  5. Mahapatra, A.; Mishra, T.K.; Sa, P.K.; Majhi, B. Background Subtraction and Human Detection in Outdoor Videos Using Fuzzy Logic. In Proceedings of the IEEE International Conference on Fuzzy Systems, Hyderabad, India, 7–10 July 2013; pp. 1–7.
  6. Ghiass, R.S.; Arandjelović, O.; Bendada, H.; Maldague, X. Infrared Face Recognition: A Literature Review. In Proceedings of the International Joint Conference on Neural Networks, Dallas, TX, USA, 4–9 August 2013; pp. 1–10.
  7. Bertozzi, M.; Broggi, A.; Rose, M.D.; Felisa, M.; Rakotomamonjy, A.; Suard, F. A Pedestrian De-Tector Using Histograms of Oriented Gradients and a Support Vector Machine Classifier. In Proceedings of the IEEE Conference on Intelligent Transportation Systems, Seattle, WA, USA, 30 September–3 October 2007; pp. 143–148.
  8. Li, Z.; Zhang, J.; Wu, Q.; Geers, G. Feature Enhancement Using Gradient Salience on Thermal Image. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications, Sydney, Australia, 1–3 December 2010; pp. 556–562.
  9. Chang, S.L.; Yang, F.T.; Wu, W.P.; Cho, Y.A.; Chen, S.W. Nighttime Pedestrian Detection Using Thermal Imaging Based on HOG Feature. In Proceedings of the International Conference on System Science and Engineering, Macao, China, 8–10 June 2011; pp. 694–698.
  10. Bertozzi, M.; Broggi, A.; Caraffi, C.; Rose, M.D.; Felisa, M.; Vezzoni, G. Pedestrian Detection by Means of Far-Infrared Stereo Vision. Comput. Vis. Image Underst. 2007, 106, 194–204. [Google Scholar] [CrossRef]
  11. St-Laurent, L.; Prévost, D.; Maldague, X. Thermal Imaging for Enhanced Foreground-Background Segmentation. In Proceedings of the International Conference on Quantitative InfraRed Thermography, Padova, Italy, 27–30 June 2006.
  12. Lin, C.F.; Lin, S.F.; Hwang, C.H.; Chen, Y.C. Real-Time Pedestrian Detection System with Novel Thermal Features at Night. In Proceedings of the IEEE International Instrumentation and Measurement Technology Conference, Montevideo, Uruguay, 12–15 May 2014; pp. 1329–1333.
  13. Zhao, J.; Cheung, S.C.S. Human Segmentation by Fusing Visible-light and Thermal Imaginary. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 1185–1192.
  14. Chen, Y.; Han, C. Night-Time Pedestrian Detection by Visual-Infrared Video Fusion. In Proceedings of the World Congress on Intelligent Control and Automation, Chongqing, China, 25–27 June 2008; pp. 5079–5084.
  15. Li, W.; Zheng, D.; Zhao, T.; Yang, M. An Effective Approach to Pedestrian Detection in Thermal Imagery. In Proceedings of the International Conference on Natural Computation, Chongqing, China, 29–31 May 2012; pp. 325–329.
  16. Neagoe, V.E.; Ciotec, A.D.; Barar, A.P. A Concurrent Neural Network Approach to Pedestrian Detection in Thermal Imagery. In Proceedings of the International Conference on Communications, Bucharest, Romania, 21–23 June 2012; pp. 133–136.
  17. Zhang, L.; Wu, B.; Nevatia, R. Pedestrian Detection in Infrared Images Based on Local Shape Features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8.
  18. Olmeda, D.; Armingol, J.M.; Escalera, A.D.L. Discrete Features for Rapid Pedestrian Detection in Infrared Images. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012; pp. 3067–3072.
  19. Wang, W.; Zhang, J.; Shen, C. Improved Human Detection and Classification in Thermal Images. In Proceedings of the IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 2313–2316.
  20. Wang, W.; Wang, Y.; Chen, F.; Sowmya, A. A Weakly Supervised Approach for Object Detection Based on Soft-Label Boosting. In Proceedings of the IEEE Workshop on Applications of Computer Vision, Tempa, FL, USA, 15–17 January 2013; pp. 331–338.
  21. Davis, J.W.; Sharma, V. Robust Detection of People in Thermal Imagery. In Proceedings of the International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004; pp. 713–716.
  22. Davis, J.W.; Keck, M.A. A Two-Stage Template Approach to Person Detection in Thermal Imagery. In Proceedings of the IEEE Workshop on Applications of Computer Vision, Breckenridge, CO, USA, 5–7 January 2005; pp. 364–369.
  23. Latecki, L.J.; Miezianko, R.; Pokrajac, D. Tracking Motion Objects in Infrared Videos. In Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance, Como, Italy, 15–16 September 2005; pp. 99–104.
  24. Davis, J.W.; Sharma, V. Background-Subtraction Using Contour-Based Fusion of Thermal and Visible Imagery. Comput. Vis. Image Underst. 2007, 106, 162–182. [Google Scholar] [CrossRef]
  25. Davis, J.W.; Sharma, V. Fusion-Based Background-Subtraction Using Contour Saliency. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops, San Diego, CA, USA, 25 June 2005; pp. 1–9.
  26. Dai, C.; Zheng, Y.; Li, X. Layered Representation for Pedestrian Detection and Tracking in Infrared Imagery. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops, San Diego, CA, USA, 25 June 2005; pp. 1–8.
  27. Dai, C.; Zheng, Y.; Li, X. Pedestrian Detection and Tracking in Infrared Imagery Using Shape and Appearance. Comput. Vis. Image Underst. 2007, 106, 288–299. [Google Scholar] [CrossRef]
  28. Calafut, M. Multiple-Object Tracking in the Infrared, Final Project (EE368) of Stanford University; Stanford University: Stanford, CA, USA; pp. 1–6.
  29. Li, J.; Gong, W. Real Time Pedestrian Tracking Using Thermal Infrared Imagery. J. Comput. 2010, 5, 1606–1613. [Google Scholar]
  30. Niblack, W. An Introduction to Digital Image Processing, 1st ed.; Prentice Hall: Englewood Cliffs, NJ, USA, 1986. [Google Scholar]
  31. OTCBVS Benchmark Dataset Collection. Available online: http://www.cse.ohio-state.edu/otcbvs-bench/ (accessed on 13 December 2014).
  32. ICI 7320 Scientific Specifications. Available online: http://www.infraredcamerasinc.com/Thermal-Cameras/Fix-Mounted-Thermal-Cameras/ICI7320_S_fix-mounted_thermal_camera.html (accessed on 13 December 2014).
  33. Davis, J.W.; Sharma, V. Background-Subtraction in Thermal Imagery Using Contour Saliency. Int. J. Comput. Vis. 2007, 71, 161–181. [Google Scholar] [CrossRef]
  34. Dagless, E.L.; Ali, A.T.; Cruz, J.B. Visual Road Traffic Monitoring and Data Collection. In Proceedings of the IEEE-IEE Vehicle Navigation and Information Systems Conference, Ottawa, ON, Canada, 12–15 October 1993; pp. 146–149.
  35. Elhabian, S.Y.; El-Sayed, K.M.; Ahmed, S.H. Moving Object Detection in Spatial Domain Using Background Removal Techniques-State-of-Art. Recent Pat. Comput. Sci. 2008, 1, 32–54. [Google Scholar] [CrossRef]
  36. Zheng, Y.; Fan, L. Moving Object Detection Based on Running Average Background and Temporal Difference. In Proceedings of the International Conference on Intelligent Systems and Knowledge Engineering, Hangzhou, China, 15–16 November 2010; pp. 270–272.
  37. Malviya, A.; Bhirud, S.G. Visual Infrared Video Fusion for Night Vision Using Background Estimation. J. Comput. 2010, 2, 66–69. [Google Scholar]
  38. Olmeda, D.; Premebida, C.; Nunes, U.; Armingol, J.M.; Escalera, A.D.L. Pedestrian Detection in Far Infrared Images. Integr. Comput. Aided Eng. 2013, 20, 347–360. [Google Scholar]
  39. Everingham, M.; Gool, L.V.; Williams, C.K. I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  40. Sensitivity and Specificity. Available online: http://en.wikipedia.org/wiki/Sensitivity_and_specificity (accessed on 13 December 2014).

Share and Cite

MDPI and ACS Style

Jeon, E.S.; Choi, J.-S.; Lee, J.H.; Shin, K.Y.; Kim, Y.G.; Le, T.T.; Park, K.R. Human Detection Based on the Generation of a Background Image by Using a Far-Infrared Light Camera. Sensors 2015, 15, 6763-6788. https://doi.org/10.3390/s150306763

AMA Style

Jeon ES, Choi J-S, Lee JH, Shin KY, Kim YG, Le TT, Park KR. Human Detection Based on the Generation of a Background Image by Using a Far-Infrared Light Camera. Sensors. 2015; 15(3):6763-6788. https://doi.org/10.3390/s150306763

Chicago/Turabian Style

Jeon, Eun Som, Jong-Suk Choi, Ji Hoon Lee, Kwang Yong Shin, Yeong Gon Kim, Toan Thanh Le, and Kang Ryoung Park. 2015. "Human Detection Based on the Generation of a Background Image by Using a Far-Infrared Light Camera" Sensors 15, no. 3: 6763-6788. https://doi.org/10.3390/s150306763

APA Style

Jeon, E. S., Choi, J. -S., Lee, J. H., Shin, K. Y., Kim, Y. G., Le, T. T., & Park, K. R. (2015). Human Detection Based on the Generation of a Background Image by Using a Far-Infrared Light Camera. Sensors, 15(3), 6763-6788. https://doi.org/10.3390/s150306763

Article Metrics

Back to TopTop