[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (8)

Search Parameters:
Keywords = ACF detector

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 39331 KiB  
Article
Fast Rock Detection in Visually Contaminated Mining Environments Using Machine Learning and Deep Learning Techniques
by Reinier Rodriguez-Guillen, John Kern and Claudio Urrea
Appl. Sci. 2024, 14(2), 731; https://doi.org/10.3390/app14020731 - 15 Jan 2024
Cited by 2 | Viewed by 1596
Abstract
Advances in machine learning algorithms have allowed object detection and classification to become booming areas. The detection of objects, such as rocks, in mining operations is affected by fog, snow, suspended particles, and high lighting. These environmental conditions can stop the development of [...] Read more.
Advances in machine learning algorithms have allowed object detection and classification to become booming areas. The detection of objects, such as rocks, in mining operations is affected by fog, snow, suspended particles, and high lighting. These environmental conditions can stop the development of mining work, which entails a considerable increase in operating costs. It is vital to select a machine learning algorithm that is accurate, fast, and contributes to lower operational costs because of the aforementioned environmental situations. In this study, the Viola-Jones algorithm, Aggregate Channel Features (ACF), Faster Regions with Convolutional Neural Networks (Faster R-CNN), Single-Shot Detector (SSD), and You Only Look Once (YOLO) version 4 were analyzed, considering the precision metrics, recall, AP50, and average detection time. In our preliminary tests, we have observed that the differences between YOLO v4 and the latest versions are not substantial for the specific problem of rock detection addressed in our article. Therefore, YOLO v4 is an appropriate and representative choice for evaluating the effectiveness of existing methods in our study. The YOLO v4 algorithm performed the best overall, whereas the SSD algorithm performed the fastest. The results indicate that the YOLO v4 algorithm is a promising candidate for detecting rocks with visual contamination in mining operations. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Viola-Jones algorithm [<a href="#B29-applsci-14-00731" class="html-bibr">29</a>].</p>
Full article ">Figure 2
<p>ACF algorithm [<a href="#B30-applsci-14-00731" class="html-bibr">30</a>].</p>
Full article ">Figure 3
<p>Faster R-CNN structure [<a href="#B31-applsci-14-00731" class="html-bibr">31</a>].</p>
Full article ">Figure 4
<p>SSD structure [<a href="#B10-applsci-14-00731" class="html-bibr">10</a>].</p>
Full article ">Figure 5
<p>YOLO v4 structure [<a href="#B32-applsci-14-00731" class="html-bibr">32</a>].</p>
Full article ">Figure 6
<p>Units of ResNet and MobileNet. (<b>a</b>) Basic unit of ResNet [<a href="#B33-applsci-14-00731" class="html-bibr">33</a>]. (<b>b</b>) Basic unit of MobileNet [<a href="#B35-applsci-14-00731" class="html-bibr">35</a>].</p>
Full article ">Figure 7
<p>Database with visual contamination [<a href="#B25-applsci-14-00731" class="html-bibr">25</a>]. (<b>a</b>) Normal. (<b>b</b>) Fog. (<b>c</b>) Rain. (<b>d</b>) Snow. (<b>e</b>) High illumination.</p>
Full article ">Figure 8
<p>Labeling problems. (<b>a</b>) Database label. (<b>b</b>) Our label.</p>
Full article ">Figure 9
<p>Methodology for the training and testing of the Viola-Jones detector [<a href="#B42-applsci-14-00731" class="html-bibr">42</a>].</p>
Full article ">Figure 10
<p>Results of the Viola-Jones detector. (<b>a</b>) Feature Haar-like. (<b>b</b>) Feature LBP. (<b>c</b>) Feature HOG.</p>
Full article ">Figure 11
<p>Results of the ACF detector. (<b>a</b>) Exp-1. (<b>b</b>) Exp-2. (<b>c</b>) Exp-3. (<b>d</b>) Exp-4.</p>
Full article ">Figure 12
<p>Results of the detectors in database. (<b>a</b>) ACF detector. (<b>b</b>) Viola-Jones detector. (<b>c</b>) Faster R-CNN detector. (<b>d</b>) SSD detector. (<b>e</b>) YOLO v4 detector.</p>
Full article ">Figure 13
<p>Results of the detectors out of the database. (<b>a</b>) ACF detector. (<b>b</b>) Viola-Jones detector. (<b>c</b>) Faster R-CNN detector. (<b>d</b>) SSD detector. (<b>e</b>) YOLO v4 detector.</p>
Full article ">Figure 14
<p>AP of the deep learning algorithms.</p>
Full article ">Figure 15
<p>P-R curve for the deep learning algorithms.</p>
Full article ">Figure 16
<p>Average detection time.</p>
Full article ">Figure 17
<p>Rock detection in various evaluated scenarios of visual contamination.</p>
Full article ">Figure 18
<p>YOLO v4 algorithm failure.</p>
Full article ">
23 pages, 19373 KiB  
Article
Development of a Hybrid Method for Multi-Stage End-to-End Recognition of Grocery Products in Shelf Images
by Ceren Gulra Melek, Elena Battini Sonmez, Hakan Ayral and Songul Varli
Electronics 2023, 12(17), 3640; https://doi.org/10.3390/electronics12173640 - 29 Aug 2023
Cited by 1 | Viewed by 2290
Abstract
Product recognition on grocery shelf images is a compelling task of object detection because of the similarity between products, the presence of the different scale of product sizes, and the high number of classes, in addition to constantly renewed packaging and added new [...] Read more.
Product recognition on grocery shelf images is a compelling task of object detection because of the similarity between products, the presence of the different scale of product sizes, and the high number of classes, in addition to constantly renewed packaging and added new products’ difficulty in data collection. The use of conventional methods alone is not enough to solve a number of retail problems such as planogram compliance, stock tracking on shelves, and customer support. The purpose of this study is to achieve significant results using the suggested multi-stage end-to-end process, including product detection, product classification, and refinement. The comparison of different methods is provided by a traditional computer vision approach, Aggregate Channel Features (ACF) and Single-Shot Detectors (SSD) are used in the product detection stage, and Speed-up Robust Features (SURF), Binary Robust Invariant Scalable Key points (BRISK), Oriented Features from Accelerated Segment Test (FAST), Rotated Binary Robust Independent Elementary Features (BRIEF) (ORB), and hybrids of these methods are used in the product classification stage. The experimental results used the entire Grocery Products dataset and its different subsets with a different number of products and images. The best performance was achieved with the use of SSD in the product detection stage and the hybrid use of SURF, BRISK, and ORB in the product classification stage, respectively. Additionally, the proposed approach performed comparably or better than existing models. Full article
(This article belongs to the Special Issue Application of Machine Learning in Graphics and Images)
Show Figures

Figure 1

Figure 1
<p>Sample images representing the challenges of grocery product recognition problem: (<b>a</b>,<b>b</b>) different viewing angle shelf image in SKU-110K [<a href="#B13-electronics-12-03640" class="html-bibr">13</a>]. (<b>c</b>) Blurred shelf image in grocery products [<a href="#B9-electronics-12-03640" class="html-bibr">9</a>]. (<b>d</b>,<b>e</b>) Different shelf design images in SKU-110K [<a href="#B13-electronics-12-03640" class="html-bibr">13</a>]. (<b>f</b>) Cluttered background shelf image in SKU-110K [<a href="#B13-electronics-12-03640" class="html-bibr">13</a>]. (<b>g</b>,<b>h</b>) Product images having high packaging design similarity among different product types in grocery products [<a href="#B9-electronics-12-03640" class="html-bibr">9</a>].</p>
Full article ">Figure 2
<p>Multi-stage end-to-end recognition process. (Red frames indicate the bounding boxes obtained from each step.)</p>
Full article ">Figure 3
<p>The steps of the proposed traditional computer vision approach in Stage-1: (<b>a</b>) original shelf image; (<b>b</b>) shelf image with detected shelf lines after first step; (<b>c</b>) shelf image with detected product regions after second step; (<b>d</b>) shelf image with completed product regions after third step.</p>
Full article ">Figure 4
<p>(<b>a</b>) The steps of the ACF in Stage-1. (<b>b</b>) The steps of the SSD in Stage-1. (Red frames indicate the obtained bounding boxes predictions.)</p>
Full article ">Figure 5
<p>The sample images of the datasets: (<b>a</b>,<b>b</b>) sample shelf images of SKU-110K [<a href="#B13-electronics-12-03640" class="html-bibr">13</a>]; (<b>c</b>,<b>d</b>) sample shelf images of the Grocery Dataset [<a href="#B12-electronics-12-03640" class="html-bibr">12</a>]; (<b>e</b>,<b>f</b>) sample shelf images of GP-20 [<a href="#B10-electronics-12-03640" class="html-bibr">10</a>]; (<b>g</b>,<b>h</b>) sample shelf images of GP-181 [<a href="#B11-electronics-12-03640" class="html-bibr">11</a>]; (<b>i</b>,<b>j</b>) sample shelf images of Grocery Products [<a href="#B9-electronics-12-03640" class="html-bibr">9</a>]; (<b>k</b>,<b>l</b>) sample template images of Grocery Products [<a href="#B9-electronics-12-03640" class="html-bibr">9</a>]. (Red frames indicates the ground truth information of annotation file of datasets).</p>
Full article ">
20 pages, 6342 KiB  
Article
Aggregate Channel Features and Fast Regions CNN Approach for Classification of Ship and Iceberg
by Sivapriya Sethu Ramasubiramanian, Suresh Sivasubramaniyan and Mohamed Fathimal Peer Mohamed
Appl. Sci. 2023, 13(12), 7292; https://doi.org/10.3390/app13127292 - 19 Jun 2023
Cited by 1 | Viewed by 1768
Abstract
Detection and classification of icebergs and ships in synthetic aperture radar (SAR) images play a vital role in marine surveillance systems even though available adaptive threshold methods give satisfying results on detection and classification for ships and icebergs, including techniques of convolutional neural [...] Read more.
Detection and classification of icebergs and ships in synthetic aperture radar (SAR) images play a vital role in marine surveillance systems even though available adaptive threshold methods give satisfying results on detection and classification for ships and icebergs, including techniques of convolutional neural networks (CNNs), but need more accuracy and precision. An efficient and accurate method was developed to detect and classify the ship and icebergs. Hence, the research method proposed locating and classifying both ships and icebergs in a given SAR image with the help of deep learning (DL) and non-DL methods. A non-DL method utilized here was the aggregate channel features (ACF) detector, which extracts region proposals from huge SAR images. The DL object detector called fast regions CNN (FRCNN) detects objects accurately from the result of ACF since the ACF method avoids unwanted regions. The novelty of this study was that ACF-FRCNN concentrates only on accurately classifying ships and icebergs. The proposed ACF-FRCNN method gave a better performance in terms of loss (18.32%), accuracy (96.34%), recall (98.32%), precision (95.97%), and the F1 score (97.13%). Compared to other conventional methods, the combined effect of ACF and FRCNN increased the speed and quality of the detection of ships and icebergs. Thus, the ACF-FRCNN method is considered a novel method for over 75 × 75 resolution ship and iceberg SAR images. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Top-down pathway, (<b>b</b>) bottom-up pathway.</p>
Full article ">Figure 1 Cont.
<p>(<b>a</b>) Top-down pathway, (<b>b</b>) bottom-up pathway.</p>
Full article ">Figure 2
<p>Proposed ACF detector model.</p>
Full article ">Figure 3
<p>ResNet 50 architecture model.</p>
Full article ">Figure 4
<p>Proposed FRCNN model.</p>
Full article ">Figure 5
<p>Proposed Framework’s Architecture Diagram.</p>
Full article ">Figure 6
<p>Generated Image and Training Fashion.</p>
Full article ">Figure 7
<p>HoG of SAR image.</p>
Full article ">Figure 8
<p>Output of ACF detectors.</p>
Full article ">Figure 9
<p>Output of FRCNN with class of iceberg.</p>
Full article ">Figure 10
<p>Output of FRCNN with class of ship.</p>
Full article ">Figure 11
<p>Percentage of loss parameter comparison among proposed and existing methods.</p>
Full article ">Figure 12
<p>Percentage of accuracy parameter comparison among proposed and existing methods.</p>
Full article ">Figure 13
<p>Performance metric graph for detection.</p>
Full article ">
18 pages, 4050 KiB  
Article
Adsorption Characteristics of Activated Carbon Fibers in Respirator Cartridges for Toluene
by Jo Anne G. Balanay and Jonghwa Oh
Int. J. Environ. Res. Public Health 2021, 18(16), 8505; https://doi.org/10.3390/ijerph18168505 - 12 Aug 2021
Cited by 6 | Viewed by 2422
Abstract
Respirator use has been shown to be associated with overall discomfort. Activated carbon fiber (ACF) has potential as an alternative adsorbent for developing thinner, lightweight, and efficient respirators due to its larger surface area, microporosity, and fabric form. The purpose of this pilot [...] Read more.
Respirator use has been shown to be associated with overall discomfort. Activated carbon fiber (ACF) has potential as an alternative adsorbent for developing thinner, lightweight, and efficient respirators due to its larger surface area, microporosity, and fabric form. The purpose of this pilot study was to determine the adsorption characteristics of commercially available ACF in respirator cartridges with varying ACF composition for toluene protection. Seven ACF types (one cloth, six felt) with varying properties were tested. Seven ACF cartridge configurations with varying ACF composition were challenged with five toluene concentrations (20–500 ppm) at constant air temperature (23 °C), relative humidity (50%), and air flow (32 LPM). Breakthrough curves were obtained using photoionization detectors. Breakthrough times (10%, 50%, and 5 ppm) and adsorption capacities were compared among ACF cartridge configurations to determine their suitable application in respiratory protection. Results showed that ACF cartridges containing the densest ACF felt types had the longest average breakthrough times (e.g., ~250–270 min to reach 5 ppm breakthrough time) and those containing ACF felt types with the highest specific surface areas had the highest average adsorption capacity (~450–470 mg/g). The ACF cartridges demonstrated breakthrough times of <1 h for 500 ppm toluene and 8–16 h for 20 ppm toluene. The ACF cartridges are more reliable for use at low ambient toluene concentrations but still have potential for use at higher concentrations for short-term protection. ACF felt forms with appropriate properties (density of ~0.07 g/cm3; specific surface area of ~2000 m2/g) have shown promising potential for the development of lighter and thinner respirators for protection against toluene. Full article
Show Figures

Figure 1

Figure 1
<p>Experimental setup for breakthrough determination for toluene across an ACF respirator cartridge.</p>
Full article ">Figure 2
<p>SEM images of activated carbon fibers (ACF) at 200× magnification. ACF felt types (AF1 to AF6); ACF cloth type (AC1).</p>
Full article ">Figure 3
<p>Breakthrough curves by cartridge type at 5 toluene concentrations: (<b>a</b>) 20 ppm, (<b>b</b>) 100 ppm, (<b>c</b>) 200 ppm, (<b>d</b>) 300 ppm, (<b>e</b>) 500 ppm. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">Figure 4
<p>Breakthrough curves for (<b>a</b>) CS1 (100% AF1) and (<b>b</b>) CS6 (100% AF6) cartridge types by toluene concentration.</p>
Full article ">Figure 5
<p>Breakthrough times as a function of toluene concentration by cartridge type. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">Figure 6
<p>Average breakthrough times by cartridge type. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">Figure 7
<p>Toluene 5 ppm breakthrough time by cartridge type and toluene concentration. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">Figure 8
<p>Average adsorption capacity by cartridge type. Shown with error bars. The letters above each bar represent differences or similarities between cartridge types. Cartridge types with the same letter are not significantly different by average comparison. Cartridge types with a different letter are significantly different by average comparison. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">Figure 9
<p>Adsorption capacity by toluene concentration and cartridge type. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">Figure 10
<p>Toluene adsorption isotherms at 23 °C by cartridge type over the relative pressure (P/P0) range 0.000–0.015. Cartridge type (ACF composition): CS1 (100% AF1); CS2 (100% AF2); CS3 (100% AF3); CS4 (100% AF4); CS5 (100% AF5); CS6 (100% AF6); CC7 (71% AF1/29% AC1).</p>
Full article ">
20 pages, 4201 KiB  
Article
Tuning of Classifiers to Speed-Up Detection of Pedestrians in Infrared Images
by Karol Piniarski, Paweł Pawłowski and Adam Dąbrowski
Sensors 2020, 20(16), 4363; https://doi.org/10.3390/s20164363 - 5 Aug 2020
Cited by 7 | Viewed by 2980
Abstract
This paper presents an experimental evaluation of real-time pedestrian detection algorithms and their tuning using the proposed universal performance index. With this index, the precise choice of various parameters is possible. Moreover, we determined the best resolution of the analysis window, which is [...] Read more.
This paper presents an experimental evaluation of real-time pedestrian detection algorithms and their tuning using the proposed universal performance index. With this index, the precise choice of various parameters is possible. Moreover, we determined the best resolution of the analysis window, which is much lower than the initial window. By such means, we can speed-up the processing (i.e., reduce the classification time by 74%). There are cases in which we increased both the processing speed and the classification accuracy. We made experiments with various baseline detectors and datasets in order to confirm versatility of the proposed ideas. The analyzed classifiers are those typically applied to detection of pedestrians, namely: aggregated channel feature (ACF), deep convolutional neural network (CNN), and support vector machine (SVM). We used a suite of five precisely chosen night (and day) IR vision datasets. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>General pedestrian detection scheme.</p>
Full article ">Figure 2
<p>Processing scheme for tuning pedestrian classification with proposed performance index.</p>
Full article ">Figure 3
<p>Test bed for comparison of tested classifiers.</p>
Full article ">Figure 4
<p>CVC-09 dataset of pedestrians: (<b>a</b>) day-time positive samples, (<b>b</b>) night-time positive samples, (<b>c</b>) day-time negative samples, (<b>d</b>) night-time negative samples.</p>
Full article ">Figure 5
<p>Distribution of pedestrian heights (in pixels) in CVC-09 dataset.</p>
Full article ">Figure 6
<p>NTPD dataset pedestrian (positive) samples.</p>
Full article ">Figure 7
<p>Two illustrative images from OSU dataset.</p>
Full article ">Figure 8
<p>Three positive samples in various resolutions: 64 × 128, 56 × 112, 48 × 96, 40 × 80, 32 × 64, 24 × 48, 16 × 32; original images are in the CVC-09 dataset.</p>
Full article ">Figure 9
<p>Detection rate and processing time as functions of image resolutions: HOG + SVM classifier (left column), ACF detector (middle column), CNN (right column) for the following datasets: LSIFIR (first row: <b>a</b>–<b>c</b>), OSU (second row: <b>d</b>–<b>f</b>).</p>
Full article ">Figure 10
<p>Detection rate and processing time as functions of image resolutions: HOG+SVM classifier (left column), ACF detector (middle column), CNN (right column) for the following datasets: NTPD (first row: <b>a</b>–<b>c</b>), CVC-09 night-time (second row: <b>d</b>–<b>f</b>), CVC-09 day-time (third row: <b>g</b>–<b>i</b>).</p>
Full article ">Figure 11
<p>Performance indices as functions of image resolutions (values on <span class="html-italic">x</span>-axis refer to particular test sets in <a href="#sensors-20-04363-t004" class="html-table">Table 4</a>): for (<b>a</b>,<b>b</b>) <math display="inline"><semantics> <mi>w</mi> </semantics></math> = 0.92, for (<b>c</b>,<b>d</b>) <math display="inline"><semantics> <mi>w</mi> </semantics></math> = 0.95, for (<b>e</b>,<b>f</b>) <math display="inline"><semantics> <mi>w</mi> </semantics></math> = 0.98, with <math display="inline"><semantics> <mi>w</mi> </semantics></math> being the weight of accuracy for various datasets and classifiers indicated with different colors as explained in the legend.</p>
Full article ">
20 pages, 7441 KiB  
Article
Efficient Vehicle Detection and Distance Estimation Based on Aggregated Channel Features and Inverse Perspective Mapping from a Single Camera
by Jong Bae Kim
Symmetry 2019, 11(10), 1205; https://doi.org/10.3390/sym11101205 - 26 Sep 2019
Cited by 21 | Viewed by 5611
Abstract
In this paper a method for detecting and estimating the distance of a vehicle driving in front using a single black-box camera installed in a vehicle was proposed. In order to apply the proposed method to autonomous vehicles, it was required to reduce [...] Read more.
In this paper a method for detecting and estimating the distance of a vehicle driving in front using a single black-box camera installed in a vehicle was proposed. In order to apply the proposed method to autonomous vehicles, it was required to reduce the throughput and speed-up the processing. To do this, the proposed method decomposed the input image into multiple-resolution images for real-time processing and then extracted the aggregated channel features (ACFs). The idea was to extract only the most important features from images at different resolutions symmetrically. A method of detecting an object and a method of estimating a vehicle’s distance from a bird’s eye view through inverse perspective mapping (IPM) were applied. In the proposed method, ACFs were used to generate the AdaBoost-based vehicle detector. The ACFs were extracted from the LUV color, edge gradient, and orientation (histograms of oriented gradients) of the input image. Subsequently, by applying IPM and transforming a 2D input image into 3D by generating an image projected in three dimensions, the distance between the detected vehicle and the autonomous vehicle was detected. The proposed method was applied in a real-world road environment and showed accurate results for vehicle detection and distance estimation in real-time processing. Thus, it was showed that our method is applicable to autonomous vehicles. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of the proposed method.</p>
Full article ">Figure 2
<p>Flowchart of vehicle detection and distance estimation.</p>
Full article ">Figure 3
<p>Feature channel decomposition process for an image.</p>
Full article ">Figure 4
<p>Flowchart of the vehicle detector using ACFs.</p>
Full article ">Figure 5
<p>Training vehicle samples for ACF-based AdaBoost classifier.</p>
Full article ">Figure 6
<p>Results of vehicle detection with the center point of the bounding box.</p>
Full article ">Figure 7
<p>Results of vehicle tracking: (<b>a</b>) pixel-matching region for vehicle tracking in the detected vehicle region; (<b>b</b>) pixel-matching process [<a href="#B47-symmetry-11-01205" class="html-bibr">47</a>].</p>
Full article ">Figure 8
<p>Results of the camera 3D position and pattern board images for extracting the camera parameters: (<b>a</b>) camera parameter estimation using pattern boards; (<b>b</b>) camera parameter estimation errors; (<b>c</b>) position of pattern boards in the 3D space.</p>
Full article ">Figure 9
<p>Mapping of the 3D and 2D coordinates.</p>
Full article ">Figure 10
<p>Results of the inverse perspective transformed image.</p>
Full article ">Figure 11
<p>Comparison of detection performance for a detected region.</p>
Full article ">Figure 12
<p>Results of vehicle average detection rate (<b>a</b>) and log average detection error rate according to vehicle verification step (<b>b</b>).</p>
Full article ">Figure 13
<p>Experiment results of vehicle distance estimation.</p>
Full article ">Figure 14
<p>Results of the proposed method.</p>
Full article ">
11 pages, 1928 KiB  
Article
Enhancing Multi-Camera People Detection by Online Automatic Parametrization Using Detection Transfer and Self-Correlation Maximization
by Rafael Martín-Nieto, Álvaro García-Martín, José M. Martínez and Juan C. SanMiguel
Sensors 2018, 18(12), 4385; https://doi.org/10.3390/s18124385 - 11 Dec 2018
Cited by 3 | Viewed by 2729
Abstract
Finding optimal parametrizations for people detectors is a complicated task due to the large number of parameters and the high variability of application scenarios. In this paper, we propose a framework to adapt and improve any detector automatically in multi-camera scenarios where people [...] Read more.
Finding optimal parametrizations for people detectors is a complicated task due to the large number of parameters and the high variability of application scenarios. In this paper, we propose a framework to adapt and improve any detector automatically in multi-camera scenarios where people are observed from various viewpoints. By accurately transferring detector results between camera viewpoints and by self-correlating these transferred results, the best configuration (in this paper, the detection threshold) for each detector-viewpoint pair is identified online without requiring any additional manually-labeled ground truth apart from the offline training of the detection model. Such a configuration consists of establishing the confidence detection threshold present in every people detector, which is a critical parameter affecting detection performance. The experimental results demonstrate that the proposed framework improves the performance of four different state-of-the-art detectors (DPM , ACF, faster R-CNN, and YOLO9000) whose Optimal Fixed Thresholds (OFTs) have been determined and fixed during training time using standard datasets. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Framework overview.</p>
Full article ">Figure 2
<p>Overview of the proposed technique: (<b>a</b>) shows two detection bounding box examples; (<b>b</b>) schematizes the geometric process; and (<b>c</b>) contains a representation of the resulting cylinders (green), the original bounding box (blue, very tilted due to the angle between the cameras’ viewpoints), and the resulting bounding boxes (red). (<b>a</b>,<b>c</b>) are cropped versions for visualization purposes.</p>
Full article ">Figure 3
<p>Visual example of the correlation between two cameras: (<b>a</b>) Camera 1 (camera under analysis) and detections in red color and (<b>b</b>) Camera 2 and detections in blue color. (<b>c</b>) Camera under analysis (Camera 1) with original red detections <math display="inline"><semantics> <msub> <mi>D</mi> <mn>1</mn> </msub> </semantics></math> and transferred blue ones <math display="inline"><semantics> <msub> <mi>D</mi> <mn>2</mn> </msub> </semantics></math>. In this case, the optimal thresholds according to the correlation between both cameras are <math display="inline"><semantics> <mrow> <mn>0</mn> <mo>.</mo> <mn>4</mn> <mo>&gt;</mo> <msubsup> <mi>τ</mi> <mrow> <mn>1</mn> </mrow> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </msubsup> <mo>≤</mo> <mn>0</mn> <mo>.</mo> <mn>8</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mn>0.2</mn> <mo>&gt;</mo> <msubsup> <mi>τ</mi> <mrow> <mn>2</mn> </mrow> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </msubsup> <mo>≤</mo> <mn>0.3</mn> </mrow> </semantics></math>, respectively. All the images are cropped versions for visualization purposes.</p>
Full article ">Figure 3 Cont.
<p>Visual example of the correlation between two cameras: (<b>a</b>) Camera 1 (camera under analysis) and detections in red color and (<b>b</b>) Camera 2 and detections in blue color. (<b>c</b>) Camera under analysis (Camera 1) with original red detections <math display="inline"><semantics> <msub> <mi>D</mi> <mn>1</mn> </msub> </semantics></math> and transferred blue ones <math display="inline"><semantics> <msub> <mi>D</mi> <mn>2</mn> </msub> </semantics></math>. In this case, the optimal thresholds according to the correlation between both cameras are <math display="inline"><semantics> <mrow> <mn>0</mn> <mo>.</mo> <mn>4</mn> <mo>&gt;</mo> <msubsup> <mi>τ</mi> <mrow> <mn>1</mn> </mrow> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </msubsup> <mo>≤</mo> <mn>0</mn> <mo>.</mo> <mn>8</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mn>0.2</mn> <mo>&gt;</mo> <msubsup> <mi>τ</mi> <mrow> <mn>2</mn> </mrow> <mrow> <mn>1</mn> <mo>,</mo> <mn>2</mn> </mrow> </msubsup> <mo>≤</mo> <mn>0.3</mn> </mrow> </semantics></math>, respectively. All the images are cropped versions for visualization purposes.</p>
Full article ">Figure 4
<p>PETS2009 (<b>a</b>) and EPFL-RLC (<b>b</b>) view planes of each camera (five and three cameras, respectively) and the common field of view of all cameras.</p>
Full article ">Figure 5
<p>Visual examples of the faster R-CNN [<a href="#B13-sensors-18-04385" class="html-bibr">13</a>] detections of all cameras transferred to the evaluation viewpoint. Each camera bounding box is represented with a different color. Sequence PETS S2-L1, Frame 102 (<b>a</b>), and sequence PETS S3MF1, Frame 78 (<b>b</b>).</p>
Full article ">Figure 5 Cont.
<p>Visual examples of the faster R-CNN [<a href="#B13-sensors-18-04385" class="html-bibr">13</a>] detections of all cameras transferred to the evaluation viewpoint. Each camera bounding box is represented with a different color. Sequence PETS S2-L1, Frame 102 (<b>a</b>), and sequence PETS S3MF1, Frame 78 (<b>b</b>).</p>
Full article ">
21 pages, 6395 KiB  
Article
Training-Based Methods for Comparison of Object Detection Methods for Visual Object Tracking
by Ahmad Delforouzi, Bhargav Pamarthi and Marcin Grzegorzek
Sensors 2018, 18(11), 3994; https://doi.org/10.3390/s18113994 - 16 Nov 2018
Cited by 7 | Viewed by 5101
Abstract
Object tracking in challenging videos is a hot topic in machine vision. Recently, novel training-based detectors, especially using the powerful deep learning schemes, have been proposed to detect objects in still images. However, there is still a semantic gap between the object detectors [...] Read more.
Object tracking in challenging videos is a hot topic in machine vision. Recently, novel training-based detectors, especially using the powerful deep learning schemes, have been proposed to detect objects in still images. However, there is still a semantic gap between the object detectors and higher level applications like object tracking in videos. This paper presents a comparative study of outstanding learning-based object detectors such as ACF, Region-Based Convolutional Neural Network (RCNN), FastRCNN, FasterRCNN and You Only Look Once (YOLO) for object tracking. We use an online and offline training method for tracking. The online tracker trains the detectors with a generated synthetic set of images from the object of interest in the first frame. Then, the detectors detect the objects of interest in the next frames. The detector is updated online by using the detected objects from the last frames of the video. The offline tracker uses the detector for object detection in still images and then a tracker based on Kalman filter associates the objects among video frames. Our research is performed on a TLD dataset which contains challenging situations for tracking. Source codes and implementation details for the trackers are published to make both the reproduction of the results reported in this paper and the re-use and further development of the trackers for other researchers. The results demonstrate that ACF and YOLO trackers show more stability than the other trackers. Full article
(This article belongs to the Section Physical Sensors)
Show Figures

Figure 1

Figure 1
<p>Challenges in two videos: First line (David) and second line (Panda) show illumination change, out-of-plane rotation, appearance change and background clutter.</p>
Full article ">Figure 2
<p>Block diagram of the online tracker. In the first frame, from this selected object, synthetic data is generated. Then a detector (i.e., RCNN, FastRCNN, FasterRCNN or ACF) is trained using the generated data and applied to the first segment of frames <math display="inline"><semantics> <msub> <mi>G</mi> <mn>1</mn> </msub> </semantics></math> to detect the objects of interest in them. The detected objects and the synthetic data added to the training vector <math display="inline"><semantics> <msub> <mi>T</mi> <mi>v</mi> </msub> </semantics></math> which is then used to update the detector. This process is continued until the end of the video.</p>
Full article ">Figure 3
<p>Synthetic data for the training of the detectors with their objects of interest, the first frame of “Pedestrain1” (<b>a</b>), this frame with additive salt and pepper noise, with noise density 0.008 (<b>b</b>), rotated version of (<b>a</b>) with −10 and 9 degree (<b>c</b>,<b>d</b>), rotated version of the enhanced image (<b>a</b>) using histogram equalization (<b>e</b>,<b>f</b>) and using contrast adjustment (<b>g</b>,<b>h</b>).</p>
Full article ">Figure 4
<p>Block diagram of the offline tracker. All frames are fed to YOLO and Kalman filter. The offline tracker outputs the YOLO response which has maximum IoU with the estimated pose of Kalman filter. The YOLO response also updates the Kalman filter.</p>
Full article ">Figure 5
<p>Intermediate results of the offline tracker: (<b>a</b>) Three frames of Volkswagen (<b>b</b>) The results of YOLO detector on the frames (<b>c</b>) Applying of the Kalman filter on the detected objects of (<b>b</b>,<b>d</b>) The final results of the offline tracker.</p>
Full article ">Figure 6
<p>The snapshots of videos in our experiments. The figures from (<b>a</b>–<b>j</b>) show David, Jump, Pedestrain1, Pedestrain2, Pedestrain3, Car, Motocross, VW, Carchase and Panda respectively. The ground truth is shown on each image with one green rectangle.</p>
Full article ">Figure 7
<p>Tracker comparison in terms of F-measure.</p>
Full article ">Figure 8
<p>The trackers performance versus the synthetic data length (left number), the training iteration number (middle number) and the training vector <math display="inline"><semantics> <msub> <mi>T</mi> <mi>v</mi> </msub> </semantics></math> length (right number).</p>
Full article ">Figure 9
<p>In the ACF tracker, the updating process was removed. The results of the online tracker and the ablative study are compared.</p>
Full article ">Figure 10
<p>Visualization results of the trackers are shown and compared. The results of RCNN, FastRCNN, FasterRCNN, ACF, YOLO and Ground truth are shown with red, yellow, blue, magenta, cyan and green frames respectively. The figures from (<b>a</b>–<b>j</b>) show Pedestrain1, VW, Motocross, David, Panda, Carchase, Car, Pedestrain2, Pedestrain3 and Jump respectively. Each column includes two randomly selected frames of the same video.</p>
Full article ">Figure 11
<p>The trackers comparison in terms of the running time.</p>
Full article ">
Back to TopTop