[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (29)

Search Parameters:
Keywords = automatic photo identification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 2171 KiB  
Article
Individual Cow Recognition Based on Ultra-Wideband and Computer Vision
by Aruna Zhao, Huijuan Wu, Daoerji Fan and Kuo Li
Animals 2025, 15(3), 456; https://doi.org/10.3390/ani15030456 - 6 Feb 2025
Viewed by 398
Abstract
This study’s primary goal is to use computer vision and ultra-wideband (UWB) localisation techniques to automatically mark numerals in cow photos. In order to accomplish this, we created a UWB-based cow localisation system that involves installing tags on cow heads and placing several [...] Read more.
This study’s primary goal is to use computer vision and ultra-wideband (UWB) localisation techniques to automatically mark numerals in cow photos. In order to accomplish this, we created a UWB-based cow localisation system that involves installing tags on cow heads and placing several base stations throughout the farm. The system can determine the distance between each base station and the cow using wireless communication technology, which allows it to determine the cow’s current location coordinates. The study employed a neural network to train and optimise the ranging data gathered in the 1–20 m range in order to solve the issue of significant ranging errors in conventional UWB positioning systems. The experimental data indicates that the UWB positioning system’s unoptimized range error has an absolute mean of 0.18 m and a standard deviation of 0.047. However, when using a neural network-trained model, the ranging error is much decreased, with an absolute mean of 0.038 m and a standard deviation of 0.0079. The average root mean square error (RMSE) of the positioning coordinates is decreased to 0.043 m following the positioning computation utilising the optimised range data, greatly increasing the positioning accuracy. This study used the conventional camera shooting method for image acquisition. Following image acquisition, the system extracts the cow’s coordinate information from the image using a perspective transformation method. This allows for accurate cow identification and number labelling when compared to the location coordinates. According to the trial findings, this plan, which integrates computer vision and UWB positioning technologies, achieves high-precision cow labelling and placement in the optimised system and greatly raises the degree of automation and precise management in the farming process. This technology has many potential applications, particularly in the administration and surveillance of big dairy farms, and it offers a strong technical basis for precision farming. Full article
(This article belongs to the Section Animal System and Management)
Show Figures

Figure 1

Figure 1
<p>Deployment of experiments.</p>
Full article ">Figure 2
<p>Block diagram of UWB positioning system hardware.</p>
Full article ">Figure 3
<p>Positioning algorithms.</p>
Full article ">Figure 4
<p>BP neural network structure: x is input and y is output.</p>
Full article ">Figure 5
<p>Examples of transmission transformations: (<b>a</b>) original image; (<b>b</b>) transformed image.</p>
Full article ">Figure 6
<p>Results for the training and test sets: (<b>a</b>) the RMSE of the training set; (<b>b</b>) the RMSE of the test set.</p>
Full article ">Figure 7
<p>Process of individual identification: (<b>a</b>) schematic of selected areas; (<b>b</b>) results of yolo testing; (<b>c</b>) coordinate conversion results; (<b>d</b>) target identification results.</p>
Full article ">
28 pages, 1699 KiB  
Review
Overview and Comparison of Deep Neural Networks for Wildlife Recognition Using Infrared Images
by Peter Sykora, Patrik Kamencay, Roberta Hlavata and Robert Hudec
AI 2024, 5(4), 2801-2828; https://doi.org/10.3390/ai5040135 - 6 Dec 2024
Viewed by 1253
Abstract
There are multiple uses for single-channel images, such as infrared imagery, depth maps, and others. To automatically classify objects in such images, an algorithm suited for single-channel image processing is required. This study explores the application of deep learning techniques for the recognition [...] Read more.
There are multiple uses for single-channel images, such as infrared imagery, depth maps, and others. To automatically classify objects in such images, an algorithm suited for single-channel image processing is required. This study explores the application of deep learning techniques for the recognition of wild animals using infrared images. Traditional methods of wildlife monitoring often rely on visible light imaging, which can be hindered by various environmental factors such as darkness, fog, and dense foliage. In contrast, infrared imaging captures the thermal signatures of animals, providing a robust alternative for wildlife detection and identification. We test a Convolutional Neural Network (CNN) model specifically designed to analyze infrared images, leveraging the unique thermal patterns emitted by different animal species. The model is trained and tested on a diverse dataset of infrared images, demonstrating high accuracy in distinguishing between multiple species. In this paper, we also present a comparison of several well-known artificial neural networks on this data. To ensure accurate testing, we introduce a new dataset containing infrared photos of Slovak wildlife, specifically including classes such as bear, deer, boar, and fox. To complement this dataset, the Fashion MNIST dataset was also used. Our results indicate that deep learning approaches significantly enhance the capability of infrared imaging for wildlife monitoring, offering a reliable and efficient tool for conservation efforts and ecological studies. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of neural network architectures.</p>
Full article ">Figure 2
<p>Preview of the convolution layer.</p>
Full article ">Figure 3
<p>MaxPooling, step of <math display="inline"><semantics> <mrow> <mn>2</mn> <mo>×</mo> <mn>2</mn> </mrow> </semantics></math> and stride of <math display="inline"><semantics> <mrow> <mn>2</mn> <mo>×</mo> <mn>2</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 4
<p>Preview of the recurrent layer.</p>
Full article ">Figure 5
<p>Data normalisation using Batch normalisation.</p>
Full article ">Figure 6
<p>Preview of the flatten layer.</p>
Full article ">Figure 7
<p>Example effect of Dropout layer.</p>
Full article ">Figure 8
<p>Diagram of VGG19 architecture.</p>
Full article ">Figure 9
<p>Diagram of ResNet50 architecture.</p>
Full article ">Figure 10
<p>Diagram of Xception architecture.</p>
Full article ">Figure 11
<p>Diagram of MobileNet architecture.</p>
Full article ">Figure 12
<p>Diagram of DenseNet architecture.</p>
Full article ">Figure 13
<p>Example of IR animal dataset.</p>
Full article ">Figure 14
<p>Example of Fashion-MNIST dataset.</p>
Full article ">Figure 15
<p>Evaluation criteria.</p>
Full article ">
16 pages, 2064 KiB  
Article
Approach for Tattoo Detection and Identification Based on YOLOv5 and Similarity Distance
by Gabija Pocevičė, Pavel Stefanovič, Simona Ramanauskaitė and Ernest Pavlov
Appl. Sci. 2024, 14(13), 5576; https://doi.org/10.3390/app14135576 - 26 Jun 2024
Viewed by 1528
Abstract
The large number of images in the different areas and the possibilities of technologies lead to various solutions in automatization using image data. In this paper, tattoo detection and identification were analyzed. The combination of YOLOv5 object detection methods and similarity measures was [...] Read more.
The large number of images in the different areas and the possibilities of technologies lead to various solutions in automatization using image data. In this paper, tattoo detection and identification were analyzed. The combination of YOLOv5 object detection methods and similarity measures was investigated. During the experimental research, various parameters have been investigated to determine the best combination of parameters for tattoo detection. In this case, the influence of data augmentation parameters, the size of the YOLOv5 models (n, s, m, l, x), and the three main hyperparameters of YOLOv5 were analyzed. Also, the efficiency of the most popular similarity distances cosine and Euclidean was analyzed in the tattoo identification process with the purpose of matching the detected tattoo with the person’s tattoo in the database. Experiments have been performed using the deMSI dataset, where images were manually labeled to be suitable for use by the YOLOv5 algorithm. To validate the results obtained, the newly collected tattoo dataset was used. The results have shown that the highest average accuracy of all tattoo detection experiments has been obtained using the YOLOv5l model, where [email protected]:0.95 is equal to 0.60, and [email protected] is equal to 0.79. The accuracy for tattoo identification reaches 0.98, and the F-score is up to 0.52 when the highest cosine similarity tattoo is associated. Meanwhile, to ensure that no suspects will be missed, the cosine similarity threshold value of 0.15 should be applied. Then, photos with higher similarity scores should be analyzed only. This would lead to a 1.0 recall and would reduce the manual tattoo comparison by 20%. Full article
(This article belongs to the Special Issue Computer Vision in Automatic Detection and Identification)
Show Figures

Figure 1

Figure 1
<p>Main principle of automated suspect identification model development based on tattoo photos.</p>
Full article ">Figure 2
<p>Sample of the deMSI tattoo dataset used in the research.</p>
Full article ">Figure 3
<p>Evaluation of YOLOv5l with the highest mAP@0.5:0.95.</p>
Full article ">Figure 4
<p>Tattoo identification accuracy metrics modeling results based on selected threshold values for cosines similarity and Euclidean distance.</p>
Full article ">Figure 5
<p>Tattoo identification accuracy scores for cosine similarity and Euclidean distance when only the most similar reference tattoo is taken as the model prediction.</p>
Full article ">
19 pages, 5691 KiB  
Article
Development of Technologies for the Detection of (Cyber)Bullying Actions: The BullyBuster Project
by Giulia Orrù, Antonio Galli, Vincenzo Gattulli, Michela Gravina, Marco Micheletto, Stefano Marrone, Wanda Nocerino, Angela Procaccino, Grazia Terrone, Donatella Curtotti, Donato Impedovo, Gian Luca Marcialis and Carlo Sansone
Information 2023, 14(8), 430; https://doi.org/10.3390/info14080430 - 1 Aug 2023
Cited by 5 | Viewed by 4121
Abstract
Bullying and cyberbullying are harmful social phenomena that involve the intentional, repeated use of power to intimidate or harm others. The ramifications of these actions are felt not just at the individual level but also pervasively throughout society, necessitating immediate attention and practical [...] Read more.
Bullying and cyberbullying are harmful social phenomena that involve the intentional, repeated use of power to intimidate or harm others. The ramifications of these actions are felt not just at the individual level but also pervasively throughout society, necessitating immediate attention and practical solutions. The BullyBuster project pioneers a multi-disciplinary approach, integrating artificial intelligence (AI) techniques with psychological models to comprehensively understand and combat these issues. In particular, employing AI in the project allows the automatic identification of potentially harmful content by analyzing linguistic patterns and behaviors in various data sources, including photos and videos. This timely detection enables alerts to relevant authorities or moderators, allowing for rapid interventions and potential harm mitigation. This paper, a culmination of previous research and advancements, details the potential for significantly enhancing cyberbullying detection and prevention by focusing on the system’s design and the novel application of AI classifiers within an integrated framework. Our primary aim is to evaluate the feasibility and applicability of such a framework in a real-world application context. The proposed approach is shown to tackle the pervasive issue of cyberbullying effectively. Full article
(This article belongs to the Special Issue Computer Vision, Pattern Recognition and Machine Learning in Italy)
Show Figures

Figure 1

Figure 1
<p>Annual publication trend from 2000 to 2022 for documents focusing on the intersection of artificial intelligence and well-being. The graph illustrates the increasing interest in this research area over time.</p>
Full article ">Figure 2
<p>Logo of the project “BullyBuster—A framework for bullying and cyberbullying action detection by computer vision and artificial intelligence methods and algorithms” (<b>a</b>). The project has been included in the Global Top 100 list of AI projects by IRCAI (<b>b</b>).</p>
Full article ">Figure 3
<p>The BullyBuster project framework effectively integrates contributions from artificial intelligence, technology, law, and psychology.</p>
Full article ">Figure 4
<p>Number of individuals who fall into a personality index for each risk level. Range 1 corresponds to a low risk, range 2 to a moderate risk, and range 3 to a high risk for bullying behaviors or victimization.</p>
Full article ">Figure 5
<p>Example of a sequence depicting a scene of panic extracted from the MED dataset [<a href="#B44-information-14-00430" class="html-bibr">44</a>]; the individuals are at such a distance as to allow the analysis of the overall behavior of the crowd without allowing personal identification.</p>
Full article ">Figure 6
<p>Final dashboards of two use cases of the BullyBuster framework, which allows the analysis of the behavior of a group of individuals to evaluate the risk of (cyber)bullying actions.</p>
Full article ">
11 pages, 1890 KiB  
Article
Optical Multi-Parameter Measuring System for Fluid and Air Bubble Recognition
by Valentina Bello, Elisabetta Bodo and Sabina Merlo
Sensors 2023, 23(15), 6684; https://doi.org/10.3390/s23156684 - 26 Jul 2023
Cited by 2 | Viewed by 2304
Abstract
Detection of air bubbles in fluidic channels plays a fundamental role in all that medical equipment where liquids flow inside patients’ blood vessels or bodies. In this work, we propose a multi-parameter sensing system for simultaneous recognition of the fluid, on the basis [...] Read more.
Detection of air bubbles in fluidic channels plays a fundamental role in all that medical equipment where liquids flow inside patients’ blood vessels or bodies. In this work, we propose a multi-parameter sensing system for simultaneous recognition of the fluid, on the basis of its refractive index and of the air bubble transit. The selected optofluidic platform has been designed and studied to be integrated into automatic pumps for the administration of commercial liquid. The sensor includes a laser beam that crosses twice a plastic cuvette, provided with a back mirror, and a position-sensitive detector. The identification of fluids is carried out by measuring the displacement of the output beam on the detector active surface and the detection of single air bubbles can be performed with the same instrumental scheme, exploiting a specific signal analysis. When a bubble, traveling along the cuvette, crosses the readout light beam, radiation is strongly scattered and a characteristic fingerprint shape of the photo-detected signals versus time is clearly observed. Experimental testing proves that air bubbles can be successfully detected and counted. Their traveling speed can be estimated while simultaneously monitoring the refractive index of the fluid. Full article
(This article belongs to the Special Issue Feature Papers in Physical Sensors 2023)
Show Figures

Figure 1

Figure 1
<p>Experimental detection of air bubbles. (<b>a</b>) Bubbles injected by means of a syringe interrupt the laser beam twice (front view). (<b>b</b>) Block schematic of the PSD signal processing circuit. (<b>c</b>) Preliminary setup for bubble detection including a camera (top view).</p>
Full article ">Figure 2
<p>Experimental results obtained during testing of water-glucose dilutions with different concentrations and RI. The sum signal (blue trace) has an almost constant value for every sample; on the other hand, the difference signal (orange trace) keeps changing because of the light beam displacement occurring when the RI of the fluid in the cuvette increases.</p>
Full article ">Figure 3
<p>Experimental results obtained during testing of air bubbles: sum signal (blue trace) and difference signal (orange trace) generated by the PSD. Every time an air bubble interrupts the light beam, both signals exhibit a very peculiar shape: each black asterisk indicates the passage on one bubble. The dashed circles with the arrow indicate the reference variable.</p>
Full article ">Figure 4
<p>Experimental results relative to air bubble detection generated by the PSD: sum signal (blue trace) and difference signal (orange trace). Red rectangles in each subset indicate the signal portions corresponding to the frames on the side. Every time an air bubble interrupts the light beam, both signals exhibit very peculiar spikes. In particular, the pictures and the PSD signals of five subsets illustrate, respectively, the following steps: (<b>a</b>) the air bubble is generated, and it is about to leave the connector, (<b>b</b>) the bubble crosses the incident beam (the black asterisk indicates the local peak of the sum signal due to the crossing of the central part of the bubble; the green asterisks indicate the double-negative-positive peak transition of the difference signal due to the crossing of the central part of the bubble), (<b>c</b>) the bubble is travelling between the incident and reflected beam, (<b>d</b>) the bubble crosses the reflected beam: black and green arrows represent the time required to bubble for crossing the incident and reflected beam, respectively, (<b>e</b>) and the bubble has left the cuvette.</p>
Full article ">Figure 5
<p>Calculation of air bubble velocity. (<b>a</b>) Distance Δ<span class="html-italic">x</span> travelled by the bubble between the incident and reflected beams. (<b>b</b>) Calculation of the time interval Δ<span class="html-italic">τ</span>.</p>
Full article ">
16 pages, 2659 KiB  
Article
Generalized Spoof Detection and Incremental Algorithm Recognition for Voice Spoofing
by Jinlin Guo, Yancheng Zhao and Haoran Wang
Appl. Sci. 2023, 13(13), 7773; https://doi.org/10.3390/app13137773 - 30 Jun 2023
Cited by 2 | Viewed by 1835
Abstract
Highly deceptive deepfake technologies have caused much controversy, e.g., artificial intelligence-based software can automatically generate nude photos and deepfake images of anyone. This brings considerable threats to both individuals and society. In addition to video and image forgery, audio forgery poses many hazards [...] Read more.
Highly deceptive deepfake technologies have caused much controversy, e.g., artificial intelligence-based software can automatically generate nude photos and deepfake images of anyone. This brings considerable threats to both individuals and society. In addition to video and image forgery, audio forgery poses many hazards but lacks sufficient attention. Furthermore, existing works have only focused on voice spoof detection, neglecting the identification of spoof algorithms. It is of great value to recognize the algorithm for synthesizing spoofing voices in traceability. This study presents a system combining voice spoof detection and algorithm recognition. In contrast, the generalizability of the spoof detection model is discussed from the perspective of embedding space and decision boundaries to face the voice spoofing attacks generated by spoof algorithms that are not available in the training set. This study presents a method for voice spoof algorithms recognition based on incremental learning, taking into account data flow scenarios where new spoof algorithms keep appearing in reality. Our experimental results on the LA dataset of ASVspoof show that our system can improve the generalization of spoof detection and identify new voice spoof algorithms without catastrophic forgetting. Full article
Show Figures

Figure 1

Figure 1
<p>The framework of the system.</p>
Full article ">Figure 2
<p>Schematic of domain adaptation. <math display="inline"><semantics><mrow><msub><mi>C</mi><mn>1</mn></msub><mo> </mo><mi>and</mi><mo> </mo><msub><mi>C</mi><mn>2</mn></msub></mrow></semantics></math> denote classification boundaries. The dashed ellipse indicates the source domain, and the solid ellipse indicates the target domain. (<b>a</b>) indicates maximizing the shaded area. (<b>b</b>) indicates that the classifier is maximized. (<b>c</b>) indicates that the divergence is minimized. (<b>d</b>) denotes the final optimization objective with more robust decision boundary.</p>
Full article ">Figure 3
<p>Schematic of OC-Softmax embedding vector distribution. The red and blue dots indicate the two classes of sample features. The dashed line indicates the classification decision boundary. The arrow indicates the optimization direction of the target class embedding vector.</p>
Full article ">Figure 4
<p>Schematic for prototype training. Different colored circles indicate different classes of prototypes. The size of the circle indicates the size of the value.</p>
Full article ">Figure 5
<p>Schematic of incremental learning. The circle in the figure represents the features. Triangles, rectangles, and different colors represent category labels, sample data, and different categories, respectively.</p>
Full article ">Figure 6
<p>CM score histogram: (<b>a</b>) AM-Softmax. (<b>b</b>) Sigmoid. (<b>c</b>) Domain adaptation. (<b>d</b>) OC-Softmax.</p>
Full article ">Figure 7
<p>(<b>a</b>) Visualization of the t-SNE of the learned A1–A9 category features distribution. (<b>b</b>) The t-SNE visualization of the learned category A1–A9 and for the learned category A10–A17 features distribution t-SNE visualization. The numbers in the figure indicate the category numbers of the forgery methods.</p>
Full article ">Figure 8
<p>Classification confusion matrix after the last iteration of training.</p>
Full article ">
22 pages, 6992 KiB  
Article
AHF: An Automatic and Universal Image Preprocessing Algorithm for Circular-Coded Targets Identification in Close-Range Photogrammetry under Complex Illumination Conditions
by Hang Shang and Changying Liu
Remote Sens. 2023, 15(12), 3151; https://doi.org/10.3390/rs15123151 - 16 Jun 2023
Cited by 3 | Viewed by 1815
Abstract
In close-range photogrammetry, circular-coded targets (CCTs) are a reliable method to solve the issue of image correspondence. Currently, the identification methods for CCTs are very mature, but complex illumination conditions are still a key factor restricting identification. This article proposes an adaptive homomorphic [...] Read more.
In close-range photogrammetry, circular-coded targets (CCTs) are a reliable method to solve the issue of image correspondence. Currently, the identification methods for CCTs are very mature, but complex illumination conditions are still a key factor restricting identification. This article proposes an adaptive homomorphic filtering (AHF) algorithm to solve this issue, utilizing homomorphic filtering (HF) to eliminate the influence of uneven illumination. However, HF parameters vary with different lighting types. We use a genetic algorithm (GA) to carry out global optimization and take the identification result as the objective function to realize automatic parameter adjustment. This is different from the optimization strategy of traditional adaptive image enhancement methods, so the most significant advantage of the proposed algorithm lies in its automation and universality, i.e., users only need to input photos without considering the type of lighting conditions. As a preprocessing algorithm, we conducted experiments combining advanced commercial photogrammetric software and traditional identification methods, respectively. We cast stripe- and lattice-structured light to create complex lighting conditions, including uneven lighting, dense shadow areas, and elliptical light spots. Experiments showed that our algorithm significantly improves the robustness and accuracy of CCT identification methods under complex lighting conditions. Given the perfect performance under stripe-structured light, this algorithm can provide a new idea for the fusion of close-range photogrammetry and structured light. This algorithm helps to improve the quality and accuracy of photogrammetry and even helps to improve the decision making and planning process of photogrammetry. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Two representative coded targets: (<b>a</b>) GCT. (<b>b</b>) Schneider CCT.</p>
Full article ">Figure 2
<p>Proposed algorithm flow framework.</p>
Full article ">Figure 3
<p>Experimental setup.</p>
Full article ">Figure 4
<p>Illumination conditions. (<b>a</b>) Normal illumination; (<b>b</b>) striped structured light; (<b>c</b>) lattice-structured light.</p>
Full article ">Figure 5
<p>AM and AHF-AM identification results. (<b>a</b>) Identification of AM under normal lighting conditions; (<b>b</b>) identification of AM under striped structured light; (<b>c</b>) identification of AM under lattice-structured light; (<b>d</b>) identification of AHF-AM under striped lighting conditions; (<b>e</b>) identification of AHF-AM under lattice lighting conditions.</p>
Full article ">Figure 6
<p>TI and AHF-TI identification results. (<b>a</b>) Identification of TI under normal lighting conditions; (<b>b</b>) identification of TI under striped structured light; (<b>c</b>) identification of TI under lattice-structured light; (<b>d</b>) identification of AHF-TI under normal lighting conditions; (<b>e</b>) identification of AHF-TI under striped structured lighting conditions; (<b>f</b>) identification of AHF-TI under lattice-structured lighting conditions.</p>
Full article ">Figure 7
<p>Processing effect of HF with different parameter values on CCT illuminated by stripe-structured light. (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.2, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.25; (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.2, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.55; (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.2, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.85; (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 1.8, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.85; (<b>e</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.0, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.85; (<b>f</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.3, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.85.</p>
Full article ">Figure 8
<p>Effect of HF with different parameter values on CCT identification under stripe-structured light illumination. (<b>a</b>) Original image <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.0, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.05; (<b>b</b>) original image <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.5, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.60; (<b>c</b>) original image <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 3.9, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.85; (<b>d</b>) AM’s identification of (<b>a</b>); (<b>e</b>) AM’s identification of (<b>b</b>); (<b>f</b>) AM’s identification of (<b>c</b>); (<b>g</b>) TI’s identification of (<b>a</b>); (<b>h</b>) TI’s identification of (<b>b</b>); (<b>i</b>) TI’s identification of (<b>c</b>).</p>
Full article ">Figure 9
<p>The processing effect of HF with different parameter values on CCTs illuminated by lattice-structured light. (<b>a</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.4, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.32; (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.6, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.50; (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>H</mi> </mrow> </semantics></math> = 2.8, <math display="inline"><semantics> <mrow> <mi>γ</mi> <mi>L</mi> </mrow> </semantics></math> = 0.80.</p>
Full article ">
16 pages, 2143 KiB  
Article
A Method for Detecting Pathologies in Concrete Structures Using Deep Neural Networks
by Joel de Conceição Nogueira Diniz, Anselmo Cardoso de Paiva, Geraldo Braz Junior, João Dallyson Sousa de Almeida, Aristofanes Correa Silva, António Manuel Trigueiros da Silva Cunha and Sandra Cristina Alves Pereira da Silva Cunha
Appl. Sci. 2023, 13(9), 5763; https://doi.org/10.3390/app13095763 - 7 May 2023
Cited by 5 | Viewed by 3429
Abstract
Pathologies in concrete structures, such as cracks, splintering, efflorescence, corrosion spots, and exposed steel bars, can be visually evidenced on the concrete surface. This paper proposes a method for automatically detecting these pathologies from images of the concrete structure. The proposed method uses [...] Read more.
Pathologies in concrete structures, such as cracks, splintering, efflorescence, corrosion spots, and exposed steel bars, can be visually evidenced on the concrete surface. This paper proposes a method for automatically detecting these pathologies from images of the concrete structure. The proposed method uses deep neural networks to detect pathologies in these images. This method results in time savings and error reduction. The paper presents results in detecting the pathologies from wide-angle images containing the overall structure and also for the specific pathology identification task for cropped images of the region of the pathology. Identifying pathologies in cropped images, the classification task could be performed with 99.4% accuracy using cross-validation and classifying cracks. Wide images containing no, one, or several pathologies in the same image, the case of pathology detection, could be analyzed with the YOLO network to identify five pathology classes. The results for detection with YOLO were measured with mAP, mean Average Precision, for five classes of concrete pathology, reaching 11.80% for fissure, 19.22% for fragmentation, 5.62% for efflorescence, 27.24% for exposed bar, and 24.44% for corrosion. Pathology identification in concrete photos can be optimized using deep learning. Full article
Show Figures

Figure 1

Figure 1
<p>Proposed method as diagram.</p>
Full article ">Figure 2
<p>Pathologies in Concrete [<a href="#B17-applsci-13-05763" class="html-bibr">17</a>].</p>
Full article ">Figure 3
<p>YOLO Detection pipeline. The input image was divided into S × S grids. Predicted bounding boxes and confidence and a class probability map for each grid cell are generated. The final detection is generated [<a href="#B23-applsci-13-05763" class="html-bibr">23</a>].</p>
Full article ">Figure 4
<p>YOLO Architecture [<a href="#B23-applsci-13-05763" class="html-bibr">23</a>].</p>
Full article ">Figure 5
<p>CNN Architecture and configuration.</p>
Full article ">Figure 6
<p>Training with YOLOv4 on the CODEBRIM dataset [<a href="#B17-applsci-13-05763" class="html-bibr">17</a>].</p>
Full article ">Figure 7
<p>Image of concrete pathologies before detection with YOLOv4 [<a href="#B17-applsci-13-05763" class="html-bibr">17</a>].</p>
Full article ">Figure 8
<p>Image of concrete pathologies after detection with YOLOv4 [<a href="#B17-applsci-13-05763" class="html-bibr">17</a>].</p>
Full article ">Figure 9
<p>Images of the two types of classes in the dataset [<a href="#B18-applsci-13-05763" class="html-bibr">18</a>].</p>
Full article ">Figure 10
<p>Accuracy obtained from CNN in one of the cross-validation steps.</p>
Full article ">Figure 11
<p>Loss obtained from CNN.</p>
Full article ">Figure 12
<p>ROC AUC.</p>
Full article ">Figure 13
<p>Classification result for two images from the two classes of the [<a href="#B18-applsci-13-05763" class="html-bibr">18</a>] dataset.</p>
Full article ">
16 pages, 5709 KiB  
Article
Performance Analysis of Deep Learning Algorithms in Diagnosis of Malaria Disease
by K. Hemachandran, Areej Alasiry, Mehrez Marzougui, Shahid Mohammad Ganie, Anil Audumbar Pise, M. Turki-Hadj Alouane and Channabasava Chola
Diagnostics 2023, 13(3), 534; https://doi.org/10.3390/diagnostics13030534 - 1 Feb 2023
Cited by 34 | Viewed by 5569
Abstract
Malaria is predominant in many subtropical nations with little health-monitoring infrastructure. To forecast malaria and condense the disease’s impact on the population, time series prediction models are necessary. The conventional technique of detecting malaria disease is for certified technicians to examine blood smears [...] Read more.
Malaria is predominant in many subtropical nations with little health-monitoring infrastructure. To forecast malaria and condense the disease’s impact on the population, time series prediction models are necessary. The conventional technique of detecting malaria disease is for certified technicians to examine blood smears visually for parasite-infected RBC (red blood cells) underneath a microscope. This procedure is ineffective, and the diagnosis depends on the individual performing the test and his/her experience. Automatic image identification systems based on machine learning have previously been used to diagnose malaria blood smears. However, so far, the practical performance has been insufficient. In this paper, we have made a performance analysis of deep learning algorithms in the diagnosis of malaria disease. We have used Neural Network models like CNN, MobileNetV2, and ResNet50 to perform this analysis. The dataset was extracted from the National Institutes of Health (NIH) website and consisted of 27,558 photos, including 13,780 parasitized cell images and 13,778 uninfected cell images. In conclusion, the MobileNetV2 model outperformed by achieving an accuracy rate of 97.06% for better disease detection. Also, other metrics like training and testing loss, precision, recall, fi-score, and ROC curve were calculated to validate the considered models. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

Figure 1
<p>Infected cell images.</p>
Full article ">Figure 2
<p>Uninfected cell images.</p>
Full article ">Figure 3
<p>Block diagram of CNN model.</p>
Full article ">Figure 4
<p>ReLU Algorithm.</p>
Full article ">Figure 5
<p>Block diagram of MobileNetV2 model.</p>
Full article ">Figure 6
<p>Block diagram of ResNet50 model.</p>
Full article ">Figure 7
<p>Confusion Matrix of CNN model.</p>
Full article ">Figure 8
<p>Confusion Matrix of MobileNetV2 model.</p>
Full article ">Figure 9
<p>Confusion Matrix of ResNet50 model.</p>
Full article ">Figure 10
<p>Accuracy and Loss of CNN model.</p>
Full article ">Figure 11
<p>Accuracy and Loss of MobileNetV2 model.</p>
Full article ">Figure 12
<p>Accuracy and Loss of ResNet50 model.</p>
Full article ">Figure 13
<p>ROC Curve of CNN model.</p>
Full article ">Figure 14
<p>ROC Curve of MobileNetV2 model.</p>
Full article ">Figure 15
<p>ROC Curve of ResNet50 model.</p>
Full article ">Figure 16
<p>Precision of all considered models.</p>
Full article ">Figure 17
<p>Recall of all considered models.</p>
Full article ">Figure 18
<p>f1-score of all considered models.</p>
Full article ">
19 pages, 3614 KiB  
Article
An Effective Approach for Automatic River Features Extraction Using High-Resolution UAV Imagery
by Marco La Salandra, Rosa Colacicco, Pierfrancesco Dellino and Domenico Capolongo
Drones 2023, 7(2), 70; https://doi.org/10.3390/drones7020070 - 18 Jan 2023
Cited by 20 | Viewed by 3490
Abstract
The effects of climate change are causing an increase in the frequency and extent of natural disasters. Because of their morphological characteristics, rivers can cause major flooding events. Indeed, they can be subjected to variations in discharge in response to heavy rainfall and [...] Read more.
The effects of climate change are causing an increase in the frequency and extent of natural disasters. Because of their morphological characteristics, rivers can cause major flooding events. Indeed, they can be subjected to variations in discharge in response to heavy rainfall and riverbank failures. Among the emerging methodologies that address the monitoring of river flooding, those that include the combination of Unmanned Aerial Vehicle (UAV) and photogrammetric techniques (i.e., Structure from Motion-SfM) ensure the high-frequency acquisition of high-resolution spatial data over wide areas and so the generation of orthomosaics, useful for automatic feature extraction. Trainable Weka Segmentation (TWS) is an automatic feature extraction open-source tool. It was developed to primarily fulfill supervised classification purposes of biological microscope images, but its usefulness has been demonstrated in several image pipelines. At the same time, there is a significant lack of published studies on the applicability of TWS with the identification of a universal and efficient combination of machine learning classifiers and segmentation approach, in particular with respect to classifying UAV images of riverine environments. In this perspective, we present a study comparing the accuracy of nine combinations, classifier plus image segmentation filter, using TWS, also with respect to human photo-interpretation, in order to identify an effective supervised approach for automatic river features extraction from UAV multi-temporal orthomosaics. The results, which are very close to human interpretation, indicate that the proposed approach could prove to be a valuable tool to support and improve the hydro-geomorphological and flooding hazard assessments in riverine environments. Full article
Show Figures

Figure 1

Figure 1
<p>Study area: (<b>A</b>) image highlighting the Basilicata region (red boundaries) in southern Italy and the location of the area analyzed (green dot); (<b>B</b>) Focus of the catchment area of the Basento river (light blue boundaries) and the location of the river reach investigated (green dot); (<b>C</b>) UAV surveyed area of the Basento river reach near the industrial area of Ferrandina (MT) (green polygon). The base map for all images is Google Earth.</p>
Full article ">Figure 2
<p>Orthomosaics covering different lengths of reach of the Basento river, generated by UAV images acquired in different time interval (2019, 2020, 2021 and 2022) and with different weather conditions (sunny and cloudy).</p>
Full article ">Figure 3
<p>Scheme of the main steps of the automatic river features extraction approach. The box in magenta indicates the process of generating UAV multi-temporal orthomosaics; the boxes in light blue indicate the filters used in combination with classifiers (green boxes) in the pre-processing phase; the classification process includes the box in yellow, which indicates the image segmentation process in TWS by leveraging the training data (light gray box at the top right) and the derived outputs (light gray boxes); the boxes in light green indicate all the post-processing operations applied to the probability maps (make binary images, calculation of riverbed water area and misclassified water area); the box in orange indicate the accuracy assessment phase of the classified and digitized (white box) riverbed water-class area.</p>
Full article ">Figure 4
<p>Classified images showing the water, terrain, dense vegetation and background classes, generated by each combination (classifier plus image filter) and for each orthomosaic (2019, 2020, 2021, 2022).</p>
Full article ">Figure 5
<p>Binary probability maps showing the water class generated by each combination (classifier plus image filter) and for each orthomosaic (2019, 2020, 2021, 2022).</p>
Full article ">Figure 6
<p>Orthomosaics of a reach of the Basento river generated by UAV images acquired in different time interval (2019, 2020, 2021 and 2022) and with different weather conditions (sunny and cloudy). In red is shown the manually digitized water class area.</p>
Full article ">Figure 7
<p>Binary probability maps showing the water class covering exclusively the riverbed area, generated by each combination (classifier plus image filter) and for each orthomosaic (2019, 2020, 2021, 2022).</p>
Full article ">Figure 8
<p>Binary probability maps showing the misclassified water class over the whole image, generated by each combination (classifier plus image filter) and for the different orthomosaics (2019, 2020, 2021, 2022).</p>
Full article ">
13 pages, 2238 KiB  
Article
CNN Based Image Classification of Malicious UAVs
by Jason Brown, Zahra Gharineiat and Nawin Raj
Appl. Sci. 2023, 13(1), 240; https://doi.org/10.3390/app13010240 - 24 Dec 2022
Cited by 9 | Viewed by 2599
Abstract
Unmanned Aerial Vehicles (UAVs) or drones have found a wide range of useful applications in society over the past few years, but there has also been a growth in the use of UAVs for malicious purposes. One way to manage this issue is [...] Read more.
Unmanned Aerial Vehicles (UAVs) or drones have found a wide range of useful applications in society over the past few years, but there has also been a growth in the use of UAVs for malicious purposes. One way to manage this issue is to allow reporting of malicious UAVs (e.g., through a smartphone application) with the report including a photo of the UAV. It would be useful to able to automatically identify the type of UAV within the image in terms of the manufacturer and specific product identification using a trained image classification model. In this paper, we discuss the collection of images for three popular UAVs at different elevations and different distances from the observer, and using different camera zoom levels. We then train 4 image classification models based upon Convolutional Neural Networks (CNNs) using this UAV image dataset and the concept of transfer learning from the well-known ImageNet database. The trained models can classify the type of UAV contained in unseen test images with up to approximately 81% accuracy (for the Resnet-18 model), even though 2 of the UAVs represented in the UAV image dataset are visually similar, and the fact that the UAV image dataset contains images of UAVs that are a significant distance from the observer. This provides a motivation to expand the study in the future to include more UAV types and other usage scenarios (e.g., UAVs carrying loads). Full article
(This article belongs to the Special Issue Deep Neural Network: Algorithms and Applications)
Show Figures

Figure 1

Figure 1
<p>UAVs Employed in Current Study.</p>
Full article ">Figure 2
<p>Distance <span class="html-italic">d</span> and Elevation <span class="html-italic">h</span> Parameters Used for Image Capture.</p>
Full article ">Figure 3
<p>Effect of Different Camera Zoom Levels On Size and Image Background (UAV: DJI Mavic Air, <span class="html-italic">h</span> = 5 m, <span class="html-italic">d</span> = 10 m, images reduced to 256 × 256 resolution).</p>
Full article ">Figure 4
<p>Effect of Different Elevations <span class="html-italic">h</span> on Size, View Angle, and Image Background (UAV: DJI Phantom 4, 10× Zoom, <span class="html-italic">d</span> = 10 m, images reduced to 256 × 256 resolution).</p>
Full article ">Figure 5
<p>Structure of the Resnet-18 Model.</p>
Full article ">Figure 6
<p>Per Class Metrics for Prediction of Test Images Using Trained Resnet-18 Model.</p>
Full article ">Figure 7
<p>Per Class Metrics for Prediction of Test Images Using Trained MobileNet v2 Model.</p>
Full article ">
22 pages, 601 KiB  
Review
Facial Age Estimation Using Machine Learning Techniques: An Overview
by Khaled ELKarazle, Valliappan Raman and Patrick Then
Big Data Cogn. Comput. 2022, 6(4), 128; https://doi.org/10.3390/bdcc6040128 - 26 Oct 2022
Cited by 18 | Viewed by 13175
Abstract
Automatic age estimation from facial images is an exciting machine learning topic that has attracted researchers’ attention over the past several years. Numerous human–computer interaction applications, such as targeted marketing, content access control, or soft-biometrics systems, employ age estimation models to carry out [...] Read more.
Automatic age estimation from facial images is an exciting machine learning topic that has attracted researchers’ attention over the past several years. Numerous human–computer interaction applications, such as targeted marketing, content access control, or soft-biometrics systems, employ age estimation models to carry out secondary tasks such as user filtering or identification. Despite the vast array of applications that could benefit from automatic age estimation, building an automatic age estimation system comes with issues such as data disparity, the unique ageing pattern of each individual, and facial photo quality. This paper provides a survey on the standard methods of building automatic age estimation models, the benchmark datasets for building these models, and some of the latest proposed pieces of literature that introduce new age estimation methods. Finally, we present and discuss the standard evaluation metrics used to assess age estimation models. In addition to the survey, we discuss the identified gaps in the reviewed literature and present recommendations for future research. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of training a typical age estimation model.</p>
Full article ">
22 pages, 8530 KiB  
Article
Study on the Weak Interlayer Identification Method Based on Borehole Photo-Acoustic Combined Measurement: Application to a Landslide Case Study
by Jinchao Wang, Junfeng Huang, Hui Min, Feng Wang, Yiteng Wang and Zengqiang Han
Appl. Sci. 2022, 12(20), 10545; https://doi.org/10.3390/app122010545 - 19 Oct 2022
Cited by 3 | Viewed by 1890
Abstract
A weak interlayer is the key factor in controlling slope stability. It is of great significance to effectively identify the weak interlayer in the study of spatial and temporal distribution law and the internal structure characteristics of a landslide. Considering the limitations of [...] Read more.
A weak interlayer is the key factor in controlling slope stability. It is of great significance to effectively identify the weak interlayer in the study of spatial and temporal distribution law and the internal structure characteristics of a landslide. Considering the limitations of traditional optical imaging and wave speed test methods, this paper presents a weak interlayer identification method based on borehole photo-acoustic combination measurement. By using the combination of optical imaging and acoustic wave scanning, the multi-source data collection of borehole rock wall and borehole surrounding rock is realized, which effectively captures the comprehensive response characteristics of the weak interlayer. This paper first constructs a multi-source data acquisition technology based on the borehole photo-acoustic combination measurement to realize the visualization of the image information and acoustic data of the target area on the borehole rock wall. Subsequently, the optical image features and the acoustic response characteristics of the weak interlayer are clarified based on the optical image and the acoustic scanning data. The hole wall texture characteristic response function, hole wall integrity characteristic response function, hole wall acoustic characteristic response function and hole wall contour characteristic response function are constructed. Finally, the landslide weak interlayer identification method considering the texture characteristics, complete characteristics, acoustic response characteristics and contour characteristics of the borehole rock wall is proposed, which effectively distinguishes the types of rock mass structural surface and realizes the automatic identification of the weak interlayer. Combined with the case analysis, the correctness and reliability of the present method are verified. The results show that the method can identify the weak interlayer and provide scientific basis for landslide management, which can provide a feasible and effective new way to identify the landslide weak interlayer in practical engineering, with a good application prospect and promotion value. Full article
Show Figures

Figure 1

Figure 1
<p>Schematic diagram of working principle.</p>
Full article ">Figure 2
<p>Schematic diagram of borehole image expansion. (<b>a</b>) Acquired image, (<b>b</b>) center coordinate, (<b>c</b>) expansion diagram.</p>
Full article ">Figure 3
<p>Schematic diagram of acoustic wave scanning data processing. (<b>a</b>) Scanning data, (<b>b</b>) circular matrix, (<b>c</b>) circumferential visualization.</p>
Full article ">Figure 4
<p>Optical image and grid division. (<b>a</b>) Optical unfolded image, (<b>b</b>) grid division of optical unfolded image.</p>
Full article ">Figure 5
<p>Acoustic wave scanning data and its matrix processing. (<b>a</b>) Horizontal section acoustic wave scanning data, (<b>b</b>) schematic diagram of matrix processing.</p>
Full article ">Figure 6
<p>Identification block diagram of weak interlayer.</p>
Full article ">Figure 7
<p>Physical diagram of instrument and equipment. (<b>a</b>) Complete system, (<b>b</b>) optical imaging measurement module, (<b>c</b>) acoustic scanning measurement module.</p>
Full article ">Figure 8
<p>Partial data collected. (<b>a</b>) Video image screenshot of rock wall, (<b>b</b>) rock wall acoustic wave scanning data.</p>
Full article ">Figure 9
<p>Rock wall expansion image.</p>
Full article ">Figure 10
<p>Rock wall optical image characteristics. (<b>a</b>) Hole wall texture feature image, (<b>b</b>) hole wall integrity feature image.</p>
Full article ">Figure 11
<p>Acoustic scanning characteristics of rock walls. (<b>a</b>) Hole wall acoustic feature image, (<b>b</b>) hole wall contour feature image.</p>
Full article ">Figure 12
<p>Reconstructed image and binary image. (<b>a</b>) Fusion image, (<b>b</b>) binarized image.</p>
Full article ">Figure 13
<p>Image comparison before and after well wall fusion. (<b>a</b>) Optical image (<b>b</b>) processed image.</p>
Full article ">Figure 14
<p>Comparison of recognition before and after hole wall image fusion. (<b>a</b>) Optical image recognition, (<b>b</b>) fusion hole wall image recognition.</p>
Full article ">Figure 15
<p>3D images before and after borehole wall fusion. (<b>a</b>) 3D images before fusion, (<b>b</b>) 3D graphics after fusion.</p>
Full article ">Figure 16
<p>Numerical comparison of different methods.</p>
Full article ">Figure 17
<p>Before and after borehole wall fusion. (<b>a</b>) Pre-fusion image (N-S azimuth), (<b>b</b>) pre-fusion image (S-N azimuth), (<b>c</b>) post-fusion image (N-S azimuth), (<b>d</b>) post-fusion image (S-N azimuth).</p>
Full article ">Figure 18
<p>3D formation images. (<b>a</b>) Formation image before fusion, (<b>b</b>) formation image after fusion.</p>
Full article ">Figure 19
<p>Probability value distribution of the identified results by the different methods. (<b>a</b>) Image recognition method, (<b>b</b>) this article’s recognition method, (<b>c</b>) sound speed recognition method.</p>
Full article ">
14 pages, 11362 KiB  
Technical Note
Mapping Asbestos-Cement Corrugated Roofing Tiles with Imagery Cube via Machine Learning in Taiwan
by Teng-To Yu, Yen-Chun Lin, Shyh-Chin Lan, Yu-En Yang, Pei-Yun Wu and Jo-Chi Lin
Remote Sens. 2022, 14(14), 3418; https://doi.org/10.3390/rs14143418 - 16 Jul 2022
Cited by 5 | Viewed by 3014
Abstract
Locating and calculating the number of asbestos-cement corrugated roofing tiles is the first step in the demolition process. In this work, archived image cubes of Taiwan served as the fundamental data source used via machine learning approach to identify the existence of asbestos-cement [...] Read more.
Locating and calculating the number of asbestos-cement corrugated roofing tiles is the first step in the demolition process. In this work, archived image cubes of Taiwan served as the fundamental data source used via machine learning approach to identify the existence of asbestos-cement corrugated roofing tiles with more than 85% accuracy. An adequate quantity of ground-truth data covering all the types of roofs via aerial hyperspectral scan was the key to success for this study. Twenty randomly picked samples from the ground-truth group were examined by X-ray refraction detection to ensure correct identification of asbestos-cement corrugated roofing tiles with remote sensing. To improve the classifying accuracy ratio, two different machine learning algorithms were applied to gather the target layers individually using the same universal training model established from 400 ground-truth samples. The agreement portions within the overlapping layers of these two approaches were labeled as the potential targets, and the pixel growth technique was performed to detect the roofing boundary and create the polygon layer with size information. Exacting images from aerial photos within the chosen polygon were compared to up-to-date Sentinel-1 images to find the temporal disagreements and remove the mismatched buildings, identified as non-asbestos roofs, from the database to reflect the actual condition of present data. This automatic matching could be easily performed by machine learning to resolve the information lag while using archived data, which is an essential issue when detecting targets with non-simultaneous acquired images over a large area. To meet the 85% kappa accuracy requirement, the recurring processes were applied to find the optimal parameters of the machine learning model. Meanwhile, this study found that the support vector machine method was easier to handle, and the convolution neuro network method offered better accuracy in automatic classification with a universal training model for vast areas. This work demonstrated a feasible approach using low-cost and low-resolution archived images to automatically detect the existence of asbestos-cement corrugated roofing tiles over large regions. The entire work was completed within 16 months for an area of 36,000 km2, and the detected number of asbestos-cement corrugated roofing tiles was more than three times the initial estimation by statistics method from two small-area field surveys. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>Left</b>) The elevation and distribution of major cities in Taiwan. (<b>Right</b>) Samples of available imagery and various roofing types.</p>
Full article ">Figure 2
<p>Flow chart and decision branches among various-origin data for this work.</p>
Full article ">Figure 3
<p>Sample locations of the asbestos-cement corrugated roofing tiles as well as equipment for ground multispectral scanning.</p>
Full article ">Figure 4
<p>Applied TensorFlow stages of CNN.</p>
Full article ">Figure 5
<p>Specification and metadata of the source imagery.</p>
Full article ">Figure 6
<p>Overlapping two classifications and pixel growing to detect the existence of roofing area.</p>
Full article ">Figure 7
<p>Confusion matrix of two machine learning methods via aerial photo and satellite image.</p>
Full article ">Figure 8
<p>Distributed location of training and validation points for SVM.</p>
Full article ">Figure 9
<p>Distributed location of training and validation points for CNN.</p>
Full article ">Figure 10
<p>Locations and quantities of detected asbestos-cement corrugated roofing tiles in Taiwan.</p>
Full article ">
16 pages, 3724 KiB  
Article
Image-Based Automatic Individual Identification of Fish without Obvious Patterns on the Body (Scale Pattern)
by Dinara Bekkozhayeva and Petr Cisar
Appl. Sci. 2022, 12(11), 5401; https://doi.org/10.3390/app12115401 - 26 May 2022
Cited by 7 | Viewed by 6325
Abstract
The precision fish farming concept has been widely investigated in research and is highly desirable in aquaculture as it creates opportunities for precisely controlling and monitoring fish cultivation processes and increasing fish welfare. The automatic identification of individual fish could be one of [...] Read more.
The precision fish farming concept has been widely investigated in research and is highly desirable in aquaculture as it creates opportunities for precisely controlling and monitoring fish cultivation processes and increasing fish welfare. The automatic identification of individual fish could be one of the keys to enabling individual fish treatment. In a previous study, we already demonstrated that the visible patterns on a fish’s body can be used for the non-invasive individual identification of fishes from the same species (with obvious skin patterns, such as salmonids) over long-term periods. The aim of this study was to verify the possibility of using fully-automatic non-invasive photo-identification of individual fish based on natural marks on the fish’s body without any obvious skin patterns. This approach is an alternative to stressful invasive tagging and marking techniques. Scale patterns on the body and operculum, as well as lateral line shapes, were used as discriminative features for the identification of individuals in a closed group of fish. We used two fish species: the European seabass Dicentrarchus labrax and the common carp Cyprinus carpio. The identification method was tested on four experimental data sets for each fish species: two separate short-term data sets (pattern variability test) and two long-term data sets (pattern stability test) for European seabass (300 individual fish) and common carp (32 individual fish). The accuracy of classification was 100% for both fish species in both the short-term and long-term experiments. According to these results, the methods used for automatic non-invasive image-based individual-fish identification can also be used for fish species without obvious skin patterns. Full article
(This article belongs to the Topic Computer Vision and Image Processing)
Show Figures

Figure 1

Figure 1
<p>Examples of seabass images. (<b>Top</b>) example of the interrupted lateral line. (<b>Middle</b>) example of a fish with scratches visible on its body. (<b>Bottom</b>) example of a missing upper-tail part.</p>
Full article ">Figure 2
<p>(<b>Left image</b>) seabass data collection design. Digital camera photographing the lateral view of the fish. Image of the seabass (<b>middle</b>, <b>bottom</b>) and carp (<b>middle</b>, <b>top</b>) on the uniform green background. This image was used for automatic fish localization. (<b>Right image</b>) experimental design for carp imaging using a tent with controlled lighting conditions.</p>
Full article ">Figure 3
<p>Identification scheme.</p>
Full article ">Figure 4
<p>Identification scheme. (<b>Left images</b>), fish detected in the original image (black and white mask) what is the output of the of fish detection procedure; (<b>Right images</b>), ROI (ROI(LL) for seabass (bottom image) and ROI1 for carp (upper image) localized in the fish bounding box by the defined percentage of fish width and height.</p>
Full article ">Figure 5
<p>Different seabass ROIs that were used for identification.</p>
Full article ">Figure 6
<p>Position of ROIs for carp identification.</p>
Full article ">Figure 7
<p>HOG visualization for ROIs. The left image is the ROI1 for carp resized to 64 by 64 pixels, where the white holes are the scales of the fish, and black lines are the connecting parts of the scales. The right image represents the ROI(LL) for seabass resized to 64 by 64 pixels, where the black line is the lateral line. The rectangle in the left-top corner of both images visualizes the HOG gradients used for pattern parametrization. The gradience are tiny white lines, which correspond to the orientation of the edges in the image. For common carp it follows the shape of the scale and for the seabass it follows the lateral line.</p>
Full article ">Figure 8
<p>Differences in ROIs for different fish (seabass).</p>
Full article ">
Back to TopTop