[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
sensors-logo

Journal Browser

Journal Browser

Artificial Intelligence in Computer Vision: Methods and Applications2nd Edition

A special issue of Sensors (ISSN 1424-8220). This special issue belongs to the section "Sensing and Imaging".

Deadline for manuscript submissions: 25 February 2025 | Viewed by 4721

Special Issue Editors


E-Mail Website
Guest Editor
Department of Mechanical Engineering, The Catholic University of America, Washington, DC 20064, USA
Interests: optics; mechanics; robotics; computer vision
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Spree3D, Alameda, CA 94502, USA
Interests: computer vision; computational photography; machine learning

E-Mail Website
Guest Editor
Neuroimaging Research Branch, National Institute on Drug Abuse, National Institutes of Health, Baltimore, MD 21224, USA
Interests: computer vision; machine learning; deep learning; computer hardware; neuroimaging
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
U.S. Army Research Laboratory, 2201 Aberdeen Boulevard, Aberdeen, MD 21005, USA
Interests: machine learning

Special Issue Information

Dear Colleagues,

In recent years, there has been high interest in the research and development of artificial intelligence techniques. In the meantime, computer vision methods have been enhanced and extended to encompass an astonishing number of novel sensors and measurement systems. As artificial intelligence spreads over almost all fields of science and engineering, computer vision remains one of its primary application areas. Notably, incorporating artificial intelligence into computer vision-based sensing and measurement techniques has led to numerous unprecedented performances, such as high-accuracy object detection, image segmentation, human pose estimation, and real-time 3D sensing, which cannot be fulfilled using conventional methods.

This Special Issue aims to cover the recent advancements in computer vision that involve using artificial intelligence methods, with a particular interest in sensors and sensing. Both original research and review articles are welcome. Typical topics include but are not limited to the following:

  • Physical, chemical, biological, and healthcare sensors and sensing techniques with deep learning approaches;
  • Localization, mapping, and navigation techniques with artificial intelligence;
  • Artificial intelligence-based recognition of objects, scenes, actions, faces, gestures, expressions, and emotions, as well as object relations and interactions;
  • 3D imaging and sensing with deep learning schemes;
  • Accurate learning with simulation datasets or with a small number of training labels for sensors and sensing;
  • Supervised and unsupervised learning for sensors and sensing;
  • Broad computer vision methods and applications that involve using deep learning or artificial intelligence.

Dr. Zhaoyang Wang
Dr. Minh P. Vo
Dr. Hieu Nguyen
Dr. John Hyatt
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Sensors is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • artificial intelligence
  • deep learning
  • computer vision
  • smart sensors
  • intelligent sensing
  • 3D imaging and sensing
  • localization and mapping
  • navigation and positioning

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Related Special Issue

Published Papers (4 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

19 pages, 8290 KiB  
Article
Multi-Scale Contrastive Learning with Hierarchical Knowledge Synergy for Visible-Infrared Person Re-Identification
by Yongheng Qian and Su-Kit Tang
Sensors 2025, 25(1), 192; https://doi.org/10.3390/s25010192 - 1 Jan 2025
Viewed by 439
Abstract
Visible-infrared person re-identification (VI-ReID) is a challenging cross-modality retrieval task to match a person across different spectral camera views. Most existing works focus on learning shared feature representations from the final embedding space of advanced networks to alleviate modality differences between visible and [...] Read more.
Visible-infrared person re-identification (VI-ReID) is a challenging cross-modality retrieval task to match a person across different spectral camera views. Most existing works focus on learning shared feature representations from the final embedding space of advanced networks to alleviate modality differences between visible and infrared images. However, exclusively relying on high-level semantic information from the network’s final layers can restrict shared feature representations and overlook the benefits of low-level details. Different from these methods, we propose a multi-scale contrastive learning network (MCLNet) with hierarchical knowledge synergy for VI-ReID. MCLNet is a novel two-stream contrastive deep supervision framework designed to train low-level details and high-level semantic representations simultaneously. MCLNet utilizes supervised contrastive learning (SCL) at each intermediate layer to strengthen visual representations and enhance cross-modality feature learning. Furthermore, a hierarchical knowledge synergy (HKS) strategy for pairwise knowledge matching promotes explicit information interaction across multi-scale features and improves information consistency. Extensive experiments on three benchmarks demonstrate the effectiveness of MCLNet. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) The traditional supervised learning paradigm only imposes supervision on the last layer of a neural network. (<b>b</b>) Deep supervised learning involves training the last and intermediate layers concurrently. (<b>c</b>) Grad-CAM [<a href="#B21-sensors-25-00192" class="html-bibr">21</a>] visualization of attention maps at different feature extraction stages. Deeper red colors signify higher weights.</p>
Full article ">Figure 2
<p>Illustration of the proposed MCLNet framework. MCLNet first decouples input IR and RGB images into modality-specific and modality-shared features. It then applies a generalized mean pooling (GeM) layer to generate a feature vector, followed by a batch normalization (BN) layer for identity inference. Meanwhile, the projection head maps multi-scale low-level features to the embedding space, where circles represent logits from the final layer and intermediate layers. Here, supervised contrastive learning (SCL) jointly supervises high-level semantics and low-level details while introducing a hierarchical knowledge synergy (HKS) strategy, using pairwise knowledge matching to enhance information consistency across supervised branches.</p>
Full article ">Figure 3
<p>Analysis of trade-off coefficient <math display="inline"><semantics> <msub> <mi>λ</mi> <mrow> <mi>M</mi> <mi>R</mi> </mrow> </msub> </semantics></math>. Re-identification rates at Rank-1 (%) and mAP (%).</p>
Full article ">Figure 4
<p>The distribution of intra-person and inter-person similarities on the two search modes of the SYSU-MM01 dataset.</p>
Full article ">Figure 5
<p>Visualization results of different modality images of randomly selected two identities. Grad-CAM [<a href="#B21-sensors-25-00192" class="html-bibr">21</a>] visualization of the attention maps of B, B + H, B + S, and MCLNet methods are performed, respectively. Deeper red colors signify higher weights.</p>
Full article ">Figure 6
<p>Top-10 retrieved results of some example queries with the MCLNet on SYSU-MM01 and RegDB. The green and red bounding boxes indicate query results matching the same identity and different identities from the gallery, respectively.</p>
Full article ">
27 pages, 12241 KiB  
Article
SURABHI: Self-Training Using Rectified Annotations-Based Hard Instances for Eidetic Cattle Recognition
by Manu Ramesh and Amy R. Reibman
Sensors 2024, 24(23), 7680; https://doi.org/10.3390/s24237680 - 30 Nov 2024
Viewed by 458
Abstract
We propose a self-training scheme, SURABHI, that trains deep-learning keypoint detection models on machine-annotated instances, together with the methodology to generate those instances. SURABHI aims to improve the keypoint detection accuracy not by altering the structure of a deep-learning-based keypoint detector model but [...] Read more.
We propose a self-training scheme, SURABHI, that trains deep-learning keypoint detection models on machine-annotated instances, together with the methodology to generate those instances. SURABHI aims to improve the keypoint detection accuracy not by altering the structure of a deep-learning-based keypoint detector model but by generating highly effective training instances. The machine-annotated instances used in SURABHI are hard instances—instances that require a rectifier to correct the keypoints misplaced by the keypoint detection model. We engineer this scheme for the task of predicting keypoints of cattle from the top, in conjunction with our Eidetic Cattle Recognition System, which is dependent on accurate prediction of keypoints for predicting the correct cow ID. We show that the final cow ID prediction accuracy on previously unseen cows also improves significantly after applying SURABHI to a deep-learning detection model with high capacity, especially when available training data are minimal. SURABHI helps us achieve a top-6 cow recognition accuracy of 91.89% on a dataset of cow videos. Using SURABHI on this dataset also improves the number of cow instances with correct identification by 22% over the baseline result from fully supervised training. Full article
Show Figures

Figure 1

Figure 1
<p>Updated cow recognition system block diagram with keypoint rectifier block (highlighted with cross-hatching).</p>
Full article ">Figure 2
<p>Abstract representation of good/easy, bad, and hard instances using Cartesian quadrants with Venn diagrams inside them.</p>
Full article ">Figure 3
<p>All ten keypoints labeled in top view. Image source: [<a href="#B1-sensors-24-07680" class="html-bibr">1</a>].</p>
Full article ">Figure 4
<p>Our visualization tool with an example of the cow recognition system correctly predicting the cow ID for the cow in the given image from the holding area. The two images on the top right of the figure are template-aligned images—one on the left produced from the given image and the other on the right fetched from the cattlog (cattle catalog). The rotated bounding box computed from the instance mask, the keypoint skeleton, and the predicted cow ID (‘2203’) are all overlaid on top of the cow instance.</p>
Full article ">Figure 5
<p>Keypoint rectification flowchart.</p>
Full article ">Figure 6
<p>Some examples of keypoint predictions before and after successful keypoint rectification using the Iterative Mode. The number of rules passed out of the 21 rules of RulesChecker1 is overlaid as ‘KP_conf’. (<b>a</b>) Example where only one keypoint (center–back) is corrected by placing it on a second-order polynomial curve fit through the spine points that are correctly detected. (<b>b</b>) Case that Iterative Mode was designed to handle—first, the hip connector is corrected, then the withers, and then the two corrected keypoints together with the tail–head keypoint are used to interpolate/correct the misplaced center–back.</p>
Full article ">Figure 7
<p>Some examples of keypoint predictions before and after successful keypoint rectification using the SBF Mode. The number of rules passed out of the 21 rules of RulesChecker1 is overlaid as ‘KP_conf’. (<b>a</b>) Case where both the pin bones are misplaced. The worst-hit keypoint will usually be the tail–head as it is involved in broken rules associated with both pin bones. SBF Mode removes the second and third worst-hit keypoints, which are the two pin bones, and interpolates new ones in their places. (<b>b</b>) Case where SBF Mode removes the left hip bone and the hip-connector keypoints and corrects them. This is a case where Iterative Mode fails, going against expectation, and SBF Mode comes to aid.</p>
Full article ">Figure 7 Cont.
<p>Some examples of keypoint predictions before and after successful keypoint rectification using the SBF Mode. The number of rules passed out of the 21 rules of RulesChecker1 is overlaid as ‘KP_conf’. (<b>a</b>) Case where both the pin bones are misplaced. The worst-hit keypoint will usually be the tail–head as it is involved in broken rules associated with both pin bones. SBF Mode removes the second and third worst-hit keypoints, which are the two pin bones, and interpolates new ones in their places. (<b>b</b>) Case where SBF Mode removes the left hip bone and the hip-connector keypoints and corrects them. This is a case where Iterative Mode fails, going against expectation, and SBF Mode comes to aid.</p>
Full article ">Figure 8
<p>Evaluation strategy for whole-day experiments: The cut-videos from Summer22 Day1 and Day2 are split into SetA and SetB each. Day1-SetA videos are used for training the keypoint detector model. The evaluation results from the three other subsets are combined to obtain new insights as shown.</p>
Full article ">Figure 9
<p>Classification of training data. Each node is a set of instances. The subsets at children nodes of <math display="inline"><semantics> <mrow> <mi>M</mi> <mi>a</mi> <mi>c</mi> <mi>h</mi> <mi>i</mi> <mi>n</mi> <mi>e</mi> <mi>A</mi> <mi>n</mi> <mi>n</mi> <mi>o</mi> <mi>t</mi> <mi>a</mi> <mi>t</mi> <mi>e</mi> <mi>d</mi> </mrow> </semantics></math> dataset do not necessarily partition the sets at their parent nodes. The <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>t</mi> <mi>r</mi> <mi>i</mi> <mi>c</mi> <mi>t</mi> </mrow> </semantics></math> sets are in fact subsets of the sibling <math display="inline"><semantics> <mrow> <mi>l</mi> <mi>e</mi> <mi>n</mi> <mi>i</mi> <mi>e</mi> <mi>n</mi> <mi>t</mi> </mrow> </semantics></math> nodes. We show them as siblings instead of showing a parent-child relationship as we are explaining only the classification.</p>
Full article ">Figure 10
<p>Plot of results from half-day experiments: evaluation on seen and unseen cows on a fresh day.</p>
Full article ">Figure 11
<p>Plot of results from half-day experiments: evaluation on unseen cows on distinct days.</p>
Full article ">
22 pages, 12107 KiB  
Article
Deep Learning-Based Classification of Macrofungi: Comparative Analysis of Advanced Models for Accurate Fungi Identification
by Sifa Ozsari, Eda Kumru, Fatih Ekinci, Ilgaz Akata, Mehmet Serdar Guzel, Koray Acici, Eray Ozcan and Tunc Asuroglu
Sensors 2024, 24(22), 7189; https://doi.org/10.3390/s24227189 - 9 Nov 2024
Viewed by 1182
Abstract
This study focuses on the classification of six different macrofungi species using advanced deep learning techniques. Fungi species, such as Amanita pantherina, Boletus edulis, Cantharellus cibarius, Lactarius deliciosus, Pleurotus ostreatus and Tricholoma terreum were chosen based on their ecological [...] Read more.
This study focuses on the classification of six different macrofungi species using advanced deep learning techniques. Fungi species, such as Amanita pantherina, Boletus edulis, Cantharellus cibarius, Lactarius deliciosus, Pleurotus ostreatus and Tricholoma terreum were chosen based on their ecological importance and distinct morphological characteristics. The research employed 5 different machine learning techniques and 12 deep learning models, including DenseNet121, MobileNetV2, ConvNeXt, EfficientNet, and swin transformers, to evaluate their performance in identifying fungi from images. The DenseNet121 model demonstrated the highest accuracy (92%) and AUC score (95%), making it the most effective in distinguishing between species. The study also revealed that transformer-based models, particularly the swin transformer, were less effective, suggesting room for improvement in their application to this task. Further advancements in macrofungi classification could be achieved by expanding datasets, incorporating additional data types such as biochemical, electron microscopy, and RNA/DNA sequences, and using ensemble methods to enhance model performance. The findings contribute valuable insights into both the use of deep learning for biodiversity research and the ecological conservation of macrofungi species. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of datasets utilized for training AI algorithms, presented from a macroscopic perspective.</p>
Full article ">Figure 2
<p>Validation accuracy.</p>
Full article ">Figure 3
<p>ROC curve.</p>
Full article ">Figure 4
<p>İmages without Grad-CAM visualization.</p>
Full article ">Figure 5
<p>ConvNeXt Grad-CAM visualization.</p>
Full article ">Figure 6
<p>EfficientNet Grad-CAM visualization.</p>
Full article ">Figure 7
<p>DenseNet121, InceptionV3, and InceptionResNetV2 Grad-CAM visualization.</p>
Full article ">Figure 8
<p>MobileNetV2, ResNet152, and Xception Grad-CAM visualization.</p>
Full article ">Figure 9
<p>Different levels of Gaussian white noise [<a href="#B40-sensors-24-07189" class="html-bibr">40</a>].</p>
Full article ">Figure 10
<p>DenseNet121 and MobileNetV2 Grad-CAM visualization on SNR-10 noisy images.</p>
Full article ">
23 pages, 1025 KiB  
Article
Adversarial Examples on XAI-Enabled DT for Smart Healthcare Systems
by Niddal H. Imam
Sensors 2024, 24(21), 6891; https://doi.org/10.3390/s24216891 - 27 Oct 2024
Viewed by 1182
Abstract
There have recently been rapid developments in smart healthcare systems, such as precision diagnosis, smart diet management, and drug discovery. These systems require the integration of the Internet of Things (IoT) for data acquisition, Digital Twins (DT) for data representation into a digital [...] Read more.
There have recently been rapid developments in smart healthcare systems, such as precision diagnosis, smart diet management, and drug discovery. These systems require the integration of the Internet of Things (IoT) for data acquisition, Digital Twins (DT) for data representation into a digital replica and Artificial Intelligence (AI) for decision-making. DT is a digital copy or replica of physical entities (e.g., patients), one of the emerging technologies that enable the advancement of smart healthcare systems. AI and Machine Learning (ML) offer great benefits to DT-based smart healthcare systems. They also pose certain risks, including security risks, and bring up issues of fairness, trustworthiness, explainability, and interpretability. One of the challenges that still make the full adaptation of AI/ML in healthcare questionable is the explainability of AI (XAI) and interpretability of ML (IML). Although the study of the explainability and interpretability of AI/ML is now a trend, there is a lack of research on the security of XAI-enabled DT for smart healthcare systems. Existing studies limit their focus to either the security of XAI or DT. This paper provides a brief overview of the research on the security of XAI-enabled DT for smart healthcare systems. It also explores potential adversarial attacks against XAI-enabled DT for smart healthcare systems. Additionally, it proposes a framework for designing XAI-enabled DT for smart healthcare systems that are secure and trusted. Full article
Show Figures

Figure 1

Figure 1
<p>Number of research papers identified and reasons for exclusion.</p>
Full article ">Figure 2
<p>Proposed DTs framework and layer.</p>
Full article ">Figure 3
<p>Feature Importance.</p>
Full article ">Figure 4
<p>SHAP Explanation.</p>
Full article ">Figure 5
<p>RF Feature Importance After the Label-flipping Attack.</p>
Full article ">Figure 6
<p>SHAP Explanation After the Label-flipping Attack.</p>
Full article ">
Back to TopTop